Message boards :
Number crunching :
GPU FLOPS: Theory vs Reality
Message board moderation
Previous · 1 . . . 7 · 8 · 9 · 10 · 11 · 12 · 13 . . . 20 · Next
Author | Message |
---|---|
-= Vyper =- Send message Joined: 5 Sep 99 Posts: 1652 Credit: 1,065,191,981 RAC: 2,537 |
Actually one of my hosts is using a Cuda10 driver and the Cuda10 app.. :) on 4 x 750TIs https://setiathome.berkeley.edu/show_host_detail.php?hostid=8570185 _________________________________________________________________________ Addicted to SETI crunching! Founder of GPU Users Group |
zoom3+1=4 Send message Joined: 30 Nov 03 Posts: 66301 Credit: 55,293,173 RAC: 49 |
I have been waiting for a new scan, with data for the RTX 20xx cards. Are we getting there? 2080 = smokes, see 2080Ti 2080Ti = Keep a Fire Extinguisher handy Savoir-Faire is everywhere! The T1 Trust, T1 Class 4-4-4-4 #5550, America's First HST |
Oddbjornik Send message Joined: 15 May 99 Posts: 220 Credit: 349,610,548 RAC: 1,728 |
Actually one of my hosts is using a Cuda10 driver and the Cuda10 app.. :) on 4 x 750TIsBut the 750TI has compute capability 5.0, so how does it run Cuda 10? Is the software backward compatible, so I can put the Cuda 10 software on my GTX 680 system (while the RTX 2070 is still in the mail) and have it chug along just like that? Or, perhaps more likely, is there something here that I totally don't understand? |
Wiggo Send message Joined: 24 Jan 00 Posts: 36642 Credit: 261,360,520 RAC: 489 |
The 680 is an older Kepler based product where as the 750 is a newer Maxwell based product so I doubt it work be a good choice.Actually one of my hosts is using a Cuda10 driver and the Cuda10 app.. :) on 4 x 750TIsBut the 750TI has compute capability 5.0, so how does it run Cuda 10? Cheers. |
rob smith Send message Joined: 7 Mar 03 Posts: 22508 Credit: 416,307,556 RAC: 380 |
Provided the underlying hardware of the GPU is supported by Cuda 10 then it will run. In terms of computation there is probably no advantage in running Cuda 10 just now as nobody has released optimised the applications yet (but they are working on them) Bob Smith Member of Seti PIPPS (Pluto is a Planet Protest Society) Somewhere in the (un)known Universe? |
Oddbjornik Send message Joined: 15 May 99 Posts: 220 Credit: 349,610,548 RAC: 1,728 |
Provided the underlying hardware of the GPU is supported by Cuda 10 then it will run. In terms of computation there is probably no advantage in running Cuda 10 just now as nobody has released optimised the applications yet (but they are working on them)A-ha, so when it says on this page that "CUDA SDK 10.0 [has] support for compute capability 3.0 – 7.5 (Kepler, Maxwell, Pascal, Volta, Turing)", that means the 680, which has CC 3.0, has a fair chance of working with Cuda 10 software, even though it probably won't be very efficient (?). I have been confusing the version levels of the compute capability (most recent version is 7.5 with the Turing cards) and the SDK (10.0 is the latest and greatest). |
Keith Myers Send message Joined: 29 Apr 01 Posts: 13164 Credit: 1,160,866,277 RAC: 1,873 |
Provided the underlying hardware of the GPU is supported by Cuda 10 then it will run. In terms of computation there is probably no advantage in running Cuda 10 just now as nobody has released optimised the applications yet (but they are working on them)A-ha, so when it says on this page that "CUDA SDK 10.0 [has] support for compute capability 3.0 – 7.5 (Kepler, Maxwell, Pascal, Volta, Turing)", that means the 680, which has CC 3.0, has a fair chance of working with Cuda 10 software, even though it probably won't be very efficient (?). I'm not sure what we are talking about. If we are talking about the Linux "special apps", then yes they do in fact have compute capability limits. Depends on how the apps were compiled on what platfrom and compiler. This is from the CA post explaining the platform requirements. Check the list of supported GPUs here, https://en.wikipedia.org/wiki/CUDA#GPUs_supported The CUDA 6.0 App requires at least CC=3.5 The CUDA 9.0 Apps requires at least CC=5.0 The CUDA 9.2 App requires at least CC=6.1 Compiled & Tested in Ubuntu. Read the README_x41p_xxxx.txt file in docs for best use, the CUDA Libraries are included. The CUDA 6 & 9 Apps will run in Ubuntu 14.04.1 and higher. The CUDA 9.2 App requires Ubuntu 16.04. The CUDA 6.0 Special App is for the older Kepler CC 3.5 GPUs that might not work well with CUDA 7.5 and above. The CUDA 9.0 App is for most normal systems and is tuned to also run on Maxwell GPUs. Place the expanded files in the setiathome.berkeley.edu folder, and set file permissions if using the Repository version of BOINC. The GTX 750Ti is a CC 5.0 card and will run on the CUDA 9.0 app. Not sure how it is running on the CUDA 10.0 application. There are two versions of the CUDA 10.0 app out there I believe, the original and maybe a reworked one by Petri and then CA one by TBar. [Edit] So went back to the original TBar post for the CUDA10 app and see he has compiled it to work with 750Ti because he is using sm_75 code; It has sm_75 code, so it's a little bigger. It will work with the 750Ti & higher in 14.04.1 and higher. So that is how the GTX 750Ti is working with the CUDA10 driver and app. So the original post from CA back in the summer is a little outdated. Seti@Home classic workunits:20,676 CPU time:74,226 hours A proud member of the OFA (Old Farts Association) |
Oddbjornik Send message Joined: 15 May 99 Posts: 220 Credit: 349,610,548 RAC: 1,728 |
[Edit] So went back to the original TBar post for the CUDA10 app and see he has compiled it to work with 750Ti because he is using sm_75 code; And that was the last piece of information I needed in order to understand how it all connects. Thank you, Keith. This means I'll wait for the RTX 2070 to arrive before I proceed. |
Shaggie76 Send message Joined: 9 Oct 09 Posts: 282 Credit: 271,858,118 RAC: 196 |
I'll try to remember to start a scan before I retire for the night -- it ties up my laptop for a few hours. There might be enough 2080's to qualify by now. |
Shaggie76 Send message Joined: 9 Oct 09 Posts: 282 Credit: 271,858,118 RAC: 196 |
Or not -- the servers keep timing out and I can even finish downloading the host database. I'll try to remember in a few days to check again. |
Keith Myers Send message Joined: 29 Apr 01 Posts: 13164 Credit: 1,160,866,277 RAC: 1,873 |
Or not -- the servers keep timing out and I can even finish downloading the host database. I'll try to remember in a few days to check again. Will be appreciated Shaggie whenever the servers allow it. My 2080 won't show up though because it is running the special app and not OpenCL. Seti@Home classic workunits:20,676 CPU time:74,226 hours A proud member of the OFA (Old Farts Association) |
juan BFP Send message Joined: 16 Mar 07 Posts: 9786 Credit: 572,710,851 RAC: 3,799 |
Is possible to make the same table but instead of the ones who use OpenCL builds shows the ones who run the special CUDA builds? |
Keith Myers Send message Joined: 29 Apr 01 Posts: 13164 Credit: 1,160,866,277 RAC: 1,873 |
Is possible to make the same table but instead of the ones who use OpenCL builds shows the ones who run the special CUDA builds? I believe Shaggie answered that question once and said it is not possible. Seti@Home classic workunits:20,676 CPU time:74,226 hours A proud member of the OFA (Old Farts Association) |
lunkerlander Send message Joined: 23 Jul 18 Posts: 82 Credit: 1,353,232 RAC: 4 |
Is possible to make the same table but instead of the ones who use OpenCL builds shows the ones who run the special CUDA builds? If not, you can probably infer how Nvidia GPUs' performance compares to one another. I also like to click on the computers tab from some of the top SETI hosts and look at individual tasks performed from their PC. You can get a good idea how many seconds tasks take to complete with various GPUs by doing this. |
Shaggie76 Send message Joined: 9 Oct 09 Posts: 282 Credit: 271,858,118 RAC: 196 |
I finished aggregating the stats but imgur is having technical issues right now and I can't upload it tonight. As you would expect the RTX 2080 Ti steals the performance crown and the 2070 sets a new record for the performance/watt. I'll try to post tomorrow. |
Keith Myers Send message Joined: 29 Apr 01 Posts: 13164 Credit: 1,160,866,277 RAC: 1,873 |
Whenever you can get it posted, it will be most appreciated by everyone writing out their XMAS shopping lists. Seti@Home classic workunits:20,676 CPU time:74,226 hours A proud member of the OFA (Old Farts Association) |
-= Vyper =- Send message Joined: 5 Sep 99 Posts: 1652 Credit: 1,065,191,981 RAC: 2,537 |
If u take "My host" into the equation then dont forget that my throughput is lower due to 185W set instead of 260-265W as it is per default. It's also approx 2-3 seconds slower than at default. Thanks for your superb list Shaggie _________________________________________________________________________ Addicted to SETI crunching! Founder of GPU Users Group |
Shaggie76 Send message Joined: 9 Oct 09 Posts: 282 Credit: 271,858,118 RAC: 196 |
At last I have data: I'm sorry that the image isn't as sharp as I'd like -- I've transition to a high-DPI laptop and when I convert the graphs to images Excel does weird things. I really should find a better way to finish off the data -- maybe some Perl GD module would be more consistent. For reference here's the number of hosts and tasks analyzed for the top few cards -- there aren't a lot of 20x0 cards in play yet but they've done enough work that the I feel comfortable with the results . |
Bill Send message Joined: 30 Nov 05 Posts: 282 Credit: 6,916,194 RAC: 60 |
Wow, the 2070 jumps off the page! This is the first time I have read this post, so please forgive me if I ask questions that have been asked before: 1. I feel like there is a lot of background with how you compile this data. For example, is the credit/hour and credit/watt-hour calculated all-time, or within a timeframe? Do you have a running list of notes somewhere? 2. I think I browsed one of your earlier posts, and a different NVIDIA card (the 970)? at one point had a higher credit/watt-hour rating. I'm curious what changed for that. 3. Do we know what CPU was used in tandem with the GPU credits? I know the CPU provides a minor role in the crunching of the WU, but I wonder if there is a significant difference between one CPU and another. This is neat to see these types of statistics. I am building a rig right now, and since it is my first build I am starting small with no GPUs. I saw some older GPUs for sale online under $50 (like a GeForce GTX 650). Obviously I would still need to check compatibility with any Boinc projects I want to crunch, but it seemed like a way to increase the number of GPU WUs to crunch on the cheap. I digress, that is probably a conversation for another post. Seti@home classic: 1,456 results, 1.613 years CPU time |
Tom M Send message Joined: 28 Nov 02 Posts: 5126 Credit: 276,046,078 RAC: 462 |
If you can find a gtx 750Ti for maybe $70 you will be happier than with the Gtx 650. (at least I think so. I am happy enough I have 3 of them :) Tom A proud member of the OFA (Old Farts Association). |
©2024 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.