GTX is dead long live RTX !!

Message boards : Number crunching : GTX is dead long live RTX !!
Message board moderation

To post messages, you must log in.

Previous · 1 · 2 · 3 · 4 · 5 · 6 · 7 · 8 . . . 12 · Next

AuthorMessage
Profile Keith Myers Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 29 Apr 01
Posts: 13164
Credit: 1,160,866,277
RAC: 1,873
United States
Message 1956537 - Posted: 20 Sep 2018, 20:50:13 UTC

Ask Petri to post his first 30 lines of his Event Log to show how BOINC reports his Titan V?
Seti@Home classic workunits:20,676 CPU time:74,226 hours

A proud member of the OFA (Old Farts Association)
ID: 1956537 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1956540 - Posted: 20 Sep 2018, 21:38:53 UTC - in response to Message 1956537.  

Ask Petri to post his first 30 lines of his Event Log to show how BOINC reports his Titan V?
Wrong way round. I'm trying to re-write BOINC so that it tells Petri the truth.
ID: 1956540 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1956542 - Posted: 20 Sep 2018, 21:46:43 UTC - in response to Message 1956535.  

yes, 64 cores/SM.

straight from nvidia: https://devblogs.nvidia.com/nvidia-turing-architecture-in-depth/

and confirmation of Volta here: http://images.nvidia.com/content/volta-architecture/pdf/volta-architecture-whitepaper.pdf
Thanks, that's exactly the sort of thing I was looking for, and I've found both references.

But I'm a bit worried by the second one. Table 1, page 10: Tesla P100, GP100 (Pascal) is shown with 64 cores per SM. Observations?
ID: 1956542 · Report as offensive
Ian&Steve C.
Avatar

Send message
Joined: 28 Sep 99
Posts: 4267
Credit: 1,282,604,591
RAC: 6,640
United States
Message 1956576 - Posted: 21 Sep 2018, 0:50:54 UTC

my only guess is that it's because theyre talking about the GP100 (Tesla) card, and not the mainstream geforce gaming cards that we are all used to. You'll notice that the Tesla card also sports HBM2 where the 1080ti has GDDR5X.

even though it's pascal, the GP100 is a different beast.

https://forums.anandtech.com/threads/gp100-and-gp104-are-different-architectures.2473319/
Seti@Home classic workunits: 29,492 CPU time: 134,419 hours

ID: 1956576 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1956613 - Posted: 21 Sep 2018, 7:31:16 UTC - in response to Message 1956576.  
Last modified: 21 Sep 2018, 7:40:57 UTC

Thanks. That does seem definitive. The picture is completed by https://docs.nvidia.com/cuda/pascal-tuning-guide/index.html:

The specific compute capabilities of GP100 and GP104 are 6.0 and 6.1, respectively. The GP102 architecture is similar to GP104.
(footnote 2). So we could detect and use that information, though I'm not sure the BOINC developers will want to bother.

Edit - let me bookmark https://en.wikipedia.org/wiki/CUDA#GPUs_supported as a handy reference list of compute capabilities.
ID: 1956613 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1956627 - Posted: 21 Sep 2018, 11:50:14 UTC

OK, I've submitted BOINC code via #2707 so that the RTX cards should display the correct GFLOPS peak from BOINC version 7.14 onwards. Earlier versions will show double the correct speed for RTX, Titan V, and Tesla P100 - hardware wizards please verify.

This project arose from an email from Ray Hinchliffe, author of SIV. BOINC and SIV use slightly different techniques for calculating peak flops: BOINC can look up the number of SMs (shader multiplexes), but needs to be told how many cores are assigned to each SM (my changes above). Ray has a technique for detecting the total number of cores directly. We compare notes periodically to make sure we're getting the same answer.

If anyone out there has a P100, Titan V, or even RTX, it would be interesting to know whether the current BOINC versions (anything later than 2014) give a different peak flops value from SIV, and (in due course) whether BOINC v7.14 converges them again.
ID: 1956627 · Report as offensive
Profile Keith Myers Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 29 Apr 01
Posts: 13164
Credit: 1,160,866,277
RAC: 1,873
United States
Message 1956667 - Posted: 21 Sep 2018, 21:14:00 UTC

More benchmark tests available, this time concentrating on compute performance of the RTX 2080Ti. The Phoronix.com tests had some good CUDA and OpenCL tests that probably are similar enough to make some conclusions about how the 2080Ti would perform on similar Seti work. Also some interesting performance per dollar charts comparing RTX 2080Ti versus GTX 1080Ti.

NVIDIA GeForce RTX 2080 Ti Shows Very Strong Compute Performance Potential
Seti@Home classic workunits:20,676 CPU time:74,226 hours

A proud member of the OFA (Old Farts Association)
ID: 1956667 · Report as offensive
Grant (SSSF)
Volunteer tester

Send message
Joined: 19 Aug 99
Posts: 13736
Credit: 208,696,464
RAC: 304
Australia
Message 1956670 - Posted: 21 Sep 2018, 21:53:21 UTC - in response to Message 1956667.  

More benchmark tests available, this time concentrating on compute performance of the RTX 2080Ti.

Other than a test here or there, it's often twice as fast as a GTX 1080Ti (and even faster than that in 1 or 2 tests).

And looking at the cost/performance comparison, even with the RTX 2080Ti's obscene pricing it actually compares favourably with the GTX 1080Ti- it's behind for most tests, but dead level for one or 2. And this is running code that isn't optimised in any way for the new architecture.

While you can't really justify the RTX2080Ti price for gaming (other than bragging rights) till there are games that can really make use of it's Ray Tracing feautres, when it comes to compute workloads its improved performance will pay for itself in no time (for heavy use scenarios). And it would do even better with software written specifically for it.
Very impressive piece of hardware.
Grant
Darwin NT
ID: 1956670 · Report as offensive
Profile Keith Myers Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 29 Apr 01
Posts: 13164
Credit: 1,160,866,277
RAC: 1,873
United States
Message 1956674 - Posted: 21 Sep 2018, 22:06:37 UTC - in response to Message 1956670.  

I agree and why I wanted to hold off any "is it worth it" comments until I saw compute performance tests since its gaming performance is moot in our use case. I think it will come out pretty good regarding credit per watt once one of these show up in Shaggie's charts. It is only pulling about 20-40 watts more than the 1080Ti in these tests. I would like to see how it compared to the Titan V but it looks like Phoronix never did a compute test on that card.
Seti@Home classic workunits:20,676 CPU time:74,226 hours

A proud member of the OFA (Old Farts Association)
ID: 1956674 · Report as offensive
Profile Zalster Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 27 May 99
Posts: 5517
Credit: 528,817,460
RAC: 242
United States
Message 1956681 - Posted: 21 Sep 2018, 22:23:00 UTC - in response to Message 1956674.  

pull the trigger...pull the trigger.... ;)
ID: 1956681 · Report as offensive
Grant (SSSF)
Volunteer tester

Send message
Joined: 19 Aug 99
Posts: 13736
Credit: 208,696,464
RAC: 304
Australia
Message 1956687 - Posted: 21 Sep 2018, 22:40:14 UTC - in response to Message 1956681.  

pull the trigger...pull the trigger.... ;)

If I had $2,000 lying around I certainly would, unfortunately...
Grant
Darwin NT
ID: 1956687 · Report as offensive
Profile Keith Myers Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 29 Apr 01
Posts: 13164
Credit: 1,160,866,277
RAC: 1,873
United States
Message 1956697 - Posted: 21 Sep 2018, 23:01:37 UTC

I'm going to have to wait on pulling the trigger too. Still want to put a TR 2950X build in front of it. But if and when I do pull the trigger, it will be either a EVGA AIO RTX 2080Ti or a custom block version.
Seti@Home classic workunits:20,676 CPU time:74,226 hours

A proud member of the OFA (Old Farts Association)
ID: 1956697 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1956903 - Posted: 23 Sep 2018, 10:22:02 UTC

First live RTX 2080 has been tested. Current versions of BOINC do report the wrong value for peak flops, but my revision has been accepted and will be included in BOINC v7.14. That code has been tested in the real world too, and gives the correct flops value.
ID: 1956903 · Report as offensive
Profile Brent Norman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester

Send message
Joined: 1 Dec 99
Posts: 2786
Credit: 685,657,289
RAC: 835
Canada
Message 1956905 - Posted: 23 Sep 2018, 10:46:21 UTC - in response to Message 1956903.  

May we ask who has one, so that we can look at the performance of it?
ID: 1956905 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1956906 - Posted: 23 Sep 2018, 10:52:59 UTC - in response to Message 1956905.  

May we ask who has one, so that we can look at the performance of it?
The testing was done by Ray Hinchliffe, as mentioned in message 1956627. He provides BOINC-specific extensions to SIV, but I don't know if he runs SETI.
ID: 1956906 · Report as offensive
Profile Brent Norman Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester

Send message
Joined: 1 Dec 99
Posts: 2786
Credit: 685,657,289
RAC: 835
Canada
Message 1956907 - Posted: 23 Sep 2018, 10:55:11 UTC - in response to Message 1956906.  

OK, thanks.
ID: 1956907 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1956918 - Posted: 23 Sep 2018, 13:05:48 UTC
Last modified: 23 Sep 2018, 13:06:05 UTC

Did anyone knows if there is any RTX2080 actualy crunching SETI? So we could follow the performance in the real world.
ID: 1956918 · Report as offensive
Profile Keith Myers Special Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 29 Apr 01
Posts: 13164
Credit: 1,160,866,277
RAC: 1,873
United States
Message 1956950 - Posted: 23 Sep 2018, 16:28:13 UTC - in response to Message 1956918.  

We're still waiting on a Turing card to show up here. Supposed to be one next week on Einstein by Archae86. In that thread Mikey says someone already has one running on PrimeGrid. Says it is faster but not OMG faster. Same generational improvement seen on every new card family release. Someone else did an analysis of the efficiency of the Turing CUDA core compared to the Pascal core on PrimeGrid tasks and came up with 46 -50% faster.
Seti@Home classic workunits:20,676 CPU time:74,226 hours

A proud member of the OFA (Old Farts Association)
ID: 1956950 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1956989 - Posted: 23 Sep 2018, 20:55:44 UTC
Last modified: 23 Sep 2018, 20:56:48 UTC

Let's wait.
ID: 1956989 · Report as offensive
Profile zoom3+1=4
Volunteer tester
Avatar

Send message
Joined: 30 Nov 03
Posts: 65747
Credit: 55,293,173
RAC: 49
United States
Message 1957068 - Posted: 24 Sep 2018, 3:40:03 UTC - in response to Message 1956903.  

First live RTX 2080 has been tested. Current versions of BOINC do report the wrong value for peak flops, but my revision has been accepted and will be included in BOINC v7.14. That code has been tested in the real world too, and gives the correct flops value.

That's good, but the price is about to go up and it has nothing to do with Nvidia or AMD, Intel, Asus, Gigabyte, MSI, PNY, etc, etc, etc.

Buying that used Asus R3E motherboard a while back that needs a new cpu socket is looking mighty good now. :)
The T1 Trust, PRR T1 Class 4-4-4-4 #5550, 1 of America's First HST's
ID: 1957068 · Report as offensive
Previous · 1 · 2 · 3 · 4 · 5 · 6 · 7 · 8 . . . 12 · Next

Message boards : Number crunching : GTX is dead long live RTX !!


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.