GPU Wars 2014: Postponed to 2015?

Message boards : Number crunching : GPU Wars 2014: Postponed to 2015?
Message board moderation

To post messages, you must log in.

Previous · 1 · 2 · 3 · 4 · Next

AuthorMessage
Gone

Send message
Joined: 31 May 99
Posts: 150
Credit: 125,779,206
RAC: 0
United Kingdom
Message 1480272 - Posted: 21 Feb 2014, 16:56:53 UTC

Its only been running 30 mins and it had some tasks done by a 560 TI during build testing last night.

I started it off with 3 tasks for a while, but it is now down to 2 tasks at 98 % GPU Load
Spec is:
MSI GeForce GTX 750 Ti TWIN FROZR GAMING OC 2GB GDDR5 1085MHz Core, 1163MHz Boost, 5400MHz Memory, 640 CUDA Cores, DVI-D, HDMI, DSub, PCIe 3.0, 3 Year Warranty

Running at 35Deg C

All stock speeds, Lunatics x41 Cuda 42.
ID: 1480272 · Report as offensive
Profile jason_gee
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 24 Nov 06
Posts: 7489
Credit: 91,093,184
RAC: 0
Australia
Message 1480380 - Posted: 21 Feb 2014, 20:15:24 UTC

The interesting part will be to find out what this is:
computeCap 5.0

Given that Kepler Class is 3.0, and BigK (GK110) is 3.5

What I'll do is wade through the unreleased Cuda 6 documentation for hints, though I don't recall seeing mention of it in the early access stuff, I admit I wasn't really looking for that though. Probably should be some in the release candidate...
"Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions.
ID: 1480380 · Report as offensive
Gone

Send message
Joined: 31 May 99
Posts: 150
Credit: 125,779,206
RAC: 0
United Kingdom
Message 1480443 - Posted: 21 Feb 2014, 23:23:32 UTC

@Jason
If you need me to try anything, just let me know.

Reg
ID: 1480443 · Report as offensive
Profile jason_gee
Volunteer developer
Volunteer tester
Avatar

Send message
Joined: 24 Nov 06
Posts: 7489
Credit: 91,093,184
RAC: 0
Australia
Message 1480445 - Posted: 21 Feb 2014, 23:33:15 UTC - in response to Message 1480443.  

@Jason
If you need me to try anything, just let me know.

Reg


Well it's a bit of a mystery at the moment what features this compute capability has, and how to use them ;) I'll keep digging.
"Living by the wisdom of computer science doesn't sound so bad after all. And unlike most advice, it's backed up by proofs." -- Algorithms to live by: The computer science of human decisions.
ID: 1480445 · Report as offensive
Profile shizaru
Volunteer tester
Avatar

Send message
Joined: 14 Jun 04
Posts: 1130
Credit: 1,967,904
RAC: 0
Greece
Message 1480712 - Posted: 22 Feb 2014, 19:01:58 UTC

I know it's early days still but how does it compare against the 560Ti?
ID: 1480712 · Report as offensive
Gone

Send message
Joined: 31 May 99
Posts: 150
Credit: 125,779,206
RAC: 0
United Kingdom
Message 1480749 - Posted: 22 Feb 2014, 21:11:40 UTC

It seems to be 2x faster ! Sadly no tasks at moment.
ID: 1480749 · Report as offensive
Grant (SSSF)
Volunteer tester

Send message
Joined: 19 Aug 99
Posts: 13736
Credit: 208,696,464
RAC: 304
Australia
Message 1480786 - Posted: 22 Feb 2014, 23:22:51 UTC - in response to Message 1480749.  
Last modified: 22 Feb 2014, 23:27:02 UTC

It seems to be 2x faster ! Sadly no tasks at moment.

Very impressive, especially when you consider the power requirements.

GTX 750Ti 60W
GTX 560Ti 210W

Double the processing, for 1/3 of the power.


EDIT- looks like I've found my new video card. If the price is right, might as well get 2 of them to replace my present 560Ti (fan bearings have had it).
Grant
Darwin NT
ID: 1480786 · Report as offensive
Profile HAL9000
Volunteer tester
Avatar

Send message
Joined: 11 Sep 99
Posts: 6534
Credit: 196,805,888
RAC: 57
United States
Message 1480796 - Posted: 22 Feb 2014, 23:55:33 UTC - in response to Message 1480786.  

It seems to be 2x faster ! Sadly no tasks at moment.

Very impressive, especially when you consider the power requirements.

GTX 750Ti 60W
GTX 560Ti 210W

Double the processing, for 1/3 of the power.


EDIT- looks like I've found my new video card. If the price is right, might as well get 2 of them to replace my present 560Ti (fan bearings have had it).

I would guess that factory OC card may be using more than the default 60w configuration rating. With it pumping out 2183 vs 1306 GFLOPS of the stock spec card at least.
No doubt it is still more efficient that the older cards.
SETI@home classic workunits: 93,865 CPU time: 863,447 hours
Join the [url=http://tinyurl.com/8y46zvu]BP6/VP6 User Group[
ID: 1480796 · Report as offensive
Lionel

Send message
Joined: 25 Mar 00
Posts: 680
Credit: 563,640,304
RAC: 597
Australia
Message 1481033 - Posted: 23 Feb 2014, 20:52:11 UTC - in response to Message 1480187.  

Hi, I just took delivery of 3 x 750 Ti and will fire them up in a couple of hours.
I will post back here if they work or not.

Any suggestions as to which version I should run Cuda 42 or Cuda 50 ?

Reg



In reading articles on this card a little while ago, there was no specific mention of either a 6 or 8 pin connector and there was mention that the card drew all its power from the PCI-E bus (60 watts). If the inferences that I read were not wrong and there is no other connector for power on each card (and all 3 are in the one machine) you may be running the risk of melting the 24 pin connector, or damaging the board in some other way.
ID: 1481033 · Report as offensive
archae86

Send message
Joined: 31 Aug 99
Posts: 909
Credit: 1,582,816
RAC: 0
United States
Message 1481083 - Posted: 23 Feb 2014, 22:55:41 UTC - in response to Message 1481033.  

In reading articles on this card a little while ago, there was no specific mention of either a 6 or 8 pin connector and there was mention that the card drew all its power from the PCI-E bus (60 watts).

Among the several GTX 750 and 750 ti cards I've looked over in the last few days, some do have am 6-pin additional power connector (widespread on expressly overclocked 750 Ti cards), and some do not (unknown or scarce on base spec 750 cards) .

References I've seen suggest that the design limit on card draw from the PCIe connector is 75 watts, and I think it likely that all the cards delivered without the 6-pin connector are expected to stay well under that limit under normal operating conditions. As I think the GPU chip supervises its own power consumption, possibly it may be difficult greatly to exceed that even under moderate user abuse.

I would hope the people who developed the PCIe specs and the designers of the motherboards both contemplated that someone might actually plug something into all of the offered slots.

These things are FAR lower power consuming than the higher end sort of gaming cards of the recent past.
ID: 1481083 · Report as offensive
Profile shizaru
Volunteer tester
Avatar

Send message
Joined: 14 Jun 04
Posts: 1130
Credit: 1,967,904
RAC: 0
Greece
Message 1481100 - Posted: 24 Feb 2014, 0:56:15 UTC - in response to Message 1480749.  

I know it's early days still but how does it compare against the 560Ti?

It seems to be 2x faster !


If that's true then it's the biggest news since the move to the colo.

I was going to say I'm surprised the forums haven't caught fire but then the cynic in me tapped my shoulder and promptly reminded me that there is hardly ever any number crunching done in the Number Crunching forum:)

But are u sure? Because before you posted I had a look at your tasks and for some reason I came to the conclusion that the 750ti can do about 80% of what the 560ti can do. Specs would be closer to these numbers too but... specs and Seti never seem to go hand in hand.

Hopefully I'm wrong and my guess on how many tasks you were crunching on each card was off:)
ID: 1481100 · Report as offensive
Gone

Send message
Joined: 31 May 99
Posts: 150
Credit: 125,779,206
RAC: 0
United Kingdom
Message 1481198 - Posted: 24 Feb 2014, 11:57:53 UTC
Last modified: 24 Feb 2014, 12:02:16 UTC

Taken from the Boinc Manager event log just now .....


GeForce GTX 560 Ti 1312 GFLOPS peak) (That's a Ti 448 too)

GeForce GTX 750 Ti 2183 GFLOPS peak)



24/02/2014 11:49:14 | | CUDA: NVIDIA GPU 0: GeForce GTX 560 Ti (driver version 334.89, CUDA version 6.0, compute capability 2.0, 1280MB, 1163MB available, 1312 GFLOPS peak)
24/02/2014 11:49:14 | | CUDA: NVIDIA GPU 1: GeForce GTX 750 Ti (driver version 334.89, CUDA version 6.0, compute capability 5.0, 2048MB, 1948MB available, 2183 GFLOPS peak)
24/02/2014 11:49:14 | | OpenCL: NVIDIA GPU 0: GeForce GTX 560 Ti (driver version 334.89, device version OpenCL 1.1 CUDA, 1280MB, 1163MB available, 1312 GFLOPS peak)
24/02/2014 11:49:14 | | OpenCL: NVIDIA GPU 1: GeForce GTX 750 Ti (driver version 334.89, device version OpenCL 1.1 CUDA, 2048MB, 1948MB available, 2183 GFLOPS peak)



I am using PCIe x1 adaptors to run the 750TIs which have their own molex connector to power the PCI e x16 slot.



ID: 1481198 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1481202 - Posted: 24 Feb 2014, 13:13:16 UTC

Did you try to run the benchmark tools? It could be interesting to see the real performance board on SETI.

You could DL them at: http://lunatics.kwsn.net/index.php?module=Downloads;catd=5

If you do, try first with the GPu direct connected to the PCI-E 3.0 and after that with it conected to the extender. That´s will give you, first the real performance of the GPU and the second how much you loose by using the adaptor (if you loose anything of course).

And please share with us the findings.
ID: 1481202 · Report as offensive
Profile shizaru
Volunteer tester
Avatar

Send message
Joined: 14 Jun 04
Posts: 1130
Credit: 1,967,904
RAC: 0
Greece
Message 1481234 - Posted: 24 Feb 2014, 15:04:21 UTC

With my relatively new (and self-proclaimed) role as Cassandra of these boards I'm going to go ahead and guess that Boinc is just "seeing" the flops wrong (it's not like it runs any benchmarks to get that number, just blurts out whatever it's told AFAIK).

The Gflops for the 448 are spot on but the Gflops for the 750ti should be around 1300 (according to Wiki anyway, which can be a bit dodgy of course).

I do hope I'm wrong though.
ID: 1481234 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1481240 - Posted: 24 Feb 2014, 15:24:02 UTC - in response to Message 1481234.  
Last modified: 24 Feb 2014, 15:43:18 UTC

With my relatively new (and self-proclaimed) role as Cassandra of these boards I'm going to go ahead and guess that Boinc is just "seeing" the flops wrong (it's not like it runs any benchmarks to get that number, just blurts out whatever it's told AFAIK).

The Gflops for the 448 are spot on but the Gflops for the 750ti should be around 1300 (according to Wiki anyway, which can be a bit dodgy of course).

I do hope I'm wrong though.

I was thinking the same. The speed (peak flops) of an NVidia card depends - critically - on the number of shaders or cuda cores (pick your own terminology).

For some absurd reason, NVidia makes it stupidly difficult to detect that number programmatically. You can ask the API for the number of 'Streaming Multiprocessors' on the card, and it'll tell you. But ask it how many shaders per SM? Nada.

Instead, you have to know (in advance) that it's some function of the Compute Capability of the card. The original NV cards were 8 shaders per SM: Fermis jumped it to 32, some Keplers went up to 48 - but now it seems to have become variable. Accordingly to the (possibly the same) Wiki article, these little Maxwells seem to be back down to 16 S per SM - in which case, BOINC might over-estimate them by 2x or 3x.

Unless our resident NV guru can tell us that NVidia have introduced a new API call at last?

Edit - COPROC_NVIDIA::set_peak_flops() in /lib/coproc.cpp seems to be stuck in a timewarp. Don't believe anything you read there.
ID: 1481240 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1481243 - Posted: 24 Feb 2014, 15:47:03 UTC
Last modified: 24 Feb 2014, 16:25:32 UTC

That´s exactly why i ask him to run the benchmarks so we could see their performance in the "real SETI world AP/MB".
ID: 1481243 · Report as offensive
Gone

Send message
Joined: 31 May 99
Posts: 150
Credit: 125,779,206
RAC: 0
United Kingdom
Message 1481255 - Posted: 24 Feb 2014, 16:23:44 UTC

I will build a host with 2 x 750s in , one on PCIe x16 and one via x1 and then run benchmarks.
I am seeing varying run times at the moment running 2 tasks per card on both 560 Ti and 750 Ti
For me the real beauty of these cards is that they need no extra power cables.

Reg
ID: 1481255 · Report as offensive
draco
Volunteer tester

Send message
Joined: 6 Dec 05
Posts: 119
Credit: 3,327,457
RAC: 0
Latvia
Message 1481269 - Posted: 24 Feb 2014, 17:04:26 UTC - in response to Message 1481255.  

I will build a host with 2 x 750s in , one on PCIe x16 and one via x1 and then run benchmarks.
I am seeing varying run times at the moment running 2 tasks per card on both 560 Ti and 750 Ti
For me the real beauty of these cards is that they need no extra power cables.

Reg


sorry, dont understand.
cards TDP is about 60w.
PCIe power possibilities is around 75W.
if you use one card, then all is ok.
if you want use two cards - there is no possibility to use it without extra power cables?
ID: 1481269 · Report as offensive
juan BFP Crowdfunding Project Donor*Special Project $75 donorSpecial Project $250 donor
Volunteer tester
Avatar

Send message
Joined: 16 Mar 07
Posts: 9786
Credit: 572,710,851
RAC: 3,799
Panama
Message 1481286 - Posted: 24 Feb 2014, 17:48:05 UTC
Last modified: 24 Feb 2014, 17:50:57 UTC

Look one of his previsous post, he is ussing this adaptor who drives the power for the second GPU from the molex connector not the PCI-e. The question is what is the impact on crunching the use of the slow PCI x1 slot will do (speed/time to crunch a WU).


ID: 1481286 · Report as offensive
draco
Volunteer tester

Send message
Joined: 6 Dec 05
Posts: 119
Credit: 3,327,457
RAC: 0
Latvia
Message 1481312 - Posted: 24 Feb 2014, 19:04:30 UTC - in response to Message 1481286.  

Look one of his previsous post, he is ussing this adaptor who drives the power for the second GPU from the molex connector not the PCI-e.


molex != extra power cable? :-O
ID: 1481312 · Report as offensive
Previous · 1 · 2 · 3 · 4 · Next

Message boards : Number crunching : GPU Wars 2014: Postponed to 2015?


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.