Post your BOINC Startup 'CUDA' Info

Message boards : Number crunching : Post your BOINC Startup 'CUDA' Info
Message board moderation

To post messages, you must log in.

Previous · 1 . . . 6 · 7 · 8 · 9 · 10 · 11 · 12 · Next

AuthorMessage
Profile Pappa
Volunteer tester
Avatar

Send message
Joined: 9 Jan 00
Posts: 2562
Credit: 12,301,681
RAC: 0
United States
Message 948306 - Posted: 19 Nov 2009, 18:26:43 UTC - in response to Message 948299.  
Last modified: 19 Nov 2009, 18:27:19 UTC

19.11.2009 03:57:09 NVIDIA GPU 0: GeForce GTS 250 (driver version 19555, CUDA version 3000, compute capability 1.1, 512MB, 470 GFLOPS peak)


These numbers are strange...--a Are there new versions released for nVidia driver and CUDA?


I guess he use a BETA/prerelease nVIDIA_driver and CUDA_version.

If this would be the final release of 195.x and 3.0 , I guess we would hear it early enough.

But from that I heared, the 'old' GPU chips wouldn't get a (big) speed up with CUDA_V3.0 (like it was between 2.2 and 2.3).
CUDA_V3.0 would be well for the new 'Fermi' (GTX3xx, maybe also for the small brothers of) series GPUs.



Thanks for your explanation...:)


Actually it is a change in the Boinc 6.10.xx series that has changed how the benchmark is preformed. The 6.6.xx Boinc version would have shown it as something like this.


19.11.2009 03:57:09 CUDA device: GeForce GTS 250 (driver version 19555, compute capability 1.1, 512MB, est. 84GFLOPS)


Regards
Please consider a Donation to the Seti Project.

ID: 948306 · Report as offensive
Crun-chi
Volunteer tester
Avatar

Send message
Joined: 3 Apr 99
Posts: 174
Credit: 3,037,232
RAC: 0
Croatia
Message 948339 - Posted: 19 Nov 2009, 21:45:04 UTC - in response to Message 948251.  

19.11.2009 03:57:09 NVIDIA GPU 0: GeForce GTS 250 (driver version 19555, CUDA version 3000, compute capability 1.1, 512MB, 470 GFLOPS peak)


These numbers are strange...--a Are there new versions released for nVidia driver and CUDA?


I guess he use a BETA/prerelease nVIDIA_driver and CUDA_version.

If this would be the final release of 195.x and 3.0 , I guess we would hear it early enough.

But from that I heared, the 'old' GPU chips wouldn't get a (big) speed up with CUDA_V3.0 (like it was between 2.2 and 2.3).
CUDA_V3.0 would be well for the new 'Fermi' (GTX3xx, maybe also for the small brothers of) series GPUs.


As always: why those guys who has CUDA 3.0 doesnot distribute dlls / suppose there is only 2 DLL like "old CUDA 2.3.
But no; they will always tell they cannot give anyone because it is prebeta ( or something like that)
It look like that two DLL.s are more secret than whole Pentagon :)

I am cruncher :)
I LOVE SETI BOINC :)
ID: 948339 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 948349 - Posted: 19 Nov 2009, 22:20:16 UTC - in response to Message 948339.  

The 3.0 DLLs have been tested at Lunatics, and showed at best a 2% performance increase over 2.3 - nothing like the 30%+ achieved by 2.3 over 2.2, or 2.2 over 2.1. So nothing worth losing your licence over.

And losing your licence is important: if they deliberately flouted the developer NDAs (Non Disclosure Agreements), they could lose their access to pre-release development tools, or closed technical message boards where problems can be reported and overcome.
ID: 948349 · Report as offensive
Profile j mercer
Avatar

Send message
Joined: 3 Jun 99
Posts: 2422
Credit: 12,323,733
RAC: 1
United States
Message 948351 - Posted: 19 Nov 2009, 22:29:36 UTC

Apply for NVIDIA’s GPU Computing registered developer program like I replied to you in your asking for copies. Here's links:

http://www.nvidia.com/object/cuda_get.html
http://nvdeveloper.nvidia.com/content/GPUComputingDeveloperApplication/frmDeveloperRegistration.asp


...
ID: 948351 · Report as offensive
Crun-chi
Volunteer tester
Avatar

Send message
Joined: 3 Apr 99
Posts: 174
Credit: 3,037,232
RAC: 0
Croatia
Message 948352 - Posted: 19 Nov 2009, 22:35:30 UTC - in response to Message 948349.  

The 3.0 DLLs have been tested at Lunatics, and showed at best a 2% performance increase over 2.3 - nothing like the 30%+ achieved by 2.3 over 2.2, or 2.2 over 2.1. So nothing worth losing your licence over.

And losing your licence is important: if they deliberately flouted the developer NDAs (Non Disclosure Agreements), they could lose their access to pre-release development tools, or closed technical message boards where problems can be reported and overcome.


Many of them are on that forum only for CUDA toolkit, and even loosing license they will not lose anything important: because that persons are not developer or something like that: they are people like ordinary one: just have lucky and get access to nvidia forum :)
I am cruncher :)
I LOVE SETI BOINC :)
ID: 948352 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 948354 - Posted: 19 Nov 2009, 22:43:47 UTC - in response to Message 948352.  

Many of them are on that forum only for CUDA toolkit, and even loosing license they will not lose anything important: because that persons are not developer or something like that: they are people like ordinary one: just have lucky and get access to nvidia forum :)

If you can write code like Jason or Raistmer then get stuck in - we still need to solve the VLAR problem.

Otherwise, there are clearly two grades of 'ordinary' people.
ID: 948354 · Report as offensive
Crun-chi
Volunteer tester
Avatar

Send message
Joined: 3 Apr 99
Posts: 174
Credit: 3,037,232
RAC: 0
Croatia
Message 948355 - Posted: 19 Nov 2009, 22:44:54 UTC - in response to Message 948349.  

The 3.0 DLLs have been tested at Lunatics, and showed at best a 2% performance increase over 2.3 - nothing like the 30%+ achieved by 2.3 over 2.2, or 2.2 over 2.1. So nothing worth losing your licence over.

And losing your licence is important:....


so when they tested 3.0 DLLS on Lunatics ,that is ok ,and if they put on Internet that is not ok :)
How interesting :)
I am cruncher :)
I LOVE SETI BOINC :)
ID: 948355 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 948357 - Posted: 19 Nov 2009, 22:58:29 UTC - in response to Message 948355.  

so when they tested 3.0 DLLS on Lunatics ,that is ok ,and if they put on Internet that is not ok :)
How interesting :)

Testing is an important part of development (as you must know, from your 'Volunteer tester' tag).

And believe me, if those tests had revealed another 30% increase, they would have been falling over themselves to develop a safe, reliable, tested solution to use the 3.0 DLLs for SETI (ensuring that the tasks validate, and hence generate credit, when compared with older versions) - while still keeping their licences to develop with 3.1, 3.2,4.0, ...
ID: 948357 · Report as offensive
Profile hiamps
Volunteer tester
Avatar

Send message
Joined: 23 May 99
Posts: 4292
Credit: 72,971,319
RAC: 0
United States
Message 948688 - Posted: 21 Nov 2009, 1:01:09 UTC

"CUDA device: GeForce GTX 275 (driver version 18585, compute capability 1.3, 896MB, est. 123GFLOPS)

CUDA device: GeForce GTX 285 (driver version 18171, CUDA version 1.3, 1024MB, est. 127GFLOPS)"

Was wondering if this is a good comparison? Seems for the money the GTX 275 is almost as fast as the GTX 285. Is there a big difference between Gddr3 and ddr3 memory? They both seem to have 240 processors. Anyone have both?
Official Abuser of Boinc Buttons...
And no good credit hound!
ID: 948688 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 948705 - Posted: 21 Nov 2009, 2:11:54 UTC
Last modified: 21 Nov 2009, 2:19:56 UTC


[http://www.nvidia.com/object/cuda_learn_products.html]
[http://www.nvidia.com/object/product_geforce_gtx_275_us.html]
[http://www.nvidia.com/object/product_geforce_gtx_285_us.html]

GTX275 - 633/1404/1134 - (123 GFLOPS ?)
GTX285 - 648/1476/1242 - (127 GFLOPS ?)

[GPU/shader/RAM]

The upper mentioned GPUs are at stock speed?

My OCed GTX260-216 have:
EVGA GTX260 Core216 SSC - 675/1458/1152 -> 112 GFLOPS
GIGABYTE GTX260(-216) SOC - 680/1500/1250 -> 117 GFLOPS
The GIGABYTE have higher speeds as a GTX285!
But O.K., 30 (240) or 27 (216) CUDA (shader) cores.

Because of CUDA..
more shader cores
more shader speed
more RAM speed
more GPU core speed (maybe)
-> more performance

ID: 948705 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 948710 - Posted: 21 Nov 2009, 2:36:38 UTC
Last modified: 21 Nov 2009, 2:37:49 UTC


BTW.

All GTX2xx GPUs have GDDR3 RAM.

The GTX285 have 512-bit and the GTX275 / GTX260-216 have 448-bit 'Memory Interface Width'.

But, if this higher 'RAM width' give so much additional performance?
ID: 948710 · Report as offensive
Profile [AF>france>pas-de-calais]symaski62
Volunteer tester

Send message
Joined: 12 Aug 05
Posts: 258
Credit: 100,548
RAC: 0
France
Message 950301 - Posted: 27 Nov 2009, 22:56:47 UTC

31/10/2009 09:07:21 NVIDIA GPU 0: GeForce 8400 GS (driver version 19062, CUDA version 2030, compute capability 1.1, 256MB, 43 GFLOPS peak)

to

27/11/2009 23:45:24 NVIDIA GPU 0: GeForce 8400 GS (driver version 19562, CUDA version 3000, compute capability 1.1, 256MB, 43 GFLOPS peak)


@+
ID: 950301 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 950396 - Posted: 28 Nov 2009, 5:53:41 UTC - in response to Message 948183.  
Last modified: 28 Nov 2009, 5:54:31 UTC

BOINC V6.6.x

Manufacturer OCed

EVGA GTX260 Core216 SSC - 675/1458/1152 -> 112 GFLOPS
GIGABYTE GTX260(-216) SOC - 680/1500/1250 -> 117 GFLOPS

[GPU/shader/RAM]


A stock GTX260-216 -> 576/1242/999
[...]

After update to BOINC V6.10.18:

The EVGA brothers :
NVIDIA GPU 0: GeForce GTX 260 (driver version 19038, CUDA version 2030, compute capability 1.3, 896MB, 630 GFLOPS peak)
NVIDIA GPU 1: GeForce GTX 260 (driver version 19038, CUDA version 2030, compute capability 1.3, 896MB, 630 GFLOPS peak)
NVIDIA GPU 2: GeForce GTX 260 (driver version 19038, CUDA version 2030, compute capability 1.3, 896MB, 630 GFLOPS peak)
NVIDIA GPU 3: GeForce GTX 260 (driver version 19038, CUDA version 2030, compute capability 1.3, 896MB, 630 GFLOPS peak)


The GIGABYTE:
NVIDIA GPU 0: GeForce GTX 260 (driver version 19038, CUDA version 2030, compute capability 1.3, 896MB, 653 GFLOPS peak)

ID: 950396 · Report as offensive
Stef

Send message
Joined: 21 Dec 08
Posts: 1
Credit: 176,496
RAC: 0
Switzerland
Message 950418 - Posted: 28 Nov 2009, 7:35:09 UTC
Last modified: 28 Nov 2009, 7:35:36 UTC

NVIDIA GPU 0: GeForce 8800 GTS 512 (driver version 19562, CUDA version 3000, compute capability 1.1, 512MB, 442 GFLOPS peak)
ID: 950418 · Report as offensive
Luke
Volunteer developer
Avatar

Send message
Joined: 31 Dec 06
Posts: 2546
Credit: 817,560
RAC: 0
New Zealand
Message 950431 - Posted: 28 Nov 2009, 9:07:53 UTC
Last modified: 28 Nov 2009, 9:08:15 UTC

Not that I use my Graphics card for crunching, but I'll post here anyway.
28/11/2009 3:17:45 p.m.		NVIDIA GPU 0: GeForce 8600M GT (driver version 18681, CUDA version 2020, compute capability 1.1, 256MB, 61 GFLOPS peak)

Hopefully the GTX260 C216 Superclock in my gaming build will do a bit better!
- Luke.
ID: 950431 · Report as offensive
Profile Mahoujin Tsukai
Volunteer tester
Avatar

Send message
Joined: 21 Jul 07
Posts: 147
Credit: 2,204,402
RAC: 0
Singapore
Message 950475 - Posted: 28 Nov 2009, 13:08:23 UTC

I'm curious,
How many GFLOPS can an Intel Core 2 Quad Q6600 do at stock speed?
ID: 950475 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 950525 - Posted: 28 Nov 2009, 17:20:42 UTC - in response to Message 950475.  

I'm curious,
How many GFLOPS can an Intel Core 2 Quad Q6600 do at stock speed?

About 2.4 per core, or 9.6 in total. Floating Point speed is very close to stock clock for the Core 2s.
ID: 950525 · Report as offensive
Josef W. Segur
Volunteer developer
Volunteer tester

Send message
Joined: 30 Oct 99
Posts: 4504
Credit: 1,414,761
RAC: 0
United States
Message 950737 - Posted: 29 Nov 2009, 6:33:18 UTC - in response to Message 950525.  

I'm curious,
How many GFLOPS can an Intel Core 2 Quad Q6600 do at stock speed?

About 2.4 per core, or 9.6 in total. Floating Point speed is very close to stock clock for the Core 2s.

For the Core 2 architecture, each core has multiple execution units, one of which can do 4 SIMD floating point operations per clock, another can do one floating point operation per clock. So 2.4 * 5 = 12 peak GFLOPS per core in terms comparable to the way the GPUs are rated.
                                                                 Joe
ID: 950737 · Report as offensive
angler

Send message
Joined: 19 Oct 00
Posts: 33
Credit: 880,214
RAC: 0
United States
Message 950795 - Posted: 29 Nov 2009, 15:00:28 UTC

NVIDIA GPU 0: GeForce 9600 GT (driver version 19107, CUDA version 2030, compute capability 1.1, 512MB, 208 GFLOPS peak
ID: 950795 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 950804 - Posted: 29 Nov 2009, 16:06:06 UTC - in response to Message 950737.  

For the Core 2 architecture, each core has multiple execution units, one of which can do 4 SIMD floating point operations per clock, another can do one floating point operation per clock. So 2.4 * 5 = 12 peak GFLOPS per core in terms comparable to the way the GPUs are rated.
                                                                 Joe

Looking for a direct comparison between the old and new BOINC estimates, I found Martin's GTS 250:

CUDA device: GeForce GTS 250 (driver version 0, CUDA version 1.1, 512MB, est. 84GFLOPS)
NVIDIA GPU 0: GeForce GTS 250 (driver version unknown, CUDA version 2030, compute capability 1.1, 512MB, 470 GFLOPS peak)

I reckon my 'est. 2.4GFLOPS' (benchmark per core) is pretty close to your '12 GFLOPS peak': it's just they way we count 'em.
ID: 950804 · Report as offensive
Previous · 1 . . . 6 · 7 · 8 · 9 · 10 · 11 · 12 · Next

Message boards : Number crunching : Post your BOINC Startup 'CUDA' Info


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.