GPU + Linux + Seti = Work???

Message boards : Number crunching : GPU + Linux + Seti = Work???
Message board moderation

To post messages, you must log in.

1 · 2 · 3 · Next

AuthorMessage
Profile ausymark

Send message
Joined: 9 Aug 99
Posts: 95
Credit: 10,175,128
RAC: 0
Australia
Message 1063848 - Posted: 6 Jan 2011, 3:07:09 UTC

Hi Team

I have recently setup my Mandriva Linux box with Boinc/Seti 6.10.58.

I am also running the nvidia173-CUDA core and CUDA-OPENCL libraries. (Have an nVidia 9800 GT)

Boinc is requesting GPU work units but isnt receiving any (has been asking for 7 days now).

My SETI account does not list any CUDA application to distribute work units for. I have only the following applications set:

SETI@home Enhanced: yes
Astropulse v5: yes
Astropulse v5.05: yes

I dont have any other options.

Suggestions please

Cheers

Mark
ID: 1063848 · Report as offensive
Profile arkayn
Volunteer tester
Avatar

Send message
Joined: 14 May 99
Posts: 4438
Credit: 55,006,323
RAC: 0
United States
Message 1063857 - Posted: 6 Jan 2011, 3:58:31 UTC - in response to Message 1063848.  

You will have to manually install the Linux CUDA app as one has not been made by the SETI admins.
http://calbe.dw70.de/seti.html

Only caveat is you will have to manually update the app if a newer one is created.

ID: 1063857 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 1063860 - Posted: 6 Jan 2011, 4:05:14 UTC - in response to Message 1063848.  
Last modified: 6 Jan 2011, 4:13:13 UTC

SETI@home have no stock Linux CUDA application (http://setiathome.berkeley.edu/apps.php).
You would need to install the application manually.
I have no experiences with Linux, you should find infos here: http://lunatics.kwsn.net/3-linux/index.0.html.

There are also optimized applications for CPU available here: http://lunatics.kwsn.net/index.php?module=Downloads;catd=1.
Strange, currently only for AstroPulse and not for SETI@home Enhanced?

Enhanced (also named MultiBeam) from here (e.g. 32bit): http://calbe.dw70.de/linux32.html?

Maybe someone other with Linux experiences could have an answer..
ID: 1063860 · Report as offensive
elgordodude
Volunteer tester

Send message
Joined: 2 Jun 10
Posts: 13
Credit: 7,550,388
RAC: 0
United States
Message 1063875 - Posted: 6 Jan 2011, 4:34:25 UTC - in response to Message 1063860.  
Last modified: 6 Jan 2011, 4:37:13 UTC

Just went through this installing an 8800 GT. The 173 won't work, you want the 260, also from what I found you need to be running in 64 bit. Most, if not all of the 32-bit Multibeam (MB) apps were removed due to massive failures.

If you can run 64bit, download the current Nvidia driver, remove all active Nvidia drivers, blacklist nouveau, install the Nvidia driver, reboot, and see what happens.

If a 260 driver installation goes well optimize from http://calbe.dw70.de/linux64.html and good luck! It took most of my holiday to figure that out.
ID: 1063875 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 20289
Credit: 7,508,002
RAC: 20
United Kingdom
Message 1063986 - Posted: 6 Jan 2011, 15:56:02 UTC - in response to Message 1063875.  

If you can run 64bit, download the current Nvidia driver, remove all active Nvidia drivers, blacklist nouveau, install the Nvidia driver, reboot, and see what happens.

If a 260 driver installation goes well optimize from http://calbe.dw70.de/linux64.html and good luck! It took most of my holiday to figure that out.


You should be using 64-bit in any case. Pointless to only use 1/2 - 1/4 of your CPU hardware with 32-bit only!

Select the 'proprietary' nVidia driver from the Mandriva Control Centre:

"Configure your computer" -> Hardware -> Set up the graphical server

Then select the nVidia driver "GeForce 6100 and later"

... And away you go!


Depending on how much graphics ram you card has, for s@h you may need to reduce your number of virtual desktops to just one if you only have 512MByte ram, rather than the default four desktops.

Check out the Lunatics optimised apps!


Happy fast crunchin',
Martin

See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 1063986 · Report as offensive
Profile ausymark

Send message
Joined: 9 Aug 99
Posts: 95
Credit: 10,175,128
RAC: 0
Australia
Message 1064399 - Posted: 7 Jan 2011, 21:04:54 UTC - in response to Message 1063986.  

Hi Team

I have installed the latest Mandriva 2010.2 x64

After some fussing about I have my Nvidia 9800 GT card as being recognised as a CUDA capable device - yaaaay

However what is now happening is that seti@home is trying to download the:

[error] No URL for file transfer of libcudart.so.3
[error] No URL for file transfer of libcufft.so.3

files.

I am sure there is a fix for this which I am currently investigating.

So thanks for all your help so far, and hopefully I can get this fixed so that I can get on with crunching :)

Cheers

Mark :)
ID: 1064399 · Report as offensive
geyser

Send message
Joined: 7 Oct 04
Posts: 8
Credit: 64,645,821
RAC: 201
United States
Message 1067269 - Posted: 16 Jan 2011, 8:14:25 UTC - in response to Message 1064399.  

I am working through the same problem as you.

If you run ldd on the CUDA executable and there are no errors
then you can modify your app_info.xml and remove the sections
about the libraries and the errors:

[error] No URL for file transfer of libcudart.so.3
[error] No URL for file transfer of libcufft.so.3

should go away.

In my case the configuration in the app_innfo.xml file,
after removing those unneeded library sections, was in
good enough state to pull down work units but the processed
too fast so now I am trying to find a way to debug the CUDA
app to see what I need to do next.
ID: 1067269 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 20289
Credit: 7,508,002
RAC: 20
United Kingdom
Message 1067395 - Posted: 16 Jan 2011, 19:08:55 UTC - in response to Message 1067269.  
Last modified: 16 Jan 2011, 19:09:45 UTC

I am working through the same problem as you.

If you run ldd on the CUDA executable and there are no errors
then you can modify your app_info.xml and remove the sections
about the libraries and the errors:

[error] No URL for file transfer of libcudart.so.3
[error] No URL for file transfer of libcufft.so.3

should go away.

Try putting in soft links "ln -s" for those in the boinc directory?

You'll likely need something like:

(Go into a command line terminal)
cd to_wherever_it_is_you_have_boinc
ln -s /usr/lib64/libcudart.so.3
ln -s /usr/lib64/libcufft.so.3

No "64" if you're only 32-bit. But then again, why would you be 32-bit?!


Happy fast crunchin',
Martin
See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 1067395 · Report as offensive
Brkovip
Avatar

Send message
Joined: 18 May 99
Posts: 274
Credit: 144,414,367
RAC: 0
United States
Message 1067467 - Posted: 16 Jan 2011, 23:14:38 UTC

I just wish they would have something 64 bit that would work with the 400 line of Nvidia cards in Linux. The 64 bit 6.08 version will just toss errors all day. :(
ID: 1067467 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 20289
Credit: 7,508,002
RAC: 20
United Kingdom
Message 1067485 - Posted: 17 Jan 2011, 0:04:08 UTC - in response to Message 1067467.  

I just wish they would have something 64 bit that would work with the 400 line of Nvidia cards in Linux. The 64 bit 6.08 version will just toss errors all day. :(

Mmmm... I need to get my nVidia 4xx running from since the Christmas break... And pick up some threads... Time flies too quick!

Happy crunchin',
Martin
See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 1067485 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 1067674 - Posted: 17 Jan 2011, 16:22:08 UTC - in response to Message 1067485.  

I just wish they would have something 64 bit that would work with the 400 line of Nvidia cards in Linux. The 64 bit 6.08 version will just toss errors all day. :(

Mmmm... I need to get my nVidia 4xx running from since the Christmas break... And pick up some threads... Time flies too quick!

Happy crunchin',
Martin

At MW@h it wasn't a big deal to unlock the CUDA app for Fermi GPUs.
Get in contact with Crunch3r, that he do it also with his S@h 6.08 CUDA app.

Please let me know if it work/ed.

ID: 1067674 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1067815 - Posted: 17 Jan 2011, 23:43:14 UTC - in response to Message 1067674.  

I just wish they would have something 64 bit that would work with the 400 line of Nvidia cards in Linux. The 64 bit 6.08 version will just toss errors all day. :(

Mmmm... I need to get my nVidia 4xx running from since the Christmas break... And pick up some threads... Time flies too quick!

Happy crunchin',
Martin

At MW@h it wasn't a big deal to unlock the CUDA app for Fermi GPUs.
Get in contact with Crunch3r, that he do it also with his S@h 6.08 CUDA app.

Please let me know if it work/ed.

The Fermi incompatibility found in SETI Windows CUDA applications v6.08 and v6.09 was nothing to do with anything called "locking".

It was, instead, the non-declaration of certain variables as "volatile" - a 'common optimisation' practiced by NVidia themselves in those early pre-Fermi CUDA builds.

See paragraph 1.2.2 of the Fermiâ„¢ Compatibility Guide v1.3 for details.
ID: 1067815 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 1067838 - Posted: 18 Jan 2011, 0:14:01 UTC - in response to Message 1067815.  
Last modified: 18 Jan 2011, 0:27:57 UTC

At MW@h it wasn't a big deal to unlock the CUDA app for Fermi GPUs.
Get in contact with Crunch3r, that he do it also with his S@h 6.08 CUDA app.

Please let me know if it work/ed.

The Fermi incompatibility found in SETI Windows CUDA applications v6.08 and v6.09 was nothing to do with anything called "locking".

It was, instead, the non-declaration of certain variables as "volatile" - a 'common optimisation' practiced by NVidia themselves in those early pre-Fermi CUDA builds.

See paragraph 1.2.2 of the Fermiâ„¢ Compatibility Guide v1.3 for details.


I was talking about the MW@h Windows CUDA app.
AFAIK & IIRC, he changed only one/two entries in the code of the stock MW@h_0.24_cuda23 app and the app worked on Fermi chips.
The..
major = x
minor = x
..part - or something similar.
I'm not a coder, so maybe I remember/declare wrong.

I don't know if this will be so easy also for his Linux S@h CUDA app.

If someone is interested to get a Linux S@h Fermi CUDA app, s/he should get in contact with Crunch3r.
Here is his S@h account: http://setiathome.berkeley.edu/show_user.php?userid=11044.
But I guess his website/forum (http://calbe.dw70.de) is the better way..

EDIT: Or maybe over the Lunatics site (http://lunatics.kwsn.net/3-linux/index.0.html).
ID: 1067838 · Report as offensive
Claggy
Volunteer tester

Send message
Joined: 5 Jul 99
Posts: 4654
Credit: 47,537,079
RAC: 4
United Kingdom
Message 1067847 - Posted: 18 Jan 2011, 0:37:09 UTC - in response to Message 1067838.  

At MW@h it wasn't a big deal to unlock the CUDA app for Fermi GPUs.
Get in contact with Crunch3r, that he do it also with his S@h 6.08 CUDA app.

Please let me know if it work/ed.

The Fermi incompatibility found in SETI Windows CUDA applications v6.08 and v6.09 was nothing to do with anything called "locking".

It was, instead, the non-declaration of certain variables as "volatile" - a 'common optimisation' practiced by NVidia themselves in those early pre-Fermi CUDA builds.

See paragraph 1.2.2 of the Fermiâ„¢ Compatibility Guide v1.3 for details.


I was talking about the MW@h Windows CUDA app.
AFAIK & IIRC, he changed only one/two entries in the code of the stock MW@h_0.24_cuda23 app and the app worked on Fermi chips.
The..
major = x
minor = x
..part - or something similar.
I'm not a coder, so maybe I remember/declare wrong.

I don't know if this will be so easy also for his Linux S@h CUDA app.

If someone is interested to get a Linux S@h Fermi CUDA app, s/he should get in contact with Crunch3r.
Here is his S@h account: http://setiathome.berkeley.edu/show_user.php?userid=11044.
But I guess his website/forum (http://calbe.dw70.de) is the better way..

EDIT: Or maybe over the Lunatics site (http://lunatics.kwsn.net/3-linux/index.0.html).


I think you'll find Crunch3r has lost interest in Seti,

Claggy
ID: 1067847 · Report as offensive
Profile Sutaru Tsureku
Volunteer tester

Send message
Joined: 6 Apr 07
Posts: 7105
Credit: 147,663,825
RAC: 5
Germany
Message 1067853 - Posted: 18 Jan 2011, 0:47:01 UTC - in response to Message 1067847.  

I think you'll find Crunch3r has lost interest in Seti,

Claggy


Maybe if a few people ask, he would do it..
..but I'm not up-to-date.

ID: 1067853 · Report as offensive
Profile aaronh
Volunteer tester
Avatar

Send message
Joined: 27 Oct 99
Posts: 169
Credit: 1,442,686
RAC: 0
United States
Message 1068387 - Posted: 19 Jan 2011, 23:43:50 UTC - in response to Message 1067815.  

The Fermi incompatibility found in SETI Windows CUDA applications v6.08 and v6.09 was nothing to do with anything called "locking".

It was, instead, the non-declaration of certain variables as "volatile" - a 'common optimisation' practiced by NVidia themselves in those early pre-Fermi CUDA builds.

See paragraph 1.2.2 of the Fermiâ„¢ Compatibility Guide v1.3 for details.


Wow, good one. Throwing a few volatile's in the SVN source led to what seem to be *valid* results (GTX460, Ubuntu 64-bit)! Sure, it's not Fermi *optimised* but first steps are first steps...
ID: 1068387 · Report as offensive
Richard Haselgrove Project Donor
Volunteer tester

Send message
Joined: 4 Jul 99
Posts: 14650
Credit: 200,643,578
RAC: 874
United Kingdom
Message 1068390 - Posted: 19 Jan 2011, 23:55:22 UTC - in response to Message 1068387.  

The Fermi incompatibility found in SETI Windows CUDA applications v6.08 and v6.09 was nothing to do with anything called "locking".

It was, instead, the non-declaration of certain variables as "volatile" - a 'common optimisation' practiced by NVidia themselves in those early pre-Fermi CUDA builds.

See paragraph 1.2.2 of the Fermiâ„¢ Compatibility Guide v1.3 for details.

Wow, good one. Throwing a few volatile's in the SVN source led to what seem to be *valid* results (GTX460, Ubuntu 64-bit)! Sure, it's not Fermi *optimised* but first steps are first steps...

Congratulations - well done. Have a word with Urs Echternacht, either here or at lunatics - he can help you with test tools to confirm those validations properly. If you can put some quality assurance behind the work you've done so far, the Linux community will be grateful.

As you say, get it working - reliably and with confidence - first: optimisation is the icing on the cake.
ID: 1068390 · Report as offensive
Profile aaronh
Volunteer tester
Avatar

Send message
Joined: 27 Oct 99
Posts: 169
Credit: 1,442,686
RAC: 0
United States
Message 1068398 - Posted: 20 Jan 2011, 0:46:08 UTC - in response to Message 1068390.  

or at lunatics - he can help you with test tools to confirm those validations properly. If you can put some quality assurance behind the work you've done so far, the Linux community will be grateful.

As you say, get it working - reliably and with confidence - first: optimisation is the icing on the cake.


(BTW, anyone know why lunatics doesn't allow new accounts to be created?)
ID: 1068398 · Report as offensive
Profile ML1
Volunteer moderator
Volunteer tester

Send message
Joined: 25 Nov 01
Posts: 20289
Credit: 7,508,002
RAC: 20
United Kingdom
Message 1068401 - Posted: 20 Jan 2011, 0:53:55 UTC - in response to Message 1068398.  

(BTW, anyone know why lunatics doesn't allow new accounts to be created?)

You need to contact a lunatic offlist or via PM to be invited/added.

Happy fast crunchin',
Martin

See new freedom: Mageia Linux
Take a look for yourself: Linux Format
The Future is what We all make IT (GPLv3)
ID: 1068401 · Report as offensive
Profile aaronh
Volunteer tester
Avatar

Send message
Joined: 27 Oct 99
Posts: 169
Credit: 1,442,686
RAC: 0
United States
Message 1071112 - Posted: 27 Jan 2011, 11:05:12 UTC - in response to Message 1068390.  

Throwing a few volatile's in the SVN source led to what seem to be *valid* results (GTX460, Ubuntu 64-bit)! Sure, it's not Fermi *optimised* but first steps are first steps...

As you say, get it working - reliably and with confidence - first: optimisation is the icing on the cake.

I've validated it against quite a few results locally (computed on CPU), and have confidence. I've adjusted several things, and have gotten it to the point where it runs about 12% faster than the stock windows client (which itself seems to run a little faster than the Lunatics client? How strange) completing a WU in an average of 10:42 on my card, with windows around 12:00.

It took me a while to whittle the video ram usage down far enough to find out if it would run two WUs on 768MB of video ram. X seems to occupy a lot more than it actually needs. I'm guessing it's sort of like the way linux memory tends to fill with disk buffers/cache. It is faster to run two at once, but I think it might be a greater difference on higher end cards. The client simply has too high utilisation of my card. Perhaps one with more multiprocessors (the 460 only has 7). nvidia-smi reports 95% gpu utilisation, 60% memory utilisation. The latter is not memory in use, but rather memory transfers (cudaMemcpy)

Average run times (not including early exit (-9) WUs:
642 seconds x1 WU
1096 seconds x2 WU = 548 seconds per WU

Including (-9) WUs
484 seconds x1 WU

I've started running it locally with an app_info.xml, and thus far have 55 consecutive valid tasks, and about 90 awaiting validation. No Invalid or Errors yet, although I'm expecting one (copied the wrong binary. Caught it after the first task errored out)

Unfortunately, I've tried making a 32-bit linux version, but it has too many errors. Still tracking that down. 64-bit only, for now.

Time to clean up the code, and make it presentable.

Congratulations - well done. Have a word with Urs Echternacht, either here or at lunatics - he can help you with test tools to confirm those validations properly. If you can put some quality assurance behind the work you've done so far, the Linux community will be grateful.

I've still yet to contact anyone... argh.
ID: 1071112 · Report as offensive
1 · 2 · 3 · Next

Message boards : Number crunching : GPU + Linux + Seti = Work???


 
©2024 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.