Speaking of crunching on your GPU....

Message boards : Number crunching : Speaking of crunching on your GPU....
Message board moderation

To post messages, you must log in.

1 · 2 · Next

AuthorMessage
FMatson
Avatar

Send message
Joined: 14 Feb 01
Posts: 35
Credit: 680,946
RAC: 0
United States
Message 174546 - Posted: 6 Oct 2005, 0:45:51 UTC
Last modified: 6 Oct 2005, 0:54:16 UTC

---Cut & pasted from an article on Tom's Hardware regarding ATI's upcoming line of graphic cards:

"...The real innovation however is just an idea at this time, but promises to have a major impact on the industry, if brought to reality. Heye mentioned that ATI plans to open the hardware architecture of the X1000 to allow third party developers to write non-graphics-related applications to run on the graphics processor. The company calls this feature "dynamic load balancing."

Compared to a Pentium 4 CPU, which delivers a floating point performance of 12 GFLOPs and a bandwidth of just under 6 GByte per second, a graphics processor is a calculation monster: According to ATI, an X1800 XT chip reaches 83 GFlops and 42 GByte per second. The full performance of a graphics may not always be needed - especially in dual-graphics environments - and users will be able to relocate processing power to other applications. According to ATI, these applications could include scientific applications such as fluid dynamics, but also entertainment-related functions such as physics or 3D audio processing. Similar features have been demonstrated by academic projects in the past on ATI and Nvidia platforms, but dynamic load balancing as described by ATI officials promises a whole new use of graphics processors.

The company expects GPU specific third-party API's to become common within a few years - with one of the most promising being physics processing: ATI believes that graphics chips provide enough power to cover the features that are currently promoted by Ageia. If ATI's vision comes true, Ageia's idea for a physics board for every gaming PC may become outdated before the startup's technology has reached a critical mass of customers..."

The original article can be found at:
http://www.tomshardware.com/hardnews/20051005_090950.html
ID: 174546 · Report as offensive
Hammer

Send message
Joined: 13 Dec 02
Posts: 74
Credit: 1,773,558
RAC: 0
Croatia
Message 174622 - Posted: 6 Oct 2005, 6:03:36 UTC

I have read some info before that Stanford is working on new Folding@home client that folds with GPU-s. They said that client is already working well and that they are still optimizing it. Basically you only need any DirectX 9 card to be able to fold with it. And they also said that performance is already excelent ( 6800Ultra has double perfomance of the curret top of the line CPUs ).
Cant wait for this to be released, than my CPU would be on Boinc and my GFX will fold :)
ID: 174622 · Report as offensive
Profile Paul D. Buck
Volunteer tester

Send message
Joined: 19 Jul 00
Posts: 3898
Credit: 1,158,042
RAC: 0
United States
Message 174631 - Posted: 6 Oct 2005, 7:01:39 UTC

Now there is an incentive to buy top end GPU cards again ...

Of course, Folding@Home has to get their BOINC act together ... Though a little bit old, I have a couple pretty top end nVidia cards still ... I forget which ones ... but, that would be an incentive to add Folding@Home to my mix especially if I could indicate that it would only run on the GPU ...

Then I could turn my P4's into tripple CPU sustems ... :)
ID: 174631 · Report as offensive
Profile mikey
Volunteer tester
Avatar

Send message
Joined: 17 Dec 99
Posts: 4215
Credit: 3,474,603
RAC: 0
United States
Message 174692 - Posted: 6 Oct 2005, 13:19:26 UTC - in response to Message 174631.  

Then I could turn my P4's into tripple CPU sustems ... :)

Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip.

ID: 174692 · Report as offensive
Astro
Volunteer tester
Avatar

Send message
Joined: 16 Apr 02
Posts: 8026
Credit: 600,015
RAC: 0
Message 174712 - Posted: 6 Oct 2005, 13:49:12 UTC - in response to Message 174692.  

Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip.

that's only 250 buck a processor,....Cheap.... get me 4 of those. LOL
ID: 174712 · Report as offensive
FMatson
Avatar

Send message
Joined: 14 Feb 01
Posts: 35
Credit: 680,946
RAC: 0
United States
Message 174759 - Posted: 6 Oct 2005, 16:23:48 UTC - in response to Message 174692.  
Last modified: 6 Oct 2005, 16:24:35 UTC

Then I could turn my P4's into tripple CPU sustems ... :)

Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip.


Now take that dual core HT chip & throw it in Gigabyte's new GA-8N SLI Quad Royal MoBo...



...now add four Nvidia 7800GTs (given a working client that will work on the GPUs ala Folding@Home's beta) and you have a SEPTEM cruncher! ;)
ID: 174759 · Report as offensive
Profile Paul D. Buck
Volunteer tester

Send message
Joined: 19 Jul 00
Posts: 3898
Credit: 1,158,042
RAC: 0
United States
Message 174780 - Posted: 6 Oct 2005, 17:42:37 UTC - in response to Message 174692.  

Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip.

I already have 2 Quads ... 2 Dual Xeons ... with GPU processing I could get pentas ...
ID: 174780 · Report as offensive
Profile Mr.Pernod
Volunteer tester
Avatar

Send message
Joined: 8 Feb 04
Posts: 350
Credit: 1,015,988
RAC: 0
Netherlands
Message 174796 - Posted: 6 Oct 2005, 18:28:32 UTC

keeping in mind overheating and overclocking,
I wonder how many client errors/inaccurate results this would give....
ID: 174796 · Report as offensive
Profile mikey
Volunteer tester
Avatar

Send message
Joined: 17 Dec 99
Posts: 4215
Credit: 3,474,603
RAC: 0
United States
Message 174874 - Posted: 6 Oct 2005, 22:30:16 UTC - in response to Message 174796.  

keeping in mind overheating and overclocking,
I wonder how many client errors/inaccurate results this would give....

As you are suggesting...cooling is ABSOLUTELY important with that setup!

ID: 174874 · Report as offensive
Profile mikey
Volunteer tester
Avatar

Send message
Joined: 17 Dec 99
Posts: 4215
Credit: 3,474,603
RAC: 0
United States
Message 174875 - Posted: 6 Oct 2005, 22:31:07 UTC - in response to Message 174780.  

Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip.

I already have 2 Quads ... 2 Dual Xeons ... with GPU processing I could get pentas ...

Or with dual gpu's you could get sextas!

ID: 174875 · Report as offensive
FMatson
Avatar

Send message
Joined: 14 Feb 01
Posts: 35
Credit: 680,946
RAC: 0
United States
Message 175327 - Posted: 8 Oct 2005, 4:04:05 UTC

A couple of interesting links for anyone interested in GPU crunching...

General-Purpose Computation Using Graphics Hardware:
http://www.gpgpu.org/cgi-bin/blosxom.cgi/GPUs/index.html

ClawHMMER: A Streaming HMMer-Search Implementation
http://graphics.stanford.edu/papers/clawhmmer/hmmer.pdf
ID: 175327 · Report as offensive
Profile AndyK
Avatar

Send message
Joined: 3 Apr 99
Posts: 280
Credit: 305,079
RAC: 0
Germany
Message 176467 - Posted: 10 Oct 2005, 23:03:45 UTC
Last modified: 10 Oct 2005, 23:03:59 UTC

An interesting code sample for seti could be: FFT Demo from nVidia
btw: this will work on every GForce 6 and 7 card!

Andy
Want to know your pending credit?


The biggest bug is sitting 10 inch in front of the screen.
ID: 176467 · Report as offensive
Profile kinhull
Volunteer tester
Avatar

Send message
Joined: 3 Oct 03
Posts: 1029
Credit: 636,475
RAC: 0
United Kingdom
Message 176484 - Posted: 10 Oct 2005, 23:30:14 UTC
Last modified: 10 Oct 2005, 23:37:25 UTC

Stupid questions I know, but I'm going to ask them anyway.

If a BOINC client that runs on the GPU is successfully developed, does that mean:


    [*]you can run one or more projects on both the GPU and the CPU, at the same time
    [*]you will lose access to you screen (ie would I still be able to read documents, or surf the web, etc, whilst the GPU and/or the CPU is being used)
    [*]the CPU will be doing more work to 'overcome' the GPU and so the CPU's efficiency in crunching it's project/wu will be decreased by a certain amount
    [*]the latest Boinc Manager will supply an option to use the GPU and/or the CPU



Thanks

[edit]

I can't seem to get the

[ list=1]
[*]
[ /list]

to work

[/edit]
Join TeamACC

Sometimes I think we are alone in the universe, and sometimes I think we are not. In either case the idea is quite staggering.
ID: 176484 · Report as offensive
Profile Paul D. Buck
Volunteer tester

Send message
Joined: 19 Jul 00
Posts: 3898
Credit: 1,158,042
RAC: 0
United States
Message 176498 - Posted: 11 Oct 2005, 0:02:33 UTC - in response to Message 176484.  

Stupid questions I know, but I'm going to ask them anyway.

If a BOINC client that runs on the GPU is successfully developed, does that mean:


  1. you can run one or more projects on both the GPU and the CPU, at the same time
  2. you will lose access to you screen (ie would I still be able to read documents, or surf the web, etc, whilst the GPU and/or the CPU is being used)
  3. the CPU will be doing more work to 'overcome' the GPU and so the CPU's efficiency in crunching it's project/wu will be decreased by a certain amount
  4. the latest Boinc Manager will supply an option to use the GPU and/or the CPU




1) yes
2) No, though there is the possibility of actually writing the raw memory data as it mutates to the screen with a possibly mutating psychodelyic effect.
3) Hard to say, but, yes, you will lose some CPU, but gain a lot more by the GPU being there ... just like HT, slower for each one, two at a time, greater throughput
4) no

use list then just "*"s

ID: 176498 · Report as offensive
Profile kinhull
Volunteer tester
Avatar

Send message
Joined: 3 Oct 03
Posts: 1029
Credit: 636,475
RAC: 0
United Kingdom
Message 176499 - Posted: 11 Oct 2005, 0:05:49 UTC - in response to Message 176498.  

1) yes
2) No, though there is the possibility of actually writing the raw memory data as it mutates to the screen with a possibly mutating psychodelyic effect.
3) Hard to say, but, yes, you will lose some CPU, but gain a lot more by the GPU being there ... just like HT, slower for each one, two at a time, greater throughput
4) no

use list then just "*"s


Thanks for the reply Paul.
Join TeamACC

Sometimes I think we are alone in the universe, and sometimes I think we are not. In either case the idea is quite staggering.
ID: 176499 · Report as offensive
Profile The Pirate
Volunteer tester
Avatar

Send message
Joined: 14 Apr 00
Posts: 191
Credit: 4,929,008
RAC: 0
United States
Message 176510 - Posted: 11 Oct 2005, 0:42:09 UTC

Currently I run Rosetta and Protean and have been holding off on Folding@home until they get their BOINC project running, but if they make a GPU version I would definately be interested.

ID: 176510 · Report as offensive
Profile Chilean
Volunteer tester
Avatar

Send message
Joined: 6 Apr 03
Posts: 498
Credit: 3,200,504
RAC: 0
Chile
Message 176511 - Posted: 11 Oct 2005, 0:45:48 UTC

Just wait 5-10 years, maybe a singel PC maybe more powerfull than what SETI@Home is right now.
ID: 176511 · Report as offensive
Profile trux
Volunteer tester
Avatar

Send message
Joined: 6 Feb 01
Posts: 344
Credit: 1,127,051
RAC: 0
Czech Republic
Message 176518 - Posted: 11 Oct 2005, 1:34:00 UTC - in response to Message 176498.  
Last modified: 11 Oct 2005, 1:40:39 UTC


  1. you can run one or more projects on both the GPU and the CPU, at the same time
  2. you will lose access to you screen (ie would I still be able to read documents, or surf the web, etc, whilst the GPU and/or the CPU is being used)
  3. the CPU will be doing more work to 'overcome' the GPU and so the CPU's efficiency in crunching it's project/wu will be decreased by a certain amount
  4. the latest Boinc Manager will supply an option to use the GPU and/or the CPU


Actually, as far as I saw the possibilities of current GPU's and the possibilities of programming on them, I believe that it is easier to process only certain subroutines on the GPU. Running the entire application there may be much more challenging, unless the GPU API and programming tools do not improve. So I'd tell the answers would be rather in this way:

  1. No, the app will still run on the CPU, but will be considerably accelerated by executing one or more threads of subroutines (i.e. the FFT functions) in the available GPU(s).
  2. It depends on the way it is coded. Certainly, you won't be able playing 3D games, but I hope at least basic screen operations may be maintained without serious interferences.
  3. The CPU will do much less than now - just sending the data to subroutines in the GPU(s), receiving and processing the results. GPU's can usually process data in several pipes, so the data flow and processing will be still quite intensive, but the increase of speed should be dramatic
  4. If it is done in this way, BOINC manager will have nothing to do with the GPU. That's purely a question of the project application. Only if the entire application can be written for the GPU, changes in BOINC client would be needed for managing the multiple processors. However, as I wrote, this will not be easily possible unless there is better API, and especially also C++ compilers and debuggers for those GPU's, what is not the case today (afaik).



trux
BOINC software
Freediving Team
Czech Republic
ID: 176518 · Report as offensive
Profile Paul D. Buck
Volunteer tester

Send message
Joined: 19 Jul 00
Posts: 3898
Credit: 1,158,042
RAC: 0
United States
Message 176602 - Posted: 11 Oct 2005, 9:37:50 UTC

In 2000 when I started in SETI@Home processing time for one work unit was 32 hours on one of my PCs. Now, we are doing 1.5 to 2 times the total processing (increasing the processing time by that amount) and the total run time is down to about 2.5 hours on my machines about 30 times the throughput in 6 years. Which is not bad at all ...

Yes, there are good questions if the GPU will be the full processor or not. I would hope that it is, because I could increase the amount of work I could keep in flight ... if it still takes a "slot" in BOINC, well, there I would have to test to see if my overall throughput would increase ... if it does, well, then, a whole new way to improve performance.
ID: 176602 · Report as offensive
Profile Tern
Volunteer tester
Avatar

Send message
Joined: 4 Dec 03
Posts: 1122
Credit: 13,376,822
RAC: 44
United States
Message 176608 - Posted: 11 Oct 2005, 9:55:43 UTC - in response to Message 176602.  

Yes, there are good questions if the GPU will be the full processor or not. I would hope that it is, because I could increase the amount of work I could keep in flight ... if it still takes a "slot" in BOINC, well, there I would have to test to see if my overall throughput would increase ... if it does, well, then, a whole new way to improve performance.


My guess is that using the GPU for the vector processing would be a similar speedup to using the Altivec unit on a G4 - cut time per WU in half; maybe even better. I would suspect it would still take a "slot" on the CPU, however. If you're currently running four WUs at a time on a Xeon, there would be some competition for the GPU (or it would have to be used by only one thread... or you'd have to have 4 GPUs, which I don't think even the biggest SLI MB will allow...) The best way to think of a GPU from what I've seen is as a "super FPU", rather than as another CPU in the same box. The OS is just not going to treat it as a processor, so a thread is going to have to sit on the CPU side and call out to the GPU. Do you remember the old x86 cards for the Mac? There was code that could spawn off threads to the x86 for processing... those weren't _just_ for running Windows.

I would be amazed if the OS allows BOINC Manager to give the GPU an entire WU 'slot'. But I wouldn't be surprised if, ONCE THE DEVELOPMENT ENVIRONMENT IS THERE (big "if" to me... more likely "when" with ATIs latest, but still...) the science apps become "super optimized" - download the xyz GPU version, and SETI WUs take 20 minutes... I want to see the benchmarking for THAT!
ID: 176608 · Report as offensive
1 · 2 · Next

Message boards : Number crunching : Speaking of crunching on your GPU....


 
©2025 University of California
 
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.