Message boards :
Number crunching :
Speaking of crunching on your GPU....
Message board moderation
Author | Message |
---|---|
FMatson ![]() Send message Joined: 14 Feb 01 Posts: 35 Credit: 680,946 RAC: 0 ![]() |
---Cut & pasted from an article on Tom's Hardware regarding ATI's upcoming line of graphic cards: "...The real innovation however is just an idea at this time, but promises to have a major impact on the industry, if brought to reality. Heye mentioned that ATI plans to open the hardware architecture of the X1000 to allow third party developers to write non-graphics-related applications to run on the graphics processor. The company calls this feature "dynamic load balancing." Compared to a Pentium 4 CPU, which delivers a floating point performance of 12 GFLOPs and a bandwidth of just under 6 GByte per second, a graphics processor is a calculation monster: According to ATI, an X1800 XT chip reaches 83 GFlops and 42 GByte per second. The full performance of a graphics may not always be needed - especially in dual-graphics environments - and users will be able to relocate processing power to other applications. According to ATI, these applications could include scientific applications such as fluid dynamics, but also entertainment-related functions such as physics or 3D audio processing. Similar features have been demonstrated by academic projects in the past on ATI and Nvidia platforms, but dynamic load balancing as described by ATI officials promises a whole new use of graphics processors. The company expects GPU specific third-party API's to become common within a few years - with one of the most promising being physics processing: ATI believes that graphics chips provide enough power to cover the features that are currently promoted by Ageia. If ATI's vision comes true, Ageia's idea for a physics board for every gaming PC may become outdated before the startup's technology has reached a critical mass of customers..." The original article can be found at: http://www.tomshardware.com/hardnews/20051005_090950.html ![]() |
Hammer Send message Joined: 13 Dec 02 Posts: 74 Credit: 1,773,558 RAC: 0 ![]() |
I have read some info before that Stanford is working on new Folding@home client that folds with GPU-s. They said that client is already working well and that they are still optimizing it. Basically you only need any DirectX 9 card to be able to fold with it. And they also said that performance is already excelent ( 6800Ultra has double perfomance of the curret top of the line CPUs ). Cant wait for this to be released, than my CPU would be on Boinc and my GFX will fold :) |
![]() Send message Joined: 19 Jul 00 Posts: 3898 Credit: 1,158,042 RAC: 0 ![]() |
Now there is an incentive to buy top end GPU cards again ... Of course, Folding@Home has to get their BOINC act together ... Though a little bit old, I have a couple pretty top end nVidia cards still ... I forget which ones ... but, that would be an incentive to add Folding@Home to my mix especially if I could indicate that it would only run on the GPU ... Then I could turn my P4's into tripple CPU sustems ... :) |
![]() ![]() Send message Joined: 17 Dec 99 Posts: 4215 Credit: 3,474,603 RAC: 0 ![]() |
Then I could turn my P4's into tripple CPU sustems ... :) Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip. ![]() |
Astro ![]() Send message Joined: 16 Apr 02 Posts: 8026 Credit: 600,015 RAC: 0 |
Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip. that's only 250 buck a processor,....Cheap.... get me 4 of those. LOL |
FMatson ![]() Send message Joined: 14 Feb 01 Posts: 35 Credit: 680,946 RAC: 0 ![]() |
Then I could turn my P4's into tripple CPU sustems ... :) Now take that dual core HT chip & throw it in Gigabyte's new GA-8N SLI Quad Royal MoBo... ![]() ...now add four Nvidia 7800GTs (given a working client that will work on the GPUs ala Folding@Home's beta) and you have a SEPTEM cruncher! ;) ![]() |
![]() Send message Joined: 19 Jul 00 Posts: 3898 Credit: 1,158,042 RAC: 0 ![]() |
Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip. I already have 2 Quads ... 2 Dual Xeons ... with GPU processing I could get pentas ... |
![]() ![]() Send message Joined: 8 Feb 04 Posts: 350 Credit: 1,015,988 RAC: 0 ![]() |
keeping in mind overheating and overclocking, I wonder how many client errors/inaccurate results this would give.... |
![]() ![]() Send message Joined: 17 Dec 99 Posts: 4215 Credit: 3,474,603 RAC: 0 ![]() |
keeping in mind overheating and overclocking, As you are suggesting...cooling is ABSOLUTELY important with that setup! ![]() |
![]() ![]() Send message Joined: 17 Dec 99 Posts: 4215 Credit: 3,474,603 RAC: 0 ![]() |
Just go out and get one of the new P4D 840EE chips, around $1,000.00US, and you could have a QUAD cruncher! It is a dual core chip with HT enabled on BOTH chips and 1 meg L2 cache for each chip. Or with dual gpu's you could get sextas! ![]() |
FMatson ![]() Send message Joined: 14 Feb 01 Posts: 35 Credit: 680,946 RAC: 0 ![]() |
A couple of interesting links for anyone interested in GPU crunching... General-Purpose Computation Using Graphics Hardware: http://www.gpgpu.org/cgi-bin/blosxom.cgi/GPUs/index.html ClawHMMER: A Streaming HMMer-Search Implementation http://graphics.stanford.edu/papers/clawhmmer/hmmer.pdf ![]() |
![]() ![]() Send message Joined: 3 Apr 99 Posts: 280 Credit: 305,079 RAC: 0 ![]() |
An interesting code sample for seti could be: FFT Demo from nVidia btw: this will work on every GForce 6 and 7 card! Andy Want to know your pending credit? The biggest bug is sitting 10 inch in front of the screen. |
![]() ![]() Send message Joined: 3 Oct 03 Posts: 1029 Credit: 636,475 RAC: 0 ![]() |
Stupid questions I know, but I'm going to ask them anyway. If a BOINC client that runs on the GPU is successfully developed, does that mean: [*]you can run one or more projects on both the GPU and the CPU, at the same time [*]you will lose access to you screen (ie would I still be able to read documents, or surf the web, etc, whilst the GPU and/or the CPU is being used) [*]the CPU will be doing more work to 'overcome' the GPU and so the CPU's efficiency in crunching it's project/wu will be decreased by a certain amount [*]the latest Boinc Manager will supply an option to use the GPU and/or the CPU
I can't seem to get the [/edit] Join TeamACC Sometimes I think we are alone in the universe, and sometimes I think we are not. In either case the idea is quite staggering. |
![]() Send message Joined: 19 Jul 00 Posts: 3898 Credit: 1,158,042 RAC: 0 ![]() |
Stupid questions I know, but I'm going to ask them anyway. 1) yes 2) No, though there is the possibility of actually writing the raw memory data as it mutates to the screen with a possibly mutating psychodelyic effect. 3) Hard to say, but, yes, you will lose some CPU, but gain a lot more by the GPU being there ... just like HT, slower for each one, two at a time, greater throughput 4) no use list then just "*"s |
![]() ![]() Send message Joined: 3 Oct 03 Posts: 1029 Credit: 636,475 RAC: 0 ![]() |
1) yes Thanks for the reply Paul. Join TeamACC Sometimes I think we are alone in the universe, and sometimes I think we are not. In either case the idea is quite staggering. |
![]() ![]() Send message Joined: 14 Apr 00 Posts: 191 Credit: 4,929,008 RAC: 0 ![]() |
|
![]() Send message Joined: 6 Apr 03 Posts: 498 Credit: 3,200,504 RAC: 0 ![]() |
|
![]() ![]() Send message Joined: 6 Feb 01 Posts: 344 Credit: 1,127,051 RAC: 0 ![]() |
Actually, as far as I saw the possibilities of current GPU's and the possibilities of programming on them, I believe that it is easier to process only certain subroutines on the GPU. Running the entire application there may be much more challenging, unless the GPU API and programming tools do not improve. So I'd tell the answers would be rather in this way:
trux BOINC software Freediving Team Czech Republic |
![]() Send message Joined: 19 Jul 00 Posts: 3898 Credit: 1,158,042 RAC: 0 ![]() |
In 2000 when I started in SETI@Home processing time for one work unit was 32 hours on one of my PCs. Now, we are doing 1.5 to 2 times the total processing (increasing the processing time by that amount) and the total run time is down to about 2.5 hours on my machines about 30 times the throughput in 6 years. Which is not bad at all ... Yes, there are good questions if the GPU will be the full processor or not. I would hope that it is, because I could increase the amount of work I could keep in flight ... if it still takes a "slot" in BOINC, well, there I would have to test to see if my overall throughput would increase ... if it does, well, then, a whole new way to improve performance. |
![]() ![]() Send message Joined: 4 Dec 03 Posts: 1122 Credit: 13,376,822 RAC: 44 ![]() ![]() |
Yes, there are good questions if the GPU will be the full processor or not. I would hope that it is, because I could increase the amount of work I could keep in flight ... if it still takes a "slot" in BOINC, well, there I would have to test to see if my overall throughput would increase ... if it does, well, then, a whole new way to improve performance. My guess is that using the GPU for the vector processing would be a similar speedup to using the Altivec unit on a G4 - cut time per WU in half; maybe even better. I would suspect it would still take a "slot" on the CPU, however. If you're currently running four WUs at a time on a Xeon, there would be some competition for the GPU (or it would have to be used by only one thread... or you'd have to have 4 GPUs, which I don't think even the biggest SLI MB will allow...) The best way to think of a GPU from what I've seen is as a "super FPU", rather than as another CPU in the same box. The OS is just not going to treat it as a processor, so a thread is going to have to sit on the CPU side and call out to the GPU. Do you remember the old x86 cards for the Mac? There was code that could spawn off threads to the x86 for processing... those weren't _just_ for running Windows. I would be amazed if the OS allows BOINC Manager to give the GPU an entire WU 'slot'. But I wouldn't be surprised if, ONCE THE DEVELOPMENT ENVIRONMENT IS THERE (big "if" to me... more likely "when" with ATIs latest, but still...) the science apps become "super optimized" - download the xyz GPU version, and SETI WUs take 20 minutes... I want to see the benchmarking for THAT! |
©2025 University of California
SETI@home and Astropulse are funded by grants from the National Science Foundation, NASA, and donations from SETI@home volunteers. AstroPulse is funded in part by the NSF through grant AST-0307956.