Posts by -= Vyper =-


log in
41) Message boards : Number crunching : 2 or 3 WU's On A GTX 470 (Message 1075780)
Posted 1288 days ago by Profile -= Vyper =-Project donor
There is no definite answer to this one.

But we can presume this would be the most accurate, run 2 WU's if you run the card at stock clocks and if you OC your graphics card you can try to run 3 Wu's.

What you can test if you are into OC'ing is to see if PCI-E speed has a remarkable impact on crunching speed.

For instance if you run three WU's at once in a OC configuration on the 470 you would perhaps get around 23 to 24 minutes for a bunch of 0.4x AR wu's..
Then go to the bios and try to set PCI-E speed to 110 Mhz instead of 100 (Don't set it much higher than that) and see if you lower your times to perhaps 22,3 - 23,3 minutes for the same 0.4x AR wu's..

If the difference is negliable then set it back to 100 because then you've made a proof of concept test that shows that setting the PCI-E speed gives you a speed boost to the GPU operations or perhaps not.

Hope you get some ideas of direction for a better throughput.

Kind regards Vyper
42) Message boards : Number crunching : Combine gtx275 & gtx570 in one system? (Message 1074349)
Posted 1293 days ago by Profile -= Vyper =-Project donor
I wouldn't dare to do it!

Atleast 750W to be on the safe side even if the powerdraw is about 500W when running full speed.
Go for a Corsair or Thermaltake PSU , choose one which have atleast 2 x PCI-E 8 connectors and 2 x PCI-E 6 connectors to be absolutely sure.

//Vyper
43) Message boards : Number crunching : Running mixed nvidia hardware most optimal on the same host (Message 1074045)
Posted 1293 days ago by Profile -= Vyper =-Project donor
I had struggles getting the most out of my system when running a GTX285 and a GTX480 in the same system.
The 285 decreased it's productivity when running two WU's at once compared to the GTX480 which loved it.
In a guide on my blog i made two separate boinc work folders with separate executables to suite my needs to increase the throughput on that host.
Head over to see screenshoots and a small guide.

With this way i could set my GTX285 to use a specific executable and run only 1WU at once and the 480 ran another executable and two at once instead.

http://vyper.kafit.se

Kind regards Vyper

P.S In my example i had a 295 but most people know the drill anyhow D.S
44) Message boards : Number crunching : Virtual Machine and CUDA Questions??? (Message 1070784)
Posted 1303 days ago by Profile -= Vyper =-Project donor
I'm sending a "Shutting Up" to myself.
Thanks for digging BeNt..

//Vyper
45) Message boards : Number crunching : Virtual Machine and CUDA Questions??? (Message 1070750)
Posted 1303 days ago by Profile -= Vyper =-Project donor
Q1: I'm not really sure but i doubt it. Don't think noone has invented graphics bypass directly to PCI-e adresses.

Q2: No! Again if someone/company hasn't invented a way of running 64bit hosts on a 32bit system. It is possible if you have 64bit all the way.
example:

HOST-HYPERVISOR-VIRTUAL MACHINE:

You need to have a:
64-64-64 configuration to make it work.

In your case you have:
64-32-64 and that is not possible from what i know.

Kind regards Vyper

P.S Could you please answer personal messages that i have wrote to you? D.S
46) Message boards : Cafe SETI : A bit of better news for the kittyman. (Message 1068503)
Posted 1309 days ago by Profile -= Vyper =-Project donor
Mark!

Do you know how often your powercompany "polls" your powerline to measure how much KWh you are using at that particular moment or do they simply collect the data once a month?

If so, you cannot choose a method of poweroptimizing your electricity in a particular timeframe more for one month to the next.
The only thing for certain is that you lower the amount of powerdraw for that particular timeframe that your powercompany measure you.

Example once a month is perhaps 1500KWh regular usage for you, if you turn it off during a special period that sums up to for instance a day a week that equals to about 1288KWh a month of consumed electricity.

When your powercompany then sends a bill they sum it up to 1288KWh equals that months average powercost if you have a non-fixed rate.

That means that how much you try to optimize and turn of during prime times the average for that month wouldn't change a single digit at all , and if you turn it off during the primetime the only thing you would do is reducing the cost for what your powercompany needs to pay for purchasing electricity for themselves to deliver that to you because their powerpolling is not once a month it perhaps polls every minute or so.
It would only increase their own revenues.

So if you want to lower a powercompanys revenues , turn the computer off during the cheaper period and only run it at the expensive peaks instead :P ..
Because as i said they got polled and billed more compared to you as customer for your powercompany.

This what it "seems" to be the way it works here in Sweden until i've heard otherwise from other safe sources.

Kind regards Vyper
47) Message boards : Number crunching : GPU Users Group Server Donation (Message 1065732)
Posted 1317 days ago by Profile -= Vyper =-Project donor
Well, our server Synergy is now up and running.
Hope we all can make use of it and further stabilize the infrastructure further!

Thank you Todd for your time and effort.

Kind regards Vyper
48) Message boards : Number crunching : PCIe speed and CUDA performance (Message 1064992)
Posted 1319 days ago by Profile -= Vyper =-Project donor
Thank you , thank you.

I'm not sure that i'm 100% correct in what i describe but atleast that gives a small hum of what is going on in terms of what happens when involving different parts of your system.
Everything that eventually can be precalculated or expanded to a easy to follow grid pointer system to make sure that least amount of data possible needed to travel through the slow PCI-E bus is almost certainly a win-win situation.
Cpu can do other work along with gpu not needing to be fed with "what now then" parameters, and if that is not avoidable it simply isn't.

Simply said i presume the system do it's best if as much preparation of data and code is made before the transfers occur to the gpu.

I just couldn't stop myself from making something human referable to what happens inside the computer system at that time of writing.

Kind regards Vyper
49) Message boards : Number crunching : PCIe speed and CUDA performance (Message 1064944)
Posted 1320 days ago by Profile -= Vyper =-Project donor
Don't know man..

Perhaps around half an hour or so. I had it all in my mind.
No. I have not taken drugs :P

Kind regards Vyper
50) Message boards : Number crunching : PCIe speed and CUDA performance (Message 1064934)
Posted 1320 days ago by Profile -= Vyper =-Project donor
Man!

I've should've been a novelist instead:

http://www.imdb.com/title/tt0060196/
=
The CPU, The GPU & The coder

:)

Think twice before you click and send the driver on it's way!

Regards Vyper
51) Message boards : Number crunching : PCIe speed and CUDA performance (Message 1064931)
Posted 1320 days ago by Profile -= Vyper =-Project donor
I haven't read throught the whole entire thread but i've read the latest posts.

And for what it seems the discussion is almost always around the bandwidth issue.
The only thing where bandwidth is an issue is when you start a new seti@home WU when the cpu needs to prepare the data, don't really know what it does but it seems like it expands it (CPU RAM Usage grows steadily), when that has leveled at around the 70-100Mb it seems to move the data over to the GPU Ram , that require bandwidth because it's a lot of data.
When the data is uploaded to GPU ram it seems to start to crunch numbers, and thus the s@h executable needs to feed the data with memory pointers and small "when this block is done move to the next block of data" parameters.

Those parameters needs to be communicated through the "slow" PCI-E bus, bandwidth is not an issue there but the overall speed of the highway.
--------------------
Think of a highway with almost no stops that leads from Washington D.C to Chicago.. It's 16 lanes in each direction.
When you have alot of shipments you send 10000 trucks from Washington to Chicago, more trucks can fit the highway and thus it arrives at it's destination faster and almost all at once.
But unfortunately the telephone and email has not been invented at this time so to know what to do with the shipment you send a driver back from Chicago to Washington with a letter which says "Thanks for the goods, where should we send this?".
The driver is going there on the highway , he has no friends and it's all silent and dull on the highway.
A couple of hours later the driver gets back to the office in Washington, the manager (CPU) opens the letter and reads "Thanks for the goods, where should we send this?" , Manager thinks OMG why didn't i think of that and quickly writes a letter which says "All goods from the 10000 Lorrys should be sent to Factory xxx at xxx road in Calgary".

All is good the driver heads back to Chicago, as usual he sees hardly none traffic at all and thinks why haven't anyone figured out a way to quickly send messages to eachother at different locations because he has a lot of time on the way to start to think of other things.
Starting to get annoyed about that he has to drive back and forth to deliver messages but he knows that it needs the delivery to be confirmed at it's destination (TCP) and if something happens on the way with the information they need to send another guy to do it's work again (TCP RESEND), he starts to sigh thinking if there is a good way to make use of foolproof deliverys with pidgins (UDP) but he quickly finds out, Crap! That isn't doable..
He's finally back to Chicago and heads to the office (GPU) in which the boss quickly opens the letter and smiles at start .. then.. he takes a paper and pen and starts to write .. "By which type, by truck, by rail, by flight" takes an envelope folds the paper and put it into the envelope and says.. "Sorry you need to go to main office ASAP"..

The truck driver starts to hate hes life as he starts to chugg the loooong way back again to Washington from Chicago curcing why the hell the trucks has a speedlimiter set att 100MPH, if the truck had a speed limiter set at atleast 110MPH it would save him atleast 45 minutes getting there, the traffic (bandwidth) is not an issue instead it's the speed of the highway that annoys him greatly as he curses why the hell someone hasn't invited faster trucks and those stupid guys at the American Road Agency (don't know the name) hasn't upped the speed limit at the highway's to atleast 110, it shouldn't do any harm!

Well times fly by (processing time), the load is still there at Chicago and hasn't moved anywhere due to undetailed information at start..
The driver gets back to the office in Washington (CPU) and the manager opens up and reads.
Here we go again, up with that paper he writes "By flight" folds that paper and put it into a new envelope and says "Well by this time i think you know what to do, huh?!" (cache) , then he smiles back to you and you gave out a long *sigh* and headed back to the truck and up to Chicago once again..

During that trip the driver starts to think.. Hey, have i seen this road before? . He hates that they haven't invented foolproof deliverys of pidgins once again (UDP Resend.. Hmm hey can't compute.. error) and quickly bins that idea. He utterly hates speedlimiter on trucks along the speedlimit on the roads and thinks of ways to speed it up a bit (overclocking).

Times fly by faster and he starts to think that soon this shipment could be of to it's real location.. For once he starts to grin as he arrives to Chicago.
When he arrives he rushes up to the office thinking Yes, finally we could be heading away with the goods , knocks on the door to the boss (GPU) , no answer?
... He knocks again but more firmly this time.. The boss coughes and finally says come in, apparantly he fell asleep (Gpu idle clockdown) but hey he is a human after all. Driver gives the letter to the boss with a smile and thinks Yes, Calgary.. But... WTF.... he grabs a new stupid paper and writes.. "Departure?" .... OMG.. he can't be serious shouldn't it be the nearest departure, it should be.
Buuut in this company it's all about confirmation.. So with the drivers head in a low position he says to the boss "Yea yea, i know the drill back to main office once again" (Cache).

The drivers head was blank, he now could hardly breathe, if only this loong trip could be more funny and speedier (overclocking).
Time flies by, and by the time the driver thinks of ways to improve things but with no other around he has trouble venting his ideas so he starts to close up to himself.

The truck is now back at washington, he drags himself up the stairs and thinks "this is it", he's exhausted.
Knocking on the door he almost immediately here the voice that states "come in", man those managers are fast i had barely touched the door in a split microsecond before he said come in.
Driver hands the letter once again and hopefully thinks that this is the last time he needs to go back Chicago but he's not overly convinced.
Manager writes four letters on the envelope and sighed "Why can't anyone think for themselves", he put that in a fresh new envelope and sealed it and gave to the driver..

With a already low head position the driver takes this letter and he mumbles cursing he's own miserably life, instead of a lorry driver he should've been a cook instead (software coder) or perhaps a braindead thing that makes all the food eatable like a stove or something (compiler).

Naah, of we go..

At that time when he travels that looong dreaded way back to Chicago he comes up with an idea.
Instead of having a huge lorry in which he takes a small envelope with him all the time, the highway should have two express lanes in each directions which the speed limit was 800Mph. If that were possible he could take another vehicle for those small packages which hasn't contributed to any sort of increased weight and be able to travel between Washington and Chicago at only one eights of that time required.
He shines up and thinks hey this is not a bad idea, slow crappy bulgy lorrys in this wide lane and a parallell high speed vehicle lane.
This is brilliant! But who the f*** listens to a small worker like me.
Everything is about the money and reducing costs these days and he quickly finds out that no company in the world would be willing to invest that money and time in producing it and he quickly finds himself stuck in time again!
He cries out "Why can't anyone with a high position enough come up with this, we need to think outside the box really!!" , he's frustrated.
The cargo at Chicago has been stopped for days now, and for what?
No detailed information at first so we can do it right from the beginning which in terms lead to fewer delays (software engineering and optimising)..

Time flies by , albeit slowly ofcourse.
The driver is back to Chicago and with heavy steps he walks to the bosses office and knocks. No answer! Exactly as last time he thinks. He waits, knocks on the door again.. No answer?!! "What?!" , knocks on the door and waits. Still no answer!! Bah, he puts he's hand to the knob and carefully opens the door, and that sight he wishes he would never have seen at all!
There he was the boss, hung from a beam in the roof and the driver was scared stiff (Driver hang ... Driver/Boss! :) ) .. Holy crap! How the f* did this happen? He quickly backs out and closes the door and wishes with all hes heart that this was a dream.
He starts to go from there to get anwers! A few steps away he notices that he has dropped the envelope and he turns around and walks into the door to the office and there he was again, sitting in the chair like nothing ever has happened with a cut gallow around hes neck staring at him at he enters the door shouting. "When you work at this company, you ALWAYS knock before entering a room!"
The driver thought he was dreaming (driver restart) but quickly responds "I'm sorry sir! Here are the last orders from Washington"

The boss yanks the letter from the driver and frowns, he opens up and reads.
The face slowly starts to smile saying "Well let's get to it".. Go to the other lorrys and wait for departure to the airport.

The driver shines up and starts to smile and starts to rush out from the door running towards the other trucks thinking "Atlast!!" , but he halts thinking what was that?
He heads back in to the room asking for permission to speak as he behaved so badly..
Driver: I'm just curious, what was it it said in that letter that could confirm this, i started getting used to the highway instead of doing what it sayed that i was employed for.
Boss: Well with this magic word, i could take care of the decisions much easier so for what it's worth for you that magic word was. ASAP
The driver stared at the boss and gasped "So with this magic letters or word my trips back and forth could have been saved if it had come directly from the beginning? (Software optimising, rescheduling)
Boss: Yes

The driver was so furious, he wants to kick the software developers in the a** but at the same time he was glad to came up with that idea with a high speed communincating lane parallell to the highway and hopefully someone makes that idea come to life in the future if he evers dare to speak of it..

Driver: "I'm on my way sir, thank you for not firing me for not beeing polite to knock at that door"
Boss: "Anytime, but always wait for an answer before entering my office in the future, got it?" (Interupt management)
Driver: "Most certainly Boss"

The driver smiled and turned again walking towards hes awaiting friends..

-------------------------

Well what can we make out of this then?!
Sorry for my little story but it points out that how matter how wide the PCI-E bus is if it only has a small payload that should arrive at it's destination.
If you increase the speed on the PCI-E you lower overall computing time because until the GPU knows what to do next it halts out and waits for further instructions.

And as i told in my little story until someone makes a parallell small payload high speed communication lane the impacts on PCI-E should matter somewhat to different degrees.

I presume that VHAR's benefits more of a faster PCI-E than for instance a regular MAR wu, because the GPU itself is more occupied working with it's data compared to constant "what to do next" messages on the PCI-E bus in VHAR case.


Sorry once again for making such a huge post but i wanted to make a small story out of this discussion and couldn't brace myself :)

Kind regards Vyper
52) Message boards : Number crunching : System Tuning (Message 1058210)
Posted 1339 days ago by Profile -= Vyper =-Project donor
I saw that you had around 120Mhz on the PCI-e..

That's too much as you noticed and make it prone to hanging, for my own part there was no problem at 115 Mhz but i really think that you could set it at 110Mhz if you want to.

Those GPU's must be waiting alot to be fed with data if you occupy the GPU's and bus that much with 3 Wu's / GPU each..

Kind regards Vyper
53) Message boards : Number crunching : Collected data of various Cuda s@h executables public/nonpublic (Alpha/Beta) (Message 1054632)
Posted 1350 days ago by Profile -= Vyper =-Project donor
Very interesting, but some explanation needed...

How do you calculate the efficiency for 2 WUs vs 1 WU?

Is that using hyperthreading with one WU per virtual core? Or 2 WUs on one core?

There's one very obvious 'interesting' result for your 'fermipfForceSerial'... Can you explain further?

I've seen comments about running multiple WUs in parallel on nVidia GPUs... Good idea or bad?


Hi

Well the efficiency is measured by running two exactly same benchmarks on one gpu at the same time!
If you take the total time taken in seconds which of course is higher than running only one single WU and compare that with the dual wu seconds required.

So the formula is as follows:

86400 / 600 seconds = 144 results / day on a 1WU per GPU config..
86400 / 900 seconds = 96 results / multiplied per app on a single GPU config..

In this testbed i had two parallel Wu's = 192 results / day.

The numbers are only example of measurement.
And you see that particular config has a 33.3% increase in throughput RAC wise.
The interesting phenomena with fermipfforceserial is that it suddenly exited when the test result wasn't done and thus that test exited faster than the other ones, but the fermipfforceserial tests in a 2WU configuration is not considered valid at all.
So the point is that with the newer Fermi architecture there is some headroom in each GPU to make it possible and increase the output / day.
In that terms i would consider it a good thing..

Kind regards Vyper

54) Message boards : Number crunching : Collected data of various Cuda s@h executables public/nonpublic (Alpha/Beta) (Message 1054598)
Posted 1350 days ago by Profile -= Vyper =-Project donor
Hi

When the servers was down i was bored and made a benchmark with various full length wu's and made some statistics over that benchmark.
The last two pages contains a summary of one executable which wouldn't be considered accurate but other than that. No problem at all.
You even see the benefit in numbers of running more than one wu on that particular card.
Consider this as a base of what i can compare data with in the future when running newer compiled Cuda 3.2 beta exes and other performance enhancements with the Cuda executables.

Head over to my blog to download the file!

http://vyper.kafit.se

Kind regards Vyper
55) Message boards : Number crunching : Manually uploading client_state.xml to s@h (Message 1042026)
Posted 1407 days ago by Profile -= Vyper =-Project donor
Sweet news!
I can't wait for it to be incorporated.
Thanks..

//Vyper
56) Message boards : Number crunching : Greetings (Message 1040196)
Posted 1412 days ago by Profile -= Vyper =-Project donor
Yoo wassup! :)

Regards Vyper
57) Message boards : Number crunching : Hardware Donation threads - an alternative solution (Message 1040173)
Posted 1412 days ago by Profile -= Vyper =-Project donor
Yeah but i mean that for instance if the whole facility was moved to a bunker like this where they have people supporting if a hdd fails, machine dies etc 24/7 , extremely ventilated and electricity included..
Perhaps to host the whole seti@home park with hands on 24/7 would perhaps cost slightly below the 100000$ mark / year.
That would leave alot of the fundings to better things so the staff can focus on remotecontrolling their systems and take some salary to keep it running and have cash for spare parts on the shelf incase a SAS drive fails or anything.
I recon that even for that sum of money there would be a dedicated 1000/1000 line to the whole park and if more speed is needed then cash is the only obstacle.

Well well, we can only hope for a better uptime when new servers are in the pipe too..

Our company has a customer which are about to move alot of machines to this place over from the UK because the stability/uptime is incredible he states.

Kind regards Vyper
58) Message boards : Number crunching : Hardware Donation threads - an alternative solution (Message 1040144)
Posted 1412 days ago by Profile -= Vyper =-Project donor
Holy sh***t!

I noticed the fees for internet charges..
Well!

Our company has a connection to another company here in Sweden.
I think the fee's for a 100/100 connection would lay around 30000$ a year and are situated in a a fireproof bunker with extraordinary uptime and they have their own dieselpower distributor and alot of redundancy fibrepaths in Sweden so if one net fails it's hardly even noticeable.

Two cents for costs.

Kind regards Vyper

59) Message boards : Number crunching : Quiting (Message 1038846)
Posted 1414 days ago by Profile -= Vyper =-Project donor

I understand your feelings. Hopefully everything will be sorted out in the end hopefully when the new server has been deployed..

Remember to drop by from time to time to check upon us.

Kind regards Vyper
60) Message boards : Number crunching : Closed *SETI/BOINC Milestones [ v2.0 ] - XXI* Closed (Message 1037843)
Posted 1419 days ago by Profile -= Vyper =-Project donor
My first negative milestone..

Some strange machine came out of nowhere and snatched the #1 top computer spots for me in a way i can't figure out for now.
Perhaps that fellow is running multiple wu's at a time with the 258.96 driver..

Well if doesn't fall downwards in the long run we've finally seen Fermi architecture overcome GT200 series in terms of parallelism because the timings report much slower progresstime / wu compared to my setup..

Well done!!

Kind regards Vyper


Previous 20 · Next 20

Copyright © 2014 University of California