Message boards : Graphics cards (GPUs) : New application version
Author | Message |
---|---|
We are having problems with the Windows application which seems to fail on a certain type of workunits. | |
ID: 5674 | Rating: 0 | rate: / Reply Quote | |
Thanks for the update. Will this version reduce the Windows cpu usage? | |
ID: 5675 | Rating: 0 | rate: / Reply Quote | |
No, it will just increase process priority as in this version there are already enough other changes. | |
ID: 5677 | Rating: 0 | rate: / Reply Quote | |
No, it will just increase process priority as in this version there are already enough other changes. That's certainly fine, thanks. | |
ID: 5678 | Rating: 0 | rate: / Reply Quote | |
If all works, the reduced cpu version will follow in two days. Yes, this is what I'm waiting for. I hope it will be much better than the current one, it's a mess on a slow card like mine. ;-) ____________ Member of BOINC@Heidelberg and ATA! | |
ID: 5679 | Rating: 0 | rate: / Reply Quote | |
If all works, the reduced cpu version will follow in two days. WOn't change a ting for the slow cards ... It will only reduce the laod on the CPU allowing more production towards CPU intense projects if you are tunning them. So, for example, my productivity on the i7 and quad should rise some because I run n+1 and with the extra load on the processor supporting GPU grid I "lose" some capacity for other projects ... ____________ | |
ID: 5684 | Rating: 0 | rate: / Reply Quote | |
I think DrNow means that the high cpu load especially hurts on his slow card, as the "well, you're getting enough credits from your GPU to compensate" doesn't work as well as with e.g. a GTX 260. | |
ID: 5687 | Rating: 0 | rate: / Reply Quote | |
WOn't change a ting for the slow cards ... You would wonder... Having a 9600GT on an X2 and the same time an app with high CPU load reduces the capability to crunch on other WUs a lot! I give you some examples from my WU-list: This was the last one with app 6.45 and look at the CPU-load, not even 5 minutes!!! With this app I made the most credits because with the 2+1 setting the GPU almost crunched on it's own, it worked bravely. :-) 6.48 already made the CPU-load higher and it recuded the crunching overall massively for me, the setting 2+1 didn't worked that good anymore, was the same like not to use it. The CPU-load with 6.52 was better, but still too high. And with 6.55 it didn't changed that much, but due to my problem with the "unspecified launch failure" I had to recude the clock rate and I guess, this influences the CPU-load as well a bit. Only positive thing with the higher CPU-load I see in the GPU temperature, it isn't that high as earlier anymore. I'm thrilled how the new app will work... ;-) ____________ Member of BOINC@Heidelberg and ATA! | |
ID: 5706 | Rating: 0 | rate: / Reply Quote | |
Yesterday I made my first WU aplication version 6.60 and it made less use of the GPU (the temperature is 7ยบ lower than the same GPU processing a v.6.55 WU) and the points per hour generated were ~380 while with the v6.55 there are ~550 (so with older WUs I get ~50% more points). | |
ID: 5737 | Rating: 0 | rate: / Reply Quote | |
Hmmm... Still get a 6.55 WU... | |
ID: 5743 | Rating: 0 | rate: / Reply Quote | |
only get the 6.55 too | |
ID: 5746 | Rating: 0 | rate: / Reply Quote | |
only get the 6.55 too I don't think it is a released application. Which is a shame as I am eagerly awaiting the upgrade so that I can get higher GPU production on my other projects. The shame of it for me is that the GPU Grid connected machines are my number 1 and 3 producers ... But, I can wait ... I guess ... sigh ... :) ____________ | |
ID: 5749 | Rating: 0 | rate: / Reply Quote | |
I'm now crunching my second v6.60 WU. The other three WUs in cache are v6.55. This one seems that it will last like the one I already done (~8.5 hours in a OC gtx280). If I switch to a v6.55 suspending the v6.60 being crunched, in five minutes the GPU goes from 58ยบ to 65ยบ with the fan fixed to 70%. | |
ID: 5752 | Rating: 0 | rate: / Reply Quote | |
I'm now crunching my second v6.60 WU. The other three WUs in cache are v6.55. This one seems that it will last like the one I already done (~8.5 hours in a OC gtx280). If I switch to a v6.55 suspending the v6.60 being crunched, in five minutes the GPU goes from 58ยบ to 65ยบ with the fan fixed to 70%. Well, someone is lucky ... all I am getting are 6.55 tasks ... :) Maybe we have to clear out the queue of the tasks generated with 6.55? Or just doing a limited run test to see if it is working? I guess I have to wait, sniff, sniff ... ____________ | |
ID: 5767 | Rating: 0 | rate: / Reply Quote | |
ยทยทยทWell, someone is luckyยทยทยท Lucky? I suppose you are joking.... I'm happy that now I only have v6.55 WUs! | |
ID: 5772 | Rating: 0 | rate: / Reply Quote | |
All the 6.60 crashed in our Windows system, so we removed it for now. | |
ID: 5777 | Rating: 0 | rate: / Reply Quote | |
All the 6.60 crashed in our Windows system, so we removed it for now. Thanks for that GDF, some Projects don't bother to Test New Applications & just release the Junk Wu's & let you spend countless hours processing them before you catch it ... | |
ID: 5779 | Rating: 0 | rate: / Reply Quote | |
New applications uploaded. | |
ID: 5836 | Rating: 0 | rate: / Reply Quote | |
New applications uploaded. How can we recognize the new applications? Do they have a new version? | |
ID: 5839 | Rating: 0 | rate: / Reply Quote | |
..... I have a couple showing version 6.61. | |
ID: 5840 | Rating: 0 | rate: / Reply Quote | |
..... I have a couple showing version 6.61. Ok, shall I kill my 2 "old" WUs in the queue? | |
ID: 5841 | Rating: 0 | rate: / Reply Quote | |
.... Hmm - just watching my first 6.61 WU going thru - it is making more use of the CPU than the 6.55 WUs. About a 10% uplift. | |
ID: 5843 | Rating: 0 | rate: / Reply Quote | |
.... Hmm - just watching my first 6.61 WU going thru - it is making more use of the CPU than the 6.55 WUs. About a 10% uplift. Do you think the overall duration will be shorter? | |
ID: 5845 | Rating: 0 | rate: / Reply Quote | |
The higher priority -> higher cpu usage should result in better GPU utilization, which would mean slightly shorter calculation times, especially with 2+1 / 4+1 on a dual or quad core, respectively. Is anyone already seeing higher GPU temps? I'm still working through my 6.55s.. or am letting my GPU do the dirty work for me ;) | |
ID: 5851 | Rating: 0 | rate: / Reply Quote | |
The higher priority -> higher cpu usage should result in better GPU utilization, which would mean slightly shorter calculation times, especially with 2+1 / 4+1 on a dual or quad core, respectively. Is anyone already seeing higher GPU temps? I'm still working through my 6.55s.. or am letting my GPU do the dirty work for me ;) Me too... I have 6.61 tasks queued but, none in work yet ... Um, not to be difficult, but the goal for many of us was to REDUCE the CPU usage. With my 280 card I am fine with the processing time, but, am not so fine with the high CPU usage ... especially since it is to the best of my understanding mostly doing a polling loop, which might be necessary, but is hardly productive use of the CPU. Increasing the priority of the task and wasting more CPU time polling seems to me to be a step backwards. It does help GPU Grid slightly, but at the cost of every other BOINC project I am running (or will be running) alongside the GPU tasks. While on the subject. It occurs to me that the architecture of the GPU Grid application is not correct. There should be one polling thread that then dispatches to the service threads as needed. What I mean is this, if I have a 2 or more GPU system you will launch (based on earlier observations) two application instances both of which poll their individual GPU ... In that this is not really productive use of the CPU when I have two tasks doing nothing it gets ugly and when I have more than that it is REALLY bad. With the dispatch scheme, the one thread is in the polling loop and it polls the GPUs one after the other checking to see if the GPU needs grooming, and if so, should wake the service thread to service the GPU ... less overhead in that only one thread is in idle poll mode ... While on the subject, I have 1G cards for the most part and the memory load seems to run about 50% ... again, is is possible that this could be tailored to available memory so that those with more than 512M would have larger loads and thus lesser demands for re-fills? These musing are from when I had two GPUs in the i7 system and there was so much usage of the CPU that the i7 was running in essentially 7 + 2 mode though it was trying to run as 8 + 2 ... I know GPU Grid wants to maximize the productivity for this project, but it should not do it at the expense of the other BOINC projects ... I am not sure what the CPU load is for SETI@Home (anyone out there running both projects?) but if it is minimal, perhaps we need to collaborate with them? ____________ | |
ID: 5853 | Rating: 0 | rate: / Reply Quote | |
6.61 (CUDA) app uses 40% of my CPU (dualcore, so 80% of a single core) - even worse than 6.55!! We had a version with acceptable CPU usage (6.56), release that via app_info.xml (as has been requested numerous times). 40% CPU usage is not within an acceptable limit anymore for me, so until that is sorted, I'll only run it on my PS3. | |
ID: 5854 | Rating: 0 | rate: / Reply Quote | |
The higher priority -> higher cpu usage should result in better GPU utilization, which would mean slightly shorter calculation times, especially with 2+1 / 4+1 on a dual or quad core, respectively. Is anyone already seeing higher GPU temps? I'm still working through my 6.55s.. or am letting my GPU do the dirty work for me ;) More thoughts ... I think I agree with Digi421 ... 6.55 was high on CPU, the next release lowered CPU usage and now we look like we are going to see a substantial increase ... though I am still doing my 6.55 tasks ... Next question is why are we polling and not using the IRQ? If we used IRQ when the GPU was in need there would be no need for an idle poll loop. Suggestion: An option to select "nice" or GPU performance. I posted this on BOINC Dev:
____________ | |
ID: 5855 | Rating: 0 | rate: / Reply Quote | |
I am currently crunching two 6.61 WU's, and they are using 20% and 23% of a cpu core on my quad. That equates to 80-92% of a single core each, as opposed to the 6.55 WU's that were anywhere from 40%-80% of a single core. So, from my experience the cpu-usage in the best case of 6.61 is about the worst case for 6.55. | |
ID: 5857 | Rating: 0 | rate: / Reply Quote | |
6.61 (CUDA) app uses 40% of my CPU (dualcore, so 80% of a single core) - even worse than 6.55!! Indeed... Just started my first 6.61 WU and am bitterly disappointed. My system (AMD X2 5200 and 9600GT) uses almost a complete core (about 90%!)for crunching on the GPUGrid task and becomes even more sluggish than before. :-( I think I stop crunching here 'til a new version comes out, it's not worth it. ____________ Member of BOINC@Heidelberg and ATA! | |
ID: 5858 | Rating: 0 | rate: / Reply Quote | |
6.61 (CUDA) app uses 40% of my CPU (dualcore, so 80% of a single core) - even worse than 6.55!! Yes, this way goes in the wrong direction, we need an app thats using only a little bit of a cpu core. | |
ID: 5859 | Rating: 0 | rate: / Reply Quote | |
Yes, this way goes in the wrong direction, we need an app thats using only a little bit of a cpu core. Actually we need the choice ... I prefer to run a mix of CPU / GPU tasks with the emphasis on CPU tasks as that is my current weight ... but some that are GPU Grid exclusive may want to have this mode of operation that gets the most out of the GPU card. I can make an argument for either direction which is why there should be the option. Personally I would give up GPU performance for less load on the CPU ... a real elegant solution would actually have three or four levels of performance ... anyway, though it is early days I still see lots of room for improvement and I think I need to make some tests with SaH's application to see how theirs loads the system ... with the thought of course, if they can have a light load, why cannot we have the same here? The answer is likely to be "because" ... but ... this is one reason why I can hardly wait for other projects to start to provide GPU applications so we can start to have some CHOICE ... and vote with our feet ... of course, I have yet to see a project that really takes the desires of participants to heart ... I mean I asked CPDN to have an option to only DL one task per computer so I would not have the things hanging around for years ... their answer was to abort them ... sigh ... waste ... ____________ | |
ID: 5860 | Rating: 0 | rate: / Reply Quote | |
Oops. GDF - it looks like we need 6.55 back or preferably 6.56 - or even 6.next if you can fix 6.56 for all.... | |
ID: 5861 | Rating: 0 | rate: / Reply Quote | |
| |
ID: 5862 | Rating: 0 | rate: / Reply Quote | |
I have checked my latest result with 6.61 and there is no speed increase, one bad issue is that the system getting sluggerish. | |
ID: 5865 | Rating: 0 | rate: / Reply Quote | |
New Linux App (6.59) | |
ID: 5866 | Rating: 0 | rate: / Reply Quote | |
As said in this thread the Windows low cpu version will come in a few days. | |
ID: 5869 | Rating: 0 | rate: / Reply Quote | |
... Actually the change with the Linux CPU usage started with the new WU types... Some of them need more CPU than the old GPUTEST WUs, and some less or the same as the GPUTEST ones. ____________ pixelicious.at - my little photoblog | |
ID: 5871 | Rating: 0 | rate: / Reply Quote | |
As said in this thread the Windows low cpu version will come in a few days. Well, the change caught us off guard and I missed the note that we would get a high CPU usage version to test the changes on all systems (though it seems that there are still problems with 64-Bit windows XP at least, see PoorBoy's other thread). Going WAYYYYY back I did find the little teensy tiny note you wrote ... When ETA announced with glee the high CPU use version, well, the glee seemed to be for the wrong reason ... or at least it struck me as such. I want to support GPU Grid, but not at the cost of one CPU per GPU ... that said, did you see my other questions below? I am running my first 6.61 version tasks (on the i7) and the CPU use is up, not as bad as I feared, though up from 3-6% to solid 6 with occasional 7% load. I have a 295 card arriving tomorrow and I was going to try it in a Linux box that I started up for some other reasons to see what the load is there ... (if I can get it to work at all, as someone else mentioned, my linux skills are, ahem, sub-par) ... ____________ | |
ID: 5873 | Rating: 0 | rate: / Reply Quote | |
On a GTX280 in 3+1 I get this times for a 6.61 WU with 2478 credits in relation to a 6.55 WU with 2435 credits | |
ID: 5874 | Rating: 0 | rate: / Reply Quote | |
On a GTX280 in 3+1 I get this times for a 6.61 WU with 2478 credits in relation to a 6.55 WU with 2435 credits ON my first task completed with 6.61 I have similar results ... {edit - add} CPU usage on the Q9300 has doubled. Now it takes almost the full core. An interesting question... does CUDA work best on systems using HT because the CPU core can make more rational decisions on use of the CPU's capacities? On the Q9300 of course, there is no HT so it is more all or nothing on the use of the core on the chip ... Not sure what to make of this yet ... but it is an interesting side note... I am still curious as to the load that exists on Linux with a 695 card with two GPU cores ... on the i7 it would take a full core to support the two GPU cores with the current application. The older application there would have been at least a few percentage "free" that another task could still run. ____________ | |
ID: 5875 | Rating: 0 | rate: / Reply Quote | |
The new windows application 6.61 certainly seems to sucking up huge amounts of CPU time. | |
ID: 5879 | Rating: 0 | rate: / Reply Quote | |
Well, I didn't announce anything, I just interpreted GDFs comments on 6.61. Seems like I may have been wrong regarding a possible speed up due to the higher priority.. but then it was a question to the guys already running it, not a statement. | |
ID: 5883 | Rating: 0 | rate: / Reply Quote | |
Lots of reported errors with 6.61. | |
ID: 5889 | Rating: 0 | rate: / Reply Quote | |
Why we have to use polling at all? Sure, the concept seems very ill-conceived, but that's actually what nVidia proposed for CUDA. Not sure I was trying to give the impression that it was ill-conceived. Sometimes the use of fast acting polling loop with a large amount of sleeping time are sometimes a good enough solution to the problem. Now that we have the realities and some experience I was asking the question so that maybe we can move in the direction were we are using a more sophisticated mode of operation. If you look closely I was asking broader questions than that. I know it CAN seem like carping if you are in the mood to read it that way. What I am doing is brainstorming what I can see, what I can infer and compare that with what I have experienced as a systems engineer ... Like the memory load. I have been pondering the memory load and the CPU load and I mused about the difference betwixt SaH and GPU Grid and speculated that the SaH CPU load was a lot lower. Of course that is likely because the whole of the SaH task can be loaded into VRAM and then the program executed. GPU Gird appears larger and thus we have load and store actions. THAT said, I also pondered on the faster cards with 1G or more VRAM if we could lower that CPU load if we used bigger slices of data, slices sized for the amount of available VRAM. Using IRQ and or a single polling task which dispatches to "grooming" tasks that do the other operations means that we only have one task that sucks up CPU while doing the polling of ALL of the GPUs and only when grooming needs to be done do we run the other threads. Operationally what would happen is that when BOINC Launched a task it would check to see if there was a polling task, if there was not, one would be launched ... if there was one, it would be used ... in other words, with one GPU we have two tasks running ... for two, three ... and for that lucky guy with 3 GTX 295s, 7 ... Anyway, all this in an attempt to make things better for all of us. Wasted CPU is wasted CPU ... if this IS the only way to get er done ... then we will live with it ... but, I think we should be wanting to look at all of the possibilities ... ____________ | |
ID: 5890 | Rating: 0 | rate: / Reply Quote | |
Lots of reported errors with 6.61. Sorry, wrong link. Correct link GDF's Response to issue | |
ID: 5893 | Rating: 0 | rate: / Reply Quote | |
Finally a version that runs fine with 4+1 Tasks (4 for CPU 1 for ACEMD). | |
ID: 5904 | Rating: 0 | rate: / Reply Quote | |
Finally a version that runs fine with 4+1 Tasks (4 for CPU 1 for ACEMD). YOu be lucky ... I just tried it and my results stay the same ... 22% CPU usage ... which means essentially one full core used only for baby sitting ... ____________ | |
ID: 5908 | Rating: 0 | rate: / Reply Quote | |
I can just point out: | |
ID: 5919 | Rating: 0 | rate: / Reply Quote | |
Hi Paul, | |
ID: 5924 | Rating: 0 | rate: / Reply Quote | |
Oh, and my 6.61 result for a 2479 credit WU, running BOINC 6.5.0 under XP 32 and 4+1: | |
ID: 5925 | Rating: 0 | rate: / Reply Quote | |
CPU time. Based on my current RAC of 5000 that would give me 5430 RAC. Not bad. But has your other project(s) suffered more than GPU has gained??? You rarely get anything for nothing! Phoneman1 | |
ID: 5926 | Rating: 0 | rate: / Reply Quote | |
CPU time. Based on my current RAC of 5000 that would give me 5430 RAC. Not bad. Yes ... And this is especially acute if GPU Grid is not a high interest project for you. I don't know, there are, what, 211 protean folding projects out there? If not that many, maybe only 197 ... :) Anyway, the question is not directly winners and losers on a project by project basis, but science as a whole ... The change in 6.55 to 6.61 increased by 50% the load on the CPUs, this means that for every hour of GPU Grid I lose half an hour for some other project just due to idle polling. I know they are working on it and I am NOT ranting ... I am just expressing a little frustration in the way things are working and though I have been dedicating more resources here than I normally would it is because this is a new technology I want to see get off the ground. THAT said, I will also say that at this time I like the project's responsiveness with ETA "G" whatsis name stopping by to see what is on our minds (if anything, sometimes mine is a blank) ... but I do not want it to be forgotten that "playing nice" is important ... Just as an example, if GPU Grid cannot get the CPU load down for their project tasks ... well I stop work for project all the time ... or reduce the share they get ... the only advantage they have at the moment is that they are the only science project around using the GPU on BOINC ... SETI@Home is not doing science, they are exploring, whole nother animal ... Anyway, in the last month I have gone from a 9800 GT to add a GTX 280 AND just now a GTX 295 ...which experience I will talk about in another thread soon ... ____________ | |
ID: 5928 | Rating: 0 | rate: / Reply Quote | |
Hi Paul, Hi ... No harm, no foul (Fowl?) ... I worry about communication sometimes because people read into what I write sometimes ... and I know I miss a lot of the detail because I am so literal minded ... a regular Sar Trek Mr. Spock type ... that is me ... The video cards do have an IRQ assigned to them which is why I asked about using the IRQs instead. And I forgot about your comment regarding one polling loop for all GPUs. Surely seems like a good idea, but may not be neccessary if they can get the cpu usage down as expected. The only problem I see with that approach is that in BOINC every task is insulated from the others. This is by design and it's actually a strength.. it lets you utilize multi cores rather easily and protects you from all kinds of strange interaction (which you usually get when you parallelize outside a core). One would have to program this carefully, but it seems manageable nevertheless. If they are going to continue to use a polling loop this is still a good idea. The point being that we reduce the number of poling loops to the minimum required, which is one. As to BOINC and task isolation this is not necessarily a show stopper. The first task starts, checks to see if there is a polling task alive, if not it starts on, otherwise it continues after regestering with the live polling loop. In between times it is servicing its GPU the task would sleep ... the polling loop would not strictly be a BOINC task, the BOINC task would be the task that you currently have sans the polling loop ... Regarding the VRAM: right now the WUs use about 70 - 80 MB of VRAM. More complex models could be simulated, which would require more memory, but then they time per step would go up and the screen would get really sluggish even on GT200 cards. I guess that's not what people want right now ;) Similarly if one told the GPU to compute several steps within one polling interval (to reduce the polling frquency) and I don't know if it would be possible at all. bottom line: I don't see potential for improvement by using more VRAM. Just a thought ... when I look at the application using one of the nvidia tools I would see about 50% memory load ... sometimes the way to reduce loads like that is to upload more stuff if there is room ... the bottom line is I want more efficiency ... :) ____________ | |
ID: 5929 | Rating: 0 | rate: / Reply Quote | |
When the beastly software gets out of control, don't fear to slay it. ;-) Meanwhile, I wonder why the Windows port of GPUGRID loads the CPU so... Today I see loads of 12% to 14% per core on a 2.4 GHz Phenom under Linux 2.6.28 and nvidia-kernel 180.22 with GPUGRID application version 6.59 and GTX 260 GPU hardware. Could it be that my GPUs are underutilized? Or is the Windows software that inefficient? | |
ID: 5931 | Rating: 0 | rate: / Reply Quote | |
Could it be that my GPUs are underutilized? You are now a witness to a great truth ... Microsoft does not actually write software that is all that good. :) Though the load of 12-14% is also pretty high as these things SHOULD be going. On a quad that is a 50% CPU load (25% of the total power of the system) ... On my windows quad I am seeing about 22% load and on the i7 about 7% per GPU task (I now have 3 GPU CPUs running, two in the new GTX 295 and the GTX 280) ... This is why i was suggesting a single polling loop for all GPU Grid processes with the processes registered with BOINC being the processes that groom the individual GPU core. I tried to get a GPU going on the linux box and I guess I am too stupid because I was not able to do so. When I D/L the driver package it told me I needed other things and it was just too much of a mess that I said the heck with it for the moment. I may play with it later, but I fell out of love with DOS style command lines back about the time I bought my Lisa ... Anyway, you are losing half the processing power of your core just to see if the GPU needs attention. the real cost should be about 1% ... anything above that is actually quite high. And if it is per GPU core, that 3% figure for my system for example, would mean that 6% would be really, really wasted overhead. I don't mind 3% overhead that much, but 3% times 3 cores is 9% of the system CPU in a idle detection loop ... not good ... So, I am back to my IRQ, single polling thread, or the ability to select "Nice" vs Performance for the GPU ... ____________ | |
ID: 5933 | Rating: 0 | rate: / Reply Quote | |
Got a 6.61 WU. 30-40% of CPU on Athlon X2. GTX 260. BOINC 6.5.0. It's terrible. | |
ID: 5934 | Rating: 0 | rate: / Reply Quote | |
Tried a few 6.61 units on my Athlon X2 3800+, 9500GT, BOINC 6.50, 32-bit Vista Home Premium. The GPU tasks took over one full core (never dropping below 47% of total CPU usage, going as high as 73%, and most typically in the range of 52-56%). With GPU task running, the machine is virtually unusable as it takes several seconds to switch between simple tasks such as a Mozilla web page tab and an already open BOINC manager. Unfortunately, I am forced to suspend crunching GPUGRID on this box at least until another app version is available. | |
ID: 5956 | Rating: 0 | rate: / Reply Quote | |
Something is going on with the project. | |
ID: 5959 | Rating: 0 | rate: / Reply Quote | |
Something is going on with the project. The project is still in the early days and this is a major complaint with the GPU Grid application. The CPU usage is much higher than desired / expected. The project is aware and is promising a newer version of the application that uses less CPU though that has not happened yet. The latest release 6.61 almost doubled the CPU useage over 6.55 where we had hoped that the opposite would happen, that the usage of 6.55 would be halved. The idea was that the GPU run times would be lowered. THough that has not been my experience to this point. Most of us agree that the theroetical ideal would be for negligible CPU use by the GPU tasks on the order of 1% to 3% of the core usage where the lowest I have experienced is 3% of total system resource usage. Sadly, we are stuck in the situation that if we want to keep the GPU spinning we have to pay for it with significant reductions in the productivity of the CPU projects we support. ____________ | |
ID: 5961 | Rating: 0 | rate: / Reply Quote | |
Scott, are your cpu usage numbers set to 100% = 1 core? Otherwise they would look strange. And the very slow responsiveness you're seeing is due to your relatively slow card (32 shaders vs recommended 50+). Nothing has changed here during the different client versions. But the WUs have changed, now we also crunch more complex models, which take even longer to process (and thus make the lag worse) | |
ID: 5962 | Rating: 0 | rate: / Reply Quote | |
Meanwhile, I wonder why the Windows port of GPUGRID loads the CPU so... No, your Linux should be fine! Trying a short explanation: the GPU crunches for some 10th of milliseconds and when it's done the CPU needs to intervene. If it doesn't the GPU runs dry. Under linux the project uses "nanosleep", don't know if it's a function or library or whatever.. anyway, it allows to send the cpu task into sleep mode and wake it up at a precisely controlled time. There is no nanosleep for windows and the underlying reason seems to be that the smallest time steps the windows scheduler knows is 1 ms, so the timing control is much less efficient (and the polling has to be more agressive, otherwise GPU performance suffers). MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 5964 | Rating: 0 | rate: / Reply Quote | |
Scott, are your cpu usage numbers set to 100% = 1 core? Otherwise they would look strange. Was writing the earlier posts fairly quickly before having to leave home to do some work. Anyway, the percentage usage is from the windows task manager processes list. And the very slow responsiveness you're seeing is due to your relatively slow card (32 shaders vs recommended 50+). Nothing has changed here during the different client versions. But the WUs have changed, now we also crunch more complex models, which take even longer to process (and thus make the lag worse) While the effects on former app versions was more evident on the slower card (vs. for example my 9600GSO), these were always relatively modest effects. Also, with 6.55 app versions, this box crunched a few of the different types (US, USPME, GPUTEST, JAN) of work with no real differences to note other than ms/time step and overall run time. It is only with the 6.61 app version that the machine is so badly affected. | |
ID: 5970 | Rating: 0 | rate: / Reply Quote | |
Well, here is an interesting twist: | |
ID: 5976 | Rating: 0 | rate: / Reply Quote | |
6.59 is the linux equivalent of 6.61 for win ;) Anyway, the percentage usage is from the windows task manager processes list. You have a dual core cpu, so the default windows behaviour should be to show 100% as both cores under full load and 50% for one core under full load. To exceed 50% one has to go multithreaded. When you say you se cpu usages of up to 73% that would mean 1 and a half core are used. Nobody else reported something like that before for GPU-Grid. And regarding the other issue: so it seems the different WU types are not causing your massive slow down and it looks like the new client is really to blame for this. MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 5996 | Rating: 0 | rate: / Reply Quote | |
6.59 is the linux equivalent of 6.61 for win ;) Ah...got it!
I have tried to test it to reproduce the more than 50% load, and haven't been able to do so. Maybe it was just something screwy with that particular workunit. I downloaded a new one and it has a load of 46-50% and is fairly constant. Unfortunately, it still has the continued slow-down problems. Also, I think you were correct in your initial assessment of more difficult work. The last unit I downloaded gives time to completion estimates that are twice what any of the pre-6.61 units were. Old work would averaged about 2.5% complete per hour on my 9500GT, but the current work is running at less than half that speed. | |
ID: 5998 | Rating: 0 | rate: / Reply Quote | |
They're trying to compensate for the increased model complexity by including less steps in each WU, so that the total completion time is more or less constant (12h on a 9800GT). This may or may not work well on different hardware, i.e. slower cards could have a disproportionately slower memory interface or something like that. But I guess a factor of 2 is not expected. | |
ID: 6001 | Rating: 0 | rate: / Reply Quote | |
...i.e. slower cards could have a disproportionately slower memory interface or something like that. But I guess a factor of 2 is not expected. I thought about that as a possibility, especially since the 9500GT is a 128-bit card. However, that should mean that on my 9600GSO (192-bit) I would see a more modest effect, but so far it seems to be no different than before the 6.61 apps (both have GDDR3, with the 9500GT having a bit faster memory clock). I am actually wondering if there could be a CPU pattern emerging. It seems that several of the above notes in this thread, including mine, are with AMD machines and not Intel's (indeed, my 9600GSO is in an Intel box). There are some significant architectural differences between those two, but I have no idea how they might play into differences in CPU - GPU interactions? | |
ID: 6002 | Rating: 0 | rate: / Reply Quote | |
I'd say the GPU-CPU interaction is, although time critical, rather limited. I.e. any differences would emerge from differences in the chipsets and drivers rather than the CPUs themselfs. | |
ID: 6003 | Rating: 0 | rate: / Reply Quote | |
While the effects on former app versions was more evident on the slower card (vs. for example my 9600GSO), these were always relatively modest effects. Also, with 6.55 app versions, this box crunched a few of the different types (US, USPME, GPUTEST, JAN) of work with no real differences to note other than ms/time step and overall run time. It is only with the 6.61 app version that the machine is so badly affected. V6.61 is a BIG step backwards here too. More CPU usage, noticeably poorer video responsiveness. Instead of improving, things are getting worse. V6.56 worked great on Win32, everything since (including the downgrade to v6.55) has been a step backward for the users IMO. | |
ID: 6004 | Rating: 0 | rate: / Reply Quote | |
I wish developers to be lucky and smart ot less than f@h founders! Wish GPUGRID luck in this New Year! | |
ID: 6018 | Rating: 0 | rate: / Reply Quote | |
So, what about "low CPU" app for Windowzz? | |
ID: 6048 | Rating: 0 | rate: / Reply Quote | |
Well, it appears to have definitely been the 6.61 app. The 6.62 app is now crunching away happily on the 9500GT at the same or better speeds than were observed with 6.55. The CPU load is around 3%, and the machine is fairly smooth in usage with the severe slowdown seen under 6.61 now absent. | |
ID: 6146 | Rating: 0 | rate: / Reply Quote | |
My machine is now again usable since v6.62. With v6.61 it was useless except for crunching. CPU usage has gone from 22% on the quad to 1%. Thanks, v6.62 is GREAT! | |
ID: 6149 | Rating: 0 | rate: / Reply Quote | |
6.62 is amazing...running in +/- 3-4 mins of Q9950 CPU time and 2x GTX 260 Core 216 Superclocked video cards. My CPU contributions are climbing again. 6.62 looks like a winner to me! Thanks GDF & Co. | |
ID: 6151 | Rating: 0 | rate: / Reply Quote | |
Message boards : Graphics cards (GPUs) : New application version