Message boards : News : On new fatty WUs
Author | Message |
---|---|
A batch of 500 WUs *_TRYP_* have been sent and will take TWICE as much to compute than normal.We are talking of 10 nanoseconds of simulation per WU! (for what most people still use big fat supercomputers). We need them for a very interesting publication we are preparing related to this experiment. Many thanks for the effort! | |
ID: 18415 | Rating: 0 | rate: / Reply Quote | |
So will take the 10 *meta_pYEEI*. | |
ID: 18416 | Rating: 0 | rate: / Reply Quote | |
Link dont work. Link are:http://www."http.com//www.youtube.com/watch?v=XnjmHYUvFW4%22 | |
ID: 18417 | Rating: 0 | rate: / Reply Quote | |
Link dont work. Link are:http://www."http.com//www.youtube.com/watch?v=XnjmHYUvFW4%22 You mean: http://www.youtube.com/watch?v=XnjmHYUvFW4 The sad thing about these work units is the 56% GPU usage (on a GTX 480) | |
ID: 18418 | Rating: 0 | rate: / Reply Quote | |
Link dont work. Link are:http://www."http.com//www.youtube.com/watch?v=XnjmHYUvFW4%22 To url: Yes. My gtx 460 gpu usage are 80 % | |
ID: 18419 | Rating: 0 | rate: / Reply Quote | |
I have one that was already 25% done when I saw this post and turned HT off, shaders turned up to 1756 (mv 1138)... bring on ALL 10 ms WUs!!! | |
ID: 18420 | Rating: 0 | rate: / Reply Quote | |
Different Task types utilize the GPU to different extents as the calculations differ. I have seen tasks from around 53% to as much as 93%. | |
ID: 18422 | Rating: 0 | rate: / Reply Quote | |
I've completed two new 'fatty' WUs on two different computers. | |
ID: 18428 | Rating: 0 | rate: / Reply Quote | |
I have one that was already 25% done when I saw this post and turned HT off, shaders turned up to 1756 (mv 1138)... bring on ALL 10 ms WUs!!! Oops. Sorry, i read wrong theard. :-) My WUs are IBUCH and these use 80% of gpu. | |
ID: 18429 | Rating: 0 | rate: / Reply Quote | |
I like this wu: | |
ID: 18430 | Rating: 0 | rate: / Reply Quote | |
It's not. | |
ID: 18432 | Rating: 0 | rate: / Reply Quote | |
ok | |
ID: 18439 | Rating: 0 | rate: / Reply Quote | |
I've completed two new 'fatty' WUs on two different computers. If you look at the difference between your Q9550 and Q6600 you will get an idea of the importance of the CPU; 40min less on the faster CPU. The different GPU clock rates are as a result of changing the clock speed. | |
ID: 18443 | Rating: 0 | rate: / Reply Quote | |
hello to all ;) | |
ID: 18446 | Rating: 0 | rate: / Reply Quote | |
If you look at the difference between your Q9550 and Q6600 you will get an idea of the importance of the CPU; 40min less on the faster CPU. I realized this difference. I concluded this project should renamed to GPUCPUgrid :) I have an idea: GPUgrid should be a multi CPU threaded project. In this way a GPUgrid task would get the sufficient CPU power when needed, because GPUgrid runs at higher priority than my other project (Rosetta). Maybe we shouldn't dedicate one CPU core to each GPU then. But if we have to, and the credit gain is more than my other CPU project can earn, I will detach from the other project. As far as I can recall, other kinds of WUs do not depend on the CPU speed to this extent. | |
ID: 18448 | Rating: 0 | rate: / Reply Quote | |
hello to all ;) The apps are mostly compiled using cuda (2.3 from memory). The fact that you have a more up to date driver that supports up to cuda 3.1 means its also backwards compatible. The main changes for cuda 3.1 are targeted to the fermi-based cards (GTX4xx) and don't offer any advantage to the older cards such your GTX260. There is no reason to recompile them for cuda 3.1 if its not going to use any of the new features and still work on older cards. There is a cuda 3.1 app for the fermi-based cards. There are only a few wu available at the moment as the project is switching to a new type of work unit. See the news item on the front page for more details about that. The specific limit of 2 wu is so the project can get results back quickly. They use the results from one set of wu to generate the next set, so unless they come back quickly it holds things up. It also stops people hogging the work units and allows more people to get some. As for optimising the guys do that regularly. If you run a tool call GPU-Z it will show you how hard the graphics card is working. Typically this is quite high. As they say on the home page we are doing work that normally is run on a super computer. ____________ BOINC blog | |
ID: 18450 | Rating: 0 | rate: / Reply Quote | |
You should adjust bonus lines accordingly if you know something like this is coming. The 'sophisticated' BOINC management makes sure at times my 9800GT will get those monster WUs while the GTX 295 crunches usual WUs. | |
ID: 18452 | Rating: 0 | rate: / Reply Quote | |
OK, thank you Mark ! Hi everybody, | |
ID: 18455 | Rating: 0 | rate: / Reply Quote | |
I completed one of the _revlo_TRYP_ Work Units on my GTX470: | |
ID: 18457 | Rating: 0 | rate: / Reply Quote | |
ignasi, | |
ID: 18476 | Rating: 0 | rate: / Reply Quote | |
I've completed two new 'fatty' WUs on two different computers. This is not 1.8M steps. It's still 2.5M steps but the last restart till it finished was of 1.8M steps. (maybe the message could be clearer). gdf | |
ID: 18477 | Rating: 0 | rate: / Reply Quote | |
OK, thank you Mark ! The main problem with OpenCL is it doesn't have the library support and developer tools that CUDA has. CUDA has been around longer. Eventually that will change. There is an FFT libray supplied with CUDA that a number of the projects make use of. From what I gather there isn't an equivilent one with OpenCL. From the project point of view OpenCL would be the way to go as you only need one app (however you have to compile with each manufacturers compiler to work on that brand). The coding is different so the apps that have been developed need to be rewritten to work under OpenCL (which all takes time and effort). It will get there eventually, it just takes a long time for OpenCL to catch up to CUDA. ____________ BOINC blog | |
ID: 18478 | Rating: 0 | rate: / Reply Quote | |
This is not 1.8M steps. It's still 2.5M steps but the last restart till it finished was of 1.8M steps. (maybe the message could be clearer). Oh, now it's perfectly clear. Thank you. I overclocked my CPU by 20% since then, and the GPU usage had risen 10%. So I guess it would be better to upgrade my CPU (+MB and RAM of course) to achieve higher GPU usage. | |
ID: 18479 | Rating: 0 | rate: / Reply Quote | |
Have all the Fatty wu's been sent out & how big are the files that get uploaded? We are talking of 10 nanoseconds of simulation per WU I don't understand. 10 nanoseconds of simulation per WU isn't very long at all. how come the Fatty units are taking twice as long? | |
ID: 18488 | Rating: 0 | rate: / Reply Quote | |
I thought most of the 500 had been sent and returned but there are a few still available/running. | |
ID: 18489 | Rating: 0 | rate: / Reply Quote | |
Each of those 500 will be reissued 10 times to achieve 100 ns of total simulation time each. | |
ID: 18490 | Rating: 0 | rate: / Reply Quote | |
Ignasi, | |
ID: 18496 | Rating: 0 | rate: / Reply Quote | |
Ignasi was abroad, but what error are you referring to? | |
ID: 18498 | Rating: 0 | rate: / Reply Quote | |
gdf, | |
ID: 18501 | Rating: 0 | rate: / Reply Quote | |
I'm curious. I'm currently crunching TRYP WU 1859786 which I got at the end of the server outage. Its info page says its's not a resend but a first-time d/l. Is this part of the original batch of 500 or a new batch? | |
ID: 18523 | Rating: 0 | rate: / Reply Quote | |
To quote GDF, "Each of those 500 will be reissued 10 times to achieve 100 ns of total simulation time each". | |
ID: 18526 | Rating: 0 | rate: / Reply Quote | |
Skgiven & Ignasi, | |
ID: 18546 | Rating: 0 | rate: / Reply Quote | |
@ftpd | |
ID: 18547 | Rating: 0 | rate: / Reply Quote | |
Motivated to make a systemic suggestion here. Thanks, | |
ID: 18550 | Rating: 0 | rate: / Reply Quote | |
That's a funny error - it looks like a driver bug affecting GTX9800 (which is what a GTS250 is, too). What driver version do you have? | |
ID: 18553 | Rating: 0 | rate: / Reply Quote | |
Well, if it’s a driver bug then the same bug in driver 19713 is still there in driver 25896, and effects both the 9800 GT (511MB) and the 1GB version of the GTS250, suggesting this one is not related to memory size. | |
ID: 18554 | Rating: 0 | rate: / Reply Quote | |
@skgiven, | |
ID: 18556 | Rating: 0 | rate: / Reply Quote | |
Hi Ton, At that time only gpugrid was running, gts250 can only do one job.Thanks, I was just trying to determine if Collatz or other GPU tasks were overwriting your GPUGrid WU memory, can occasionally happen if you run two GPU projects on the same computer. They don’t run at the same time, but Boinc can stop one running, and suspend it in memory, to let the other run. Sometimes when this happens it can corrupt the task, but I doubt that this is the case (you only ran one Collatz task on that GPU, probably when there was a task shortage). RNA World was running using CPU.Unless it was running Beta tasks it would not mess up the GPUGrid task. If you had lots of RNA failures on the same system, then it would be fair to say it could mess up the system and cause problems for GPU tasks (as they also have to use the CPU too). Later that week two other WU also were aborted. HIVPR?Well, I expect in your case you were not video editing or playing a computer game, or you would have said. So I think so. I see you are having to User Aborting these tasks, and I can fully understand why, they all crashed on that card. A real pain for the cruncher to sit over a system that way. I'm not keen on this situation. As Matt said, there is a driver "related" bug: (SWAN : FATAL : Failure executing kernel sync [transpose_float2] [700] Assertion failed: 0, file swanlib_nv.cpp, line 121) I just wanted to make sure there was not something else. However the same problem has been raised in two other threads, http://www.gpugrid.net/forum_thread.php?id=2274 http://www.gpugrid.net/forum_thread.php?id=2278 We can’t do any more than ignasi did - inform the developers, and throw in a few more suggestions, if only to treat the symptoms. The researchers have been working on new applications for a while now, 5% faster last I heard. Hopefully they can find a work around for this, as well as make the app much faster for GTX460’s and so on. | |
ID: 18570 | Rating: 0 | rate: / Reply Quote | |
@skgiven, | |
ID: 18571 | Rating: 0 | rate: / Reply Quote | |
I've discovered a new type fatty WU on one of my computers called *_IBUCH_1_pYEEI_long_*. It's running for 4 hours 15 minutes and completed 55% (GTX 480 @ 800MHz, 63% GPU usage, SWAN_SYNC=0, C2Q 9550 @ 3.71GHz). | |
ID: 19719 | Rating: 0 | rate: / Reply Quote | |
I've discovered a new type fatty WU on one of my computers called *_IBUCH_1_pYEEI_long_*. It's running for 4 hours 15 minutes and completed 55% (GTX 480 @ 800MHz, 63% GPU usage, SWAN_SYNC=0, C2Q 9550 @ 3.71GHz). It's finished in 7 hours and 47 minutes. Details here. | |
ID: 19722 | Rating: 0 | rate: / Reply Quote | |
I had a _pYEEI_long_ back in June, and posted about it. Nobody took any notice then, either. | |
ID: 19724 | Rating: 0 | rate: / Reply Quote | |
Only have 145KASHIF_HIVPR_no_boundxxxx types and take 3:30 hours, depending on de # of steps, some take longer, 4 hours. CPU is an C2EX9650 @ 3.5GHz. | |
ID: 19725 | Rating: 0 | rate: / Reply Quote | |
These are different tasks from those released in June; even different apps. | |
ID: 19730 | Rating: 0 | rate: / Reply Quote | |
These are different tasks from those released in June; even different apps. That's great, but now they came out of the blue. They may include some of the new subroutines being worked on (as with the GIANNI_DHFR1000), intended to increase turnover, or they might just need to get through more work, for a paper. Perhaps these are over-taxing the Fermi controller when the cards are overclocked, but they seem to work on GT200 series cards that are OC'd. These *_IBUCH_1_pYEEI_long_* WUs not as stressing to the card as the GIANNI_DHFRs, they run at 30% lower GPU usage, therefore generating much less heat so overclocking is not an issue with those. BTW I've installed a larger cooler (Arctic Cooling Xtreme Plus + WR004 kit) on each of my GTX480s, so the larger heat dissipation of GIANNI_DHFRs is not an issue anymore, even when overclocked to 800MHz. The double length is clearly just to annoy me, I have 6 GT240's. Not just to you... I was shocked when I saw that it's been running for more than 4 hours, and it's barely completed the half of the WU. The first thing came to my mind was that my brand new coolers didn't do their job so well, and the card's been locked up. But then I saw the progress indicator is running, and it's eased my mind a little, so I could read the name of the WU :) So the "_long_" suffix was the right explanation for the long running time. I've posted about it to warn my fellow crunchers, once the crew didn't. BTW some *_IBUCH_GCY_* WUs are appearing as well, those are very long too (about 5h30m to complete on my GTX480s) | |
ID: 19732 | Rating: 0 | rate: / Reply Quote | |
New type of long WUs been arrived: *-IBUCH_?_BCC_long_* | |
ID: 19743 | Rating: 0 | rate: / Reply Quote | |
p9-IBUCH_1_pYEEI_long_101130-2-10-RND3169_0 | |
ID: 19744 | Rating: 0 | rate: / Reply Quote | |
New type of long WUs been arrived: *-IBUCH_?_BCC_long_* No problem on your foureighty. I have the half of that: a twofourty and 1 BCC_long running. Estimated runtime: 33 hours and 50 minutes and a 30% CPU-use. GPU-load: 62% Fan-speed: 41% Temperature: 63C. | |
ID: 19745 | Rating: 0 | rate: / Reply Quote | |
I had a _long_ IBUCH task on a GT240 (32h), and another aet1_IBUCH task that was also equally lengthy (33h). aet :) | |
ID: 19746 | Rating: 0 | rate: / Reply Quote | |
Dear all, | |
ID: 19748 | Rating: 0 | rate: / Reply Quote | |
Is there a way to tell the server not to send me any of the new fatty workunits? My 9800 GT, running nearly 24/7, can complete the normal workunits in time, but not any workunits that try to take twice as long with the same allowed time to completion. Even now, about half of the workunits it tries complete with computation errors. | |
ID: 19757 | Rating: 0 | rate: / Reply Quote | |
Bring em on ... I love when you guys push the limits ... I'm working on a IBUCH_RED right now!!! | |
ID: 19761 | Rating: 0 | rate: / Reply Quote | |
Remember to keep your cache level low; no point in having a task sitting waiting for half a day before it starts. | |
ID: 19764 | Rating: 0 | rate: / Reply Quote | |
My cache level was already low enough to delay getting a new workunit to perhaps 1 hour before the last one finishes. | |
ID: 19767 | Rating: 0 | rate: / Reply Quote | |
One of the big ones cancelled after 18 hrs on GTX295. | |
ID: 19771 | Rating: 0 | rate: / Reply Quote | |
Dear all, Just completed one of your test WUs - hope this is what you expected to see. | |
ID: 19773 | Rating: 0 | rate: / Reply Quote | |
We'll do it through a command we have never used that will be tested over the weekend.It looks like the normal 50% bonus was applied. Not sure if the never used before command will be run subsequent to the task completion. I guess this will be a manual test. Richards task, Claimed credit 127.270717592593, Granted credit 190.906076388889. We will see if these numbers change. | |
ID: 19776 | Rating: 0 | rate: / Reply Quote | |
Remember to keep your cache level low; no point in having a task sitting waiting for half a day before it starts. I do not understand the cache level issue. You get two WU's and that is all. One computes and one is in cache. There is no way to change that it seems to me. On WCG I have a 3 day cache. But on GPUGrid I have not seen any cache management function. | |
ID: 19780 | Rating: 0 | rate: / Reply Quote | |
Remember to keep your cache level low; no point in having a task sitting waiting for half a day before it starts. While you could do it here, this is a Boinc wide setting and will apply to all projects. As local settings override this you might as well just set it on Boinc Manager, Advanced Preferences, Network Usage, Additional Work Buffer, set to 0.1days or 0.01 if you prefer. As you can see there is no GPU tab or separate GPU cache setting on Boinc and the default cache setting of 0.50days is bad for this project. For a high end Fermi you should still get away with a 3day cache, if the GPU is optimized for performance (Free CPU core/thread, swan_sync=0, always use GPU...); you can only have one running task (per GPU) and one queued task, no matter how many days cache you set, so if your high end Fermi finishes a task in under 10h then even the queued task will finished in 20h. So 0.5days or 9days makes no odds. You just have to consider the consequences of a 4h local Internet outage, system restarts and the effect of your use on the system. On the other hand for an entry level GPU such as a GT240, these long tasks take around 30 to 32h, so if you have a high cache the queued task will have sat waiting to run for 32h before starting, and therefore will not finish until 64h after it was sent. So you are better off with a 0.01day cache. Ideally tasks get returned within 24h and if not then within 48h. Beyond that you are late and your efforts are of less value to the project. After about 4 days your efforts become worthless to the project and might even slow it down. After 5days you don't even get any credit. Personally I would prefer this credit cut-off was 4days, to better reflect the reality of the situation. | |
ID: 19783 | Rating: 0 | rate: / Reply Quote | |
Remember to keep your cache level low; no point in having a task sitting waiting for half a day before it starts. @skgiven: pls check your PM. | |
ID: 19784 | Rating: 0 | rate: / Reply Quote | |
After about 4 days your efforts become worthless to the project and might even slow it down. After 5days you don't even get any credit. Personally I would prefer this credit cut-off was 4days, to better reflect the reality of the situation. My 7-year old desktop finally hit the dust, and now I have a replacement with a CUDA-card! So without reading the small print I attached to gpugrid. 4½ days later my WU completed - and now I learn that I'm probably wasting the projects time. So maybe it's good idea to shorten the deadline to prevent GT218-hopefuls like me getting in the way :-) | |
ID: 19805 | Rating: 0 | rate: / Reply Quote | |
Your NVidia ION only has 16 cuda cores, it is well short of the recommended minimum of 96 cuda cores. | |
ID: 19807 | Rating: 0 | rate: / Reply Quote | |
For this, we will grant 40% BONUS in credits with the *long* WUs. We'll do it through a command we have never used that will be tested over the weekend. If successful, we will use it for ALL NEW *long* SUBMISSIONS starting NEXT WEEK. Looks like those never used before commands are still not working. I've just completed another *_long_* monster wu, normal 50% bonus been applied on it. | |
ID: 19881 | Rating: 0 | rate: / Reply Quote | |
Looks like those never used before commands are still not working. Hi Retvari Zoltan, This WU you are linking is actually a from a batch that is two weeks old, about to finish: 9 of 10 steps completed. By new ones submitted this week we meant brand new long batch submissions. Which I am about to do in the next couple of days. Thanks, ignasi | |
ID: 19887 | Rating: 0 | rate: / Reply Quote | |
By new ones submitted this week we meant brand new long batch submissions. Ok. I can't wait to meet those :) | |
ID: 19890 | Rating: 0 | rate: / Reply Quote | |
First batch of 500 is out. | |
ID: 19900 | Rating: 0 | rate: / Reply Quote | |
First batch of 500 is out. Negative. We encourage you to discontinue any IBUCH_AcPYEEIP_long. They will compute well but the experiment is not correct. My fault. I appologise. | |
ID: 19901 | Rating: 0 | rate: / Reply Quote | |
We have cancelled all UNSENT WUs for the wrong batch. | |
ID: 19902 | Rating: 0 | rate: / Reply Quote | |
First batch of 500 is out. I didn't mean to rush you in any way. I've aborted 2 of those. | |
ID: 19905 | Rating: 0 | rate: / Reply Quote | |
Message boards : News : On new fatty WUs