Message boards : Number crunching : NATHAN_FAX3 and FAX4 discussion
Author | Message |
---|---|
These NATHAN_FAX3 workunits take really long time to finish. | |
ID: 23855 | Rating: 0 | rate: / Reply Quote | |
My overclocked (850MHz) GTX 580 (supported by a Core i7 980X running at 32*134MHz (4.288GHz)) finished its first NATHAN_FAX3 in 12 hours 55 minutes and 20 seconds. | |
ID: 23858 | Rating: 0 | rate: / Reply Quote | |
These NATHAN_FAX3 workunits take really long time to finish...Don't be surprised if a NATHAN_FAX3 takes more than 30 hours to finish on a lesser card (GTX 460, GTX 560) Yep...My stock GTX 570 is almost 13 hours into this bad boy and the BOINC manager shows 54% completion. Bye-bye time bonus? ____________ | |
ID: 23859 | Rating: 0 | rate: / Reply Quote | |
They look closer to 4times as long. Perhaps slightly protracted, but better than 4h! | |
ID: 23861 | Rating: 0 | rate: / Reply Quote | |
Yep...My stock GTX 570 is almost 13 hours into this bad boy and the BOINC manager shows 54% completion. Bye-bye time bonus? These NATHAN_FAX3 workunits could run much faster on your PC if you set the SWAN_SYNC=0 environmental setting, and free up one CPU core. | |
ID: 23863 | Rating: 0 | rate: / Reply Quote | |
They look closer to 4times as long. Perhaps slightly protracted, but better than 4h! Yes, the NATHAN_FA series has 500.000 steps, while the NATHAN_FAX series has 2.000.000 steps. That's exactly 4 times. But it's not the end of the story, because the time per step of the NATHAN_FAX is also slightly higher (23.2ms) than NATHAN_FAn's (21.8ms). | |
ID: 23867 | Rating: 0 | rate: / Reply Quote | |
These NATHAN_FAX3 workunits could run much faster on your PC if you set the SWAN_SYNC=0 environmental setting, and free up one CPU core. Thanks, RZ. I'll free up one CPU and see how it goes. Is SWAN_SYNC applicable in Windows hosts? ____________ | |
ID: 23869 | Rating: 0 | rate: / Reply Quote | |
Is SWAN_SYNC applicable in Windows hosts? Yes, it still seems to make some difference on Windows, but probably more for Linux (different apps). Remember to free a CPU core/thread (for each GPU), or it makes no difference. ____________ FAQ's HOW TO: - Opt out of Beta Tests - Ask for Help | |
ID: 23871 | Rating: 0 | rate: / Reply Quote | |
Is SWAN_SYNC applicable in Windows hosts? Sure. Start button -> type systempropertiesadvanced in the search box, press enter -> press the environmental variables button near the bottom of the window -> press the new button near the buttom (under the "system variables") -> type swan_sync to the upper box (name), and 0 (zero) to the lower box (value) Press OK three times. After that, you need to restart the BOINC client (stop the scientific applications on close), or you can restart Windows. | |
ID: 23872 | Rating: 0 | rate: / Reply Quote | |
Our times: | |
ID: 23873 | Rating: 0 | rate: / Reply Quote | |
Okay... I set swan_sync=0 and have freed up a CPU. The acemd process is now using a full core (25% on my C2Q) and about 230MB RAM. Is that what I should expect? | |
ID: 23875 | Rating: 0 | rate: / Reply Quote | |
Okay... I set swan_sync=0 and have freed up a CPU. The acemd process is now using a full core (25% on my C2Q) and about 230MB RAM. Is that what I should expect? Yes. Our 1 GPU task consumes about 97% of one cpu core (i7-2600k). | |
ID: 23876 | Rating: 0 | rate: / Reply Quote | |
On my GTX 570: 23-plus hours, 95K. It was reported 29.5 hours after it was delivered, so I guess I didn't get the time bonus... :P | |
ID: 23885 | Rating: 0 | rate: / Reply Quote | |
I just heard a tiny scream from my GTX 260. | |
ID: 23890 | Rating: 0 | rate: / Reply Quote | |
Also, GPU usage is below 90% again, while the old Nathans finally got up to 97% on a Fermi (GTX 570). | |
ID: 23893 | Rating: 0 | rate: / Reply Quote | |
First NATHAN_FAX3 unit completed here and it took almost 40 hours with a GTX 560 Ti (384 cores). Bit too much IMO. Don't want to free a CPU-core in combination with the swan setting since I run a CPU project as well. | |
ID: 23895 | Rating: 0 | rate: / Reply Quote | |
wow 115MB upload? This will kill my Mobil connection i think, where three BOINC Machines are connected on some bad connection O.o | |
ID: 23897 | Rating: 0 | rate: / Reply Quote | |
First NATHAN_FAX3 unit completed here and it took almost 40 hours with a GTX 560 Ti (384 cores). Bit too much IMO. Don't want to free a CPU-core in combination with the swan setting since I run a CPU project as well. I run CPU project too. Freeing cores doesn't mean CPU task are going to slow down. In fact I experienced that they are running faster. The reason is that the CPU with so much work get stucked and slows down (slowing GPU too). Freeing cores don't mean they are not going to be used!! It means less task are being exectuted at a time. 8 cores doesn't necessary means x8. You have to consider too that GPU usually takes 0,2-0,4 CPU but when the CPU is not been in use (because is waiting the GPU to return results) will make the others CPUs task run faster. This task took 31,64 hours on a GTX 560 Ti without SWAN_SYNC but freeing 2 cores. Now I chaged it to experience the difference. ____________ HOW TO - Full installation Ubuntu 11.10 | |
ID: 23900 | Rating: 0 | rate: / Reply Quote | |
I just heard a tiny scream from my GTX 260. lol...maybe your 260 can get with my 275s to form a PTSD support group... | |
ID: 23901 | Rating: 0 | rate: / Reply Quote | |
@Damaraland: I guess you have an Intel CPU with hyper threading switched on. If you free up one HT core, the other process on the shared core will speed up. But at the moment I run 4 PSP-PRP tasks, this project benefits too little from HT so it is switched off. | |
ID: 23902 | Rating: 0 | rate: / Reply Quote | |
I think one should get the bonus anyway whether you finish these behemoth tasks within 24 hours or not... +1 ____________ | |
ID: 23903 | Rating: 0 | rate: / Reply Quote | |
Well, the posted estimate for long runs is 8-12 hours for the fastest cards. This was a gross overestimate until now, as my 260 would complete most of them in 8 hours, and some of the biggest ones would take 12-15 hours. Now it seems we've gone in the other direction, taking well over half a day to complete on 5xx series cards. If the size of the tasks is putting such a demand even on the most powerful cards, it seems reasonable that the bonus cutoffs would be extended a bit, maybe to 36/72 hours for 50%/25%, or something along those lines. Obviously you don't want to go crazy with bonus points, but it's something to consider if we're going to be modeling larger molecular systems that take considerably longer to complete. | |
ID: 23906 | Rating: 0 | rate: / Reply Quote | |
I've had 2 NATHAN_FAX3 tasks complete. Both tasks were run with a dedicated core. One task was run without SWAN_SYNC being set, and the other task was run with SWAN_SYNC=0 (restarted BOINC after the change). The results were the same though. | |
ID: 23908 | Rating: 0 | rate: / Reply Quote | |
there does seem to be some credit disparity for the FAX3 wu in terms of time vs credit rewarded but im not sure. fair enough, so long as the work keeps coming! | |
ID: 23909 | Rating: 0 | rate: / Reply Quote | |
Perhaps we need a short, medium and long queues. All the current longs except FAX3 go to the medium queue. We'll need an option to select which type of work to allow as well. Deselect the long queue by default and put a note next to the long queue option to suggest GTX570 or better. | |
ID: 23910 | Rating: 0 | rate: / Reply Quote | |
Perhaps we need a short, medium and long queues. All the current longs except FAX3 go to the medium queue. We'll need an option to select which type of work to allow as well. Deselect the long queue by default and put a note next to the long queue option to suggest GTX570 or better. I proposed something similar. Response: out of the question because maintenaince costs (it seems with the team they have they can only handle 3 queues. It may also be possible to limit the long queue to certain speed cards by checking the est flops and memory values returned in the scheduler request (assuming it does pass it across). I agree with this. I proposed it too. I think the program that distributes the task should me smarter. It wouldn't need to look in the Flops, I think the best would be that the server just look the computer configuration and determine if it's slow or fast. ____________ HOW TO - Full installation Ubuntu 11.10 | |
ID: 23912 | Rating: 0 | rate: / Reply Quote | |
Also, GPU usage is below 90% again, while the old Nathans finally got up to 97% on a Fermi (GTX 570). look over to primegrid and the Genefer World Record tasks. I finished two tasks on my gtx 580 and both of them took easily over 75 hours. Thats long and part of why i switched back for a bit gpugrid. (And to get the next batch here ;-)) ____________ | |
ID: 23916 | Rating: 0 | rate: / Reply Quote | |
Target acquired. FAX3!! | |
ID: 23919 | Rating: 0 | rate: / Reply Quote | |
41hrs in and still have 21 to go?!?!?!?! gts450, core 980, shader 1960, ram 2050 hooked upto a 9950be with slight overclock. had just switched over to long wu's, regular long wu's ran 1~3hrs longer than short ones till i got this fax3. miffed at the idea that im loseing the 24hr bonus plus losing credits from the wu's i could have ran waiting for this to finish. | |
ID: 23920 | Rating: 0 | rate: / Reply Quote | |
34 hours with my GTX275 @633 1134 1521 (shaders slightly over-clocked). | |
ID: 23924 | Rating: 0 | rate: / Reply Quote | |
Would be nice to build a table of GPU that are qualified for the long queue in it's new configuration. We don't know their exact model numbers yet :) | |
ID: 23926 | Rating: 0 | rate: / Reply Quote | |
SWAN_SYNC doesn't make any effect. In both cases I freed 2 cores out of 8. Linux 3.0.0-16-generic. i7-2600K CPU @ 3.40GHz | |
ID: 23927 | Rating: 0 | rate: / Reply Quote | |
For the new apps I have not tested the benefit of using SWAN_SYNC. I will start now for Windows, but it could do with being tested for Linux as well (especially). Performance differences may fluctuate by task type, so several task types would need to be looked at. Remember to restart for the changes to be applied, and to use capitals for Linux. | |
ID: 23928 | Rating: 0 | rate: / Reply Quote | |
Pff O.o Over 5! times more computingtime but only ~2,8time more credits O.o Kick it over 100k and IĀ“m back happy ^^ Little disapointing, i had fun with the "new" 285 for only 2 days :( | |
ID: 23929 | Rating: 0 | rate: / Reply Quote | |
got a FAX4 wu just now. is there any significant difference between this and a FAX3? | |
ID: 23950 | Rating: 0 | rate: / Reply Quote | |
My GTX260 completed the FAX3 WU in 36.5 hours. The FAX4 it's working on now looks like it will finish in around 22.5 hours. I think upload is still going to put it over the 24h limit, but anyone with a newer card should probably be okay now. | |
ID: 23967 | Rating: 0 | rate: / Reply Quote | |
I think upload is still going to put it over the 24h limit, but anyone with a newer card should probably be okay now. Still figuring out what I did wrong with SWAN_SYNC (it's off now) Because upload didn't make 24h on GTX 260, but I did on GTX 560 GTX 260 22,7 h FAX4 GTX 560 Ti 19,82 h FAX4 ____________ HOW TO - Full installation Ubuntu 11.10 | |
ID: 23968 | Rating: 0 | rate: / Reply Quote | |
Hi all, | |
ID: 23970 | Rating: 0 | rate: / Reply Quote | |
The researchers stated that they will actively review task sizes. If for example they see higher failure rates they will most likely make changes to reduce runtime. In the mean time if any crunchers don't like the duration or experiences failures, crunch some of the normal tasks. While credit will be lower you will get badges for contributing to different research papers ;) | |
ID: 23971 | Rating: 0 | rate: / Reply Quote | |
On my GTX 260 it's looking like it's going to take about 50 hours to complete a NATHAN_FAX WU. Typically long runs would take 20 - 25 hours. | |
ID: 23972 | Rating: 0 | rate: / Reply Quote | |
I get a FAX3 done in just under 45 hours on my GTX 460 and a FAX4 in 28 hours. A bit too big for my taste... No chance of a time bonus here. Until now I always got it... | |
ID: 23975 | Rating: 0 | rate: / Reply Quote | |
There are two time bonuses; one for <24h (50%) and one for <48h (25%). | |
ID: 23978 | Rating: 0 | rate: / Reply Quote | |
There are two time bonuses; one for <24h (50%) and one for <48h (25%). I'm not really bothered by all this credit/time wise. However, if this project doesn't want to lose others perfectly good crunching machines I would raise the credits massively. On my overclocked GT460's on long wu's I used to get around 1 cobblestone for 1 second computing and now with Nathan WU's on the same cards I am only getting 1 cobblestone for 2 seconds computing. It seem there is serious credit deflation on this project and thet isn't good for the amount of computing power your going to get and keep. There also seems serious discrimination against machines with cards that aren't top rank. I think attitudes need to be rebalanced with requirments on this project. None of this matters much to myself as I am more concerned that enough power is left over to run other things. ____________ Radio Caroline, the world's most famous offshore pirate radio station. Great music since April 1964. Support Radio Caroline Team - Radio Caroline | |
ID: 23981 | Rating: 0 | rate: / Reply Quote | |
By my point of wiev there is much bigger difference between the longest a and the shortest GPU tasks on Seti (runnig KWSN package: 2x on GPU CC1.1, 3x on GPU CC1.3 and on Fermi GPU makes me no sense to waste cycles and electricity for Seti), than in long queue on GPUGRID. On the other you side are right comparing to DistrRTgen or PG PSA manual tpsieving, which are giving great credit (5x - 8x more than NATHAN_CB1 series). | |
ID: 23987 | Rating: 0 | rate: / Reply Quote | |
I run long tasks on a GTX 460 and have come to the conclusion that it is better for me to abort these FAX3 and FAX4 units.... | |
ID: 24005 | Rating: 0 | rate: / Reply Quote | |
The method that I use to estimate completion times is this: | |
ID: 24054 | Rating: 0 | rate: / Reply Quote | |
For GPUGrid tasks, the time taken and % completed, can be used to accurately measure the total time. So your formula is sound. If after 1h a task is 10% completed, then it will take 10h to complete the task. | |
ID: 24056 | Rating: 0 | rate: / Reply Quote | |
The researchers stated that they will actively review task sizes. If for example they see higher failure rates they will most likely make changes to reduce runtime. In the mean time if any crunchers don't like the duration or experiences failures, crunch some of the normal tasks. While credit will be lower you will get badges for contributing to different research papers ;) Adding to this comment from skgiven, I just want to reiterate that indeed we are evaluating many factors with respect to these long work units. Your comments/complaints are not falling on deaf ears, so keep them coming. Additionally, we are looking at ways to reduce the size of uploads for the users, thought it is not clear at this point if we will be able to do that. Please be patient with the implementation of changes/improvements. Betting slip, nenym: Interesting to hear your analysis about how we compare to other projects and long vs short. I don't think we're interested in getting into a credit war with anyone, but you're absolutely correct that it's important for us to remain relevant. | |
ID: 24064 | Rating: 0 | rate: / Reply Quote | |
Additionally, we are looking at ways to reduce the size of uploads for the users, thought it is not clear at this point if we will be able to do that. Please be patient with the implementation of changes/improvements. I tried to compress one of the workunits and it compressed by a factor of ~5:1. Maybe the project is already set up to take advantage of BOINC's built in compression (http://boinc.berkeley.edu/trac/wiki/FileCompression) but I can't tell from the logs on my end. If you're not doing any compression then it looks like there's a considerable benefit to be had. | |
ID: 24067 | Rating: 0 | rate: / Reply Quote | |
Your comments/complaints are not falling on deaf ears, so keep them coming. Additionally, we are looking at ways to reduce the size of uploads for the users, thought it is not clear at this point if we will be able to do that. Please be patient with the implementation of changes/improvements. Here goes my point of view: 1) The easiest way: put 4 queues of work long (>12 h), medium (12-8), short (<8) and beta. Everybody happy and the system prepared to new Keppler cards. In this scenario I might choose short & medium, but anyone could decide upon their % dedication on this project and hardware. 2) The advaced way. Make a decent Job Planner. This might consume a lot of ressources, but I see it profitable. http://boinc.berkeley.edu/trac/wiki/SchedMatch I can't believe that BOINC guys didn't work on this before with the huge differencies in hardware. Maybe the system is too adapted for SETI???? ____________ HOW TO - Full installation Ubuntu 11.10 | |
ID: 24073 | Rating: 0 | rate: / Reply Quote | |
Additionally, we are looking at ways to reduce the size of uploads for the users, thought it is not clear at this point if we will be able to do that. Please be patient with the implementation of changes/improvements. This may also reduce or eliminate the problems that I am having in uploading finished WUs. If the completed task upload files are not compressed, this could explain why WUs only upload after multiple retries. (Please note that I am not the only GPUGRID cruncher that is experiencing this exact problem - please see this thread.) I think I can reasonably say that I have done everything from my end that is possible to resolve this problem, however, I was unsuccessful at resolving it. I have had prior experience in uploading / downloading uncompressed files. My experience is that uncompressed files are prone to transmission error. IP will retry transmission, and if over a dubious connection like the one that I seem to have between my ISP and the GPUGRID server, retries may further exacerbate the problem. In addition, this would cut the size of the upload files considerably, and lessen the time that it takes to transmit them. As such, I am highly in favor of having completed WU upload files compressed. ____________ | |
ID: 24123 | Rating: 0 | rate: / Reply Quote | |
Hello, i pushed my GTX 460 to the limits, full voltage, at 900 chipset clock, hopefully it wont burn up, hopefully it will fall into the first 24 hour bonus time. It will be very very close to 24 hours, my upload is at max. 45 KB/s, when it happen i will get 75.000 points on FAX4 , arenĀ“t i ? | |
ID: 24277 | Rating: 0 | rate: / Reply Quote | |
close, 71400 is the exact value :) | |
ID: 24279 | Rating: 0 | rate: / Reply Quote | |
too close , i think, the WU got crashed , i lowered the chipset clocking about 10 Mhz and will see. | |
ID: 24287 | Rating: 0 | rate: / Reply Quote | |
Message boards : Number crunching : NATHAN_FAX3 and FAX4 discussion