The only thing that is kind of suspicious to me is that you suspend boinc while running the tool. In the basic configuration boinc leaves the applications in memory confronting those with unexpected changes in the client_state.xml. Advise is to stop boinc completely and not suspend it. But i don't think this will cause the disk-full errors though....
Marius.........This morning found 2 of my host computers had run out of CUDA work. CPU's had lot's of work. I am incommunicado with Berkeley due to bandwidth maxed out at Berkeley. I set the config file as follows.[Settings]...TrueAngleRate=1...After running ReSchedule I observed several work units were aborted by the AutoKill function of V11 CUDA app. Is it possible that in the reassignment of work from CPU's to CUDA that some VLARVHAR work was moved but should not have been?
But surely your program stops BOINC first thus removing the (suspended) apps from memory. I certainly saw the manager window empty of tasks.
Quote from: MarkJ on 04 Jul 2009, 07:51:49 amI did notice that the rebranded work units immediately went into "running, High Priority" mode. Seems their estimated time was 20+ hours. That estimate is dropping like a stone as it starts crunching them so it will work it out, but they will be finished by the time the TSI has come up (60 mins).If tasks are estimated at 20 hours, but finishing in less than 1 hour, you have a horribly out-of-kilter Duration Correction Factor. With no VLAR handler in place, I would have expected a sawtooth waveform with maybe a factor of 4x between peak and valley, but 20x? No wonder you were seeing preemptions and EDF on boinc_alpha.I suggest a sanity-check on the FLOPs estimates in your app_info: once the after-effects of your CUDA/VLAR processing have worked their way out of your system, estimates for all SETI work (MB/CPU, MB/CUDA, and AP) should be accurate within a few %.
I did notice that the rebranded work units immediately went into "running, High Priority" mode. Seems their estimated time was 20+ hours. That estimate is dropping like a stone as it starts crunching them so it will work it out, but they will be finished by the time the TSI has come up (60 mins).
CPU tasks: 470 (470 VLAR/VHAR)GPU tasks: 15 (0 VLAR/VHAR)Starting BOINC serviceBOINC service startedand I'm trying to get a 50-50 split on the work units.All this is run from a batch file with this commmand.........
OK.....thanks for the explanation.If you reschedule 100% to GPU does that not leave the CPU's cold??
Yes, didn't i tell you you would see an akward rise in V*AR . V*AR are always forced on the cpu no matter what happends to avoid vlarkill, then as a bonus it tries to confirm to the 50% rule but it cannot do this unfortunately.
Suggestions for future versions:Perhaps only VLAR should be forced to CPU. Allow VHAR to go either way, default to CPU however. Maybe the test function could show the task counts by AR group (VLAR/VHAR/other) before and after with current settings.
Quote from: samuel7 on 05 Jul 2009, 11:29:36 amSuggestions for future versions:Perhaps only VLAR should be forced to CPU. Allow VHAR to go either way, default to CPU however. Maybe the test function could show the task counts by AR group (VLAR/VHAR/other) before and after with current settings. Internally i know exactly those groups, but if you are using Raistmer's MB_6.08_mod_CUDA_V11_VLARKill_refined.exe al vhar's will be automaticly killed AFAIK (true angle rate > 1.127), so i assume you are running a different tool what can handle a VHAR?Or i'm i wrong here? (i use Raistmer's tool for several years now and don't even know what the original tools are any more)
There's no actual need to rebrand VHAR. They run adequately on the GPU, and they don't cause any screen stuttering or other computer-use problems. Raistmer has just observed a slight efficiency gain (i.e. optimisation) by using the CPU instead of the GPU for VHAR, but it's not enough to lose any sleep over. In any case, much of the efficiency gain is lost on Quads and above when there are lots of VHAR about, because they suffer from memory bus contention.VLAR are different entirely. Their efficiency on GPU is abysmal, and the general screen delay makes the rest of the computer unusable - it's driven away many potential crunchers. Anything you can do to keep them off the GPU is a big step forward - but I prefer rebranding to killing.