Hi Christoph, Thank you for your reply! The failed jobs are not queued anymore â they have crashed (in this case, due to the insufficient disk space for their output). If the jobs were still running, I could have held and then released them to solve the problem. The question
is if I can tell HTcondor to run just those failed jobs again if the jobs have crashed and are not running anymore. Thank you, Siarhei. From: HTCondor-users <htcondor-users-bounces@xxxxxxxxxxx>
On Behalf Of Beyer, Christoph
Hi, I think as long as the jobs are still queued you can put them back in idle mode thorugh condor_qedit ?
Best christoph
Von:
"Vaurynovich, Siarhei" <siarhei.vaurynovich@xxxxxxxxxxxxx> Hello, Situation: I have a large DAG of jobs which is in the process of running. A few jobs failed but most of the jobs in the DAG keep running. From the log files, I have figured out the problem and fixed it. Please,
let me know if there is a way to tell HTCondor to try again the failed nodes (and all of their CHILD nodes, of course) without killing any of the currently running jobs in the same DAG and without waiting for the whole DAG to fail (and generate a rescue file)? From the documentation on condor_submit_dag, I can see that the following command might be a good candidate (I have sub-DAGs): condor_submit_dag -DoRecovery -do_recurse submit_file.dag Please, let me know if that is what I should do. Thank you very much for your help, Siarhei. ............................................................................ Trading instructions sent electronically to Bernstein shall not be deemed For further important information about AllianceBernstein please click here
|