[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Condor-users] exit status 100 in SchedLog



Condor has been running fine for a few months but all the jobs got killed (some restarted) suddenly yesterday. Here is the log on master node SchedLog:

7/31 15:00:03 Shadow pid 10451 for job 573.2 exited with status 100
7/31 15:00:03 match (<10.10.20.64:49539>#1175522523#195) out of jobs (cluster id 181); relinquishing
7/31 15:00:03 Sent RELEASE_CLAIM to startd on <10.10.20.64:49539>
7/31 15:00:03 Match record (<10.10.20.64:49539>, 181, -1) deleted
7/31 15:00:04 Got VACATE_SERVICE from <10.10.20.64:34423>
7/31 15:00:04 Shadow pid 9427 for job 636.0 exited with status 100
7/31 15:00:04 match (<10.10.20.76:46461>#1175522447#121) out of jobs (cluster id 636); relinquishing
7/31 15:00:04 Sent RELEASE_CLAIM to startd on <10.10.20.76:46461>
7/31 15:00:04 Match record (<10.10.20.76:46461>, 636, -1) deleted
7/31 15:00:04 Got VACATE_SERVICE from <10.10.20.76:59431>
7/31 15:00:04 In DedicatedScheduler::reaper pid 22101 has status 1024
7/31 15:00:04 Shadow pid 22101 exited with status 4
7/31 15:00:04 ERROR: Shadow exited with job exception code!

It seems the Shadows exited with status 100 or 4. What is status 100 and 4? Does it have anything to do with the network or file system?

Junjun