[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [HTCondor-users] Job not starting correctly

Dollars for donuts says you've got a memory limit problem. Signal 139 is probably 128+11 (obviously!). Linux will kill some processes with signal 11 SISSEGV when if you have a memory limit in place.

The interactive job sets a different memory request than the one you put in your job, if you put one in at all.


If you put request_memory=10000 and it works, you have the beginning of your answer.


On Mon, May 3, 2021 at 5:04 PM Peter Ellevseth <Peter.Ellevseth@xxxxxxxxxx> wrote:



We are running a commercial CFD-code via htcondor. Been doing it for years without any issued. I installed a new version of that software and want to run it via htcondor as per usual. I to this by telling condor to run a locally installed bash-script on the execute node which in turn starts the CFD-solver. I have to do it this to source some files need by the solver to start (license etc).


However, the new version is refusing to start. From the the StarterLog.slotX I see the job immediately stops with


05/03/21 23:56:33 (pid:4135578) Create_Process succeeded, pid=4135579

05/03/21 23:56:33 (pid:4135578) Process exited, pid=4135579, status=139

05/03/21 23:56:33 (pid:4135578) Got SIGQUIT. Performing fast shutdown.


If I ssh in to one of the execute nodes I can start it just and it runs as normal.


If I do condor_submit -interactive my_submit_file, I am able to run the script with ./runscript just fine.


The why wonât it start when I submit the file normally??



HTCondor-users mailing list
To unsubscribe, send a message to htcondor-users-request@xxxxxxxxxxx with a
subject: Unsubscribe
You can also unsubscribe by visiting

The archives can be found at: