hkaiser changed the topic of #ste||ar to: STE||AR: Systems Technology, Emergent Parallelism, and Algorithm Research | stellar-group.org | HPX: A cure for performance impaired parallel applications | github.com/STEllAR-GROUP/hpx | This channel is logged: irclog.cct.lsu.edu
Yorlik_ has joined #ste||ar
Yorlik has quit [Ping timeout: 248 seconds]
K-ballo has quit [Ping timeout: 246 seconds]
hkaiser has quit [Quit: Bye!]
K-ballo has joined #ste||ar
K-ballo has quit [Read error: Connection reset by peer]
K-ballo has joined #ste||ar
tufei has quit [Remote host closed the connection]
tufei has joined #ste||ar
hkaiser has joined #ste||ar
scofield_zliu has joined #ste||ar
scofield_zliu has quit [Read error: Connection reset by peer]
K-ballo1 has joined #ste||ar
K-ballo has quit [Ping timeout: 255 seconds]
K-ballo1 is now known as K-ballo
hkaiser has quit [Quit: Bye!]
diehlpk_work has joined #ste||ar
hkaiser has joined #ste||ar
hkaiser has quit [Quit: Bye!]
hkaiser has joined #ste||ar
<pansysk75[m]> I need some help attaching gdb to the main process of a multi-process hpx program on Rostam
<pansysk75[m]> One issue is, if I launch the processes using SLURM on some compute node, then I cannot login to that node, until that process is finished
<pansysk75[m]> Something like logging in on a node and then launching >1 process on that same node would work for me, if I could find out how to do that (Tried to do it with mpirun but sth is off)
<pansysk75[m]> * is finished (thus I don't have a way of attaching gdb!)
K-ballo1 has joined #ste||ar
K-ballo has quit [Ping timeout: 276 seconds]
K-ballo1 is now known as K-ballo
<pansysk75[m]> Ok, I found a way of launching 2 processes by launching the first process with --hpx:localities=2 --hpx:node=0 and the second process with --hpx:localities=2 --hpx:node=1. This will probably do the trick
<hkaiser> pansysk75[m]: you should be able to login to a node that was assigned to your application by SLURM
<hkaiser> if you use the command line options you list, then both localities run on the same physical node - that's fine if that's what you want
diehlpk_work has quit [Remote host closed the connection]