<rori>
I'm not sure where `Proj` is defined but you should have your answer there ;)
diehlpk_work has quit [Ping timeout: 256 seconds]
hkaiser has joined #ste||ar
<hkaiser>
hey ms[m]
<gonidelis[m]>
hkaiser: I pushed the ranges::transform CPO ;)
<hkaiser>
gonidelis[m]: \o/
<gonidelis[m]>
hkaiser: I think it is safe to proceed on the BinaryOperation overloads :)
<gonidelis[m]>
hkaiser: reminder: meeting in 1 hour ;)
<hkaiser>
gonidelis[m]: if you feel comfortable to do that - sure!
<hkaiser>
yes, I will be there
<hkaiser>
need coffee first, though
<gonidelis[m]>
hkaiser: Member's must indicate their personal coffee mug in order to be accepted in the meeting anyways
<gonidelis[m]>
Memebers^^
<hkaiser>
ok, deal ;-)
<ms[m]>
hkaiser: hey! sorry didn't see your message yesterday in time
<hkaiser>
np, it was late
<hkaiser>
ms[m]: would you have time for a short(-ish) chat about hpx-kokkos later today or tomorrow?
<ms[m]>
yeah, sure
<ms[m]>
either is fine
<ms[m]>
including right now
<hkaiser>
right now doesn't work, sorry - need coffee
<ms[m]>
np :P
akheir has joined #ste||ar
<hkaiser>
would 10am/17.00 work? or tomorrow 9am/16.00?
<ms[m]>
let's do it tomorrow morning (for you) then
<hkaiser>
ok
<hkaiser>
thanks a lot
<ms[m]>
👍️
<hkaiser>
Katie will send a zoom link
<ms[m]>
all right, thanks
<rori>
may I join the hpx-kokkos meeting ? :D
diehlpk_work has joined #ste||ar
<hkaiser>
rori: sure
<rori>
gonidelis: meeting?
<gonidelis[m]>
I am loggin in right now
<rori>
👍️
Yorlik has joined #ste||ar
<hkaiser>
hey Yorlik, welcome back!
<Yorlik>
Heyo!
<Yorlik>
Never been away - just lurking.
weilewei has joined #ste||ar
nanmiao11 has joined #ste||ar
bita has joined #ste||ar
<diehlpk_work>
ms[m], Can we change rostam to a working cluster again?
<diehlpk_work>
Currently, it is a build cluster and I am not sure this is what we want
<diehlpk_work>
One thing we could do is that jenkins can not use all of our GPU nodes
<hkaiser>
diehlpk_work: not sure what you mean by 'working cluster'? isn't it 'working'?
<diehlpk_work>
hkaiser, It is working, but only for jenkins
<diehlpk_work>
not for me because jenkins uses all nodes
<diehlpk_work>
Just wanted to debug octotiger on rostam, but jenkins uses all cuda nodes
<hkaiser>
diehlpk_work: akheir is working on making the jenkins jobs low priority, so that everybody should be able to quickly get access to nodes
<diehlpk_work>
We might keep geev for us solely and jenkisn can use bahram
<diehlpk_work>
and might keep at least one marvin and teo medusa nodes for us and jenkins can not use these
<diehlpk_work>
So we could use these ndoes for debugging and do not have to wait until jenkins finished
<hkaiser>
diehlpk_work: sure, we can do that - we talked about this with akheir last meeting, I believe
<diehlpk_work>
I can mention it again tomorrow
<hkaiser>
pls do
<diehlpk_work>
At least having some nodes available and jenkins can not allocate all would be a first step
<ms[m]>
diehlpk_work: yes, it's not meant to make life horrible for interactive users, this is just the initial configuration
<ms[m]>
let's try what akheir has in mind first, and if that isn't enough we can try to change it further
<diehlpk_work>
Ok, I hope it will become better
<diehlpk_work>
At least I can apply for QB and run my code there
<diehlpk_work>
hkaiser, How do I apply for QB?
<ms[m]>
yeah, indeed, please remind us if things don't improve
<hkaiser>
diehlpk_work: apply for a loni account, put me in as the sponsor
<diehlpk_work>
Ok, I will do that
<hkaiser>
then either use Dominics allocation or apply for one yourself
<hkaiser>
startup allocations are easy to apply for and will be approve immediately, I think
<diehlpk_work>
I think I will apply without DOminic, since I like to have some time to run the peridyanmic code on a large scale
<diehlpk_work>
So we have time for octotiger and my code
<akheir>
ms[m]: There is problem with some Jenkins' runs which I haven't figured out yet. The jobs hang and slurm cannot release the node.
<diehlpk_work>
akheir, Can we exclude geev from the jenkins runs?
<diehlpk_work>
So we have at least one cuda node available?
<akheir>
yes, I will do that today
<diehlpk_work>
Same for one marvin node and two or three medusa nodes?
<hkaiser>
let's create special partitions on rostam to be used by jenkins
<ms[m]>
is it possible to have a separate partitions for jenkins (cpu only and gpu) that are a subset of the other partitions
<ms[m]>
?
<hkaiser>
nod
<hkaiser>
akheir: ^^
<diehlpk_work>
Can we have a debug queue as well? So we can get a small allocation (15 minutes) with a higher priority?
<akheir>
diehlpk_work: That would be to much fragmentation on partitions. Jenkins jobs won't take that long, lower priority should solve the problem
<hkaiser>
ok, cool
<ms[m]>
👍️
<diehlpk_work>
akheir, Ok, anything what improves the current situation will be appreciated
<ms[m]>
akheir: not sure what I can do about the hung jobs, the jenkins interface doesn't show anything until jobs are completed
<akheir>
we don't hat that nodes for increasing the number of queues make a impact.
<ms[m]>
if you have any ideas on what I could look at I'm all ears (or if you have access to logs for those jobs that you'd like me to have a look at)
<akheir>
ms[m]: I have to investigate, slurm complains about open I/O files and cannot release the node, the only way to cancel the job is to reboot the node. I think this main reason for complains
<akheir>
I have to find fix for this problem first
<ms[m]>
hmm, ok
<ms[m]>
I'll try to dig around and see if there's anything that looks like it could cause that
<tiagofg[m]>
hkaiser Hello! Regarding the inheritance issue, I would like to know if the problem has a solution or not, I'm finishing my master's thesis and I really needed that information to know if I have to rewrite the code in another way or not
<tiagofg[m]>
For you guys must be a simple thing I guess.
karame_ has joined #ste||ar
<hkaiser>
tiagofg[m]: you wanted to create a small example I could look at
<karame_>
hkaiser Could you please send me the zoom ID.
<akheir>
ms[m]: I didn't see your comment about the special partition for Jenkins. Yes that's the way I have to configure the nodes, In slurm the queue and partition are the same. in order to have lower priority queue we have to create new partitions