K-ballo changed the topic of #ste||ar to: STE||AR: Systems Technology, Emergent Parallelism, and Algorithm Research | stellar.cct.lsu.edu | HPX: A cure for performance impaired parallel applications | github.com/STEllAR-GROUP/hpx | Buildbot: http://rostam.cct.lsu.edu/ | Log: http://irclog.cct.lsu.edu/
hkaiser has joined #ste||ar
K-ballo has quit [Quit: K-ballo]
jehelset has joined #ste||ar
hkaiser has quit [Quit: bye]
wash[m] has quit [Read error: No route to host]
wash[m] has joined #ste||ar
<ms[m]1> gonidelis[m]: sorry for not replying earlier... the most important ones are `-DCMAKE_BUILD_TYPE=Release -DHPX_WITH_MALLOC=tcmalloc/mimalloc/jemalloc`, `CXXFLAGS=-march=native` may also make a small difference
<gonidelis[m]> oh should mallocs make a significant difference?
K-ballo has joined #ste||ar
hkaiser has joined #ste||ar
<ms[m]1> gonidelis[m]: big difference!
<ms[m]1> the system allocator is pretty terrible when it comes to multithreading
<gonidelis[m]> ah ms[m] thanks
<gnikunj[m]> I've usually found that jemalloc works the fastest, tcmalloc a close second, and system is significantly slower than these.
<sestro[m]> Are there any benchmarks that can give me an indication for which workloads the performance of the allocators differs significantly?
<hkaiser> sestro[m]: any application should do, allocators are fundamental
<sestro[m]> Okay, I am using the system allocator right now as I use HPX in a shared library hoping not using jemalloc would not hurt too much.
<hkaiser> sestro[m]: depends on the platform, on linux you might see significant speedup
V|r has quit [Quit: ZNC 1.7.5+deb4 - https://znc.in]
<srinivasyadav227> hkaiser: there is a merge conflict on #5235, its not allowing me to make changes, so should i solve the merge conflict with another commit and push? and you told you have some comments regarding #5254, please tell if any further changes are required for #5235
<hkaiser> srinivasyadav227: best is to rebase onto master while resolving the conflicts
<hkaiser> and then force-push to your branch
<sestro[m]> hkaiser: At some point I tried using a different one but this ended in some horrible conflict between different allocators in different shared libraries/python modules. I probably did something stupid and should explore that again.
<srinivasyadav227> hkaiser: ok i will do that
Vir has joined #ste||ar
hkaiser has quit [Ping timeout: 252 seconds]
hkaiser has joined #ste||ar
<srinivasyadav227> hkaiser: Thanks for pushing ;-), i would use rebase from next time, i was not familiar with that, i only knew git merge
<srinivasyadav227> srinivasyadav227: regarding GSoC project "Add vectorization to par_unseq", i have a doubt, i should implement all theses four right? (unsequenced_policy, unsequenced_task_policy, parallel_unsequenced_policy, parallel_unsequenced_task_policy)
<hkaiser> srinivasyadav227: they are all the same, essentially
<srinivasyadav227> hkaiser: oh ok, for those policies we need to write unsequenced_exector.hpp and parallel_unsequenced_executor.hpp? something similar to (https://github.com/STEllAR-GROUP/hpx/blob/master/libs/parallelism/executors/include/hpx/executors/parallel_executor.hpp)
<srinivasyadav227> and for all these we need common vectorization backend to be implemented first right?
<hkaiser> srinivasyadav227: we already have those
<hkaiser> they do not support async execution yet, but that's orthogonal
<srinivasyadav227> oh wait, i am confused, what are the expected deliverables for "Add vectorization to par_unseq"
<diehlpk_work> ms[m]1, Do you need help to finalize the GSoD propsoal?
<ms[m]1> diehlpk_work: yeah, if you had a good idea for the budget section I'd appreciate your help
<hkaiser> srinivasyadav227: support more algorithms
<diehlpk_work> I would use the base salary there and derive the pay per hour
<diehlpk_work> And would multiply this value with the hours the technical writer will work for us
<diehlpk_work> So using the average from the above webpage yields 29.26 per hour
<diehlpk_work> Do you have time to participate during the six months of the program (April-November 2021)? Project sizes vary, but range from a commitment of 5-30 hours per week during the program.
<diehlpk_work> So I would just define how much work per week we want to see
<diehlpk_work> Using this approach the example amount reflects 14 hours per week
<diehlpk_work> ms[m]1, Let me know what you think?
bita has joined #ste||ar
<srinivasyadav227> hkaiser: oh, that means I need not implement par_unseq policy again, we already have it, i should use the existing par_unseq policy and support it for parallel alogirthms?, and currently no hpx parallel algorithm supports par_unseq
<hkaiser> yes
<hkaiser> srinivasyadav227: that is correct
shubham has joined #ste||ar
<srinivasyadav227> hkaiser: ok thanks, i think i need to spend much time on proposal and work on PRs little less till April 13, then again i would focus on PRs, is that fine?
nanmiao has joined #ste||ar
<jedi18[m]> @freenode_hkaiser:matrix.org same here, my exams are starting next week so I probably won't get time to work on any more PRs till they're over (on the bright side, this means no exams during the gsoc period)
<hkaiser> srinivasyadav227, jedi18[m]: sure, please take your time
<srinivasyadav227> hkaiser: thank you ;-)
weilewei has joined #ste||ar
shubhu_ has joined #ste||ar
<ms[m]1> diehlpk_work: yeah, that sounds reasonable
<ms[m]1> I obviously don't know how much work that project will take but I imagined something like 10-20 hours per week would be reasonable
<ms[m]1> with 15 hours per week for 12 weeks that's roughly the 5000 that we have there now
weilewei has quit [Quit: Ping timeout (120 seconds)]
<shubhu_> hello everyone
<nanmiao> =D =D
shubhu_ has quit [Quit: Leaving]
<ms[m]1> that's with networking and distributed runtime off I think
<hkaiser> ms[m]1: uhh
<hkaiser> is that on master now?
<hkaiser> it did pass on the PR :/
<hkaiser> at least I believe it did
nanmiao has quit [Quit: Connection closed]
<ms[m]1> hkaiser: looks like it was already there on the PR :/
<ms[m]1> let me know if I can help there, it might be something I need to change in the way the local runtime is started
<hkaiser> ms[m]1: could be
<hkaiser> the PR mainly changed the command line handling and the post plugin load repeated command line processing
<hkaiser> I wouldn't expect that to be different for the local case
<hkaiser> ms[m]1: I'd appreciate if you had to time to look
<ms[m]1> it's most likely not a good thing that that's handled in runtime_support_server...
<hkaiser> ms[m]1: could be
<hkaiser> we can disable the test for the local case for now
<ms[m]1> it's definitely the reason, but can we leave it for now, I'd like to have a look
<ms[m]1> if we disable it I'll never look at it
<ms[m]1> if I don't do it by the end of next week we can disable it
<hkaiser> sure
<hkaiser> ok
<hkaiser> ms[m]1: I'm working on reviving libcds, once that's done I'll try to look at the problem as well
<ms[m]1> 👍️
peltonp1 has quit [Read error: Connection reset by peer]
nanmiao has joined #ste||ar
jehelset has quit [Remote host closed the connection]
nanmiao has quit [Quit: Connection closed]
nanmiao has joined #ste||ar
<bita> hkaiser, about the LowerMatrix, I read the last example here: https://bitbucket.org/blaze-lib/blaze/wiki/Triangular%20Matrices#!the-triangular-property-is-always-enforced in the comment it says"// Throws an exception; lower matrix invariant would be violated!" I think that's the example of how it shouldn't work
<hkaiser> bita: nod
<hkaiser> I think LowerMatrix should be used on the right hand side
<bita> so is there anything I don't get or should we use a for with blaze::band?
<hkaiser> DynamicMatrix<double> m = LowerMatrix<DynamicMatrix<double>>(src);
<hkaiser> would that work?
<bita> I will check it out
<hkaiser> thanks
<hkaiser> bita: most likely you need to use LowerMatrix<CustomMatrix<>>, though
karame_ has joined #ste||ar
<bita> It only works if src is actually a LowerMatrix
<hkaiser> hmmm
<hkaiser> does it throw at runtime?
<bita> yes, as I don't see it and it wants to abort it "C:\Repos\blaze_app_1\cmake-build-debug\Debug\app1.exe (process 20600) exited with code 3."
<hkaiser> nod
<hkaiser> darn
<hkaiser> let's ask Klaus how to do that
<hkaiser> ot really sure if band is the appropriate way of extracting a diagonal matrix
<bita> Thank you. I think I saw his name the other day here. Should we email him or ...?
<hkaiser> yah, email is best - do you have his email address?
<bita> yes :+1
bita has quit [Ping timeout: 252 seconds]
shubham has quit [Quit: Connection closed for inactivity]
nanmiao has quit [Quit: Connection closed]
karame_ has quit [Ping timeout: 240 seconds]
jehelset has joined #ste||ar
<K-ballo> github says I've only contributed 2 commits to hpx