hkaiser changed the topic of #ste||ar to: STE||AR: Systems Technology, Emergent Parallelism, and Algorithm Research | stellar.cct.lsu.edu | HPX: A cure for performance impaired parallel applications | github.com/STEllAR-GROUP/hpx | Buildbot: http://rostam.cct.lsu.edu/ | Log: http://irclog.cct.lsu.edu/
hkaiser has quit [Ping timeout: 260 seconds]
nikunj has joined #ste||ar
nikunj has quit [Client Quit]
nikunj has joined #ste||ar
nanashi55 has quit [Ping timeout: 268 seconds]
nanashi55 has joined #ste||ar
quaz0r has quit [Ping timeout: 252 seconds]
quaz0r has joined #ste||ar
quaz0r has quit [Ping timeout: 244 seconds]
nikunj has quit [Remote host closed the connection]
<simbergm>
jbjnr_: why don't you make a small test to see how the parallel for loop compares to a custom version with async/apply/latch/counting_semaphore or something like that?
<simbergm>
and regarding launching more tasks to do a memcpy or similar, you only have n cores which can do n memcpy's at a time, so why launch (many) more tasks than that?
nikunj1997 has quit [Ping timeout: 246 seconds]
rtohid has joined #ste||ar
<jbjnr_>
simbergm: you seem to be trying very hard from dissauding me from using the code I'm working on :) I would like to make simple test, but I need to work on my slides for Pavia, so shouldn't spend too much time on it. It's mostly to make the code look better rather than as a micro-optimization for performance. I don't like having to store a vector of futures that I don't really care about. I have an idea on using the apply with an
<jbjnr_>
extra template param to call a function on completeion that would serve my purposes, but requires a bit of work. If I can't sleep tonight. I'll try it.
<simbergm>
ok :) I'm just trying to make sure you don't end up reinventing the wheel
<simbergm>
if performance isn't a big concern I'm still not sure why the parallel for loop doesn't work for you
bibek has joined #ste||ar
david_pfander has quit [Ping timeout: 252 seconds]
diehlpk_work has joined #ste||ar
<diehlpk_work>
HPX was approved for fedora and should be there soon
<diehlpk_work>
So one could do sudo dnf install hpx
<diehlpk_work>
hpx-openmpi
<diehlpk_work>
hx-mpich
<diehlpk_work>
And people can play around with hpx on fedora 28/29
<diehlpk_work>
And hpx-*-examples for the examples
<diehlpk_work>
And hpx-devel for the headers and cmake stuff
<zao>
Nice.
aserio1 has joined #ste||ar
aserio has quit [Ping timeout: 250 seconds]
aserio1 is now known as aserio
hkaiser has joined #ste||ar
bibek has quit [Quit: Konversation terminated!]
dmarce1 has joined #ste||ar
<dmarce1>
Is calling hpx::new_ recursively within component constructors something that should work efficiently?
<hkaiser>
dmarce1: I don't think that calling it recursively impacts performance in any way
aserio has quit [Read error: Connection reset by peer]
aserio has joined #ste||ar
nikunj has joined #ste||ar
nikunj97 has joined #ste||ar
nikunj has quit [Ping timeout: 240 seconds]
aserio has quit [Ping timeout: 264 seconds]
jaafar has joined #ste||ar
jaafar has quit [Quit: Konversation terminated!]
jaafar has joined #ste||ar
<heller_>
dmarce1: you just might not need the members to be components. Components do come at a cost