charm AT lists.siebelschool.illinois.edu
Subject: Charm++ parallel programming system
List archive
- From: "Van Der Wijngaart, Rob F" <rob.f.van.der.wijngaart AT intel.com>
- To: Sam White <white67 AT illinois.edu>
- Cc: "charm AT cs.uiuc.edu" <charm AT cs.uiuc.edu>
- Subject: RE: [charm] Adaptive MPI
- Date: Wed, 23 Nov 2016 17:28:04 +0000
- Accept-language: en-US
Hi Sam,
The first experiment was successful, but the isomalloc example hangs. See below. Unless it is a symptom of something bigger, I am not going to worry about the latter, since I wasn’t planning to use isomalloc for heap migration anyway. My regular MPI code on which the AMPI version is based runs fine for all the parameters I have tried, but I reckon that it may contain a memory bug that manifests itself only with load balancing
Rob
rfvander@klondike:~/Cjacobi3D$ make /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -c jacobi.C /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -o jacobi jacobi.o -module CommonLBs -lm /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -c -DNO_PUP jacobi.C -o jacobi.iso.o /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -o jacobi.iso jacobi.iso.o -module CommonLBs -memory isomalloc /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -c -tlsglobal jacobi.C -o jacobi.tls.o /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -o jacobi.tls jacobi.tls.o -tlsglobal -module CommonLBs #-memory isomalloc /opt/charm/charm-6.7.0/multicore-linux64/bin/../lib/libconv-util.a(sockRoutines.o): In function `skt_lookup_ip': sockRoutines.c:(.text+0x334): warning: Using 'gethostbyname' in statically linked applications requires at runtime the shared libraries from the glibc version used for linking /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -c jacobi-get.C /opt/charm/charm-6.7.0/multicore-linux64/bin/ampicxx -o jacobi-get jacobi-get.o -module CommonLBs -lm rfvander@klondike:~/Cjacobi3D$ ./charmrun +p3 ./jacobi 2 2 2 +vp8 +balancer RotateLB +LBDebug 1 Running command: ./jacobi 2 2 2 +vp8 +balancer RotateLB +LBDebug 1 +p3
Charm++: standalone mode (not using charmrun) Charm++> Running in Multicore mode: 3 threads Converse/Charm++ Commit ID: v6.7.0-1-gca55e1d Warning> Randomization of stack pointer is turned on in kernel, thread migration may not work! Run 'echo 0 > /proc/sys/kernel/randomize_va_space' as root to disable it, or try run with '+isomalloc_sync'. CharmLB> Verbose level 1, load balancing period: 0.5 seconds CharmLB> Load balancer assumes all CPUs are same. Charm++> Running on 1 unique compute nodes (16-way SMP). Charm++> cpu topology info is gathered in 0.000 seconds. [0] RotateLB created iter 1 time: 0.142733 maxerr: 2020.200000 iter 2 time: 0.157225 maxerr: 1696.968000 iter 3 time: 0.172039 maxerr: 1477.170240 iter 4 time: 0.146178 maxerr: 1319.433024 iter 5 time: 0.123098 maxerr: 1200.918072 iter 6 time: 0.131063 maxerr: 1108.425519 iter 7 time: 0.138213 maxerr: 1033.970839 iter 8 time: 0.138295 maxerr: 972.509242 iter 9 time: 0.138113 maxerr: 920.721889 iter 10 time: 0.121553 maxerr: 876.344030
CharmLB> RotateLB: PE [0] step 0 starting at 1.489509 Memory: 72.253906 MB CharmLB> RotateLB: PE [0] strategy starting at 1.489573 CharmLB> RotateLB: PE [0] Memory: LBManager: 920 KB CentralLB: 3 KB CharmLB> RotateLB: PE [0] #Objects migrating: 8, LBMigrateMsg size: 0.00 MB CharmLB> RotateLB: PE [0] strategy finished at 1.489592 duration 0.000019 s CharmLB> RotateLB: PE [0] step 0 finished at 1.507922 duration 0.018413 s
iter 11 time: 0.152840 maxerr: 837.779089 iter 12 time: 0.136401 maxerr: 803.868831 iter 13 time: 0.138095 maxerr: 773.751705 iter 14 time: 0.139319 maxerr: 746.772667 iter 15 time: 0.139327 maxerr: 722.424056 iter 16 time: 0.141794 maxerr: 700.305763 iter 17 time: 0.142484 maxerr: 680.097726 iter 18 time: 0.141056 maxerr: 661.540528 iter 19 time: 0.153895 maxerr: 644.421422 iter 20 time: 0.198588 maxerr: 628.564089 [Partition 0][Node 0] End of program rfvander@klondike:~/Cjacobi3D$ ./charmrun +p3 ./jacobi.iso 2 2 2 +vp8 +balancer RotateLB +LBDebug 1 Running command: ./jacobi.iso 2 2 2 +vp8 +balancer RotateLB +LBDebug 1 +p3
Charm++: standalone mode (not using charmrun) Charm++> Running in Multicore mode: 3 threads ^C rfvander@klondike:~/Cjacobi3D$ ./charmrun +p3 ./jacobi.iso 2 2 2 +vp8 +balancer RotateLB +LBDebug 1 +isomalloc_sync Running command: ./jacobi.iso 2 2 2 +vp8 +balancer RotateLB +LBDebug 1 +isomalloc_sync +p3
Charm++: standalone mode (not using charmrun) Charm++> Running in Multicore mode: 3 threads From: samt.white AT gmail.com [mailto:samt.white AT gmail.com]
On Behalf Of Sam White
Can you try an example AMPI program with load balancing? You can try charm/examples/ampi/Cjacobi3D/, running with something like ' ./charmrun +p3 ./jacobi 2 2 2 +vp8 +balancer RotateLB +LBDebug 1'. You can also test that example with Isomalloc by running jacobi.iso (and as the warning in the Charm preamble output suggests, run with +isomalloc_sync). It also might help to build Charm++/AMPI with '-g' to get stacktraces. -Sam
On Wed, Nov 23, 2016 at 2:19 AM, Van Der Wijngaart, Rob F <rob.f.van.der.wijngaart AT intel.com> wrote:
|
- [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- <Possible follow-up(s)>
- Re: [charm] Adaptive MPI, Sam White, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- Message not available
- Re: [charm] Adaptive MPI, Sam White, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- Message not available
- Re: [charm] Adaptive MPI, Sam White, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/24/2016
- Re: [charm] Adaptive MPI, Phil Miller, 11/25/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/25/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/28/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/28/2016
- Re: [charm] Adaptive MPI, Sam White, 11/23/2016
- Re: [charm] Adaptive MPI, Sam White, 11/23/2016
Archive powered by MHonArc 2.6.19.