charm AT lists.siebelschool.illinois.edu
Subject: Charm++ parallel programming system
List archive
- From: "Van Der Wijngaart, Rob F" <rob.f.van.der.wijngaart AT intel.com>
- To: Sam White <white67 AT illinois.edu>
- Cc: "Miller, Philip B" <mille121 AT illinois.edu>, "charm AT cs.uiuc.edu" <charm AT cs.uiuc.edu>
- Subject: RE: [charm] Adaptive MPI
- Date: Tue, 29 Nov 2016 17:12:39 +0000
- Accept-language: en-US
I wanted to add that I observed the same performance anomaly with the code that my intern, Evangelos Georganas, wrote last year, which I compiled and ran with Charm++ version 6.7.0. My experiments are all on a single shared-memory node. Should I expect things to be more favorable for the explicit PUP method on a distributed-memory system? Thanks. From: Van Der Wijngaart, Rob F
Hi Sam,
I switched to the 6.7.1 development branch and am now able to run to completion most of the time. I did notice the following peculiar behaviors. With a smallish data set (a few tens of MB), the isomalloc version without PUP runs at about the same speed as the one with an explicit PUP routine (I am using LBRotate). With a larger data set (about 100 times the smallish set) the isomalloc version without PUP consistently runs about 3X faster than the one with the explicit PUP routine. In my pup routine if have a lot of individual pup calls (about 100). Do you observe the same in your regression tests? And are there known conditions under which pure isomalloc is faster than explicit PUP routines? The explicit PUP version runs out of memory sooner than the pure isomalloc version. When I look at the output from LBDebug, I see that the load balancer uses the same amount of memory, regardless of explicit PUP or pure isomalloc. With pure isomalloc and the larger problem I see very consistent duration of the load balancing step (LBDebug says about 0.0025s per step), whereas for the PUP version it is consistently about 1.2s.
Thanks.
Rob
From:
samt.white AT gmail.com [mailto:samt.white AT gmail.com]
On Behalf Of Sam White
That is a known issue in the 6.7.1 release. I mentioned this issue in the previous 'When to migrate' thread to try to save you this effort. As things stand, you can either use the 6.7.0 release with the old interface, or use mainline/development
charm until 6.8.0 is released in the coming weeks. The development version is tested nightly on various platforms, and contains many bug fixes and performance improvements over 6.7.0.
On Mon, Nov 28, 2016 at 5:58 PM, Van Der Wijngaart, Rob F <rob.f.van.der.wijngaart AT intel.com> wrote:
|
- RE: [charm] Adaptive MPI, (continued)
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/23/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/24/2016
- Re: [charm] Adaptive MPI, Phil Miller, 11/25/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/25/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/28/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/28/2016
- Message not available
- Re: [charm] Adaptive MPI, Sam White, 11/28/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/28/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/29/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/29/2016
- RE: [charm] Adaptive MPI, Van Der Wijngaart, Rob F, 11/28/2016
Archive powered by MHonArc 2.6.19.