Difference between revisions of "Main Page/Meeting Notes"
From Nekcem
Jump to navigationJump to search(2 intermediate revisions by the same user not shown) | |||
Line 2: | Line 2: | ||
==Notes== | ==Notes== | ||
=== 02/17/2011, Misun, Jing Fu=== | === 02/17/2011, Misun, Jing Fu=== | ||
− | * | + | * Paper |
− | ** | + | ** will keep iterating the NekCEM I/O paper in accordance to review/feedback from ICS |
+ | ** Chris suggest [http://www.ppam.pl/ PPAM] (submit: 4/30, notify: 6/15) over ICPP and Cluster | ||
+ | ** Results on Jaguar is going to take a while (allocation application turnaround time, compile NekCEM on Cray machines, tune performance, summarize results in paper) | ||
+ | ** Results from Jaguar/Lustre can go into follow-up journal paper (most likely by the end of summer) | ||
+ | |||
+ | * hybrid model | ||
+ | ** I/O w/ pthread is coming along on SMP machines, will try Intrepid soon | ||
+ | ** OpenMP should expect strong support (according to Bronis etc.), will likely be used for computation frame | ||
+ | ** pthread should be acceptable for I/O tasks | ||
+ | ** MPI task? | ||
+ | ** test sub-communicator collective performance degradation (for potential total comm split) | ||
+ | *** for all_reduce, going through subcomm would force routine go torus rather than collective tree network, causing a 10x-30x performance drop for double and integer all_reduce, compared to on MPI_COMM_WORLD; tested on 32k and 64k | ||
+ | |||
+ | * summer time frame | ||
+ | ** Flexible, Misun only absent July 26-29 |
Latest revision as of 16:16, 22 February 2011
This page records meeting notes about NekCEM.
Notes
02/17/2011, Misun, Jing Fu
- Paper
- will keep iterating the NekCEM I/O paper in accordance to review/feedback from ICS
- Chris suggest PPAM (submit: 4/30, notify: 6/15) over ICPP and Cluster
- Results on Jaguar is going to take a while (allocation application turnaround time, compile NekCEM on Cray machines, tune performance, summarize results in paper)
- Results from Jaguar/Lustre can go into follow-up journal paper (most likely by the end of summer)
- hybrid model
- I/O w/ pthread is coming along on SMP machines, will try Intrepid soon
- OpenMP should expect strong support (according to Bronis etc.), will likely be used for computation frame
- pthread should be acceptable for I/O tasks
- MPI task?
- test sub-communicator collective performance degradation (for potential total comm split)
- for all_reduce, going through subcomm would force routine go torus rather than collective tree network, causing a 10x-30x performance drop for double and integer all_reduce, compared to on MPI_COMM_WORLD; tested on 32k and 64k
- summer time frame
- Flexible, Misun only absent July 26-29