Difference between revisions of "Main Page/faq"

From Nekcem
Jump to navigationJump to search
 
(11 intermediate revisions by the same user not shown)
Line 1: Line 1:
 
This is the resource listing page for [http://www.mcs.anl.gov/~mmin/nekcem.html NekCEM].
 
This is the resource listing page for [http://www.mcs.anl.gov/~mmin/nekcem.html NekCEM].
 
==Resource Links==
 
==Resource Links==
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/PIO Public document for parallel I/O of NekCEM]  
+
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/PIO Parallel I/O of NekCEM]  
  
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/hybrid_prog Document for hybrid programming (proposed work)]
+
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/hybrid_prog Hybrid programming (proposed work)]
 +
 
 +
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/aio Sync/Async Blocking/non-blocking I/O]
 +
 
 +
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/bgQ Blue Gene/Q and Mira]
 +
 
 +
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/C_Fortran C Fortran mixed programming]
 +
 
 +
*[https://wiki.mcs.anl.gov/nekcem/index.php/Main_Page/Meeting_Notes Meeting Notes]
  
 
== Implementation ==
 
== Implementation ==
Line 29: Line 37:
 
** elemType → %4d
 
** elemType → %4d
  
=== NekCEM notes for myself===
+
=== NekCEM notes===
 
* scaling
 
* scaling
 
** strong scaling: defined as how the solution time varies with the number of processors for a fixed total problem size.
 
** strong scaling: defined as how the solution time varies with the number of processors for a fixed total problem size.
 
** weak scaling: defined as how the solution time varies with the number of processors for a fixed problem size per processor.
 
** weak scaling: defined as how the solution time varies with the number of processors for a fixed problem size per processor.
 +
* pre-compute file size
 +
** #grid point = nx * ny * nz * nelt; size = #grid point * 3 * float
 +
** cell type: 2d → 4 * #cell * int + 1* #cell * int (3d → 9)
 +
** #field = nfields * 3 * #grid point; size = #field * float;
 +
 
* .box → num elements in x,y,z
 
* .box → num elements in x,y,z
 
* .rea → input data
 
* .rea → input data
Line 44: Line 57:
  
 
* only CELL and point data need to be re-computed
 
* only CELL and point data need to be re-computed
 +
 +
* compile and run NekCEM
 +
** in a specific case, ../../bin/cleanall, ../../bin/makenek, ../../bin/nek "case_name" #proc (e.g. in cylwave, ../../bin/nek cylwave 4)
 +
 
== To-do List ==
 
== To-do List ==
 
*More tests on BG/P for config with ng = M and 1< nf < M  
 
*More tests on BG/P for config with ng = M and 1< nf < M  
Line 51: Line 68:
 
*Parallel I/O for reading .rea file
 
*Parallel I/O for reading .rea file
  
== Some notes ==
+
== Miscellaneous notes ==
 
* Fortran generated binary file may not be correctly read in C.
 
* Fortran generated binary file may not be correctly read in C.
 
* -lstdc++ for link
 
* -lstdc++ for link

Latest revision as of 13:28, 19 May 2011

This is the resource listing page for NekCEM.

Resource Links

Implementation

I/O code

  • I/O functions were initiated from cem_out function of cem_dg.F (and cem_dg2.F).
  • Implementation of parallel I/O routine were defined in vtkbin.c and rbIO_nekcem.c
  • vtkcommon.c and vtkcommon.h serve as a place to hold common functions as well as global variables.
  • cem_out_fields3 (in cem_dg.F)
    • openfile3(dumpno, nid) !vtkbin.c
    • vtk_dump_header3
      • writeheader3() !vtkbin.c
      • writenodes3() !vtkbin.c
      • write2dcells3 !vtkbin.c
        or write3dcells3 !vtkbin.c
    • vtk_dump_field3
      • writefield3 !vtkbin.c
    • close_file3 !vtkbin.c
  • Binary file → ASCII file: transfer double/float/int/read to chars then write out
    • float (4 bytes) → %18.8E
    • int (4 bytes) → %10d
    • long long (8 bytes) → %18lld
    • elemType → %4d

NekCEM notes

  • scaling
    • strong scaling: defined as how the solution time varies with the number of processors for a fixed total problem size.
    • weak scaling: defined as how the solution time varies with the number of processors for a fixed problem size per processor.
  • pre-compute file size
    • #grid point = nx * ny * nz * nelt; size = #grid point * 3 * float
    • cell type: 2d → 4 * #cell * int + 1* #cell * int (3d → 9)
    • #field = nfields * 3 * #grid point; size = #field * float;
  • .box → num elements in x,y,z
  • .rea → input data
  • SIZEu → SIZE parameters:
    • lxi ?
    • lp = #proc
    • lelx = 20 each dimension
    • lelv = alloc max # of element per proc
  • .usr → subuser.F
  • cem() in cem_dg.F is the main solver and application entry point
  • only CELL and point data need to be re-computed
  • compile and run NekCEM
    • in a specific case, ../../bin/cleanall, ../../bin/makenek, ../../bin/nek "case_name" #proc (e.g. in cylwave, ../../bin/nek cylwave 4)

To-do List

  • More tests on BG/P for config with ng = M and 1< nf < M
  • Tests on Kraken and Jaguar
  • Pthread + MPI for I/O
  • OpenMP/Pthread + MPI for NekCEM computation
  • Parallel I/O for reading .rea file

Miscellaneous notes

  • Fortran generated binary file may not be correctly read in C.
  • -lstdc++ for link
  • libF77 and libI77
  • common.h and common_c.h
  • write() in Fortran: 6 refer to screen, * is to screen as well ..