5.5 FieldConvert modules -m

FieldConvert allows the user to manipulate the Nektar++ output binary files (.chk and .fld) by using the flag -m (which stands for module). Specifically, FieldConvert has these additional functionalities

  1. C0Projection: Computes the C0 projection of a given output file;
  2. QCriterion: Computes the Q-Criterion for a given output file;
  3. L2Criterion: Computes the Lambda 2 Criterion for a given output file;
  4. addcompositeid: Adds the composite ID of an element as an additional field;
  5. fieldfromstring: Modifies or adds a new field from an expression involving the existing fields;
  6. addFld: Sum two .fld files;
  7. combineAvg: Combine two Nektar++ binary output (.chk or .fld) field file containing averages of fields (and possibly also Reynolds stresses) into single file;
  8. concatenate: Concatenate a Nektar++ binary output (.chk or .fld) field file into single file (deprecated);
  9. dof: Count the total number of DOF;
  10. equispacedoutput: Write data as equi-spaced output using simplices to represent the data for connecting points;
  11. extract: Extract a boundary field;
  12. gradient: Computes gradient of fields;
  13. halfmodetofourier: Convert HalfMode expansion to SingleMode for further processing;
  14. homplane: Extract a plane from 3DH1D expansions;
  15. homstretch: Stretch a 3DH1D expansion by an integer factor;
  16. innerproduct: take the inner product between one or a series of fields with another field (or series of fields).
  17. interpfield: Interpolates one field to another, requires fromxml, fromfld to be defined;
  18. interppointdatatofld: Interpolates given discrete data using a finite difference approximation to a fld file given an xml file;
  19. interppoints: Interpolates a field to a set of points. Requires fromfld, fromxml to be defined, and a topts, line, plane or box of target points;
  20. interpptstopts: Interpolates a set of points to another. Requires a topts, line, plane or box of target points;
  21. isocontour: Extract an isocontour of “fieldid” variable and at value “fieldvalue”. Optionally “fieldstr” can be specified for a string definition or “smooth” for smoothing;
  22. jacobianenergy: Shows high frequency energy of Jacobian;
  23. qualitymetric: Evaluate a quality metric of the underlying mesh to show mesh quality;
  24. mean: Evaluate the mean of variables on the domain;
  25. meanmode: Extract mean mode (plane zero) of 3DH1D expansions;
  26. pointdatatofld: Given discrete data at quadrature points project them onto an expansion basis and output fld file;
  27. printfldnorms: Print L2 and LInf norms to stdout;
  28. removefield: Removes one or more fields from .fld files;
  29. scalargrad: Computes scalar gradient field;
  30. scaleinputfld: Rescale input field by a constant factor;
  31. shear: Computes time-averaged shear stress metrics: TAWSS, OSI, transWSS, TAAFI, TACFI, WSSG;
  32. streamfunction: Calculates stream function of a 2D incompressible flow.
  33. surfdistance: Computes height of a prismatic boundary layer mesh and projects onto the surface (for e.g. y+ calculation).
  34. vorticity: Computes the vorticity field.
  35. wss: Computes wall shear stress field.

The module list above can be seen by running the command

FieldConvert -l

In the following we will detail the usage of each module.

5.5.1 Smooth the data: C0Projection module

To smooth the data of a given .fld file one can use the C0Projection module of FieldConvert

FieldConvert -m C0Projection test.xml test.fld test-C0Proj.fld

where the file test-C0Proj.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

The option localtoglobalmap will do a global gather of the coefficients and then scatter them back to the local elements. This will replace the coefficients shared between two elements with the coefficients of one of the elements (most likely the one with the highest id). Although not a formal projection it does not require any matrix inverse and so is very cheap to perform.

The option usexmlbcs will enforce the boundary conditions specified in the input xml file.

The option helmsmoothing=L will perform a Helmholtz smoothing projection of the form

(      (   ) )        (   )
   2    2π- 2   new     2π- 2 orig
  ∇  +   L     ^u    =   L    ^u

which can be interpreted in a Fourier sense as smoothing the original coefficients using a low pass filter of the form

             1                      2π
u^nekw =  -----2---2-^uokrig where  K0 = ---
        (1 + k ∕K 0)                  L

and so L is the length scale below which the coefficients values are halved or more. Since this form of the Helmholtz operator is not possitive definite, currently a direct solver is necessary and so this smoother is mainly of use in two-dimensions.

5.5.2 Calculate Q-Criterion: QCriterion module

To perform the Q-criterion calculation and obtain an output data containing the Q-criterion solution, the user can run

FieldConvert -m QCriterion test.xml test.fld test-QCrit.fld

where the file test-QCrit.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.3 Calculate λ2: L2Criterion module

To perform the λ2 vortex detection calculation and obtain an output data containing the values of the λ2 eigenvalue, the user can run

FieldConvert -m L2Criterion test.xml test.fld test-L2Crit.fld

where the file test-L2Crit.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.4 Add composite ID: addcompositeid module

When dealing with a geometry that has many surfaces, we need to identify the composites to assign boundary conditions. To assist in this, FieldConvert has a addcompositeid module, which adds the composite ID of every element as a new field. To use this we simply run

FieldConvert -m addcompositeid mesh.xml out.dat

In this case, we have produced a Tecplot file which contains the mesh and a variable that contains the composite ID. To assist in boundary identification, the input file mesh.xml should be a surface XML file that can be obtained through the NekMesh extract module (see section 4.4.3).

5.5.5 Add new field: fieldfromstring module

To modify or create a new field using an expression involving the existing fields, one can use the fieldfromstring module of FieldConvert

FieldConvert -m fieldfromstring:fieldstr="x+y+u":fieldname="result" \ 
file1.xml file2.fld file3.fld

In this case fieldstr is a required parameter describing a function of the coordinates and the existing variables, and fieldname is an optional parameter defining the name of the new or modified field (the default is newfield). file3.fld is the output containing both the original and the new fields, and can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.6 Sum two .fld files: addFld module

To sum two .fld files one can use the addFld module of FieldConvert

FieldConvert -m addfld:fromfld=file1.fld:scale=-1 file1.xml file2.fld \ 
file3.fld

In this case we use it in conjunction with the command scale which multiply the values of a given .fld file by a constant value. file1.fld is the file multiplied by value, file1.xml is the associated session file, file2.fld is the .fld file which is summed to file1.fld and finally file3.fld is the output which contain the sum of the two .fld files. file3.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.7 Combine two .fld files containing time averages: combineAvg module

To combine two .fld files obtained through the AverageFields or ReynoldsStresses filters, use the combineAvg module of FieldConvert

FieldConvert -m combineAvg:fromfld=file1.fld file1.xml file2.fld \ 
file3.fld

file3.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.8 Concatenate two files: concatenate module

To concatenate file1.fld and file2.fld into file-conc.fld one can run the following command

FieldConvert file.xml  file1.fld  file2.fld  file-conc.fld

where the file file-conc.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt. The concatenate module previously used for this purpose is not required anymore, and will be removed in a future release.

5.5.9 Count the number of DOF: dof module

To count the number of DOF in a solution file, one can run the following command

FieldConvert -m dof file.xml file.fld out.stdout

5.5.10 Equi-spaced output of data: equispacedoutput module

This module interpolates the output data to a truly equispaced set of points (not equispaced along the collapsed coordinate system). Therefore a tetrahedron is represented by a tetrahedral number of poinst. This produces much smaller output files. The points are then connected together by simplices (triangles and tetrahedrons).

FieldConvert -m equispacedoutput test.xml test.fld test.dat

or

FieldConvert -m equispacedouttput test.xml test.fld test.vtu

Note: Currently this option is only set up for triangles, quadrilaterals, tetrahedrons and prisms.

5.5.11 Extract a boundary region: extract module

The boundary region of a domain can be extracted from the output data using the following command line

FieldConvert -m extract:bnd=2 test.xml \ 
        test.fld test-boundary.fld

The option bnd specifies which boundary region to extract. Note this is different to NekMesh where the parameter surf is specified and corresponds to composites rather boundaries. If bnd is not provided, all boundaries are extracted to different fields. The output will be placed in test-boundary_b2.fld. If more than one boundary region is specified the extension _b0.fld, _b1.fld etc will be outputted. To process this file you will need an xml file of the same region. This can be generated using the command:

NekMesh -m extract:surf=5  test.xml test\_b0.xml

The surface to be extracted in this command is the composite number and so needs to correspond to the boundary region of interest. Finally to process the surface file one can use

FieldConvert  test\_b0.xml test\_b0.fld test\_b0.dat

This will obviously generate a Tecplot output if a .dat file is specified as last argument. A .vtu extension will produce a Paraview or VisIt output.

5.5.12 Compute the gradient of a field: gradient module

To compute the spatial gradients of all fields one can run the following command

FieldConvert -m gradient test.xml  test.fld  test-grad.fld

where the file file-grad.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.13 Convert HalfMode expansion to SingleMode for further processing: halfmodetofourier module

To obtain full Fourier expansion form a HalfMode result, use the comand:

FieldConvert -m halfmodetofourier file.xml half_mode_file.fld single_mode_file.fld

5.5.14 Extract a plane from 3DH1D expansion: homplane module

To obtain a 2D expansion containing one of the planes of a 3DH1D field file, use the command:

FieldConvert -m homplane:planeid=value file.xml file.fld file-plane.fld

If the option wavespace is used, the Fourier coefficients corresponding to planeid are obtained. The command in this case is:

FieldConvert -m homplane:wavespace:planeid=value file.xml \ 
    file.fld file-plane.fld

The output file file-plane.fld can be processed in a similar way as described in section 5.2 to visualise it either in Tecplot or in Paraview.

5.5.15 Stretch a 3DH1D expansion: homstretch module

To stretch a 3DH1D expansion in the z-direction, use the command:

FieldConvert -m homstretch:factor=value file.xml file.fld file-stretch.fld

The number of modes in the resulting field can be chosen using the command-line parameter --output-points-hom-z.

The output file file-stretch.fld can be processed in a similar way as described in section 5.2 to visualise it either in Tecplot or in Paraview.

5.5.16 Inner Product of a single or series of fields with respect to a single or series of fields: innerproduct module

You can take the inner product of one field with another field using the following command:

FieldConvert -m innerproduct:fromfld=file1.fld  file2.xml file2.fld \ 
out.stdout

This command will load the file1.fld and file2.fld assuming they both are spatially defined by files.xml and determine the inner product of these fields. The input option fromfld must therefore be specified in this module.

Optional arguments for this module are fields which allow you to specify the fields that you wish to use for the inner product, i.e.

FieldConvert -m innerproduct:fromfld=file1.fld:fields="0,1,2" file2.xml \ 
file2.fld out.stdout

will only take the inner product between the variables 0,1 and 2 in the two fields files. The default is to take the inner product between all fields provided.

Additional options include multifldids and allfromflds which allow for a series of fields to be evaluated in the following manner:

FieldConvert -m innerproduct:fromfld=file1.fld:multifldids="0-3"\ 
file2.xml  file2.fld out.stdout

will take the inner product between a file names field1_0.fld, field1_1.fld, field1_2.fld and field1_3.fld with respect to field2.fld.

Analogously including the options allfromflds, i.e.

FieldConvert -m innerproduct:fromfld=file1.fld:multifldids="0-3":\ 
allfromflds  file2.xml file2.fld out.stdout

Will take the inner product of all the from fields, i.e. field1_0.fld,field1_1.fld,field1_2.fld and field1_3.fld with respect to each other. This option essentially ignores file2.fld. Only the unique inner products are evaluated so if four from fields are given only the related trianuglar number 4 × 5∕2 = 10 of inner products are evaluated.

This option can be run in parallel.

5.5.17 Interpolate one field to another: interpfield module

To interpolate one field to another, one can use the following command:

FieldConvert -m interpfield:fromxml=file1.xml:fromfld=file1.fld \ 
        file2.xml file2.fld

This command will interpolate the field defined by file1.xml and file1.fld to the new mesh defined in file2.xml and output it to file2.fld. The fromxml and fromfld must be specified in this module. In addition there are two optional arguments clamptolowervalue and clamptouppervalue which clamp the interpolation between these two values. Their default values are -10,000,000 and 10,000,000.

Tip: This module can run in parallel where the speed is increased not only due to using more cores but also, since the mesh is split into smaller sub-domains, the search method currently adopted performs faster.

5.5.18 Interpolate scattered point data to a field: interppointdatatofld module

To interpolate discrete point data to a field, use the interppointdatatofld module:

FieldConvert -m interppointdatatofld:frompts=file1.pts file1.xml file1.fld

or alternatively for csv data:

FieldConvert -m interppointdatatofld:frompts=file1.csv file1.xml file1.fld

This command will interpolate the data from file1.pts (file1.csv) to the mesh and expansions defined in file1.xml and output the field to file1.fld. The file file.pts must be of the form:

1<?xml version="1.0" encoding="utf-8" ?> 
2<NEKTAR> 
3  <POINTS DIM="1" FIELDS="a,b,c"> 
4    1.0000 -1.0000 1.0000 -0.7778 
5    2.0000 -0.9798 0.9798 -0.7980 
6    3.0000 -0.9596 0.9596 -0.8182 
7    4.0000 -0.9394 0.9394 -0.8384 
8  </POINTS> 
9</NEKTAR>

where DIM="1" FIELDS="a,b,c specifies that the field is one-dimensional and contains three variables, a, b, and c. Each line defines a point, while the first column contains its x-coordinate, the second one contains the a-values, the third the b-values and so on. In case of n-dimensional data, the n coordinates are specified in the first n columns accordingly. An equivalent csv file is:

# x, a, b, c 
1.0000,-1.0000,1.0000,-0.7778 
2.0000,-0.9798,0.9798,-0.7980 
3.0000,-0.9596,0.9596,-0.8182 
4.0000,-0.9394,0.9394,-0.8384

In order to interpolate 1D data to a nD field, specify the matching coordinate in the output field using the interpcoord argument:

FieldConvert -m interppointdatatofld:frompts=1D-file1.pts:interpcoord=1 \ 
        3D-file1.xml 3D-file1.fld

This will interpolate the 1D scattered point data from 1D-file1.pts to the y-coordinate of the 3D mesh defined in 3D-file1.xml. The resulting field will have constant values along the x and z coordinates. For 1D Interpolation, the module implements a quadratic scheme and automatically falls back to a linear method if only two data points are given. A modified inverse distance method is used for 2D and 3D interpolation. Linear and quadratic interpolation require the data points in the .pts-file to be sorted by their location in ascending order. The Inverse Distance implementation has no such requirement.

5.5.19 Interpolate a field to a series of points: interppoints module

You can interpolate one field to a series of given points using the following command:

FieldConvert -m interppoints:fromxml=file1.xml:fromfld=\ 
        file1.fld:topts=file2.pts file2.dat

This command will interpolate the field defined by file1.xml and file1.fld to the points defined in file2.pts and output it to file2.dat. The fromxml and fromfld must be specified in this module. The format of the file file2.pts is of the same form as for the interppointdatatofld module:

1<?xml version="1.0" encoding="utf-8" ?> 
2<NEKTAR> 
3  <POINTS DIM="2" FIELDS=""> 
4    0.0 0.0 
5    0.5 0.0 
6    1.0 0.0 
7  </POINTS> 
8</NEKTAR>

Similar to the interppointdatatofld module, the .pts file can be interchanged with a .csv file (the output can also be written to .csv):

# x, y 
0.0,0.0 
0.5,0.0 
1.0,0.0

There are three optional arguments clamptolowervalue, clamptouppervalue and defaultvalue the first two clamp the interpolation between these two values and the third defines the default value to be used if the point is outside the domain. Their default values are -10,000,000, 10,000,000 and 0.

In addition, instead of specifying the file file2.pts, a module list of the form

FieldConvert -m interppoints:fromxml=file1.xml:fromfld= \ 
        file1.fld:line=npts,x0,y0,x1,y1 file2.dat

can be specified where npts is the number of equispaced points between (x0,y0) to (x1,y1). This also works in 3D, by specifying (x0,y0,z0) to (x1,y1,z1).

An extraction of a plane of points can also be specified by

FieldConvert -m interppoints:fromxml=file1.xml:fromfld=file1.fld:\ 
      plane=npts1,npts2,x0,y0,z0,x1,y1,z1,x2,y2,z2,x3,y3,z3 file2.dat

where npts1,npts2 is the number of equispaced points in each direction and (x0,y0,z0), (x1,y1,z1), (x2,y2,z2) and (x3,y3,z3) define the plane of points specified in a clockwise or anticlockwise direction.

In addition, an extraction of a box of points can also be specified by

FieldConvert -m interppoints:fromxml=file1.xml:fromfld=file1.fld:\ 
     box=npts1,npts2,npts3,xmin,xmax,ymin,ymax,zmin,zmax file2.dat

where npts1,npts2,npts3 is the number of equispaced points in each direction and (xmin,ymin,zmin) and (xmax,ymax,zmax) define the limits of the box of points.

There is also an additional optional argument cp=p0,q which adds to the interpolated fields the value of cp = (p-p0)∕q and cp0 = (p-p0 + 0.5u2)∕q where p0 is a reference pressure and q is the free stream dynamics pressure. If the input does not contain a field “p” or a velocity field “u,v,w” then cp and cp0 are not evaluated accordingly

Note: This module runs in parallel for the line, plane and box extraction of points.

5.5.20 Interpolate a set of points to another: interpptstopts module

You can interpolate one set of points to another using the following command:

FieldConvert file1.pts -m interpptstopts:topts=file2.pts file2.dat

This command will interpolate the data in file1.pts to a new set of points defined in file2.pts and output it to file2.dat.

Similarly to the interppoints module, the target point distribution can also be specified using the line, plane or box options. The optional arguments clamptolowervalue, clamptouppervalue, defaultvalue and cp are also supported with the same meaning as in interppoints.

One useful application for this module is with 3DH1D expansions, for which currently the interppoints module does not work. In this case, we can use for example

FieldConvert file1.xml file1.fld -m interpptstopts:\ 
    plane=npts1,npts2,x0,y0,z0,x1,y1,z1,x2,y2,z2,x3,y3,z3 \ 
    file2.dat

With this usage, the equispacedoutput module will be automatically called to interpolate the field to a set of equispaced points in each element. The result is then interpolated to a plane by the interpptstopts module.

Note: This module does not work in parallel.

5.5.21 Isocontour extraction: iscontour module

Extract an isocontour from a field file. This option automatically take the field to an equispaced distribution of points connected by linear simplicies of triangles or tetrahedrons. The linear simplices are then inspected to extract the isocontour of interest. To specify the field fieldid can be provided giving the id of the field of interest and fieldvalue provides the value of the isocontour to be extracted.

FieldConvert -m isocontour:fieldid=2:fieldvalue=0.5 test.xml test.fld \ 
        test-isocontour.dat

Alternatively fieldstr="u+v" can be specified to calculate the field u + v and extract its isocontour. You can also specify fieldname="UplusV" to define the name of the isocontour in the .dat file, i.e.

FieldConvert -m isocontour:fieldstr="u+v":fieldvalue=0.5:\ 
  fieldname="UplusV" test.xml  test.fld test-isocontour.dat

Optionally smooth can be specified to smooth the isocontour with default values smoothnegdiffusion=0.495, smoothnegdiffusion=0.5 and smoothiter=100. This option typically should be used wiht the globalcondense option which removes multiply defined verties from the simplex definition which arise as isocontour are generated element by element. The smooth option preivously automatically called the globalcondense option but this has been depracated since it is now possible to read isocontour files directly and so it is useful to have these as separate options.

In addition to the smooth or globalcondense options you can specify removesmallcontour=100 which will remove separate isocontours of less than 100 triangles.

Note: Currently this option is only set up for triangles, quadrilaterals, tetrahedrons and prisms.

5.5.22 Show high frequency energy of the Jacobian: jacobianenergy module

FieldConvert -m jacobianenergy file.xml file.fld jacenergy.fld

The option topmodes can be used to specify the number of top modes to keep.

The output file jacenergy.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.23 Calculate mesh quality: qualitymetric module

The qualitymetric module assesses the quality of the mesh by calculating a per-element quality metric and adding an additional field to any resulting output. This does not require any field input, therefore an example usage looks like

FieldConvert -m qualitymetric mesh.xml mesh-with-quality.dat

Two quality metrics are implemented that produce scalar fields Q:

5.5.24 Evaluate the mean of variables on the domain: mean module

To evaluate the mean of variables on the domain one can use the mean module of FieldConvert

FieldConvert -m mean file1.xml file2.fld out.stdout

This module does not create an output file which is reinforced by the out.stdout option. The integral and mean for each field variable are then printed to the stdout.

5.5.25 Extract mean mode of 3DH1D expansion: meanmode module

To obtain a 2D expansion containing the mean mode (plane zero in Fourier space) of a 3DH1D field file, use the command:

FieldConvert -m meanmode file.xml file.fld file-mean.fld

The output file file-mean.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot or in Paraview or VisIt.

5.5.26 Project point data to a field: pointdatatofld module

To project a series of points given at the same quadrature distribution as the .xml file and write out a .fld file use the pointdatatofld module:

FieldConvert -m pointdatatofld:frompts=file.pts file.xml file.fld

This command will read in the points provided in the file.pts and assume these are given at the same quadrature distribution as the mesh and expansions defined in file.xml and output the field to file.fld. If the points do not match an error will be dumped.

The file file.pts which is assumed to be given by an interpolation from another source is of the form:

1<?xml version="1.0" encoding="utf-8" ?> 
2<NEKTAR> 
3  <POINTS DIM="3" FIELDS="p"> 
4  1.70415 -0.4       -0.0182028      -0.106893 
5  1.70415 -0.395683  -0.0182028      -0.106794 
6  1.70415 -0.3875    -0.0182028      -0.106698 
7  1.70415 -0.379317  -0.0182028      -0.103815 
8  </POINTS> 
9</NEKTAR>

where DIM="3" FIELDS="p specifies that the field is three-dimensional and contains one variable, p. Each line defines a point, the first, second, and third columns contains the x,y,z-coordinate and subsequent columns contain the field values, in this case the p-value So in the general case of n-dimensional data, the n coordinates are specified in the first n columns accordingly followed by the field data. Alternatively, the file.pts can be interchanged with a csv file.

The default argument is to use the equispaced (but potentially collapsed) coordinates which can be obtained from the command.

FieldConvert file.xml file.dat

In this case the pointdatatofld module should be used without the –noequispaced option. However this can lead to problems when peforming an elemental forward projection/transform since the mass matrix in a deformed element can be singular as the equispaced points do not have a sufficiently accurate quadrature rule that spans the polynomial space. Therefore it is advisable to use the set of points given by

FieldConvert --noequispaced file.xml file.dat

which produces a set of points at the gaussian collapsed coordinates.

Finally the option setnantovalue=0 can also be used which sets any nan values in the interpolation to zero or any specified value in this option.

5.5.27 Print L2 and LInf norms: printfldnorms module

FieldConvert -m printfldnorms test.xml test.fld out.stdout

This module does not create an output file which is reinforced by the out.stdout option. The L2 and LInf norms for each field variable are then printed to the stdout.

5.5.28 Removes one or more fields from .fld files: removefield module

This module allows to remove one or more fields from a .fld file:

FieldConvert -m removefield:fieldname="u,v,p" test.xml test.fld test-removed.fld

where the file test-removed.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt. The lighter resulting file speeds up the postprocessing of large files when not all fields are required.

5.5.29 Computes the scalar gradient: scalargrad module

The scalar gradient of a field is computed by running:

FieldConvert -m scalargrad:bnd=0 test.xml test.fld test-scalgrad.fld

The option bnd specifies which boundary region to extract. Note this is different to NekMesh where the parameter surf is specified and corresponds to composites rather boundaries. If bnd is not provided, all boundaries are extracted to different fields. To process this file you will need an xml file of the same region.

5.5.30 Scale a given .fld: scaleinputfld module

To scale a .fld file by a given scalar quantity, the user can run:

FieldConvert -m scaleinputfld:scale=value test.xml test.fld test-scal.fld

The argument scale=value rescales of a factor value test.fld by the factor value. The output file file-scal.fld can be processed in a similar way as described in section 5.2 to visualise the result either in Tecplot, Paraview or VisIt.

5.5.31 Time-averaged shear stress metrics: shear module

Time-dependent wall shear stress derived metrics relevant to cardiovascular fluid dynamics research can be computed using this module. They are

To compute these, the user can run:

FieldConvert -m shear:N=value:fromfld=test_id_b0.fld \ 
    test.xml test-multishear.fld

The argument N and fromfld are compulsory arguments that respectively define the number of fld files corresponding to the number of discrete equispaced time-steps, and the first fld file which should have the form of test_id_b0.fld where the first underscore in the name marks the starting time-step file ID.

The input .fld files are the outputs of the wss module. If they do not contain the surface normals (an optional output of the wss modle), then the shear module will not compute the last metric, |WSSG|.

5.5.32 Stream function of a 2D incompressible flow: streamfunction module

The streamfunction module calculates the stream function of a 2D incompressible flow, by solving the Poisson equation

∇2ψ =  - ω

where ω is the vorticity. Note that this module applies the same boundary conditions specified for the y-direction velocity component v to the stream function, what may not be the most appropriate choice.

To use this module, the user can run

FieldConvert -m streamfunction test.xml test.fld test-streamfunc.fld

where the file test-streamfunc.fld can be processed in a similar way as described in section 5.2.

5.5.33 Boundary layer height calculation: surfdistance module

The surface distance module computes the height of a boundary layer formed by quadrilaterals (in 2D) or prisms and hexahedrons (in 3D) and projects this value onto the surface of the boundary, in a similar fashion to the extract module. In conjunction with a mesh of the surface, which can be obtained with NekMesh, and a value of the average wall shear stress, one potential application of this module is to determine the distribution of y+ grid spacings for turbulence calculations.

To compute the height of the prismatic layer connected to boundary region 3, the user can issue the command:

FieldConvert -m surfdistance:bnd=3 input.xml output.fld

Note that no .fld file is required, since the mesh is the only input required in order to calculate the element height. This produces a file output_b3.fld, which can be visualised with the appropriate surface mesh from NekMesh.

5.5.34 Calculate vorticity: vorticity module

To perform the vorticity calculation and obtain an output data containing the vorticity solution, the user can run

FieldConvert -m vorticity test.xml test.fld test-vort.fld

where the file test-vort.fld can be processed in a similar way as described in section 5.2.

5.5.35 Computing the wall shear stress: wss module

To obtain the wall shear stres vector and magnitude, the user can run:

FieldConvert -m wss:bnd=0:addnormals=1 test.xml test.fld test-wss.fld

The option bnd specifies which boundary region to extract. Note this is different to NekMesh where the parameter surf is specified and corresponds to composites rather boundaries. If bnd is not provided, all boundaries are extracted to different fields. The addnormals is an optional command argument which, when turned on, outputs the normal vector of the extracted boundary region as well as the shear stress vector and magnitude. This option is off by default. To process the output file(s) you will need an xml file of the same region.

5.5.36 Manipulating meshes with FieldConvert

FieldConvert has support for two modules that can be used in conjunction with the linear elastic solver, as shown in chapter 12. To do this, FieldConvert has an XML output module, in addition to the Tecplot and VTK formats.

The deform module, which takes no options, takes a displacement field and applies it to the geometry, producing a deformed mesh:

FieldConvert -m deform input.xml input.fld deformed.xml

The displacement module is designed to create a boundary condition field file. Its intended use is for mesh generation purposes. It can be used to calculate the displacement between the linear mesh and a high-order surface, and then produce a fld file, prescribing the displacement at the boundary, that can be used in the linear elasticity solver.

Presently the process is somewhat convoluted and must be used in conjunction with NekMesh to create the surface file. However the bash input below describes the procedure. Assume the high-order mesh is in a file called mesh.xml, the linear mesh is mesh-linear.xml that can be generated by removing the CURVED section from mesh.xml, and that we are interested in the surface with ID 123.

# Extract high order surface 
NekMesh -m extract:surf=123 mesh.xml mesh-surf-curved.xml 
 
# Use FieldConvert to calculate displacement between two surfaces 
FieldConvert -m displacement:id=123:to=mesh-surf-curved.xml \ 
    mesh-linear.xml mesh-deformation.fld 
 
# mesh-deformation.fld is used as a boundary condition inside the 
# solver to prescribe the deformation conditions.xml contains 
# appropriate Nektar++ parameters (mu, E, other BCs, ...) 
LinearElasticSolver mesh-linear.xml conditions.xml 
 
# This produces the final field mesh-linear.fld which is the 
# displacement field, use FieldConvert to apply it: 
FieldConvert-g -m deform mesh-linear.xml mesh-linear.fld mesh-deformed.xml