Skip to content 3.56 KiB
Newer Older

## Summary Version


## Purpose of Benchmark

The Alya System is a Computational Mechanics code capable of solving different physics, each one with its own modelization characteristics, in a coupled way. Among the problems it solves are: convection-diffusion reactions, incompressible flows, compressible flows, turbulence, bi-phasic flows and free surface, excitable media, acoustics, thermal flow, quantum mechanics (DFT) and solid mechanics (large strain). ALYA is written in Fortran 90/95 and parallelized using MPI and OpenMP.

* Web site:

* Code download:

* Test Case A:

* Test Case B:

## Mechanics of Building Benchmark

Cristian Morales's avatar
Cristian Morales committed
Alya builds the makefile from the compilation options defined in In order to build ALYA (Alya.x), please follow these steps after unpack the tar.gz:
Cristian Morales's avatar
Cristian Morales committed
Go to to directory: Executables/unix

cd Executables/unix

Edit (some default files can be found in directory

* Select your own MPI wrappers and paths
* Select size of integers. Default is 4 bytes, For 8 bytes, select -DI8
* Choose your metis version, metis-4.0 or metis-5.1.0_i8 for 8-bytes integers

Configure Alya: 

    ./configure -x nastin parall

Compile metis:  

    make metis4 


    make metis5
Cristian Morales's avatar
Cristian Morales committed
Finally, compile Alya:   

    make -j 8

## Mechanics of Running Benchmark

### Datasets

The parameters used in the datasets try to represent at best typical industrial runs in order to obtain representative speedups. For example, the iterative solvers are never converged to machine accuracy, but only as a percentage of the initial residual. 

The different datasets are:

    Test Case A: SPHERE_16.7M ... 16.7M sphere mesh
    Test Case B: SPHERE_132M .... 132M sphere mesh

### How to execute Alya with a given dataset

In order to run ALYA, you need at least the following input files per execution:


In our case X=sphere

To execute a simulation, you must be inside the input directory and you should submit a job like:

    mpirun Alya.x sphere

Cristian Morales's avatar
Cristian Morales committed
How to measure the performance

There are many ways to compute the scalability of Nastin module.

1. **For the complete cycle including: element assembly + boundary assembly + subgrid scale assembly + solvers, etc.**  

> In *.nsi.cvg file, column "30. Elapsed CPU time"

2. **For single kernels: element assembly, boundary assembly, subgrid scale assembly, solvers**.  Single kernels. Here, average and maximum times are indicated in *.nsi.cvg at each iteration of each time step:

>     Element assembly: 19. Ass. ave cpu time    20. Ass. max cpu time
>     Boundary assembly: 33. Bou. ave cpu time  34. Bou. max cpu time
>     Subgrid scale assembly: 31. SGS ave cpu time     32. SGS max cpu time
>     Iterative solvers: 21. Sol. ave cpu time     22. Sol. max cpu time
> Note that in the case of using Runge-Kutta time integration (the case
> of the sphere), the element and boundary assembly times are this of
> the last assembly of current time step (out of three for third order).

3. **Using overall times**. 

> At the end of *.log file, total timings are shown for all modules. In
> this case we use the first value of the NASTIN MODULE.


If you have any question regarding the runs, please feel free to contact Guillaume Houzeaux: