Problem with MPI + Quasi-3D + AWGN initialization for ICNS solver
******************* This email originates from outside Imperial. Do not click on links and attachments unless you recognise the sender. If you trust the sender, add them to your safe senders list https://spam.ic.ac.uk/SpamConsole/Senders.aspx to disable email stamping for this address. ******************* Hello everybody, I am currently running parallel quasi-3D simulations (Hom1D) with the incompressible Navier-Stokes solver within Nektar++ version 5.3.0 and the data transfer at the partition boundaries appears to be faulty when initializing the velocity field in the homogeneous z-direction with additive white Gaussian noise as can be seen here: [g853.png] I don't know if this problem still exists in the current Nektar version, but I couldn't find anything about it in the user archives and therefore wanted to draw attention to this. All the best, Alex
Hi Alex, Thanks for your email. Unfortunately, this issue exists with parallel simulations for AWGN initial conditions. I am currently figuring out a solution for this which seems to be related to enforcing C0 continuity at the partition boundaries. A workaround is to initialise an AWGN initial condition in serial (-np 1) by setting TFinal = 0 - let's say the solution field will be written to solution.fld Then start a parallel run from the solution.fld. Cheers, Chi Hin ________________________________ From: nektar-users-bounces@imperial.ac.uk <nektar-users-bounces@imperial.ac.uk> on behalf of Alexander Schukmann <alexander.schukmann@protonmail.com> Sent: 25 March 2024 10:07 To: nektar-users <nektar-users@imperial.ac.uk> Subject: [Nektar-users] Problem with MPI + Quasi-3D + AWGN initialization for ICNS solver This email from alexander.schukmann@protonmail.com originates from outside Imperial. Do not click on links and attachments unless you recognise the sender. If you trust the sender, add them to your safe senders list<https://spam.ic.ac.uk/SpamConsole/Senders.aspx> to disable email stamping for this address. Hello everybody, I am currently running parallel quasi-3D simulations (Hom1D) with the incompressible Navier-Stokes solver within Nektar++ version 5.3.0 and the data transfer at the partition boundaries appears to be faulty when initializing the velocity field in the homogeneous z-direction with additive white Gaussian noise as can be seen here: [g853.png] I don't know if this problem still exists in the current Nektar version, but I couldn't find anything about it in the user archives and therefore wanted to draw attention to this. All the best, Alex
participants (2)
- 
                
                Alexander Schukmann
- 
                
                Chan, Chi Hin