I have a code that has a time-stepping algorithm in which a distributed matrix is solved at each time step. After a number of time steps, Intel MPI crashes with the error message
MPIR_Get_contextid_sparse_group(1330): Too many communicators (0/16384 free on this process; ignore_id=0)
Fatal error in PMPI_Cart_create: Other MPI error, error stack:
I've attached a simple test case that exhibits the problem. I'm using Intel MPI 5.3.3, MKL 11.3.3, and ifort 6.0.3. I've experienced the problem on both windows and linux.
Is this a bug and/or is there a workaround? This is really a show-stopper for use as our simulations can have a huge number of time-steps and it doesn't take many simulation steps to exhaust all the communicators for some problems.
Thanks,
John