- 19 May, 2021 1 commit
-
-
Andreas Marek authored
-
- 13 May, 2021 1 commit
-
-
Andreas Marek authored
-
- 12 May, 2021 1 commit
-
-
Andreas Marek authored
It turned out that users ignored the warnings of the test programs if their MPI library did not provide "MPI_THREAD_SERIALIED" or "MPI_THREAD_MULTIPLE". Thus for safty reasons, ELPA does from now on during the call to "elpa_setup" check the provided threading level of the MPI library. If the provided level is too low, ELPA will do the following - limit the number of OpenMP threads **internal** to ELPA to 1 - print a warning about this - ignore the settings of the user via -- the OMP_NUM_THREADS variable -- calling the set-method set("omp_threads", value) These settings will **not** affect the run-time of a threaded blas and lapack library, **if** the number of threads for these libraries can be controlled independently of the OMP_NUM_THREADS variable (for example for Intel's MKL one can use MKL_NUM_THREADS)
-
- 06 May, 2021 2 commits
-
-
Andreas Marek authored
-
Andreas Marek authored
-
- 05 May, 2021 1 commit
-
-
Andreas Marek authored
-
- 03 Mar, 2021 1 commit
-
-
Andreas Marek authored
-
- 26 Feb, 2021 2 commits
-
-
Andreas Marek authored
-
Andreas Marek authored
- Rename keyword "gpu" -> "nvidia-gpu" - Add keyword "amd-gpu"
-
- 18 Feb, 2021 1 commit
-
-
Andreas Marek authored
For ELPA 1stage now you can use GPUs in the OpenMP enabled version of ELPA. Currently this implies that within ELPA1 only _ONE_ OpenMP thread is used, independent of how many threads the calling application uses
-
- 17 Feb, 2021 1 commit
-
-
Andreas Marek authored
-
- 21 Jan, 2021 1 commit
-
-
Andreas Marek authored
-
- 14 Dec, 2020 1 commit
-
-
Andreas Marek authored
-
- 25 Sep, 2020 1 commit
-
-
Andreas Marek authored
-
- 10 Aug, 2020 1 commit
-
-
Andreas Marek authored
-
- 31 Jul, 2020 2 commits
- 10 Jun, 2020 1 commit
-
-
Andreas Marek authored
-
- 05 Jun, 2020 1 commit
-
-
Wenzhe Yu authored
-
- 08 Apr, 2020 1 commit
-
-
Andreas Marek authored
-
- 07 Apr, 2020 1 commit
-
-
Andreas Marek authored
-
- 24 Mar, 2020 1 commit
-
-
Andreas Marek authored
-
- 05 Mar, 2020 2 commits
-
-
Andreas Marek authored
-
Andreas Marek authored
-
- 04 Mar, 2020 1 commit
-
-
Andreas Marek authored
-
- 03 Mar, 2020 2 commits
-
-
Andreas Marek authored
-
Andreas Marek authored
-
- 02 Mar, 2020 1 commit
-
-
Andreas Marek authored
-
- 28 Feb, 2020 1 commit
-
-
Andreas Marek authored
-
- 24 Oct, 2019 1 commit
-
-
Andreas Marek authored
-
- 23 Oct, 2019 1 commit
-
-
Andreas Marek authored
-
- 22 Oct, 2019 1 commit
-
-
Sebastian Ohlmann authored
When profiling the GPU version, NVTX can be used to highlight the corresponding regions of the code in the timeline of the profiling tool (nvvp or nsight systems). This is very useful to correlate what happens on the GPU with what part of the code we are in.
-
- 19 Oct, 2019 1 commit
-
-
Andreas Marek authored
-
- 17 Oct, 2019 1 commit
-
-
Andreas Marek authored
ELPA can now be linked against a 64bit integer version of MPI and ScalaPack. This is an experimental feature The following points are still to be done - does not work with real QR-decomposition - generalized routines return wrong results - the C tests and the C Cannon algorithm implementation do not work (no 64bit header files for MPI *at least* with Intel MPI)
-
- 19 Jun, 2019 1 commit
-
-
Carolin Penke authored
-
- 16 May, 2019 1 commit
-
-
Pavel Kus authored
also introducing --with-likwid parameter to the configure script Can be used to compile with likwid and to enable the likwid to measure performance of individual compute steps
-
- 18 Apr, 2019 1 commit
-
-
Andreas Marek authored
This solves issue #66
-
- 15 Nov, 2018 1 commit
-
-
Pavel Kus authored
-
- 15 Oct, 2018 1 commit
-
-
Pavel Kus authored
The GPU initialization is actually quite constly, e.g. on Minsky it takes roughly 0.7s. That is hurting performance for small matrices. Thus a check has been added and now GPU should be initialized only the first time.
-
- 13 Oct, 2018 1 commit
-
-
Pavel Kus authored
-