test_complex.F90 11.3 KB
Newer Older
1
2
!    This file is part of ELPA.
!
3
!    The ELPA library was originally created by the ELPA consortium,
4
5
!    consisting of the following organizations:
!
6
7
!    - Max Planck Computing and Data Facility (MPCDF), formerly known as
!      Rechenzentrum Garching der Max-Planck-Gesellschaft (RZG),
8
9
10
!    - Bergische Universität Wuppertal, Lehrstuhl für angewandte
!      Informatik,
!    - Technische Universität München, Lehrstuhl für Informatik mit
11
12
!      Schwerpunkt Wissenschaftliches Rechnen ,
!    - Fritz-Haber-Institut, Berlin, Abt. Theorie,
13
!    - Max-Plack-Institut für Mathematik in den Naturwissenschaften,
14
15
!      Leipzig, Abt. Komplexe Strukutren in Biologie und Kognition,
!      and
16
17
18
19
!    - IBM Deutschland GmbH
!
!
!    More information can be found here:
20
!    http://elpa.mpcdf.mpg.de/
21
22
!
!    ELPA is free software: you can redistribute it and/or modify
23
24
!    it under the terms of the version 3 of the license of the
!    GNU Lesser General Public License as published by the Free
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
!    Software Foundation.
!
!    ELPA is distributed in the hope that it will be useful,
!    but WITHOUT ANY WARRANTY; without even the implied warranty of
!    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
!    GNU Lesser General Public License for more details.
!
!    You should have received a copy of the GNU Lesser General Public License
!    along with ELPA.  If not, see <http://www.gnu.org/licenses/>
!
!    ELPA reflects a substantial effort on the part of the original
!    ELPA consortium, and we ask you to respect the spirit of the
!    license that we chose: i.e., please contribute any changes you
!    may have back to the original ELPA library distribution, and keep
!    any derivatives of ELPA under the same license that we chose for
!    the original distribution, the GNU Lesser General Public License.
!
!
#include "config-f90.h"
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
!>
!> Fortran test programm to demonstrates the use of
!> ELPA 1 complex case library.
!> If "HAVE_REDIRECT" was defined at build time
!> the stdout and stderr output of each MPI task
!> can be redirected to files if the environment
!> variable "REDIRECT_ELPA_TEST_OUTPUT" is set
!> to "true".
!>
!> By calling executable [arg1] [arg2] [arg3] [arg4]
!> one can define the size (arg1), the number of
!> Eigenvectors to compute (arg2), and the blocking (arg3).
!> If these values are not set default values (4000, 1500, 16)
!> are choosen.
!> If these values are set the 4th argument can be
!> "output", which specifies that the EV's are written to
!> an ascii file.
!>
62
program test_complex_double_precision
63
64
65
66
67
68

!-------------------------------------------------------------------------------
! Standard eigenvalue problem - COMPLEX version
!
! This program demonstrates the use of the ELPA module
! together with standard scalapack routines
69
!
70
71
72
73
74
! Copyright of the original code rests with the authors inside the ELPA
! consortium. The copyright of any additional modifications shall rest
! with their original authors, but shall adhere to the licensing terms
! distributed along with the original code in the file "COPYING".
!-------------------------------------------------------------------------------
75
   use precision
76
   use elpa1_legacy
77
   use elpa_utilities, only : error_unit
78
79
80
#ifdef WITH_OPENMP
   use test_util
#endif
81

Andreas Marek's avatar
Andreas Marek committed
82
83
84
85
86
   use mod_read_input_parameters
   use mod_check_correctness
   use mod_setup_mpi
   use mod_blacs_infrastructure
   use mod_prepare_matrix
87
   use elpa_mpi
88
89
90
#ifdef HAVE_REDIRECT
   use redirect
#endif
91

92
93
94
#ifdef HAVE_DETAILED_TIMINGS
 use timings
#endif
95
  use output_types
96
97
98
99
100
101
102
103
104
   implicit none

   !-------------------------------------------------------------------------------
   ! Please set system size parameters below!
   ! na:   System size
   ! nev:  Number of eigenvectors to be calculated
   ! nblk: Blocking factor in block cyclic distribution
   !-------------------------------------------------------------------------------

105
106
   integer(kind=ik)              :: nblk
   integer(kind=ik)              :: na, nev
107

108
   integer(kind=ik)              :: np_rows, np_cols, na_rows, na_cols
109

110
111
   integer(kind=ik)              :: myid, nprocs, my_prow, my_pcol, mpi_comm_rows, mpi_comm_cols
   integer(kind=ik)              :: i, mpierr, my_blacs_ctxt, sc_desc(9), info, nprow, npcol
112

113
   real(kind=rk8), allocatable    :: ev(:)
Andreas Marek's avatar
Andreas Marek committed
114

115
   complex(kind=ck8), allocatable :: a(:,:), z(:,:), as(:,:)
116

117
   complex(kind=ck8), parameter   :: CZERO = (0._rk8,0.0_rk8), CONE = (1._rk8,0._rk8)
118

119
   integer(kind=ik)              :: STATUS
120
#ifdef WITH_OPENMP
121
   integer(kind=ik)              :: omp_get_max_threads,  required_mpi_thread_level, provided_mpi_thread_level
122
#endif
123
   type(output_t)                :: write_to_file
124
   logical                       :: success
125
126
   character(len=8)              :: task_suffix
   integer(kind=ik)              :: j
127

128
129
#define DOUBLE_PRECISION_COMPLEX 1

130
   success = .true.
Andreas Marek's avatar
Andreas Marek committed
131
132
   ! read input parameters if they are provided
   call read_input_parameters(na, nev, nblk, write_to_file)
133
134
135

   !-------------------------------------------------------------------------------
   !  MPI Initialization
Andreas Marek's avatar
Andreas Marek committed
136
   call setup_mpi(myid, nprocs)
137

138
   STATUS = 0
139

Andreas Marek's avatar
Andreas Marek committed
140
#define COMPLEXCASE
141
#define ELPA1
142
#include "elpa_print_headers.X90"
143

144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
#ifdef HAVE_DETAILED_TIMINGS

   ! initialise the timing functionality

#ifdef HAVE_LIBPAPI
   call timer%measure_flops(.true.)
#endif

   call timer%measure_allocated_memory(.true.)
   call timer%measure_virtual_memory(.true.)
   call timer%measure_max_allocated_memory(.true.)

   call timer%set_print_options(&
#ifdef HAVE_LIBPAPI
                print_flop_count=.true., &
                print_flop_rate=.true., &
#endif
                print_allocated_memory = .true. , &
                print_virtual_memory=.true., &
                print_max_allocated_memory=.true.)


  call timer%enable()

168
  call timer%start("program: test_complex_double_precision")
169
170
#endif

171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
   !-------------------------------------------------------------------------------
   ! Selection of number of processor rows/columns
   ! We try to set up the grid square-like, i.e. start the search for possible
   ! divisors of nprocs with a number next to the square root of nprocs
   ! and decrement it until a divisor is found.

   do np_cols = NINT(SQRT(REAL(nprocs))),2,-1
      if(mod(nprocs,np_cols) == 0 ) exit
   enddo
   ! at the end of the above loop, nprocs is always divisible by np_cols

   np_rows = nprocs/np_cols

   if(myid==0) then
      print *
Pavel Kus's avatar
Pavel Kus committed
186
187
188
189
190
191
192
193
194
195
      print '(a)','Standard eigenvalue problem - ELPA1, COMPLEX version'
      print *
      print '((a,i0))', 'Matrix size: ', na 
      print '((a,i0))', 'Num eigenvectors: ', nev
      print '((a,i0))', 'Blocksize: ', nblk 
      print '((a,i0))', 'Num MPI proc: ', nprocs 
      print '((a))', 'Using gpu: NO'
      print '((a,i0))', 'Num gpu devices: ', 0
      print '((a))', 'Number type: complex'
      print '((a))', 'Number precision: double'
196
197
198
199
200
      print *
      print '(3(a,i0))','Number of processor rows=',np_rows,', cols=',np_cols,', total=',nprocs
      print *
   endif

Pavel Kus's avatar
Pavel Kus committed
201
  !-------------------------------------------------------------------------------
202
203
204
205
206
207
208
209
210
   ! Set up BLACS context and MPI communicators
   !
   ! The BLACS context is only necessary for using Scalapack.
   !
   ! For ELPA, the MPI communicators along rows/cols are sufficient,
   ! and the grid setup may be done in an arbitrary way as long as it is
   ! consistent (i.e. 0<=my_prow<np_rows, 0<=my_pcol<np_cols and every
   ! process has a unique (my_prow,my_pcol) pair).

Andreas Marek's avatar
Andreas Marek committed
211
212
   call set_up_blacsgrid(mpi_comm_world, my_blacs_ctxt, np_rows, np_cols, &
                         nprow, npcol, my_prow, my_pcol)
213
214
215
216
217
218

   if (myid==0) then
     print '(a)','| Past BLACS_Gridinfo.'
   end if

   ! All ELPA routines need MPI communicators for communicating within
219
   ! rows or columns of processes, these are set in elpa_get_communicators.
220

221
   mpierr = elpa_get_communicators(mpi_comm_world, my_prow, my_pcol, &
222
                                   mpi_comm_rows, mpi_comm_cols)
223
224
225
226
227
228
229
230

   if (myid==0) then
     print '(a)','| Past split communicator setup for rows and columns.'
   end if

   ! Determine the necessary size of the distributed matrices,
   ! we use the Scalapack tools routine NUMROC for that.

Andreas Marek's avatar
Andreas Marek committed
231
232
   call set_up_blacs_descriptor(na ,nblk, my_prow, my_pcol, np_rows, np_cols, &
                                na_rows, na_cols, sc_desc, my_blacs_ctxt, info)
233
234
235
236
237
238
239
240

   if (myid==0) then
     print '(a)','| Past scalapack descriptor setup.'
   end if

   !-------------------------------------------------------------------------------
   ! Allocate matrices and set up a test matrix for the eigenvalue problem

241
242
243
#ifdef HAVE_DETAILED_TIMINGS
   call timer%start("set up matrix")
#endif
244
245
246
247
248
249
   allocate(a (na_rows,na_cols))
   allocate(z (na_rows,na_cols))
   allocate(as(na_rows,na_cols))

   allocate(ev(na))

250
   call prepare_matrix_double(na, myid, sc_desc, a, z, as)
251

252
253
254
#ifdef HAVE_DETAILED_TIMINGS
   call timer%stop("set up matrix")
#endif
255
256
257
258
259
260
261
   !-------------------------------------------------------------------------------
   ! Calculate eigenvalues/eigenvectors

   if (myid==0) then
     print '(a)','| Entering one-step ELPA solver ... '
     print *
   end if
262
#ifdef WITH_MPI
263
   call mpi_barrier(mpi_comm_world, mpierr) ! for correct timings only
264
#endif
265
   success = elpa_solve_evp_complex_1stage_double(na, nev, a, na_rows, ev, z, na_rows, nblk, &
266
                               na_cols, mpi_comm_rows, mpi_comm_cols, mpi_comm_world)
267

268
269
   if (.not.(success)) then
      write(error_unit,*) "solve_evp_complex produced an error! Aborting..."
270
#ifdef WITH_MPI
Andreas Marek's avatar
Andreas Marek committed
271
      call MPI_ABORT(mpi_comm_world, 1, mpierr)
272
#endif
273
274
   endif

275
276
277
278
279
280
281
282
283
284
   if (myid==0) then
     print '(a)','| One-step ELPA solver complete.'
     print *
   end if

   if(myid == 0) print *,'Time tridiag_complex  :',time_evp_fwd
   if(myid == 0) print *,'Time solve_tridi      :',time_evp_solve
   if(myid == 0) print *,'Time trans_ev_complex :',time_evp_back
   if(myid == 0) print *,'Total time (sum above):',time_evp_back+time_evp_solve+time_evp_fwd

285
   if(write_to_file%eigenvectors) then
286
287
288
289
290
291
292
293
294
295
     write(unit = task_suffix, fmt = '(i8.8)') myid
     open(17,file="EVs_complex_out_task_"//task_suffix(1:8)//".txt",form='formatted',status='new')
     write(17,*) "Part of eigenvectors: na_rows=",na_rows,"of na=",na," na_cols=",na_cols," of na=",na

     do i=1,na_rows
       do j=1,na_cols
         write(17,*) "row=",i," col=",j," element of eigenvector=",z(i,j)
       enddo
     enddo
     close(17)
296
   endif
297
298
299
300
301
302
303
304
305
306
307
308

   if(write_to_file%eigenvalues) then
      if (myid == 0) then
         open(17,file="Eigenvalues_complex_out.txt",form='formatted',status='new')
         do i=1,na
            write(17,*) i,ev(i)
         enddo
         close(17)
      endif
   endif


309
310
311
   !-------------------------------------------------------------------------------
   ! Test correctness of result (using plain scalapack routines)

312
   status = check_correctness(na, nev, as, z, ev, sc_desc, myid)
313

Andreas Marek's avatar
Andreas Marek committed
314
315
   deallocate(a)
   deallocate(as)
316
317
318

   deallocate(z)
   deallocate(ev)
319
320

#ifdef HAVE_DETAILED_TIMINGS
321
   call timer%stop("program: test_complex_double_precision")
322
   print *," "
323
324
   print *,"Timings program: test_complex_double_precision"
   call timer%print("program: test_complex_double_precision")
325
   print *," "
326
   print *,"End timings program: test_complex_double_precision"
327
#endif
328
#ifdef WITH_MPI
329
330
   call blacs_gridexit(my_blacs_ctxt)
   call mpi_finalize(mpierr)
331
#endif
332
333
334
335
   call EXIT(STATUS)
end

!-------------------------------------------------------------------------------