!WRF:DRIVER_LAYER:MAIN ! ! "Nest up" program in WRFV2. ! ! Description: ! ! The nest up (nup.exe) program reads from wrfout_d02_<date> files for ! the nest and generates wrfout_d01_<date> files for the same periods as ! are in the input files. The fields in the output are the fields in the ! input for state variables that have 'h' and 'u' in the I/O string of ! the Registry. In other words, these are the fields that are normally ! fed back from nest->parent during 2-way nesting. It will read and ! output over multiple files of nest data and generate an equivalent ! number of files of parent data. The dimensions of the fields in the ! output are the size of the nest fields divided by the nesting ratio. ! ! Source file: main/nup_em.F ! ! Compile with WRF: compile em_real ! ! Resulting executable: ! ! main/nup.exe ! -and- ! symbolic link in test/em_real/nup.exe ! ! Run as: nup.exe (no arguments) ! ! Namelist information: ! ! Nup.exe uses the same namelist as a nested run of the wrf.exe. ! Important settings are: ! ! &time_control ! ! start_* <start time information for both domains> ! end_* <start time information for both domains> ! history_interval <interval between frames in input/output files> ! frames_per_outfile <number of frames in input/output files> ! io_form_history <2 for NetCDF> ! ! &domains ! ... ! max_dom <number of domains; must be 2> ! e_we <col 2 is size of nested grid in west-east> ! <col 1 is ignored in the namelist> ! e_sn <col 2 is size of nested grid in south-north> ! <col 1 is ignored in the namelist> ! parent_grid_ratio <col 2 is nesting ratio in both dims> ! feedback <must be 1> ! smooth_option <recommend 0> ! ! &physics ! <all options in this section should be the same ! as the run that generated the nest data> ! ! created: JM 2006 01 25 PROGRAM nup_em,66 USE module_machine USE module_domain, ONLY : domain, wrfu_timeinterval, alloc_and_configure_domain, & domain_clock_set, domain_get_current_time, domain_get_stop_time, head_grid, & domain_clock_get, domain_clockadvance USE module_domain_type, ONLY : program_name USE module_streams USE module_initialize_real, only : wrfu_initialize USE module_integrate USE module_driver_constants USE module_configure, only : grid_config_rec_type, model_config_rec USE module_io_domain USE module_utility USE module_timing USE module_wrf_error #ifdef DM_PARALLEL USE module_dm #endif ! USE read_util_module !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !new for bc USE module_bc USE module_big_step_utilities_em USE module_get_file_names #ifdef WRF_CHEM !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! ! for chemistry USE module_input_chem_data ! USE module_input_chem_bioemiss !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! #endif IMPLICIT NONE ! interface INTERFACE ! mediation-supplied SUBROUTINE med_read_wrf_chem_bioemiss ( grid , config_flags) USE module_domain TYPE (domain) grid TYPE (grid_config_rec_type) config_flags END SUBROUTINE med_read_wrf_chem_bioemiss SUBROUTINE nup ( parent_grid , nested_grid, in_id, out_id, newly_opened ) USE module_domain TYPE (domain), POINTER :: parent_grid, nested_grid INTEGER, INTENT(IN) :: in_id, out_id ! io units LOGICAL, INTENT(IN) :: newly_opened ! whether to add global metadata END SUBROUTINE nup END INTERFACE TYPE(WRFU_TimeInterval) :: RingInterval !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !new for bc INTEGER :: ids , ide , jds , jde , kds , kde INTEGER :: ims , ime , jms , jme , kms , kme INTEGER :: ips , ipe , jps , jpe , kps , kpe INTEGER :: its , ite , jts , jte , kts , kte INTEGER :: ijds , ijde , spec_bdy_width INTEGER :: i , j , k INTEGER :: time_loop_max , time_loop INTEGER :: total_time_sec , file_counter INTEGER :: julyr , julday , iswater , map_proj INTEGER :: icnt REAL :: dt , new_bdy_frq REAL :: gmt , cen_lat , cen_lon , dx , dy , truelat1 , truelat2 , moad_cen_lat , stand_lon REAL , DIMENSION(:,:,:) , ALLOCATABLE :: ubdy3dtemp1 , vbdy3dtemp1 , tbdy3dtemp1 , pbdy3dtemp1 , qbdy3dtemp1 REAL , DIMENSION(:,:,:) , ALLOCATABLE :: mbdy2dtemp1 REAL , DIMENSION(:,:,:) , ALLOCATABLE :: ubdy3dtemp2 , vbdy3dtemp2 , tbdy3dtemp2 , pbdy3dtemp2 , qbdy3dtemp2 REAL , DIMENSION(:,:,:) , ALLOCATABLE :: mbdy2dtemp2 CHARACTER(LEN=19) :: start_timestr , current_timestr , end_timestr, timestr CHARACTER(LEN=19) :: stopTimeStr !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! INTEGER :: num_veg_cat , num_soil_top_cat , num_soil_bot_cat REAL :: time INTEGER :: rc INTEGER :: loop , levels_to_process INTEGER , PARAMETER :: max_sanity_file_loop = 100 TYPE (domain) , POINTER :: keep_grid, grid_ptr, null_domain, parent_grid , nested_grid TYPE (domain) :: dummy TYPE (grid_config_rec_type) :: config_flags INTEGER :: number_at_same_level INTEGER :: time_step_begin_restart INTEGER :: max_dom , domain_id , fid , fido, fidb , idum1 , idum2 , ierr INTEGER :: status_next_var INTEGER :: debug_level LOGICAL :: newly_opened CHARACTER (LEN=19) :: date_string #ifdef DM_PARALLEL INTEGER :: nbytes INTEGER, PARAMETER :: configbuflen = 4* CONFIG_BUF_LEN INTEGER :: configbuf( configbuflen ) LOGICAL , EXTERNAL :: wrf_dm_on_monitor #endif INTEGER :: idsi, in_id, out_id INTEGER :: e_sn, e_we, pgr CHARACTER (LEN=80) :: inpname , outname , bdyname CHARACTER (LEN=80) :: si_inpname CHARACTER *19 :: temp19 CHARACTER *24 :: temp24 , temp24b CHARACTER *132 :: fname CHARACTER(len=24) :: start_date_hold CHARACTER (LEN=80) :: message integer :: ii #include "version_decl" ! Interface block for routine that passes pointers and needs to know that they ! are receiving pointers. INTERFACE SUBROUTINE med_feedback_domain ( parent_grid , nested_grid ) USE module_domain USE module_configure TYPE(domain), POINTER :: parent_grid , nested_grid END SUBROUTINE med_feedback_domain SUBROUTINE Setup_Timekeeping( parent_grid ) USE module_domain TYPE(domain), POINTER :: parent_grid END SUBROUTINE Setup_Timekeeping END INTERFACE ! Define the name of this program (program_name defined in module_domain) program_name = "NUP_EM " // TRIM(release_version) // " PREPROCESSOR" #ifdef DM_PARALLEL CALL disable_quilting #endif ! Initialize the modules used by the WRF system. Many of the CALLs made from the ! init_modules routine are NO-OPs. Typical initializations are: the size of a ! REAL, setting the file handles to a pre-use value, defining moisture and ! chemistry indices, etc. CALL init_modules(1) ! Phase 1 returns after MPI_INIT() (if it is called) #ifdef NO_LEAP_CALENDAR CALL WRFU_Initialize( defaultCalKind=WRFU_CAL_NOLEAP, rc=rc ) #else CALL WRFU_Initialize( defaultCalKind=WRFU_CAL_GREGORIAN, rc=rc ) #endif CALL init_modules(2) ! Phase 2 resumes after MPI_INIT() (if it is called) ! Get the NAMELIST data. This is handled in the initial_config routine. All of the ! NAMELIST input variables are assigned to the model_config_rec structure. Below, ! note for parallel processing, only the monitor processor handles the raw Fortran ! I/O, and then broadcasts the info to each of the other nodes. #ifdef DM_PARALLEL IF ( wrf_dm_on_monitor() ) THEN CALL initial_config ENDIF CALL get_config_as_buffer( configbuf, configbuflen, nbytes ) CALL wrf_dm_bcast_bytes( configbuf, nbytes ) CALL set_config_as_buffer( configbuf, configbuflen ) CALL wrf_dm_initialize #else CALL initial_config #endif ! And here is an instance of using the information in the NAMELIST. CALL nl_get_debug_level ( 1, debug_level ) CALL set_wrf_debug_level ( debug_level ) ! set the specified boundary to zero so the feedback goes all the way ! to the edge of the coarse domain CALL nl_set_spec_zone( 1, 0 ) ! Allocated and configure the mother domain. Since we are in the nesting down ! mode, we know a) we got a nest, and b) we only got 1 nest. NULLIFY( null_domain ) !!!! set up the parent grid (for nup_em, this is the grid we do output from) CALL nl_set_shw( 1, 0 ) CALL nl_set_shw( 2, 0 ) CALL nl_set_i_parent_start( 2, 1 ) CALL nl_set_j_parent_start( 2, 1 ) CALL nl_get_e_we( 2, e_we ) CALL nl_get_e_sn( 2, e_sn ) CALL nl_get_parent_grid_ratio( 2, pgr ) ! parent grid must cover the entire nest, which is always dimensioned a factor of 3 + 1 ! so add two here temporarily, then remove later after nest is allocated. e_we = e_we / pgr + 2 e_sn = e_sn / pgr + 2 CALL nl_set_e_we( 1, e_we ) CALL nl_set_e_sn( 1, e_sn ) CALL wrf_message ( program_name ) CALL wrf_debug ( 100 , 'nup_em: calling alloc_and_configure_domain coarse ' ) CALL alloc_and_configure_domain ( domain_id = 1 , & grid = head_grid , & parent = null_domain , & kid = -1 ) parent_grid => head_grid ! Set up time initializations. CALL Setup_Timekeeping ( parent_grid ) CALL domain_clock_set( head_grid, & time_step_seconds=model_config_rec%interval_seconds ) CALL model_to_grid_config_rec ( parent_grid%id , model_config_rec , config_flags ) CALL set_scalar_indices_from_config ( parent_grid%id , idum1, idum2 ) !!!! set up the fine grid (for nup_em, this is the grid we do input into) CALL wrf_message ( program_name ) CALL wrf_debug ( 100 , 'wrf: calling alloc_and_configure_domain fine ' ) CALL alloc_and_configure_domain ( domain_id = 2 , & grid = nested_grid , & parent = parent_grid , & kid = 1 ) ! now that the nest is allocated, pinch off the extra two rows/columns of the parent ! note the IKJ assumption here. parent_grid%ed31 = parent_grid%ed31 - 2 parent_grid%ed33 = parent_grid%ed33 - 2 CALL nl_set_e_we( 1, e_we-2 ) CALL nl_set_e_sn( 1, e_sn-2 ) write(0,*)'after alloc_and_configure_domain ',associated(nested_grid%intermediate_grid) CALL model_to_grid_config_rec ( nested_grid%id , model_config_rec , config_flags ) CALL set_scalar_indices_from_config ( nested_grid%id , idum1, idum2 ) ! Set up time initializations for the fine grid. CALL Setup_Timekeeping ( nested_grid ) ! Adjust the time step on the clock so that it's the same as the history interval CALL WRFU_AlarmGet( nested_grid%alarms(HISTORY_ALARM), RingInterval=RingInterval ) CALL WRFU_ClockSet( nested_grid%domain_clock, TimeStep=RingInterval, rc=rc ) CALL WRFU_ClockSet( parent_grid%domain_clock, TimeStep=RingInterval, rc=rc ) ! Get and store the history interval from the fine grid; use for time loop ! Initialize the I/O for WRF. CALL init_wrfio ! Some of the configuration values may have been modified from the initial READ ! of the NAMELIST, so we re-broadcast the configuration records. #ifdef DM_PARALLEL CALL get_config_as_buffer( configbuf, configbuflen, nbytes ) CALL wrf_dm_bcast_bytes( configbuf, nbytes ) CALL set_config_as_buffer( configbuf, configbuflen ) #endif ! Open the input data (wrfout_d01_xxxxxx) for reading. in_id = 0 out_id = 0 main_loop : DO WHILE ( domain_get_current_time(nested_grid) .LT. domain_get_stop_time(nested_grid) ) IF( WRFU_AlarmIsRinging( nested_grid%alarms( HISTORY_ALARM ), rc=rc ) ) THEN CALL domain_clock_get( nested_grid, current_timestr=timestr ) newly_opened = .FALSE. IF ( in_id.EQ. 0 ) THEN CALL model_to_grid_config_rec ( nested_grid%id , model_config_rec , config_flags ) CALL construct_filename2a ( fname , config_flags%history_outname , nested_grid%id , 2 , timestr ) CALL open_r_dataset ( in_id, TRIM(fname), nested_grid , & config_flags , 'DATASET=HISTORY' , ierr ) IF ( ierr .NE. 0 ) THEN WRITE(message,*)'Failed to open ',TRIM(fname),' for reading. ' CALL wrf_message(message) EXIT main_loop ENDIF CALL model_to_grid_config_rec ( parent_grid%id , model_config_rec , config_flags ) CALL construct_filename2a ( fname , config_flags%history_outname , parent_grid%id , 2 , timestr ) CALL open_w_dataset ( out_id, TRIM(fname), parent_grid , & config_flags , output_history, 'DATASET=HISTORY' , ierr ) IF ( ierr .NE. 0 ) THEN WRITE(message,*)'Failed to open ',TRIM(fname),' for writing. ' CALL wrf_message(message) EXIT main_loop ENDIF newly_opened = .TRUE. ENDIF CALL model_to_grid_config_rec ( nested_grid%id , model_config_rec , config_flags ) CALL input_history ( in_id, nested_grid , config_flags , ierr ) IF ( ierr .NE. 0 ) THEN WRITE(message,*)'Unable to read time ',timestr CALL wrf_message(message) EXIT main_loop ENDIF ! CALL nup ( nested_grid , parent_grid, in_id, out_id, newly_opened ) ! CALL model_to_grid_config_rec ( parent_grid%id , model_config_rec , config_flags ) CALL output_history ( out_id, parent_grid , config_flags , ierr ) IF ( ierr .NE. 0 ) THEN WRITE(message,*)'Unable to write time ',timestr CALL wrf_message(message) EXIT main_loop ENDIF nested_grid%nframes(history_only) = nested_grid%nframes(history_only) + 1 IF ( nested_grid%nframes(history_only) >= config_flags%frames_per_outfile ) THEN CALL model_to_grid_config_rec ( nested_grid%id , model_config_rec , config_flags ) CALL close_dataset ( in_id , config_flags , "DATASET=HISTORY" ) CALL model_to_grid_config_rec ( parent_grid%id , model_config_rec , config_flags ) CALL close_dataset ( out_id , config_flags , "DATASET=HISTORY" ) in_id = 0 out_id = 0 nested_grid%nframes(history_only) = 0 ENDIF CALL WRFU_AlarmRingerOff( nested_grid%alarms( HISTORY_ALARM ), rc=rc ) ENDIF CALL domain_clockadvance( nested_grid ) CALL domain_clockadvance( parent_grid ) ENDDO main_loop CALL model_to_grid_config_rec ( parent_grid%id , model_config_rec , config_flags ) CALL med_shutdown_io ( parent_grid , config_flags ) CALL wrf_debug ( 0 , 'nup_em: SUCCESS COMPLETE NUP_EM INIT' ) ! CALL wrf_shutdown CALL WRFU_Finalize( rc=rc ) END PROGRAM nup_em SUBROUTINE nup ( nested_grid, parent_grid , in_id, out_id, newly_opened ) 1,23 USE module_domain USE module_io_domain USE module_utility USE module_timing USE module_wrf_error ! IMPLICIT NONE ! Args TYPE(domain), POINTER :: parent_grid, nested_grid INTEGER, INTENT(IN) :: in_id, out_id ! io descriptors LOGICAL, INTENT(IN) :: newly_opened ! whether to add global metadata ! Local INTEGER :: julyr , julday , iswater , map_proj INTEGER :: icnt, ierr REAL :: dt , new_bdy_frq REAL :: gmt , cen_lat , cen_lon , dx , dy , truelat1 , truelat2 , moad_cen_lat , stand_lon REAL , DIMENSION(:,:,:) , ALLOCATABLE :: ubdy3dtemp1 , vbdy3dtemp1 , tbdy3dtemp1 , pbdy3dtemp1 , qbdy3dtemp1 REAL , DIMENSION(:,:,:) , ALLOCATABLE :: mbdy2dtemp1 REAL , DIMENSION(:,:,:) , ALLOCATABLE :: ubdy3dtemp2 , vbdy3dtemp2 , tbdy3dtemp2 , pbdy3dtemp2 , qbdy3dtemp2 REAL , DIMENSION(:,:,:) , ALLOCATABLE :: mbdy2dtemp2 INTEGER :: ids , ide , jds , jde , kds , kde INTEGER :: ims , ime , jms , jme , kms , kme INTEGER :: ips , ipe , jps , jpe , kps , kpe INTEGER :: its , ite , jts , jte , kts , kte INTERFACE SUBROUTINE med_feedback_domain ( parent_grid , nested_grid ) USE module_domain USE module_configure TYPE(domain), POINTER :: parent_grid , nested_grid END SUBROUTINE med_feedback_domain SUBROUTINE med_interp_domain ( parent_grid , nested_grid ) USE module_domain USE module_configure TYPE(domain), POINTER :: parent_grid , nested_grid END SUBROUTINE med_interp_domain END INTERFACE IF ( newly_opened ) THEN CALL wrf_get_dom_ti_integer ( in_id , 'MAP_PROJ' , map_proj , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'DX' , dx , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'DY' , dy , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'CEN_LAT' , cen_lat , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'CEN_LON' , cen_lon , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'TRUELAT1' , truelat1 , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'TRUELAT2' , truelat2 , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'MOAD_CEN_LAT' , moad_cen_lat , 1 , icnt , ierr ) CALL wrf_get_dom_ti_real ( in_id , 'STAND_LON' , stand_lon , 1 , icnt , ierr ) ! CALL wrf_get_dom_ti_real ( in_id , 'GMT' , gmt , 1 , icnt , ierr ) ! CALL wrf_get_dom_ti_integer ( in_id , 'JULYR' , julyr , 1 , icnt , ierr ) ! CALL wrf_get_dom_ti_integer ( in_id , 'JULDAY' , julday , 1 , icnt , ierr ) CALL wrf_get_dom_ti_integer ( in_id , 'ISWATER' , iswater , 1 , icnt , ierr ) ENDIF parent_grid%fnm = nested_grid%fnm parent_grid%fnp = nested_grid%fnp parent_grid%rdnw = nested_grid%rdnw parent_grid%rdn = nested_grid%rdn parent_grid%dnw = nested_grid%dnw parent_grid%dn = nested_grid%dn parent_grid%znu = nested_grid%znu parent_grid%znw = nested_grid%znw parent_grid%zs = nested_grid%zs parent_grid%dzs = nested_grid%dzs parent_grid%p_top = nested_grid%p_top parent_grid%rdx = nested_grid%rdx * 3. parent_grid%rdy = nested_grid%rdy * 3. parent_grid%resm = nested_grid%resm parent_grid%zetatop = nested_grid%zetatop parent_grid%cf1 = nested_grid%cf1 parent_grid%cf2 = nested_grid%cf2 parent_grid%cf3 = nested_grid%cf3 parent_grid%cfn = nested_grid%cfn parent_grid%cfn1 = nested_grid%cfn1 #ifdef WRF_CHEM parent_grid%chem_opt = nested_grid%chem_opt parent_grid%chem_in_opt = nested_grid%chem_in_opt #endif !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! ! Various sizes that we need to be concerned about. ids = parent_grid%sd31 ide = parent_grid%ed31 kds = parent_grid%sd32 kde = parent_grid%ed32 jds = parent_grid%sd33 jde = parent_grid%ed33 ims = parent_grid%sm31 ime = parent_grid%em31 kms = parent_grid%sm32 kme = parent_grid%em32 jms = parent_grid%sm33 jme = parent_grid%em33 ips = parent_grid%sp31 ipe = parent_grid%ep31 kps = parent_grid%sp32 kpe = parent_grid%ep32 jps = parent_grid%sp33 jpe = parent_grid%ep33 nested_grid%imask_nostag = 1 nested_grid%imask_xstag = 1 nested_grid%imask_ystag = 1 nested_grid%imask_xystag = 1 ! Interpolate from nested_grid back onto parent_grid CALL med_feedback_domain ( parent_grid , nested_grid ) parent_grid%ht_int = parent_grid%ht !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! #if 0 CALL construct_filename2( si_inpname , 'wrf_real_input_em' , parent_grid%id , 2 , start_date_char ) CALL wrf_debug ( 100 , 'med_sidata_input: calling open_r_dataset for ' // TRIM(si_inpname) ) CALL model_to_grid_config_rec ( parent_grid%id , model_config_rec , config_flags ) CALL open_r_dataset ( idsi, TRIM(si_inpname) , parent_grid , config_flags , "DATASET=INPUT", ierr ) IF ( ierr .NE. 0 ) THEN CALL wrf_error_fatal( 'real: error opening wrf_real_input_em for reading: ' // TRIM (si_inpname) ) END IF ! Input data. CALL wrf_debug ( 100 , 'nup_em: calling input_aux_model_input2' ) CALL input_aux_model_input2 ( idsi , parent_grid , config_flags , ierr ) parent_grid%ht_input = parent_grid%ht ! Close this fine grid static input file. CALL wrf_debug ( 100 , 'nup_em: closing fine grid static input' ) CALL close_dataset ( idsi , config_flags , "DATASET=INPUT" ) ! We need a parent grid landuse in the interpolation. So we need to generate ! that field now. IF ( ( parent_grid%ivgtyp(ips,jps) .GT. 0 ) .AND. & ( parent_grid%isltyp(ips,jps) .GT. 0 ) ) THEN DO j = jps, MIN(jde-1,jpe) DO i = ips, MIN(ide-1,ipe) parent_grid% vegcat(i,j) = parent_grid%ivgtyp(i,j) parent_grid%soilcat(i,j) = parent_grid%isltyp(i,j) END DO END DO ELSE IF ( ( parent_grid% vegcat(ips,jps) .GT. 0.5 ) .AND. & ( parent_grid%soilcat(ips,jps) .GT. 0.5 ) ) THEN DO j = jps, MIN(jde-1,jpe) DO i = ips, MIN(ide-1,ipe) parent_grid%ivgtyp(i,j) = NINT(parent_grid% vegcat(i,j)) parent_grid%isltyp(i,j) = NINT(parent_grid%soilcat(i,j)) END DO END DO ELSE num_veg_cat = SIZE ( parent_grid%landusef , DIM=2 ) num_soil_top_cat = SIZE ( parent_grid%soilctop , DIM=2 ) num_soil_bot_cat = SIZE ( parent_grid%soilcbot , DIM=2 ) CALL land_percentages ( parent_grid%xland , & parent_grid%landusef , parent_grid%soilctop , parent_grid%soilcbot , & parent_grid%isltyp , parent_grid%ivgtyp , & num_veg_cat , num_soil_top_cat , num_soil_bot_cat , & ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & ips , ipe , jps , jpe , kps , kpe , & model_config_rec%iswater(parent_grid%id) ) END IF DO j = jps, MIN(jde-1,jpe) DO i = ips, MIN(ide-1,ipe) parent_grid%lu_index(i,j) = parent_grid%ivgtyp(i,j) END DO END DO CALL check_consistency ( parent_grid%ivgtyp , parent_grid%isltyp , parent_grid%landmask , & ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & ips , ipe , jps , jpe , kps , kpe , & model_config_rec%iswater(parent_grid%id) ) CALL check_consistency2( parent_grid%ivgtyp , parent_grid%isltyp , parent_grid%landmask , & parent_grid%tmn , parent_grid%tsk , parent_grid%sst , parent_grid%xland , & parent_grid%tslb , parent_grid%smois , parent_grid%sh2o , & config_flags%num_soil_layers , parent_grid%id , & ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & ips , ipe , jps , jpe , kps , kpe , & model_config_rec%iswater(parent_grid%id) ) !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! ! We have 2 terrain elevations. One is from input and the other is from the ! the horizontal interpolation. parent_grid%ht_fine = parent_grid%ht_input parent_grid%ht = parent_grid%ht_int ! We have both the interpolated fields and the higher-resolution static fields. From these ! the rebalancing is now done. Note also that the field parent_grid%ht is now from the ! fine grid input file (after this call is completed). CALL rebalance_driver ( parent_grid ) ! Different things happen during the different time loops: ! first loop - write wrfinput file, close data set, copy files to holder arrays ! middle loops - diff 3d/2d arrays, compute and output bc ! last loop - diff 3d/2d arrays, compute and output bc, write wrfbdy file, close wrfbdy file ! Set the time info. print *,'current_date = ',current_date CALL domain_clock_set( parent_grid, & current_timestr=current_date(1:19) ) ! ! SEP Put in chemistry data ! #ifdef WRF_CHEM IF( parent_grid%chem_opt .NE. 0 ) then IF( parent_grid%chem_in_opt .EQ. 0 ) then ! Read the chemistry data from a previous wrf forecast (wrfout file) ! Generate chemistry data from a idealized vertical profile message = 'STARTING WITH BACKGROUND CHEMISTRY ' CALL wrf_message ( message ) CALL input_chem_profile ( parent_grid ) message = 'READING BEIS3.11 EMISSIONS DATA' CALL wrf_message ( message ) CALL med_read_wrf_chem_bioemiss ( parent_grid , config_flags) ELSE message = 'RUNNING WITHOUT CHEMISTRY INITIALIZATION' CALL wrf_message ( message ) ENDIF ENDIF #endif #endif ! Output the first time period of the data. IF ( newly_opened ) THEN CALL wrf_put_dom_ti_integer ( out_id , 'MAP_PROJ' , map_proj , 1 , ierr ) ! CALL wrf_put_dom_ti_real ( out_id , 'DX' , dx , 1 , ierr ) ! CALL wrf_put_dom_ti_real ( out_id , 'DY' , dy , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'CEN_LAT' , cen_lat , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'CEN_LON' , cen_lon , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'TRUELAT1' , truelat1 , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'TRUELAT2' , truelat2 , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'MOAD_CEN_LAT' , moad_cen_lat , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'STAND_LON' , stand_lon , 1 , ierr ) CALL wrf_put_dom_ti_integer ( out_id , 'ISWATER' , iswater , 1 , ierr ) CALL wrf_put_dom_ti_real ( out_id , 'GMT' , gmt , 1 , ierr ) CALL wrf_put_dom_ti_integer ( out_id , 'JULYR' , julyr , 1 , ierr ) CALL wrf_put_dom_ti_integer ( out_id , 'JULDAY' , julday , 1 , ierr ) ENDIF END SUBROUTINE nup SUBROUTINE land_percentages ( xland , & 2,4 landuse_frac , soil_top_cat , soil_bot_cat , & isltyp , ivgtyp , & num_veg_cat , num_soil_top_cat , num_soil_bot_cat , & ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte , & iswater ) USE module_soil_pre IMPLICIT NONE INTEGER , INTENT(IN) :: ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte , & iswater INTEGER , INTENT(IN) :: num_veg_cat , num_soil_top_cat , num_soil_bot_cat REAL , DIMENSION(ims:ime,1:num_veg_cat,jms:jme) , INTENT(INOUT):: landuse_frac REAL , DIMENSION(ims:ime,1:num_soil_top_cat,jms:jme) , INTENT(IN):: soil_top_cat REAL , DIMENSION(ims:ime,1:num_soil_bot_cat,jms:jme) , INTENT(IN):: soil_bot_cat INTEGER , DIMENSION(ims:ime,jms:jme), INTENT(OUT) :: isltyp , ivgtyp REAL , DIMENSION(ims:ime,jms:jme) , INTENT(OUT) :: xland CALL process_percent_cat_new ( xland , & landuse_frac , soil_top_cat , soil_bot_cat , & isltyp , ivgtyp , & num_veg_cat , num_soil_top_cat , num_soil_bot_cat , & ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte , & iswater ) END SUBROUTINE land_percentages SUBROUTINE check_consistency ( ivgtyp , isltyp , landmask , & 2,2 ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte , & iswater ) IMPLICIT NONE INTEGER , INTENT(IN) :: ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte , & iswater INTEGER , DIMENSION(ims:ime,jms:jme), INTENT(INOUT) :: isltyp , ivgtyp REAL , DIMENSION(ims:ime,jms:jme), INTENT(INOUT) :: landmask LOGICAL :: oops INTEGER :: oops_count , i , j oops = .FALSE. oops_count = 0 DO j = jts, MIN(jde-1,jte) DO i = its, MIN(ide-1,ite) IF ( ( ( landmask(i,j) .LT. 0.5 ) .AND. ( ivgtyp(i,j) .NE. iswater ) ) .OR. & ( ( landmask(i,j) .GT. 0.5 ) .AND. ( ivgtyp(i,j) .EQ. iswater ) ) ) THEN print *,'mismatch in landmask and veg type' print *,'i,j=',i,j, ' landmask =',NINT(landmask(i,j)),' ivgtyp=',ivgtyp(i,j) oops = .TRUE. oops_count = oops_count + 1 landmask(i,j) = 0 ivgtyp(i,j)=16 isltyp(i,j)=14 END IF END DO END DO IF ( oops ) THEN CALL wrf_debug( 0, 'mismatch in check_consistency, turned to water points, be careful' ) END IF END SUBROUTINE check_consistency SUBROUTINE check_consistency2( ivgtyp , isltyp , landmask , & 2,12 tmn , tsk , sst , xland , & tslb , smois , sh2o , & num_soil_layers , id , & ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte , & iswater ) USE module_configure USE module_optional_input INTEGER , INTENT(IN) :: ids , ide , jds , jde , kds , kde , & ims , ime , jms , jme , kms , kme , & its , ite , jts , jte , kts , kte INTEGER , INTENT(IN) :: num_soil_layers , id INTEGER , DIMENSION(ims:ime,jms:jme) :: ivgtyp , isltyp REAL , DIMENSION(ims:ime,jms:jme) :: landmask , tmn , tsk , sst , xland REAL , DIMENSION(ims:ime,num_soil_layers,jms:jme) :: tslb , smois , sh2o INTEGER :: oops1 , oops2 INTEGER :: i , j , k fix_tsk_tmn : SELECT CASE ( model_config_rec%sf_surface_physics(id) ) CASE ( SLABSCHEME , LSMSCHEME , RUCLSMSCHEME ) DO j = jts, MIN(jde-1,jte) DO i = its, MIN(ide-1,ite) IF ( ( landmask(i,j) .LT. 0.5 ) .AND. ( flag_sst .EQ. 1 ) ) THEN tmn(i,j) = sst(i,j) tsk(i,j) = sst(i,j) ELSE IF ( landmask(i,j) .LT. 0.5 ) THEN tmn(i,j) = tsk(i,j) END IF END DO END DO END SELECT fix_tsk_tmn ! Is the TSK reasonable? DO j = jts, MIN(jde-1,jte) DO i = its, MIN(ide-1,ite) IF ( tsk(i,j) .LT. 170 .or. tsk(i,j) .GT. 400. ) THEN print *,'error in the TSK' print *,'i,j=',i,j print *,'landmask=',landmask(i,j) print *,'tsk, sst, tmn=',tsk(i,j),sst(i,j),tmn(i,j) if(tmn(i,j).gt.170. .and. tmn(i,j).lt.400.)then tsk(i,j)=tmn(i,j) else if(sst(i,j).gt.170. .and. sst(i,j).lt.400.)then tsk(i,j)=sst(i,j) else CALL wrf_error_fatal ( 'TSK unreasonable' ) end if END IF END DO END DO ! Is the TMN reasonable? DO j = jts, MIN(jde-1,jte) DO i = its, MIN(ide-1,ite) IF ( ( ( tmn(i,j) .LT. 170. ) .OR. ( tmn(i,j) .GT. 400. ) ) .AND. ( landmask(i,j) .GT. 0.5 ) ) THEN print *,'error in the TMN' print *,'i,j=',i,j print *,'landmask=',landmask(i,j) print *,'tsk, sst, tmn=',tsk(i,j),sst(i,j),tmn(i,j) if(tsk(i,j).gt.170. .and. tsk(i,j).lt.400.)then tmn(i,j)=tsk(i,j) else if(sst(i,j).gt.170. .and. sst(i,j).lt.400.)then tmn(i,j)=sst(i,j) else CALL wrf_error_fatal ( 'TMN unreasonable' ) endif END IF END DO END DO ! Is the TSLB reasonable? DO j = jts, MIN(jde-1,jte) DO i = its, MIN(ide-1,ite) IF ( ( ( tslb(i,1,j) .LT. 170. ) .OR. ( tslb(i,1,j) .GT. 400. ) ) .AND. ( landmask(i,j) .GT. 0.5 ) ) THEN print *,'error in the TSLB' print *,'i,j=',i,j print *,'landmask=',landmask(i,j) print *,'tsk, sst, tmn=',tsk(i,j),sst(i,j),tmn(i,j) print *,'tslb = ',tslb(i,:,j) print *,'old smois = ',smois(i,:,j) DO l = 1 , num_soil_layers sh2o(i,l,j) = 0.0 END DO DO l = 1 , num_soil_layers smois(i,l,j) = 0.3 END DO if(tsk(i,j).gt.170. .and. tsk(i,j).lt.400.)then DO l = 1 , num_soil_layers tslb(i,l,j)=tsk(i,j) END DO else if(sst(i,j).gt.170. .and. sst(i,j).lt.400.)then DO l = 1 , num_soil_layers tslb(i,l,j)=sst(i,j) END DO else if(tmn(i,j).gt.170. .and. tmn(i,j).lt.400.)then DO l = 1 , num_soil_layers tslb(i,l,j)=tmn(i,j) END DO else CALL wrf_error_fatal ( 'TSLB unreasonable' ) endif END IF END DO END DO ! Let us make sure (again) that the landmask and the veg/soil categories match. oops1=0 oops2=0 DO j = jts, MIN(jde-1,jte) DO i = its, MIN(ide-1,ite) IF ( ( ( landmask(i,j) .LT. 0.5 ) .AND. ( ivgtyp(i,j) .NE. iswater .OR. isltyp(i,j) .NE. 14 ) ) .OR. & ( ( landmask(i,j) .GT. 0.5 ) .AND. ( ivgtyp(i,j) .EQ. iswater .OR. isltyp(i,j) .EQ. 14 ) ) ) THEN IF ( tslb(i,1,j) .GT. 1. ) THEN oops1=oops1+1 ivgtyp(i,j) = 5 isltyp(i,j) = 8 landmask(i,j) = 1 xland(i,j) = 1 ELSE IF ( sst(i,j) .GT. 1. ) THEN oops2=oops2+1 ivgtyp(i,j) = iswater isltyp(i,j) = 14 landmask(i,j) = 0 xland(i,j) = 2 ELSE print *,'the landmask and soil/veg cats do not match' print *,'i,j=',i,j print *,'landmask=',landmask(i,j) print *,'ivgtyp=',ivgtyp(i,j) print *,'isltyp=',isltyp(i,j) print *,'iswater=', iswater print *,'tslb=',tslb(i,:,j) print *,'sst=',sst(i,j) CALL wrf_error_fatal ( 'mismatch_landmask_ivgtyp' ) END IF END IF END DO END DO if (oops1.gt.0) then print *,'points artificially set to land : ',oops1 endif if(oops2.gt.0) then print *,'points artificially set to water: ',oops2 endif END SUBROUTINE check_consistency2