Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

rt bug fixes #395

Closed
wants to merge 7 commits into from
Closed

Conversation

MinsukJi-NOAA
Copy link
Contributor

Description

Bug fix in check_results function of rt_utils.sh

  • Exit if error occurs in compare_ncfile.py
  • Exit if error occurs in cmp
  • Is a change of answers expected from this PR? No
  • Are any library updates included in this PR (modulefiles etc.)? No

Bug fix for rt.sh -n option flag. This was broken when the MACHINES column of rt.conf was modified.

Issue(s) addressed

#296

Testing

Will run regression tests on supported platforms.

Copy link
Collaborator

@climbfuji climbfuji left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks ok as far as I can tell. Will check again and approve when it is time for the PR to go in.

@DeniseWorthen
Copy link
Collaborator

One thing I've noticed is that if the RT fails (for example, one test times out) then the "clean after" command does not take affect. All the fv3_X.exe and matching modules.fv3_X are left in the tests directory. Is this a design feature or is it something we like to fix?

@climbfuji
Copy link
Collaborator

climbfuji commented Jan 27, 2021 via email

@DeniseWorthen
Copy link
Collaborator

I can see where you'd use this---the trouble I have is associating the compile line w/ the test number. Short of counting each COMPILE line, is there a way to know which fv3_XX.exe is used for the failed test?

@climbfuji
Copy link
Collaborator

climbfuji commented Jan 27, 2021 via email

@MinsukJi-NOAA
Copy link
Contributor Author

I think so. The log_hera.intel/run_001 script, for example, should have a line with "cp ... fv3_N.exe RUNDIR/fv3.exe"

On Jan 27, 2021, at 7:16 AM, Denise Worthen @.***> wrote: I can see where you'd use this---the trouble I have is associating the compile line w/ the test number. Short of counting each COMPILE line, is there a way to know which fv3_XX.exe is used for the failed test? — You are receiving this because you commented. Reply to this email directly, view it on GitHub <#395 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/AB5C2RIXWSRQQE6GTFWL7RTS4AN4ZANCNFSM4WUCLANA.

That is a good suggestion. Let me look into implementing it.

@climbfuji
Copy link
Collaborator

I think so. The log_hera.intel/run_001 script, for example, should have a line with "cp ... fv3_N.exe RUNDIR/fv3.exe"

On Jan 27, 2021, at 7:16 AM, Denise Worthen @.***> wrote: I can see where you'd use this---the trouble I have is associating the compile line w/ the test number. Short of counting each COMPILE line, is there a way to know which fv3_XX.exe is used for the failed test? — You are receiving this because you commented. Reply to this email directly, view it on GitHub <#395 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/AB5C2RIXWSRQQE6GTFWL7RTS4AN4ZANCNFSM4WUCLANA.

That is a good suggestion. Let me look into implementing it.

That line is there already. From cat log_cheyenne.intel/run_001_fv3_ccpp_control_prod.log:

+ mkdir -p /glade/scratch/heinzell/FV3_RT/rt_13602/fv3_ccpp_control_prod
+ cd /glade/scratch/heinzell/FV3_RT/rt_13602/fv3_ccpp_control_prod
+ cp /glade/work/heinzell/fv3/ufs-weather-model/ufs-weather-model-gsl-develop-20210120-update-from-develop/intel/tests/fv3_1.exe fv3.exe
+ cp /glade/work/heinzell/fv3/ufs-weather-model/ufs-weather-model-gsl-develop-20210120-update-from-develop/intel/tests/modules.fv3_1 modules.fv3

@junwang-noaa
Copy link
Collaborator

junwang-noaa commented Jan 27, 2021 via email

@junwang-noaa
Copy link
Collaborator

junwang-noaa commented Jan 27, 2021 via email

@MinsukJi-NOAA
Copy link
Contributor Author

@junwang-noaa, let me look into both your questions.

@MinsukJi-NOAA
Copy link
Contributor Author

I have another question. I was running a test with shorter forecast length (24->3), I noticed RT reports that the atmos_4xdaily.nc reports ALT OK compared to baseline even though the run has fewer forecast times in the file, is that expected? I guess this could also be true if the diag_table is changed to have fewer fields, as long as the file has subset data of baseline data, the compare_netcdf will report the comparison is OK.

If forecast length is different, ALT CHECK should lead to NOT OK, because the array dimensions are different between baseline and rundir. Changes were made to add dimension check to compare_ncfile.py just now. Similarly, if the number of variables is different in the fields, it should lead to ALT CHECK NOT OK.

@MinsukJi-NOAA MinsukJi-NOAA marked this pull request as ready for review January 29, 2021 18:59
@MinsukJi-NOAA
Copy link
Contributor Author

This will be merged together with @DomHeinzeller 's PR #396.

@MinsukJi-NOAA
Copy link
Contributor Author

This will be merged together with #396.

@climbfuji
Copy link
Collaborator

@MinsukJi-NOAA I believe this was just merged as part of #396. Can you check and, if so, close the PR please? Thanks.

@MinsukJi-NOAA
Copy link
Contributor Author

Merged via #396

@MinsukJi-NOAA MinsukJi-NOAA deleted the rt-fixes branch February 24, 2021 19:20
epic-cicd-jenkins pushed a commit that referenced this pull request Apr 17, 2023
)

* Fix @ issue on LOGDIR.

* Get rid of RUN_CMD_* specification in deactivate_tasks.

* Add TEST_ALT_* directories to all machines.

* Enforce config sourcing order in setup.

* Also fix DYN/PHY dir @ situation.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants