Differing statistical results between AWS and VM instances #2481
-
Having an issue that I can't seem to solve. Running MET11.0/METplus 5.0 on a Rocky Linux 8 VM and MET11.1/METplus 5.1 on a Rocky Linux 9 AWS box and trying to replicate grid_stat results for a 3km WRF against URMA analysis. When I run METplus using the same model data and analysis files on both machines, the VM provides plausible statistical results whereas they are completely out to lunch on AWS - FBAR is the same value throughout all of the verification regions and MSESS is highly negative across the board, among other issues. Config files and input data match between machines outside of their respective directory structures. I've attached the grid_stat log from the AWS machine, and config files and a resulting forecast hour CNT output from each machine for your review. Let me know if there's anything else I can provide that might be of use. |
Beta Was this translation helpful? Give feedback.
Replies: 5 comments 7 replies
-
Chris, I see that you're seeing unexpected differences in the CNT output computed by Grid-Stat version 11.0.0 on your VM versus 11.1.0 on AWS. Thank's for passing along sample data to illustrate. I definitely agree that it looks like the AWS v11.1.0 output is out to lunch. I just picked off the first CNT line of output (for TMP/Z2) and compared the VM 11.0.0 output to the AWS 11.1.0 output, and note the following:
Looking in the METplus log file, I see that this output was generated with the following command:
I suspect that Grid-Stat is handling the URMA input just fine, but we're seeing differing behavior with the forecast data. So I'd like to start by checking how MET reads the GRIB2 forecast file: If possible please try running the following to plot the TMP/Z2 data:
My guess is that you'll see very different values plotted (and listed in the If that is indeed the case, please send me...
You can post it to our anonymous ftp site following these instructions. |
Beta Was this translation helpful? Give feedback.
-
Hey John, As expected, the VM ps looks normal while the AWS one is a constant temp of 266K. Files have been uploaded to the FTP server; I also included the model file from both machines just in case. |
Beta Was this translation helpful? Give feedback.
-
@cwmac thanks for passing along the data. Indeed, this definitely looks like a problem in the AWS build of MET. I'll note that checking the GRIB2 files you posted, the
The I'm happy to report that my local installation of MET version 11.1.0 plots the full range of data just fine. And at verbosity level 4 (
So I'm pretty confident that there's just an issue in the way MET 11.1.0 was compiled in the AWS instance. And for GRIB2 issues, I always start by looking at how the grib2c library was compiled. Please see the 4th item in this list and note the following in particular:
@jprestop is the person who knows the most of installing MET. And she uses/maintains this compile_MET_all.sh shell script to do so. The commands for compiling the G2C library start on line 490. Can you please take a look on the AWS instance to see how the G2C library was compiled? What version is being compiled? Do you see "64BIT" used anywhere? Hopefully once we get the compilation of the G2C library sorted, and recompile MET, it will behavior properly on AWS. |
Beta Was this translation helpful? Give feedback.
-
I'm compiling the dependencies via the wgrib2 package. g2clib version is 1.4.0. I put the make log on the FTP server for your review. |
Beta Was this translation helpful? Give feedback.
-
Thanks @jprestop. I have gone through the compile_met_all.sh scripts and compiled the dependencies that way; however, once I get to compiling MET, getting a ton of Python errors on the ensemble_stat set. Sent the make.log (make-20240203.log) to the server for your review. |
Beta Was this translation helpful? Give feedback.
Thank you @cwmac! Could you please update your values for
MET_PYTHON_CC
andMET_PYTHON_LD
as shown below?Then, please try recompiling MET, and let us know how it goes.