Yannick started the meeting with an update on JEDI 4. He announced that ewok had an update yesterday and should be able to run 4D Hofx now.
JEDI 3 (Anna) Benjamin has a series of PRs for BUMP major upgrade for the static B (saber#85). |
JEDI 2 (Dan) Jedi 2.1 - Interpolation
Meetings taking place with ESMF team to design the refactor and gather requirements.
Dan will prepare Yaml file with Aircraft for 3DEnVar. SOCA team exercising the new LETKF, found some new bugs. Work underway to repair.
Cycling and debugging the cylc suites. 3DVar. Static B variances are too small in the tropics. Cycled for about a week so far.
Pushed latest develop into cubed-sphere-base. Stripped out the UM-JEDI stuff, testing purely in Atlas. Able to generate the GMesh figures.
Adopted a Factory approach within the VADER repo.
Discussion with the model teams who all reported that the work should be minimal. Met Office reported an issue where the GeoVaLs may have different dimensions for the increments. Marek and Anna discussed some work arounds that should be possible. |
After Dan's update, Chris S reported that the MPAS team is now running regional DA. He said the implementation was very straightforward and commended the JEDI team for an extensible design. It was all done in configuration files with no code changes needed.
The long-term solution to this will involve a refactoring of the ObsVector class in ioda but we would not want to delay the release for this. Wojciech kindly offered to help with this by introducing a quick fix to `put_db` that is similar to the one he previously implemented for `get_db`, namely converting variables (excluding metadata) that end in `_<number>` to 2D arrays before they are written out. This should be sufficient for this release since it covers all the current known use cases. It will allow us to put out a release where the input and output files are consistent. The second development issue has to do with performance, but it is distinct from the ioda v2 performance issues that we have been discussing for thelast few months, which have been resolved. This issue was separately identified by Mark O and JJ and manifests most clearly as a slowdown in the ObsSpace constructor for obs types such as Satwinds that have millions of variables. It has been traced to a problem with hdf5 in processing arrays of variable-length strings, such as datetime. Steve and Ryan are working on a solution and expect to have one soon. We estimate that the release will occur about 3 weeks after these development issues have been resolved. Two weeks for documentation, clean-up & bugfixes after release branches have been created. One week code freeze. Also regarding the release, Mark O has made a number of pull requests this week to upgrade mpas and other repos to be compatible with ecbuild 3.6+. Together with previous PRs by Mark and others, we're now nearly ready to make the transition to using the latest ecmwf releases of ecbuild, eckit, fckit, and atlas in all containers and HPC modules. We plan to make these new containers and modules the default on all CI and HPC systems over the course of the next two weeks. Maryam has been continuing to implement the publication of test results to cdash. She started with oops but now this has been implemented for most base JEDI repositories as well as the model repositories that are currently connected to CI, namely fv3-jedi and mpas. She is also working on archiving and documenting the tools she has developed for achieving this, including an AWS lambda function, by adding them to the jedi-tools repository. Jacques is now up and running on Orion and ready to start running applictions with ewok. Summit/PGI update: fv3-bundle now compiles on Summit with PGI compilers and Mark M is now debugging some fv3-test failures. It appears that many test failures have to do with how PGI writes fortran namelists to the string arrays that are used by fms. Azure update: the c192/c96 fv3-gfs 3denvar benchmark is now running on Azure clusters (without a container) and exhibiting comparable timings to Discover and AWS. The next step is to run a more computationally intensive application to better assess the performance of the different platforms. We're targeting JEDI-GDAS-10 for this but we want to wait until the ioda v2 PRs have been merged. Debugging of Intel OneAPI Supercontainers also continues. |
Ben asked about Intel compilers in containers. Mark said that there have been ongoing discussions with NCAR/UCAR and Intel legal on this. The current status is that intel will allow the distribution of docker containers with Intel OneAPI compilers inside. Docker containers are multi-layered so it is clear which components came from intel and which did not. However, Singularity container images are flat binary files so this distinction cannot be made. So, when we update the intel containers for the release (likely next week) we will push a development container with compilers to Docker Hub. However, for Singularity, we make a shell script and a dockerfile available on GitHub with instructions on how to use them to create your own intel development container. A warning however - it does take a lot of memory - at least 20 GB - to build the intel container. We will also continue to distribute tagged containers to accompany each JEDI release.
Sergey asked if we have JEDI running on Azure clusters. Mark confirmed that yes, we do, on native modules - but not yet in multi-node singularity containers.
Ben then gave an update on OBS 1 (for Hui). They are working on the JEDI GDAS application. The have modified routines to produce GeoVals files from JEDI and use them in the Hofx application. They are also working on calibration fixes. They are using the new ewok and seeing problems with the ozone - they are working on fixes. They are also working on speeding up diagnostics and the data ingest.
Francois (OBS 2) reported that the performance enhancement work continues with SageMaker. We have requested and obtained an increase in the number of user profiles for the Sagemaker instances, and thus the number of users that can run the diagnostics. Hailing is working on EDA at full resolution.
OBS 3 (Ryan) There were several OBS3 events recently. |
There was then some discussion of an OBS issue that Greg brought up having to do with the standard for the vertical coordinate: should we establish either a top-down or a bottom-up convention? Uncertainty can make it difficult to implement filters and other functionality. It was agreed that a standard is appropriate for GeoVals but the model grids will have their own custom implementation.
David S then asked about implementing a standard set of physical constants that can be used across JEDI repos. There was some agreement that this would be useful. Mark M wondered if udunits could provide this but it was unclear if the precision of the provided constants could be adapted as needed. It was agreed that the constants would have to be coded rather than specified in yaml files. We should discuss this further, potentially in a focus topic meeting.
Guillaume then gave a SOCA update. Travis is working on the Ewok workflow with SOCA. They ran into a science issue with GDAS reanalysis. Hamideh and Ming are updating to 2018 observations. Hamideh is also working on the sea ice Jacobian. They have LETKF running with the sea ice model. Kriti is working on regional ocean DA. Guillaume is working with the B Matrix, the ocean CICE model, and 3Denvar.
LAND (Andy) FV3 Sergey, Clara and Tseganeh have ongoing work with the LETKF for snow depth, there’s now a specific test for this Need to add model orography/surface elevation to geometry so it can be used for observation localization and quality control, Dan has provided some guidance on how to do this. Clara has found an issue with H(x) that is giving small, non-zero values for snow depth where we are expecting no snow. Assuming this is a feature of the interpolation rather than a bug, but we’re checking that out. WRF-Hydro NWM Made sure everything is working with ecbuild3.6 Begun testing with IODA2, but don’t think this will a big deal for us Mainly been trying to migrate to new(ish) H(x), but this is requiring us to implement variable change, so taking a bit of time We need to do this to handle the variable change associated with partitioning the multi-layer snowpack (which will also need to be done with FV3) A lot of work has gone into understanding and designing this snowpack partitioning UCLDAS Zhichang has been working away on this, but not getting a big update on this until next Monday, so will report next time |
CRTM (Ben)
They are now finishing with the IASI coefficients which are missing short-wave (Patrick is working on this). They are also making comparisons between CRTM and RTTOV (Cheng is working on this). And, they are working on improvements to VIS and IR.