| 1 |
- the top level routine is loop_1x1_bulk.m. It does the following: |
| 2 |
[loops over years and 6-hourly records] |
| 3 |
load fields (e.g. by calling ncep_load_fields.m) |
| 4 |
compute bulks (exf_bulk_largeyeager04.m or gmaze_bulk_coare.m) |
| 5 |
compute net fluxes |
| 6 |
compute time averages (averagesFields.m) |
| 7 |
write to disk (averagesFields.m) |
| 8 |
|
| 9 |
- loop_1x1_flux.m and loop_1x1_flux_noicemask.m |
| 10 |
simply do a similar formatting of flux data (no bulk |
| 11 |
involved) to allow easy comparisons. |
| 12 |
|
| 13 |
-a few more scripts: |
| 14 |
gpcp_load_atlas.m loads/averages GPCP precip data set |
| 15 |
quickcow_load_atlas.m loads quickcow wind stress atlas |
| 16 |
plot_bulk.m a sample script to plot results |
| 17 |
domaine_global_def.m, domaine.m, ecmwf_grid.m and ncep_grid.m |
| 18 |
handle data and computational grids |
| 19 |
|
| 20 |
-the way scripts are setup right now: |
| 21 |
do not account leap years (allways do 365 days), output are daily in local dir. |
| 22 |
fields are interpolated to the 1x1 degree computational grid. |
| 23 |
atm. state/fluxes data is NCEP, ECMWF/ERA40 or CORE/LargeYeager. |
| 24 |
sst is Reynolds, runoff is Large and Yeager, Ice coverage is Hadley Center. |
| 25 |
links to data sets (from Charmaine directories and mines) are hardcoded. |
| 26 |
data sets are under /net/ross/raid* and /net/altix3700/raid*. |