| 1 | - the top level routine is loop_1x1_bulk.m. It does the following: | 
| 2 | [loops over years and 6-hourly records] | 
| 3 | load fields (e.g. by calling ncep_load_fields.m) | 
| 4 | compute bulks (exf_bulk_largeyeager04.m or gmaze_bulk_coare.m) | 
| 5 | compute net fluxes | 
| 6 | compute time averages (averagesFields.m) | 
| 7 | write to disk (averagesFields.m) | 
| 8 |  | 
| 9 | - loop_1x1_flux.m and loop_1x1_flux_noicemask.m | 
| 10 | simply do a similar formatting of flux data (no bulk | 
| 11 | involved) to allow easy comparisons. | 
| 12 |  | 
| 13 | -a few more scripts: | 
| 14 | gpcp_load_atlas.m       loads/averages GPCP precip data set | 
| 15 | quickcow_load_atlas.m   loads quickcow wind stress atlas | 
| 16 | plot_bulk.m             a sample script to plot results | 
| 17 | domaine_global_def.m, domaine.m, ecmwf_grid.m and ncep_grid.m | 
| 18 | handle data and computational grids | 
| 19 |  | 
| 20 | -the way scripts are setup right now: | 
| 21 | do not account leap years (allways do 365 days), output are daily in local dir. | 
| 22 | fields are interpolated to the 1x1 degree computational grid. | 
| 23 | atm. state/fluxes data is NCEP, ECMWF/ERA40 or CORE/LargeYeager. | 
| 24 | sst is Reynolds, runoff is Large and Yeager, Ice coverage is Hadley Center. | 
| 25 | links to data sets (from Charmaine directories and mines) are hardcoded. | 
| 26 | data sets are under /net/ross/raid* and /net/altix3700/raid*. |