1 WRF Model Version 3.0.1 (August 5, 2008)
2 http://www.mmm.ucar.edu/wrf/users/
4 ------------------------
5 WRF PUBLIC DOMAIN NOTICE
6 ------------------------
8 WRF was developed at the National Center for Atmospheric Research
9 (NCAR) which is operated by the University Corporation for
10 Atmospheric Research (UCAR). NCAR and UCAR make no proprietary
11 claims, either statutory or otherwise, to this version and
12 release of WRF and consider WRF to be in the public domain for
13 use by any person or entity for any purpose without any fee or
14 charge. UCAR requests that any WRF user include this notice on
15 any partial or full copies of WRF. WRF is provided on an "AS
16 IS" basis and any warranties, either express or implied,
17 including but not limited to implied warranties of
18 non-infringement, originality, merchantability and fitness for a
19 particular purpose, are disclaimed. In no event shall
20 UCAR be liable for any damages, whatsoever, whether direct,
21 indirect, consequential or special, that arise out of or in
22 connection with the access, use or performance of WRF, including
26 ======================================
27 This is the main directory for the WRF Version 3 source code release.
28 ======================================
30 V3.0.1.1 Release Notes:
31 -----------------------
33 Version 3.0.1.1 has only limited bug fixes compared to version 3.0.1.
34 The detailed updates for WRF-ARW can be found at
35 http://www.mmm.ucar.edu/wrf/users/wrfv3/updates-3.0.1.1.html, and
36 for WRF-NMM at http://www.dtcenter.org/wrf-nmm/users/model/wrfv3/updates.php
37 Users only need to get the new version if they are involved with one of more of the following:
41 1) Over decomposing domains
42 2) Using AIX/xlf with the "old" Kain-Fritsch cumulus scheme
43 3) Grib edition 2 output
47 1) ndown (offline 1-way nesting) with the ifort compiler
48 2) 3d sea breeze case (existing sea breeze case is 2d by default)
49 3) Multiple time period initialization for a global domain
50 4) Observation nudging using either fine grids or with dynamic initialization
54 1) Restart capability for single domain forecasts
55 2) Noise reduction on eastern boundary in areas of complex topography
56 3) Kain-Fritsch cumulus
62 This is a bug fix release. The detailed updates for WRF-ARW can be found at
63 http://www.mmm.ucar.edu/wrf/users/wrfv3/updates-3.0.1.html, and
65 http://www.dtcenter.org/wrf-nmm/users/model/wrfv3/updates.php
67 Online User's Guides have also been updated.
69 ======================================
74 - For directions on compiling WRF, see below or Users Web page.
75 - For more information on WRF V3 release, visit WRF Users home page
76 http://www.mmm.ucar.edu/wrf/users/
77 - WRF V3 works with WPS, and SI is no longer supported.
78 Please see User' Guide for WPS.
79 - WRF V3 executable does not work with wrfinput/wrfbdy produced by previous versions.
80 - WRF V3 has removed one obsolete namelist variable, dyn_opt. Edit with caution.
82 ======================================
86 - V2.0.1: May 21, 2004
87 - V2.0.2: June 3, 2004
88 - V2.0.3: Nov 12, 2004
90 - V2.1: August 4, 2005
92 - V2.1.2: Jan 27, 2006
96 ======================================
98 How to compile and run?
99 -----------------------
101 - In WRFV3 directory, type 'configure' - this will create a configure.wrf
102 file that has appropriate compile options for the supported computers.
104 Note: WRF requires netCDF library. If your netCDF library is installed in
105 some odd directory, set environment variable NETCDF before you type
106 'configure'. For example,
108 setenv NETCDF /usr/local/netcdf-pgi
110 - Type 'compile case_name' where you can find the case_names by simply typing
111 compile. The following are available:
115 compile em_heldsuarez
118 compile em_quarter_ss
120 compile em_seabreeze2d_x
121 compile em_squall2d_x
122 compile em_squall2d_y
124 - If sucessful, this will create either real.exe or ideal.exe and wrf.exe
125 in directory main/, and the appropriate executables will be linked into
126 the test directoires under test/case_name, or run/.
128 - cd to the appropriate test or run directory to run ideal/real/wrf.
130 - If it is one of the idealized cases (b_wave, hill2d_x, grav2d_x, quarter_ss,
131 squall2d_x, squall2d_y, em_les or em_heldsuarez), cd the the appropriate directory, type
135 to produce wrfinput_d01 file for wrf model. Then type
141 - If it is real-data case (real), place files from WPS (met_em.*)
142 in the appropriate directory, type
146 to produce wrfbdy_d01 and wrfinput_d01. Then type
152 - If you use mpich, type
154 mpirun -np number-of-processors wrf.exe
156 - For information on how to make nested runs, visit
157 http://www.mmm.ucar.edu/wrf/users/
160 ======================================
164 - Advanced Research WRF (ARW) solver: Eulerian mass, hydrostatic and non-hydrostatic
165 * Arakawa C-grid staggering
166 * Runge-Kutta 2nd and 3rd order timestep options
167 * scalar-conserving flux form for prognostic variables
168 * 2nd to 6th order advection options (horizontal and vertical)
169 * time-split small step for acoustic modes
170 * small step horizontally explicit, vertically implicit
171 * divergence damping option and vertical time off-centering
172 * external-mode filtering option for mass model
173 * hydrostatic option via namelist option
174 * positive-definite advection for moisture, scalar, tke and chemical tracers
175 * global modeling capability on latitude-longitude grid
178 * multiple domains and multiple nest levels
179 * supports integer nest grid ratio
180 * feedback option for both odd and even nest grid ratios
187 * Automatic move using a mid-level vortex-following algorithm
190 * microphysics (Kessler/ WRF Single Moment 3, 5 and 6 class / Lin et al./
191 Eta Ferrier / Thompson / Goddard / 2-moment Morrison / )
192 * cumulus parameterization (Kain-Fritsch with shallow convection /
193 Betts-Miller-Janjic / Grell-Devenyi ensemble / New Grell )
194 * planetary boundary layer (Yosei University / Mellor-Yamada-Janjic / ACM2 )
195 * surface layer (similarity theory MM5 / Eta)
196 * slab soil model (5-layer thermal diffusion / Noah land-surface model (4 levels) /
197 RUC LSM (6 levels) / Pleim-Xu (2 levels)
198 * Urban canopy model (works with Noah LSM)
199 * longwave radiation (RRTM, CAM)
200 * shortwave radiation (simple MM5 scheme / Goddard / CAM)
201 * sub-grid turbulence (constant K diffusion/ 2-D Smagorinsky/ predicted TKE /
202 2-D, 6th order diffusion)
203 * Rayleigh damping for w at the upper boundary layer
204 * land-use categories determine surface properties
207 * three-dimensional analysis nudging
208 * observation nudging
212 * Hierarchical software architecture that insulates scientific code
213 (Model Layer) from computer architecture (Driver Layer)
214 * Multi-level parallelism supporting shared-memory (OpenMP), distributed-memory (MPI),
215 and hybrid share/distributed modes of execution
216 * Active data registry: defines and manages model state fields, I/O,
217 nesting, configuration, and numerous other aspects of WRF through a single file,
220 Easy to extend: forcing and feedback of new fields specified by
221 editing a single table in the Registry
222 Efficient: 5-8% overhead on 64 processes of IBM
224 * Enhanced I/O options:
225 NetCDF and Parallel HDF5 formats
226 Nine auxiliary input and history output streams separately controllable through the namelist
227 Special input streams for analysis and observation nudging
228 Output file names and time-stamps specifiable through namelist
229 Special output stream for 3DVAR
230 * Efficient execution on a range of computing platforms:
231 IBM SP systems, (e.g. NCAR "bluevista","blueice","bluefire" Power5-based system)
235 IA64 MPP (HP Superdome, SGI Altix, NCSA Teragrid systems)
237 x86_64 (e.g. TACC's "Ranger", NOAA/GSD "wJet" )
238 PGI, Intel, Pathscale, gfortran, g95 compilers supported
239 Sun Solaris (single threaded and SMP)
240 Cray X1, X1e (vector), XT3/4 (Opteron)
241 Mac Intel/ppc, PGI/ifort/g95
245 * RSL_LITE: communication layer, scalable to very
246 large domains, supports nesting.
247 * I/O: NetCDF, parallel NetCDF (Argonne), HDF5, GRIB, raw binary, Quilting (asynchronous I/O), MCEL (coupling)
248 * ESMF Time Management, including exact arithmetic for fractional
249 time steps (no drift).
250 * ESMF integration - WRF can be run as an ESMF component.
251 * Improved documentation, both on-line (web based browsing tools) and in-line