The concept of the ALADIN project was proposed by Météo-France in 1990, with the aim of building a mutually beneficial collaboration with the National Meteorological Services of Central and Eastern Europe. This collaboration was to be in the field of Numerical Weather Prediction (NWP), which provides the basis for the forecasting tools of modern meteorology. The easy to translate acronym (Aire Limitée Adaptation dynamique Développement InterNational) clearly indicates the major axes of this project at its beginnings.
20 years later, as defined in the 4th Memorandum of Understanding,
The goal of the ALADIN Collaboration is to improve the value of the meteorological, hydrological and environmental warning and forecast services delivered by all Members to their users, through the operational implementation of a NWP system capable of resolving horizontal scales from the meso-beta to the meso-gamma scale and improving the prediction of severe weather phenomena such as heavy precipitation, intensive convection and strong winds.
This objective will be fulfilled through continuation and expansion of the activities of the ALADIN Consortium in the field of High Resolution Short Range Weather Forecast, including:
- Maintenance of an ALADIN System (...);
- Joint research and development activities, on the basis of the common Strategic Plan and related Work Plans, with the aim of maintaining the ALADIN System at scientific and technical state of the art level within the NWP community;
- Sharing scientific results, numerical codes, operational environments, related expertise and know-how, as necessary for all ALADIN Consortium members to conduct operational and research activities with the same tools.
About one hundred scientists, from sixteen countries, each with its own specificity in resources and knowledge base, are permanently contributing to the progress of ALADIN NWP system. They are working together on a modern code of the atmosphere that definitely deserves its proper place between the European state-of-the-art NWP models: 80 Full-Time Equivalent persons in the last years of the project’s life. This code is now operated every day in fifteen Euro-Mediterranean countries, on a huge variety of computing platforms ranging from a PC Cluster under Linux to Vector Computers.
ALADIN consortium had a number of unique successes in the past : for instance, the pluging of an existing physics parameterization in the existing code, leading to the AROME model; ALADIN is at the forefront of the gray-zone problematics with the ALARO physics; ALADIN dynamical core is remarkably stable; ...
ALADIN also allowed to build a high-level scientific team, distributed in sixteen countries that managed to reach the level of the best research centres, as witnessed by the PhD theses and publications in international journals. The General Assembly of Partners, the workshops, the meetings, the newsletters regularly offer opportunity of various exchanges within the ALADIN community.
ALADIN is preparing for the serious evolutions expected within the NWP landscape in the coming five to ten years. There is the ever-lasting question where to draw the line between resolved vs. parameterized processes. There is the question of the efficiency and the scalability of ALADIN dynamical core. There are the external drivers, such as the demands of the end users, and the evolution of the high-performance computing machines. Additionally a serious reorganization of the code is now at hand, in particular within the OOPS project. Besides that, the international meteorological context is steadily changing, specifically with the merger of the ALADIN and the HIRLAM consortia.
Table of contents Practical information on (...) Agenda and list of participant On-line presentations of (...) Posters HMG/CSSI meeting On-line registration for (...) List of registered participant
The National Meteorological Administration of Romania will host (...)
When planning your next visit to Toulouse/CNRM/GMAP, please pay attention to French public holidays and Meteo-France closing days (RTTi) !..
From January 1st, 2014, for security reasons, the access to GMAP offices is only allowed between 7:00 and 21:00 on working days. On Saturdays and on (...)
Table of contents General Assembly Program Manager, full-time Policy Advisory Committee ALADIN Local Teams Managers CSSI : Committee for Scientifi Support Team Task forces
The governance and the management of the ALADIN Consortium is defined by the 4th MoU and (...)
Table of contents 19th ALADIN General Assembly 1st joint ALADIN General (...)
For the first time, the ALADIN General Assembly and the HIRLAM Council will held a joint meeting.
Taking advantage of the opportunity of many of their Directors gathering in Reading (...)
The need for better regional and local short-range weather predictions is continuously increasing in our society, not only in the agricultural, industrial and economic fields but also in relation with outdoor leisure activities. Of prime importance is the prediction of risks, for human life and heavy damages. Not only storms must be better forecasted, but also severe weather which may cause disruptions (in water and power distribution, surface and air transportation, ..), like the occurrence of heavy rain, snowfalls, or fog. The forecast of such weather elements is to be improved, in regards to their intensity and location, with an increased precision both in time and space.
Such an increasing demand impacts directly on numerical weather prediction, since NWP models, especially high-resolution limited-area ones, are now a major tool for weather forecasting. All models and each team have now got to make headway towards very high resolution, typically horizontal scales around 1-3 km. Besides, assuming that the performances of computers will keep increasing as they did in the past, it is realistic to envisage, within 3 to 4 years, horizontal resolutions of 2-3 km for NWP operational models, though on small targeted domains first. By most partners, the main operational applications should stay a little behind, with horizontal resolutions between 5 and 8 km.
However, NWP, at such scales, has to address new problems, linked either to the increased complexity of equations required for a better description of phenomena, or to the very large amounts of data the models will need (mainly meteorological observations) and produce. And, at each step and level, one will have to find compromises between scientific ambitions and the capacity of operational forecasting suites.
The STORMNET research plan can be described following 4 main directions, which may be considered in parallel in spite of many common points or interactions.
This domain combines the 3 main historical issues of limited-area NWP.
There is now a general agreement that non-hydrostatic aspects must be taken into account at horizontal scales below 5 km, and even that a fully compressible system of equations should be used. Such features are implemented in all 4 European limited-area NWP models, though at different levels of achievement.
However, some debates are still open, between modelling groups or simply for improving any model: type of horizontal projection, finite-differences vs spectral methods, Eulerian vs semi-Lagrangian advection, from explicit to fully implicit time-stepping, choice of the vertical coordinate and the prognostic model variables, formulation of the lower and upper boundary conditions... Answering these questions, if possible, requires significant further work.
Besides, as horizontal or vertical resolution increases, new problems are likely to emerge, as concerns accuracy, stability or efficiency. Thus, some present choices for the very basic set of model equations may be reconsidered in the near future.
The driving of a limited-area model through its lateral boundary conditions by a global or a coarser limited-are model, must be carefully designed to allow a good representation of large scales, not described by the model itself. All models are presently using the same formulation, the pragmatic Davies’ relaxation scheme. It presents some known conception problems, but is quite simple and robust.
Though research in this domain started several years ago, there is still significant work required to implement a better alternative in operations, especially for spectral models. Once this difficult problem solved, more sophisticated algorithms could be addressed, if necessary.
The time-dimension of coupling is also to be reconsidered when facing very high resolution and short-range forecasts.
Model equations have also to describe the radiative and orographic forcings and exchanges with the surface, as well as the moist and subgrid-scale processes. This part of the model will undergo an important transformation, since one shall have to get rid simultaneously of two currently widely used key assumptions.
First of all, there will be less and less need to take into account details of the flow organisation internal to the mesh-box. Typically, organised convection and details of the subgrid scale underlying topography will cease to generate fluxes of magnitude comparable to those produced by explicit computations between neighbouring grid-points (this aspect deserving special attention at the intermediate resolutions). Secondly, the forcing of parameterizations will cease to be treated as horizontally homogeneous, which means that the 3 space-dimensions will play a role in the computation of turbulent and radiative fluxes. Besides, some aspects which have been neglected up to now because of the too coarse horizontal resolution, such as the impact of lakes or towns on the low atmosphere, will have to be parameterized.
This revolution in the perception that NWP code designers have of "physics" will also have a strong influence the way the parameterization codes are treated algorithmically. Old sources of inaccuracies and of latent instabilities will disappear, while one can safely anticipate that they shall be replaced by new ones, and interfacing with dynamics may have to be reconsidered. Exchanges with external experts should also be encouraged, to improve the description of air-sea exchanges, the impact of aerosols on radiation, the parameterization of run-off, etc ... or even couple directly the NWP model with another one (a procedure which is also likely to favour the emergence of problems).
The interaction of physical and large-scale forcings with dynamics will have to be revisited in some models. Two main challenges can be considered. The first one, dealing with the organisation of the time-step, should optimally allow to use various packages of physical parameterizations, and even combinations, with variants of the dynamics, while ensuring accuracy, stability and efficiency for the whole system. The second one is the progressive introduction of 3-dimensional aspects in physical parameterizations (typically for turbulence), which will lead to significantly more complex interactions between physics and dynamics.
A very good model is useless when run with bad initial conditions. Moreover the refinement of the initial state using meteorological observations, usually called data assimilation, is all the more important since the forecast model is sophisticated, and improvements of both systems must always be kept consistent. One has to note that developments in data assimilation should provide not only more reliable initial conditions for a local high-resolution weather forecast, but also analytical tools to exploit the information from local high-density observation networks. Enhanced links with nowcasting (i.e. very short-range - a few hours- forecast of extreme events) are also expected.
A wider range of horizontal scales, say 1-10 km, is considered here, since there are still significant improvements to be brought to the present operational data assimilation systems (i.e. at scales around 10 km, on average) while one starts from scratch at very high resolution, but the following list of research topics concentrates on issues specific to limited-area models.
Work will continue on the use of global-coverage data, with emphasis put on the specific pre-processing algorithms required for very dense observations and limited domains. Satellite data of various kind will be injected, at the highest possible resolution. Special attention will be paid to cloud or rain-affected data. However final products designed by specialized centres (SAF) will be used whenever possible.
Local remote-sensing observations, such as radar (reflectivities, 3d or radial winds) and GPS data, will also be assimilated. They are likely to provide useful information on humidity, and the initialization of moist variables is a crucial but very difficult issue at high resolution. The corresponding work has already started in some NWS, using more or less sophisticated assimilation tools, which allows to be confident on a future operational use.
Synoptic surface data, the most conventional observation type, will gain an extra interest. Indeed, the surface network remains a fairly cheap and dense source of informations for the surface and near-surface atmospheric conditions (e.g. snow coverage, precipitations, temperature and humidity at 2 m height, ...).
To end with, attention will be paid to the qualification of observations errors and to their representativeness.
Three major issues are considered here: the description of background errors, analysis algorithms, and the problem of initialization (filtering of spurious numerical modes).
The proper specification of the "background" error covariances is a major issue in modern data assimilation, whatever the analysis algorithm. These functions control the spatial filtering and the space propagation of analysis increments and provide balance conditions. The assumptions used up to now to prescribe them will have to be revisited at very high resolution, where balance conditions are different and moist processes play an increased part. Developments will concern the five main following aspects: sampling (error estimation techniques), covariance representations (on which space, with which level of simplification, ...), balance conditions, flow dependence, and tunings.
Though the simple nudging technique is used by a few teams, there is now a general trend to rather consider variational methods (3D-Var here). Apart from optimization issues, some topics should receive an increasing interest : time dimension (higher frequency of analyses versus semi-continuous systems), choice of adequate lateral boundary conditions, ensemble filters, and coupling of high-resolution limited-area and global synoptic-scale data assimilation systems.
Little is known about the actual need for a specific initialization, in the sense of an extra filtering of analysis fields, before launching a high resolution forecast. New diagnostic tools to evaluate the level of "noise" in analyzed fields have to be developed, while addressing the problem of the actual physical balance conditions present at very high resolution in the atmosphere.
F4D-Var techniques have been very successful and are widely used operationally for global data assimilation and also, recently, for operational data assimilation in a model at 10 km resolution, at the Japanese Meteorological Agency. Theoretical arguments are in favour of 4D-Var also on the mesoscale, knowing for example the importance of the time dimension and of moist processes, i.e. of physics. Many of the processes on the mesoscale are strongly non-linear, however, and this may be critical for 4D-Var, which makes heavy use of tangent-linear approximations. Multi-incremental methods (with re-linearizations) may help to circumvent this inherent problem.
A major issue is to settle a numerically efficient dynamical core: tangent-linear and adjoint formulations of the semi-Lagrangian time-stepping need to be developed (done already by the HIRLAM community). In fact, 4D-Var is and will stay expensive, as concerns computing cost and development work. A second important issue is to assess the minimum level of required simplified regular physics. It is likely that at least a boundary-layer description and micro-physics are needed.
Any 4D-Var system operating at 10 km or below is worth testing in research mode. These tests should also include derived applications such as gradient computations, sensitivity studies using the adjoint model, and a-posteriori retuning. Comparisons with three-dimensional assimilation schemes are certainly relevant in the early stages.
Two rather independent issues are to be considered here. Firstly, the initialization of the prognostic variables describing soil and surface (temperature and humidity).The present trend towards an off-line variational assimilation (since the associated time-scales are longer than for atmospheric process) will be maintained, trying to simultaneously go to more sophisticated algorithms, use new observation types (remote-sensing data), and keep the overall computing cost within safe limits. Secondly there will be an increasing interest for diagnostic analyses, using observations of any type close to the surface, in order to provide a fine description of the boundary layer. This will help forecasters in nowcasting, and may provide forcing to off-line models. As a first step, very basic evaluations will be required, such as the comparison of several pragmatical approaches, based on existing techniques.
This domain is of major importance, since the overall objective is the supply of improved local weather forecasts to any end-user. However, here one has also to consider the present, coarser, operational resolution, since the first following issues are at their first stage of development.
Predictability is the net result of several distinct sources of uncertainty: (i) the use of imperfect models; (ii) the use of observations that imperfectly sample the past and present state of the system, both properties that, to some extent, can be controlled, and (iii) a fundamental intrinsic property of the atmosphere itself, which, very much like other state parameters, can at best be measured but cannot be changed.
One popular way of figuring out predictability is ensemble forecasting. Other ways consist of coupling observations to the data assimilation system (adaptive observation) or to assimilate observations depending of predictability-related properties (adaptive assimilation).
The following issues should be addressed: (i) initialization of an ensemble forecast relevant to mesoscale, short-range prediction; the techniques should become able to take into account uncertainties in the humidity fields, surface conditions (soil wetness); (ii) the direct extension of forecast error representation as done in synoptic scale ensembles to convective processes is non trivial; (iii) assess the respective importance of the initial errors and those resulting from the diversity of coupling scenarios; (iv) the validation of an ensemble forecast tool, especially on the mesoscale, depends on the statistical of the ensemble with respect to some climatological reference; the usual approaches suffer from two major drawbacks related to the small number of validation cases and the kind of climatological reference; ways to improve on these methods using event-based approaches should be studied (v) the interpretation and use of an ensemble forecast is an open question; while probabilistic forecasts are directly usable in cost-loss models with explicit weather parameters and their uncertainty, their application to natural disaster management remains a challenge; (vi) since there are theoretical arguments on the inability of the current ensemble strategy (which is a kind of "clever under-sampling") to deal with extreme events, research on alternative approaches such as adaptive assimilation or observation should be extended to the mesoscale.
An obvious way to control the skill of NWP models is to compare the observations of a meteorological network with the respective model values, using more or less advanced interpolation methods. However, this method has already reached its limits at the present operational scales, especially when forecasts of precipitation and cloudiness are to be evaluated. The presently used sample of observations is quite coarse, and smooth models are favoured: typically a good forecast of the intensity of a storm with a time-shift of a few hours is penalized against the prediction of a far weaker event at the right time. The use of ensemble prediction systems cannot prevent from addressing the development of new methods for the verification of deterministic forecasts (and also for probabilistic ones). In particular, there is a need for a method that values phase errors properly; such a method will probably be using modern pattern-recognition techniques. This involves common research work with data assimilation issues, for the comparison to new observations.
A meaningful evaluation of the quality of meteorological forecasts has to consider their impact on downstream applications. The output of NWP models is a major source of information when running air-pollution, hydrological, oceanographic or sea-state models. A closer cooperation with the corresponding research teams will help in the qualification of the forecast skill of NWP models as well as in the evaluation of the relevance of new developments.
This is the upstream counterpart of verification, many developments will be common ones. Modellers will have to define new diagnostic tools, better suited to very high resolution, and to face ensuing new problems. Besides, intensive numerical experimentation, combining forecast and assimilation, will be required.
Today NWP is much more than a model code: it is an ensemble of programmes with many and deep interactions between them, possibly running on several platforms. Moreover, with the substantial increases in horizontal and vertical resolution that we experience, the amount of data we have to deal with becomes a problem in itself.
Therefore, NWP centres are facing some challenging issues which need a joint effort :
This list has been created in order to keep in touch (...)