Parallel Computations with Large-scale Air Pollution Models
Large-scale mathematical models are very powerful tools in the efforts to provide more information and more detailed information about the pollution levels, especially about pollution levels which exceed certain critical values.. However, the model used must satisfy at least two conditions: (i) it m...
Gespeichert in:
| Datum: | 2015 |
|---|---|
| Hauptverfasser: | , , |
| Format: | Artikel |
| Sprache: | Ukrainian |
| Veröffentlicht: |
PROBLEMS IN PROGRAMMING
2015
|
| Schlagworte: | |
| Online Zugang: | https://pp.isofts.kiev.ua/index.php/ojs1/article/view/19 |
| Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
| Назва журналу: | Problems in programming |
| Завантажити файл: | |
Institution
Problems in programming| id |
pp_isofts_kiev_ua-article-19 |
|---|---|
| record_format |
ojs |
| resource_txt_mv |
ppisoftskievua/cd/0a1b1abf0a0697722f22118981dd50cd.pdf |
| spelling |
pp_isofts_kiev_ua-article-192018-10-09T10:24:50Z Parallel Computations with Large-scale Air Pollution Models Параллельные вычисления на основе крупномасштабных моделей загрязнения атмосферы Паралельні обчислення на основі великомасштабних моделей забруднення атмосфери Dimov, I. Farago, I. Zlatev, Z. UDC 51.681.3 УДК 51.681.3 УДК 51.681.3 Large-scale mathematical models are very powerful tools in the efforts to provide more information and more detailed information about the pollution levels, especially about pollution levels which exceed certain critical values.. However, the model used must satisfy at least two conditions: (i) it must be verified that the model results are reliable and (ii) it should be possible to carry out different study by using the model. It is clear that comprehensive studies about relationships between different input parameters and the model results can only be carried out (a) if the numerical methods used in the model are sufficiently fast and (b) if the code runs efficiently on the available high-speed computers. Some results obtained recently by a new unified version of the Danish Eulerian Model will be presented in this paper. Крупномасштабные математические модели – очень мощный инструмент для получения более детальной информации относительно уровней загрязнений. Однако модель должна удовлетво- рить по крайней мере двум условиям: (i) результаты моделирования должны быть надежными и (ii) должна существовать возможность уточнения и изучения характеристик модели. Всестороннее изучение отношений между различными параметрами входа и результатами моделирования может быть выполнено, если (a) численные методы, используемые в модели, достаточно быстры и (b) программное обеспечение на доступных быстродействующих компь- ютерах достаточно эффективно. Представлены результаты параллельной реализации моделирования загрязнения атмосферы, полученные в новой объединенной версии датской Эйлеровой модели. Великомасштабні математичні моделі – дуже потужний інструмент для одержання більш детальної інформації щодо рівнів забруднень. Проте використовувана модель повинна задовольнити принаймні двом умовам: (i) результати моделювання повинні бути надійними і (ii) повинна існувати можливість уточнення і вивчення різноманітних характеристик моделей. Всебічне вивчення відношень між різноманітними параметрами входу і результатами моделювання може бути виконане, якщо (a) чисельні методи, використовувані в моделі, достатньо швидкі та (b) програмне забезпечення на доступних швидкодіючих комп'ютерах достатньо ефективне. Подані результати рівнобіжної реалізації моделювання забруднення атмосфери, отримані в новій об'єднаній версії датської Ейлерової моделі. PROBLEMS IN PROGRAMMING ПРОБЛЕМЫ ПРОГРАММИРОВАНИЯ ПРОБЛЕМИ ПРОГРАМУВАННЯ 2015-07-01 Article Article application/pdf https://pp.isofts.kiev.ua/index.php/ojs1/article/view/19 PROBLEMS IN PROGRAMMING; No 3 (2003) ПРОБЛЕМЫ ПРОГРАММИРОВАНИЯ; No 3 (2003) ПРОБЛЕМИ ПРОГРАМУВАННЯ; No 3 (2003) 1727-4907 uk https://pp.isofts.kiev.ua/index.php/ojs1/article/view/19/23 Copyright (c) 2015 ПРОБЛЕМИ ПРОГРАМУВАННЯ |
| institution |
Problems in programming |
| baseUrl_str |
https://pp.isofts.kiev.ua/index.php/ojs1/oai |
| datestamp_date |
2018-10-09T10:24:50Z |
| collection |
OJS |
| language |
Ukrainian |
| topic |
UDC 51.681.3 |
| spellingShingle |
UDC 51.681.3 Dimov, I. Farago, I. Zlatev, Z. Parallel Computations with Large-scale Air Pollution Models |
| topic_facet |
UDC 51.681.3 УДК 51.681.3 УДК 51.681.3 |
| format |
Article |
| author |
Dimov, I. Farago, I. Zlatev, Z. |
| author_facet |
Dimov, I. Farago, I. Zlatev, Z. |
| author_sort |
Dimov, I. |
| title |
Parallel Computations with Large-scale Air Pollution Models |
| title_short |
Parallel Computations with Large-scale Air Pollution Models |
| title_full |
Parallel Computations with Large-scale Air Pollution Models |
| title_fullStr |
Parallel Computations with Large-scale Air Pollution Models |
| title_full_unstemmed |
Parallel Computations with Large-scale Air Pollution Models |
| title_sort |
parallel computations with large-scale air pollution models |
| title_alt |
Параллельные вычисления на основе крупномасштабных моделей загрязнения атмосферы Паралельні обчислення на основі великомасштабних моделей забруднення атмосфери |
| description |
Large-scale mathematical models are very powerful tools in the efforts to provide more information and more detailed information about the pollution levels, especially about pollution levels which exceed certain critical values.. However, the model used must satisfy at least two conditions: (i) it must be verified that the model results are reliable and (ii) it should be possible to carry out different study by using the model. It is clear that comprehensive studies about relationships between different input parameters and the model results can only be carried out (a) if the numerical methods used in the model are sufficiently fast and (b) if the code runs efficiently on the available high-speed computers. Some results obtained recently by a new unified version of the Danish Eulerian Model will be presented in this paper. |
| publisher |
PROBLEMS IN PROGRAMMING |
| publishDate |
2015 |
| url |
https://pp.isofts.kiev.ua/index.php/ojs1/article/view/19 |
| work_keys_str_mv |
AT dimovi parallelcomputationswithlargescaleairpollutionmodels AT faragoi parallelcomputationswithlargescaleairpollutionmodels AT zlatevz parallelcomputationswithlargescaleairpollutionmodels AT dimovi parallelʹnyevyčisleniânaosnovekrupnomasštabnyhmodelejzagrâzneniâatmosfery AT faragoi parallelʹnyevyčisleniânaosnovekrupnomasštabnyhmodelejzagrâzneniâatmosfery AT zlatevz parallelʹnyevyčisleniânaosnovekrupnomasštabnyhmodelejzagrâzneniâatmosfery AT dimovi paralelʹníobčislennânaosnovívelikomasštabnihmodelejzabrudnennâatmosferi AT faragoi paralelʹníobčislennânaosnovívelikomasštabnihmodelejzabrudnennâatmosferi AT zlatevz paralelʹníobčislennânaosnovívelikomasštabnihmodelejzabrudnennâatmosferi |
| first_indexed |
2025-07-17T09:59:26Z |
| last_indexed |
2025-07-17T09:59:26Z |
| _version_ |
1850411598497185792 |
| fulltext |
Параллельное программирование
© I. Dimov, I. Farago, Z. Zlatev, 2003
44 ISSN 1727-4907. Проблемы программирования. 2003. № 3
UDC 51:681.3
I. Dimov, I. Farago, Z. Zlatev
PARALLEL COMPUTATIONS WITH LARGE-SCALE AIR
POLLUTION MODELS
Large-scale mathematical models are very powerful tools in the efforts to provide more
information and more detailed information about the pollution levels, especially about pol-
lution levels which exceed certain critical values.. However, the model used must satisfy at
least two conditions: (i) it must be verified that the model results are reliable and (ii) it
should be possible to carry out different study by using the model. It is clear that compre-
hensive studies about relationships between different input parameters and the model re-
sults can only be carried out (a) if the numerical methods used in the model are suffi-
ciently fast and (b) if the code runs efficiently on the available high-speed computers.
Some results obtained recently by a new unified version of the Danish Eulerian Model will
be presented in this paper.
1. Need for a Unified Air Pollution
Model
It will be possible to apply an air
pollution model in many different studies
only if it is very flexible. The work on the
development of the Danish Eulerian
Model (DEM) has been initiated in the
beginning of the 80’s. Only two species
were studied in the first version of the
model. The space domain was rather
large (the whole of Europe), but the dis-
cretization was very crude; the space do-
main was divided by use of a 32x32 grid,
which means, roughly speaking, that 150
km x 150 km grid-squares were used.
Both a two-dimensional version and a
three-dimensional version of this model
were developed. In both versions the
chemical scheme was very simple. There
was only one linear chemical reaction.
The two-dimensional version of this
model is fully described in Zlatev [27],
while the three-dimensional version was
discussed in Zlatev et al. [30]. These first
two versions were gradually upgraded to
much more advanced versions: (i) ver-
sions containing more chemical species,
(ii) versions discretized on more refined
grids and (iii) versions utilizing more ad-
vanced numerical algorithms (see Zlatev
[28] and Zlatev et al. [29]).
Recently, the different versions of
DEM were united in a common model.
Several options can easily be specified
when the new model is applied in differ-
ent scientific studies:
• the model can be used both as
a 2-D model and a 3-D model,
• different horizontal resolutions
can be specified (50 km x 50 km, 16.67
km x 16.67 km and 10 km x 10 km cells
can be used in the horizontal planes),
• different chemical schemes can
be selected (a scheme with 35 species
and a scheme with 56 species being
available at present),
• an algorithm for dividing the
large arrays into chunks can be applied
in an attempt to exploit better the cache
memory of the computer used.
The new model is called UNI-
DEM.
2. Mathematical Description of UNI-DEM
UNI-DEM is described mathemati-
cally, see [28], by a system of partial dif-
ferential equations (PDEs). The number
of equations in this system is equal to the
number of chemical species which are
treated by the model. Five physical and
chemical processes are represented in the
model with different mathematical opera-
tors. These processes are: (i) horizontal
advection (horizontal transport due to the
wind), (ii) horizontal diffusion, (iii) chem-
istry + emissions, (iv) deposition and (v)
vertical exchange (vertical advection +
vertical diffusion).
The system of PDEs by which
these five processes are described in
UNI-DEM is of the following type:
Параллельное программирование
45
( )
( ) ( )
( )
( ) ( )
( ) [ ]
1 2
1 2
, ,...,
, , , ,
, , , 0, , 1, 2,..., ,
ii i
z
i i i
x
i
y i q
i i i i
wcc cK
t z z z
uc vc cK
x y x x
cK Q c c c
y y
E x y z t c
x y z t T i q
κ κ
∂∂ ∂∂ = − + − ∂ ∂ ∂ ∂
∂ ∂ ∂∂ − − + + ∂ ∂ ∂ ∂
∂∂
+ + + ∂ ∂
+ − +
∈Ω ∈ =
(1)
where (i) Ω is the space domain on
which the system of PDEs (1) is defined
(in the 2-D options of UNI-DEM Ω is a
4800 km x 4800 km square, while a
4800 km x 4800 km x 3 km parallelepi-
ped is the space domain Ω in the 3-D
options of UNI-DEM), (ii) [ ]T,0 is the
time-interval (UNI-DEM is normally run
over a time-period covering meteorologi-
cal data for a given year) and (iii) q is
the number of chemical species involved
in the model.
The unknown functions ic in (1)
are concentrations of the involved
chemical species. The quantities u, v
and w are wind velocities along the co-
ordinate axes. xK , yK and zK are diffu-
sion coefficients. The chemical reactions
are described by the non-linear func-
tions iQ . The emissions sources in the
space domain Ω of the model are rep-
resented by the functions iE . Finally,
i1κ and i2κ are deposition coefficients
(for dry deposition and wet deposition
respectively).
The system of PDEs (1) has to be
considered with appropriate initial and
boundary conditions. The choice of such
conditions is beyond the scope of this
paper (some details can be found in
Zlatev [28]).
Different air pollution models can
be obtained from the generic form (1) of
the system of PDEs by using different
mechanisms in the description of the un-
derlying physical and chemical processes
(More details can be found in Ebel [11]
and Zlatev [28]).
3. Applying Splitting in UNI-DEM
It is difficult to treat directly the
system of PDEs (1) by which the UNI-
DEM is described mathematically. There-
fore, different kinds of operator splitting
are used in the large-scale air pollution
models. Any splitting procedure leads to
sub-models, which can be treated by dif-
ferent numerical methods. It is also easier
to choose and use standard packages for
some of the sub-models. Thus, efficient
methods can be selected for each sub-
model, which leads to better overall per-
formance when the whole model is han-
dled computationally.
An attempt to simplify the splitting
procedure used in [28] is made in UNI-
DEM. The big model is split in three sub-
models by using this procedure:
• a sub-model containing the ver-
tical exchange operator,
( )(1)(1) (1)
,
1,2,..., ,
ii i
z
wcc cK
t z z z
i q
∂ ∂ ∂∂
= − + ∂ ∂ ∂ ∂
=
(2)
• a sub-model containing the
horizontal advection and the horizontal
diffusion operators,
( ) ( )(2) (2)(2)
(2) (2)
,
1, 2,..., ,
i ii
i i
x y
uc vcc
t x y
c cK K
x x y y
i q
∂ ∂∂
= − − +
∂ ∂ ∂
∂ ∂∂ ∂
+ + ∂ ∂ ∂ ∂
=
(3)
• a sub-model containing the
chemistry, the emissions and the deposi-
tion operators,
( )
( ) ( )
(3)
(3) (3) (3)
1 2
(3)
1 2
, ,...,
, , , ,
1, 2,..., .
i
i q
i i i i
c Q c c c
t
E x y z t c
i q
κ κ
∂
= +
∂
+ − +
=
(4)
In the actual computations the
three sub-models are treated successively
at a sequence of small time sub-intervals
of length τ of [ ]T,0 . The three sub-
models, (2)-(4), are connected by using
appropriate initial conditions (the cou-
Параллельное программирование
46
pling of the sub-models obtained after
applying some splitting procedure are
discussed in Zlatev [28]).
The first of these three sub-models
is not used when any of the 2-D options
is specified.
The particular splitting procedure
proposed here, see (2)—(4), is based on
ideas discussed in Marchuk [19] and
McRae et al. [20]. Other splitting proce-
dures may also be devised. A very popu-
lar splitting procedure is that proposed
independently in 1968 by Marchuk [18]
and Strang [25].
The splitting procedure is in gen-
eral producing some errors. The order of
the splitting procedure described by (2)-
(4) is one, i.e. ( )τO . The splitting proce-
dure proposed in Marchuk [17] and
Strang [25] is of second order, ( )2τO . At
the same time, it should also be men-
tioned that the latter splitting procedure
is considerably more expensive computa-
tionally.
If some assumptions are imposed,
then the errors due to the splitting pro-
cedure may vanish. Different cases, in
which such assumptions are imposed and,
as a result of this, the errors due to the
splitting procedure disappear, are dis-
cussed in Dimov et al. [5] and Lanser and
Verwer [17].
4. Numerical Methods Used in UNI-DEM
The sub-models are to be treated
by appropriate numerical methods. The
particular numerical methods that are im-
plemented in UNI-DEM are listed below.
• The finite elements discussed in
Pepper et al. [23] are used to discretize
the spatial derivatives in the vertical ex-
change sub-model (2). The semi-discre-
tized model is handled by the ϑ -method
(see Lambert [15] or Samarskii [24]).
• The finite elements from Pepper
et al. [24] are used at present also in the
discretization of the spatial derivatives in
the horizontal advection-diffusion sub-
model (3). Many other methods were
tried: psedospectral algorithm, finite dif-
ferences, methods based on the Bott’s al-
gorithm, semi-Lagrangian algorithm, etc.
The semi-discretized sub-model is treated
by using a set of predictor-corrector (PC)
schemes with several different correctors.
The major idea is to vary the PC schemes
in an attempt both to preserve the stabil-
ity of the computations and to keep the
same time-stepsize (Zlatev [28]).
• The chemistry-emission-deposi-
tion sub-model (4) is treated by using an
improved QSSA (Quasi-Steady-State-
Approximation) algorithm. The original
QSSA algorithm, several classical time-
integration methods for stiff systems of
ODEs and a partitioning algorithm have
also been tried, see Alexandrov et al. [1].
5. Need for high-speed computers
The spatial discretization of UNI-
DEM leads to huge systems of ordinary
differential equations (especially for the
fine resolution options and for the 3-D
options). This is illustrated in Table 1 and
Table 2, where the numbers of equations
and time-steps in the advection part for
the option with 35 chemical species are
given. The numbers of time-steps for the
3-D options of UNI-DEM are the same as
the numbers of time-steps for the corre-
sponding 2-D options (which are given in
Table 2).
Table 1. Numbers of equations when six
options of UNI-DEM are used
Spatial resolution
(horizontal)
2-D option
(1-layer)
3-D option
(10 layers)
96x96 grid 322 560 3 225 600
288x288 grid 2 903 040 29 030 400
480x480 grid 8 064 000 80 640 000
Table 2. Numbers of time-steps when
UNI-DEM is used in a one-year run
Spatial resolution
(horizontal)
2-D option
(1-layer)
96x96 grid 35 520
288x288 grid 106 560
480x480 grid 213 120
In many studies the model has to
be run over many years (in, for example,
climatic studies) and/or with many hun-
dreds of different scenarios (when, for
example, the influence of the biogenic
emissions on high pollution levels is
studied).
Параллельное программирование
47
This short discussion explains why
comprehensive air pollution studies can
successfully be carried out only on mod-
ern high-speed computers and when fur-
thermore the code optimized for runs on
such computers, which is by far not an
easy task.
Moreover, it must be emphasized
that the computers which are available
for us at present are not sufficiently fast
for all options of UNI-DEM. This fact is
clearly seen when the results in Table 4
are studied.
6. Organization of the parallel
computations
It is very important to apply stan-
dard parallelization tools. This is crucial
when the code has to be moved from one
computer to another (which is often
needed in practice).
If a shared memory computer is
available, OpenMP [26] can successfully
be implemented. It is quite straightfor-
ward to use this tool. However one
should be careful in the efforts to solve
the following task: to identify as large as
possible well balanced parallel tasks.
While the programming by using
OpenMP is relatively simple, the solution
of the latter tasks might be difficult.
If the computer is with distributed
memory, then one can apply MPI [14]
(the Message Passing Interface). The im-
plementation of MPI is a much more
complicated task than the implementa-
tion of OpenMP. On the other side, the
use of MPI leads to dividing the data be-
tween the processors, which implies the
use of shorter arrays. This very often re-
sults automatically in a better utilization
of the different cache memories of the
computer used.
Very often MPI can be used even
on shared memory machines and, more-
over, may give good results on such
computers, because as mentioned above
the caches are utilized in a more effi-
cient way.
6.1. Identification of the parallel
tasks when OpenMP is to be used. The
first of the two actions is to identify the
parallel tasks in the different sub-models
(obtained as a result of applying the
splitting procedure used).
Parallel Tasks in the Vertical Ex-
change Sub-Model. In this sub-model,
the vertical exchange of each pollutant
along each vertical grid-line can be con-
sidered as a parallel task. The number of
parallel tasks is equal to the product of
the number of chemical species and the
number of vertical grid-lines. Since the
number of chemical species in UNI-DEM
is in fact 32, the numbers of
parallel tasks are 96x96x32=294912,
288x288x32=2654208 and 480x490x32=
=7372800 for the 3-D options discretized
on 96x96, 288x288 and 480x480 grids re-
spectively. The tasks are small. Therefore,
it is convenient to group several grid-
points as a bigger task in an attempt to
improve the performance.
Parallel Tasks in the Horizontal
Advection-Diffusion Sub-model. In the
horizontal advection-diffusion sub-model,
the transport of each pollutant can be
considered as a parallel task. The number
of species in the UNI-DEM is 35 (see
Section 1). However, three of them are
linear combinations of other species.
Therefore, the number of independent
species is 32. The number of parallel
tasks in the 2-D options is equal to the
number of independent species. This
means that the loading balance is not
perfect when the number of processors is
not a divisor of 32 for the 2-D options. If
the 3-D options are used, then the
number of parallel tasks become equal to
the product of the number of species and
the number of layers, i.e. this number is
now 320. Again the loading balance will
not be perfect when the number of
processors used is not a divisor of 320.
Parallel Tasks in the Chemistry-
Emission-Deposition Sub-model. In the
chemical module, the chemical reactions
at a given grid-point can be considered
as parallel tasks. This means that the
number of parallel tasks is very large
and, moreover, the loading balance is
perfect. The numbers of parallel tasks are
9216, 82944 and 230400 for the 2-D op-
tions discretized on 96x96, 288x288 and
480x480 grids respectively. The numbers
Параллельное программирование
48
of tasks will be increased by a factor of
10 when the corresponding 3-D options
are used. However, in all cases the tasks
are small. Therefore it is convenient to
group several grid-points as a bigger
task. The performance can be improved
considerably by a suitable choice of the
groups of grid-points. The particular
choices of grouping of the small tasks are
discussed in detail in Owczarz and Zlatev
[21], [22].
6.2. Implementation of MPI
options. The approach used when MPI
tools are to be implemented is based on
dividing the space domain of the model
into p sub-domains, where p is the num-
ber of processors which are to be used in
the run. Two specific modules are needed
in the MPI versions:
• a pre-processing module and
• a post-processing module.
Pre-processing Module. The first of
these modules performs a pre-processing
procedure. In this procedure the data are
divided into several parts (according to
the nodes that are to be applied) and
sent to the assigned for the job nodes. In
this way each processor will work on its
own data during the whole run.
Post-processing Module. The sec-
ond module performs a post-processing
procedure. This means that every node
prepares during the run its own output
files. When the run is finished, then the
post-processing module collects the
output data on one of the nodes and
prepares them for future applications
(visualizations, animations, etc.).
Benefits of Using the Additional
Modules. By the preparation of the two
additional modules, one avoids some
excessive communications during the
run. However, it should also be stressed
that not all communications are avoided.
The use of the pre-processor and the
post-processor is in fact equivalent to the
application of a kind of domain decom-
position. Therefore, it is clear that some
communications are needed along the
inner boundaries of the domains.
Such communications are to be carried
out only once per step and only a few
data are to be communicated. Thus, the
actual communications that are to be
carried out during the computational
process are very cheap.
6.3. Combination of MPI and
OpenMP options. On some computer
architectures, like IBM SMP, it is
efficient to use a combination of MPI and
OpenMP options. These computers
consist of several nodes. Every node
contains several processors which share
the same memory. Every mode has its
own local memory. Therefore, it is natural
to use OpenMP within every node and
MPI across the nodes. Some results from
runs on such an architecture with a
combination of OpenMP and MPI options
are presented in Owczarz and Zlatev [22].
7. Parallel runs on SUN computers
The results presented in this sec-
tion were obtained by running six options
of UNI-DEM on 8 processors of SUN
computers. The code was also run on
many other parallel computers: SGI Ori-
gin 2000, IBM SMP, Cray T3E, etc. (see
Georgiev and Zlatev [15] and Owczarz
and Zlatev [21], [22]).
7.1. Information about the com-
puters used. SUN shared memory com-
puters were used in all runs. These com-
puters are located at the High Perform-
ance Centre at the Technical University
of Denmark (DTU). The characteristics of
the computers used are listed in Table 3.
More information about these computers
can be found in http://www.hpc.dtu.dk.
7.2. Results obtained by the MPI
options and different spatial resolutions.
Both 2-D and 3-D options discretized on
all grids mentioned above (a 96x96 grid,
a 288x288 grid and a 480x480 grid) were
used in the experiments. Some results ob-
tained by using these computers are
given in Table 4. MPI programming is
used in the runs the results of which are
shown in Table 4.
The speed-ups were linear (and for
the large problems even super-linear).
This means that the computing times will
be increased with a factor approximately
equal to eight if the jobs are run on one
processor. The largest job will be com-
pleted in more than 286 days if the com-
Параллельное программирование
49
puter is continuously running in so many
days (i.e. no rebooting, no upgrading and
no crashes in at least 286 days). On the
other hand, if 32 processors are used
when the most time-consuming job is run
(the 3-D option on a 480x480 grid), then
the computing time is 192.08 CPU hours
(a speed-up of 4.46 when compared with
the corresponding run on 8 processors).
Table 4. Computing times in CPU hours
when 8 processors are used
It is clearly seen that if the refined
options (especially the 3-D refined op-
tions) are used, then more powerful proc-
essors and more processors are needed
when runs over long-time periods are to
be carried out and/or when results from
many scenarios are to be investigated.
7.3. OpenMP options versus MPI
options. It is more profitable to use MPI
than OpenMP when the UNI-DEM code
is run on SUN computers. Some results,
which illustrate this statement, are given
in Table 5. Both the OpenMP option and
the MPI option were run over a time-
interval consisting of 1000 steps and the
computing times given in Table 5 are in
seconds.
It is seen that the results obtained
by the MPI version are better than those
obtained by the OpenMP option. This
deserves some explanation remarks. In
the MPI option the largest arrays are di-
vided to p sub-domains (where p is the
number of processors) containing arrays
Table 5. Computing times in seconds
when 8 processors are used to run both
the OpenMp option and the MPI option
over a time-interval of 1000 steps
Physical processes
(and total time)
OpenMP
option
MPI
option
Advection-diffusion 2618 457
Chemistry-emission-
deposition
1339 688
Total computing
time
4011 1281
which are p times smaller. This leads
automatically to a reduction of the cache
misses and, thus, improves the perform-
ance. A general framework and theory to
overcome shared memory bottlenecks
and to increase performance by meanse
of enhanced schemes of data exchanges
in multilevel memory systems is pre-
sented in [8—10]. In Table 5 we have
p=8. If we increase the number of proc-
essors, then the MPI option will perform
even better, because the arrays in the
sub-domains become smaller. To illus-
trate that such a statement is true, we
performed the same runs (as those in
Table 5) on 32 processors. The results
are shown in Table 6. The comparison of
the speed-ups (which are given in
brackets in Table 6) shows that the effi-
ciency of the parallelization is greater
for the MPI option.
7.4. Remark about the use of com-
putational grids. Using computational
grids might resolve the difficulties arising
when the refined options are to be ap-
plied in comprehensive air pollution stud-
ies (as the studies listed in the next sec-
tion). If advanced modules for treatment
of particles and/or if meteorological fore-
cast models are attached to UNI-DEM in
Spatial resolution
(horizontal)
2-D option
(1-layer)
3-D option
(10 layers)
96x96 grid 1.41 12.39
288x288 grid 19.56 152.72
480x480 grid 98.72 856.75
Table 3. The computers available at the SUN grid at the Danish Centre for
Scientific Computing
Computer Type Power RAM Processors
Bohr SUN Fire 6800 UltraSparc-III 750 MHz 48 GB 24
Erlang SUN Fire 6800 UltraSparc-III 750 MHz 48 GB 24
Hald SUN Fire 12K UltraSparc-III 750 MHz 144 GB 48
Euler SUN Fire 6800 UltraSparc-III 750 MHz 24 GB 24
Hilbert SUN Fire 6800 UltraSparc-III 750 MHz 36 GB 24
Newton SUN Fire 15K UltraSparc-III 900 MHz 404 GB 72
Параллельное программирование
50
an attempt to improve the reliability of
the results, then the computational work
is increased very considerably. Therefore,
also here computational grids might be
very useful.
Table 6. Computing times in seconds
when 32 processors are used to run both
the OpenMp option and the MPI option
over a time-interval of 1000 steps. The
speed-ups (related to the corresponding
results obtained in the runs on 8 proces-
sors) are given in brackets
8. Comprehensive Studies with UNI-DEM
UNI-DEM has been used in many
comprehensive studies related to air pol-
lution levels in Europe and/or in differ-
ent specified areas in Europe (as, for ex-
ample, Ukraine, Denmark, etc.). The
most important of these studies are
listed below.
• Studying the impact of future
climate changes on the pollution levels in
Europe, Dimov et al. [6].
• Studying the influence of the
biogenic emissions on high ozone levels
in Europe, Geernaert and Zlatev [11].
• Economical estimates of losses
of crops that are caused by high ozone
levels were found in Dimov et al. [7] and
Zlatev et al. [31]. Until now results for
Denmark and Bulgaria have been ob-
tained. However, the same study can be
performed for any other country in
Europe.
• Long-term variations of the pol-
lution levels in Europe, Ambelas Skjoth
et al. [3], Bastrup-Birk et al. [4], Havasi
and Zlatev [15], Zlatev et al. [30].
• Studying the relationship be-
tween the human-made emissions and
the pollution levels in Europe, Ambelas
Skjøth et al. [3], Bastrup-Birk et al. [4],
Dimov et al. [7], Zlatev et al. [30].
• Studying the frequency of ap-
pearance of pollution levels that exceed
the established by EU critical levels [2]
(and, thus, might lead to damages on
plants, animals and human health),
Geernaert and Zlatev [11], Havasi and
Zlatev [15].
• Studying the contribution of
emission sources from any given area in
Physical processes
(and total time)
OpenMP
option
MPI
option
Advection-diffusion 1001 (2.62) 120
(3.81)
Chemistry-emission-
deposition
607 (2.21) 171
(4.02)
Total computing
time
1771 (2.26) 348
(3.68)
Figure 1. Distribution of the nitrogen dioxide concentrations in Europe
(the left-hand-side plot) and in Ukraine (the right-hand-side plot)
Параллельное программирование
51
Europe to its neighbouring areas (for ex-
ample, studying the contribution from the
European emission sources outside Den-
mark to the pollution levels in Denmark),
Ambelas Skjoth et al. [3], Dimov et al.
[7], Havasi and Zlatev [15], Zlatev [28]
and Zlatev et al. [30].
9. Some Results from the Studies
The distribution of nitrogen diox-
ide in Europe is shown, as an example, in
the left-hand-side plot of Figure 1. From
this plot it is seen that the levels of the
2NO concentrations in Ukraine are con-
siderably lower than the levels in the
most polluted parts of Europe.
The corresponding distribution in
Ukraine is given in the right-hand-side
plot of Figure 1 (note that the scales in the
two plots of Figure 1 are different). Some
more details about the levels of the 2NO
concentrations in Ukraine can be seen
from this plot. One can see that the levels
of the 2NO concentrations are rather high
in the area around Kiev (this should be
expected around such a big city).
NH3 + NH4 CONCENTRATIONS
IN THE PERIOD FROM 1989 TO 1998
CHANGES (RELATIVE TO 1989) IN PERCENT
TANGE CALC:
KELDSNOR CALC:
ANHOLT CALC:
The whole of Denmark:
OBS:
OBS:
OBS:
1989 1990 1991 1992 1993 1994 1995 1996 1997 1998
YEAR
50
70
90
110
C
O
N
C
E
N
T
R
A
T
IO
N
R
A
T
IO
S
Figure 2. Demonstration of the decreasing
trend for the Danish ammonia-ammonium
concentrations
Some trends of the Danish ammo-
nia-ammonium levels in the ten-year pe-
riod 1989—1998 are shown in Fig. 2. The
reduction of the ammonia-ammonium
concentrations is typical not only for
Denmark but also for many other coun-
tries in Europe in the same time period.
Similar reductions of the concentration
levels can be seen also for other species.
The reductions of the concentrations are
due to reductions of the emissions.
Acknowledgements
All runs presented in this paper
were performed on the computers of the
Danish Centre for Scientific Computing
(DCSC) under grants: CPU-1002-27 and
CPU-1101-17. The authors should like to
thank DCSC very much for giving us ac-
cess to several parallel computers.
The work of the first author was
supported by the European Commission
under project "Centre of Excellence
BIS-21" (contract number ICA1-CT-2000-
70016) and by the Ministry of Education
and Science of Bulgaria under grand
I-811.
The work of the second author
was supported partly by the National
Scientific Research Foundation (OTKA)
N. T043765.
1. Numerical integration of chemical ODE prob-
lems arising in air pollution models / V. Ale-
xandrov, A. Sameh, Y. Siddique and Z. Zlatev
// Environmental Modeling and Assessment.
– 1997. – 2. – P. 365—377.
2. Directive 2002/3/EC of the European Parlia-
ment relating to ozone in ambient air // Offi-
cial J. of the European Communities. – 2002.
– L67. – P. 14—30.
3. Studying variations of pollution levels in a
given region of Europe during a long time-
period / C. Ambelas Skjoth, A. Bastrup-Birk,
J. Brandt, Z. Zlatev // Systems Analysis Model-
ling Simulation. – 2000. – 37. – P. 297—311.
4. Studying cumulative ozone exposures in
Europe during a seven-year period / A. Bas-
trup-Birk, J. Brandt, I. Uria, Z. Zlatev // Jour-
nal of Geophysical Research. – 1997. – 102.
– P. 23917—23935.
5. Commutativity of the operators in splitting
methods for air pollution models / I. Dimov, I.
Farago, Á. Havasi, Z. Zlatev // Annal. Univ.
Sci. Budapest. – 2001. – 44. – P. 129—150.
6. Dimov I., Geernaert G., Zlatev Z. Influence of
future climate changes in Europe on ozone
critical levels // Nordic Meteorological Meet-
ing 23, 2002 (H. E. Joergensen, ed.). –
http://www.dams.dk/nmm2002/proceedings.htm.
7. Economical estimations of the losses of crops
due to high ozone levels / I. Dimov, Tz. Os-
tromsky, I. Tzvetanov and Z. Zlatev // Agri-
cultural Economics and Management. –
1999. – 5. – P. 48—52.
Параллельное программирование
52
8. Doroshenko A. Mathematical Models and Methods
for Organization of High Performance Parallel Com-
putation. — Kiev: Naukova dumka, 2000. — 177 p.
9. Doroshenko A., Godlevsky A.B., Hluchy L. Models
and Parallel Programming Abstractions for Synchro-
nization and Communication // Computers and Artifi-
cial Intelligence. — 1999. — 18, N 4. — P. 361–381.
10. Doroshenko A. Modeling synchronization and com-
munication abstraction for dynamical paralelization //
HPCN’97: Proc. Int. Conf. High-Performance Com-
puting and Networking, Vienn, Austria, Apr.1997. —
Lect. Not. in Comput. Sci. — 1997. — Vol. 1225. —
P. 752–761.
11. Ebel A. Chemical Transfer and Transport
Modelling // Transport and Chemical Trans-
formation of Pollutants in the Troposphere /
Eds. P. Borrell, P.M. Borrell. – Berlin:
Springer, 2000. – P. 85—128.
12. Geernaert G., Zlatev Z. Testing the influence
of the biogenic emissions on high ozone lev-
els in Denmark and in Europe // Proceedings
from the EUROTRAC-2 Symposium 2002 /
Eds. P. M. Midgley and M. Reuther, Weiker-
sheim: Margraf Verlag. – 2002.
13. Georgiev K., Zlatev Z. Parallel sparse matrix
algorithms for air pollution models // Parallel
and Distributed Computing Practices. –
1999. – 2. – P. 429—442.
14. Gropp W., Lusk E., Skjellum A. Using MPI:
portable programming with the message pass-
ing interface. – Cambridge, Massachusetts:
MIT Press, 1994.
15. Havasi Á., Zlatev Z. Trends in Hungarian air
pollution levels on a long time-scale // At-
mospheric Environment. – 2002. – 36. –
P. 4145—4156.
16. Lambert J. D. Numerical methods for ordinary
differential equations. – New York: Willey,
1991.
17. Lanser D., Verwer J.G. Analysis of operator
splitting in advection-diffusion-reaction prob-
lems in air pollution modeling // J. of Com-
putational and Applied Mathematics. – 1999.
– 111. – P. 201—216.
18. G.I. Marchuk: Some application of splitting-up
methods to the solution of mathematical
physics problems. Applik. Mat., Vol. 13
(1968), 46—75.
19. Marchuk G.I. Mathematical modelling for the
problem of the environment // Studies in
Mathematics and Applications: North-Hol-
land, Amsterdam, 1985. – 16.
20. McRae G.J., Goodin W.R., Seinfeld J.H. Nu-
merical solution of the atmospheric diffusion
equations for chemically reacting flows // J.
of Computational Physics. – 1982. – 45. –
P. 1—42.
21. Owczarz W., Zlatev Z. Running a large air pol-
lution model on an IBM SMP computer // In-
tern. J. of Computer Research. – 2001. – 10.
– P. – 321—330.
22. Owczarz W., Zlatev Z. Parallel matrix compu-
tations in air pollution modeling // Parallel
Computing. – 2002. – 28. – P. – 355—368.
23. Pepper D.W., Kern C.D., Long P.E. Modelling
the dispersion of atmospheric pollution using
cubic splines and chapeau functions // At-
mospheric Environment. – 1979. – 13. –
P. 223—237.
24. Samarskii A.A. Theory of difference schemes.
– Moscow: Nauka, 1977.
25. Strang G. On the construction and comparison
of difference schemes // SIAM J. on Numeri-
cal Analysis. – 1968. – 5. – P. 506—517.
26. WEB-site for OpenMP tools, http://www.-
openmp.org, 1999.
27. Zlatev Z. Mathematical model for studying
the sulphur pollution over Europe // J. of
Computational and Applied Mathematics. –
1985. – 12. P. 651—666.
28. Zlatev Z. Computer treatment of large air pol-
lution models. – Dordrecht-Boston-London:
Kluwer Academic Publishers, 1995.
29. Zlatev Z. Parallel algorithms for the chemical
part of large air pollution models. Mathemati-
cal Modelling. – 2001. – 13. – P. 22—32.
30. Zlatev Z., Berkowicz R., Prahm L.P. Three di-
mensional advection-diffusion modelling for
regional scale. Atmospheric Environment. –
1983. – 17. – P. 491—499.
31. Calculating losses of crops in Denmark
caused by high ozone levels / Z. Zlatev, I.
Dimov, Tz. Ostromsky, G. Geernaert, I. Tzve-
tanov, A. Bastrup-Birk // Environmental
Modelling and Assessment. – 2001. – 6. –
P. 35—55.
Date received 18.08.03
About authors
Dr.Sc. Ivan Dimov
director of Central Laboratory for Parallel
Processing
Bulgarian Academy of Sciences, Sofia, Bulgaria
E-mail: ivdimov@bas.bg
C.Sc. Istvan Farago
ass. professor of Department of Applied
Analysis
Eotvos Lorand University, Budapest, Hungary
E-mail: faragois@cs.elte.hu
Dr.Sc. Zlatko Zlatev
senior researacher of National Environ-
mental Research Laboratory
Roskilde, Denmark
E-mail: ZZ@DMU.dk
|