Frequency Analysis - An Overview - ScienceDirect Topics
Frequency Analysis - An Overview - ScienceDirect Topics
Frequency Analysis
Bivariate drought frequency analysis is a technique of copula-based
probabilistic analysis that connects multivariate probability
distributions to their one-dimensional (1D) marginal probability
distribution while still capturing the essential features of dependence
and correlation among the random variables (Sadri and Burn, 2012).
From: Extreme Hydrology and Climate Variability, 2019
Related terms:
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 1/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
observations themselves. The Pearson type III distribution is particularly useful for
hydrologic investigations because the third parameter, the skew, permits fitting
nonnormal samples to the distribution. When the skew is 0, the log-Pearson type
III distribution becomes a two-parameter distribution that is identical to the
lognormal distribution (U.S. Army Corps of Engineering, 1993).
Normal distribution is a symmetrical, bell-shaped, continuous distribution,
theoretically representing the distribution of accidental errors about their mean,
the Gaussian low of errors. Lognormal or Galton distribution is a transformed
normal distribution in which the variable is replaced by its logarithmic value. It is a
nonsymmetrical distribution, very convenient for the frequency analysis of many
hydrological parameters. The theory of extreme values considers the distribution of
the largest or smallest observations occurring in each group of repeated samples.
Gumbel (1941) was the first to employ extreme value theory for analysis of flood
frequency. Chow (1954) shows that Gumbel distribution is essentially a lognormal
distribution with constant skewness.
There are few commonly used methods and approaches of testing the goodness of
fit of time series data to postulated theoretical probability distributions; for
example, the (1) χ2 (chi-squared) test, (2) Kolmogorov-Smirnov test, (3) probability
plot correlation coefficient test (Adeloye and Montaseri, 2002), and (4) L-moment
goodness of fit test (Tallaksen et al., 2004).
Figure 4-29 represents graphical presentation of three probability distribution
curves (normal, lognormal, and Gumbel), defined for a measured maximum
annual discharge time series of the Kupari gauging station over the period 1951–
2007. Table 4-4 gives the values of maximum annual discharges for different
recurrence periods using the three previously mentioned distributions.
Table 4-4. Normal, Lognormal, and Gumbel Values, Defined for the Measured
Maximum Annual Discharge Time Series of the Kupari Gauging Station over the
Period 1951–2007
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 2/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
The zero flow values (dry spring) that occurred at some intermittent springs pose a
special problem for the frequency analysis of a time series of minimum annual
discharges. The problem can be solved using the theorem of total probability. The
following equation serves for calculating the probability of a discharge less or equal
to discharge Qi:
where F*(Qi) is a probability distribution function for the nonzero discharges, the
equation for which is
1
−1+𝑘 (4.36)
𝐹∗ (𝑄𝑖 ) = 𝑇 𝑘
𝑘 = 𝑁∗ / 𝑁 (4.37)
where T is the return period in year, N is the number of data (years) in time series,
N* is the number of data (years) in a time series with nonzero values.
The applicability of equation (4.36) depends on getting positive values of the
probability F*(Qi), which means that
𝑘 ≥ (𝑇 − 1) / 𝑇 (4.38)
𝑇(0) = 𝑁 / (𝑁 − 𝑁∗ ) (4.39)
Introduction
Fateh Chebana, in
Multivariate Frequency Analysis of Hydro-Meteorological Variables, 2023
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 3/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
function. Fig. 1.1 illustrates some of the aspects of multivariate HFA highlighted in
the literature and their importance and the volume of studies focused on this. It
shows that the modeling step, especially based on copulas, is most prevalent the
literature. Although copula and modeling are important and essential, many other
aspects and steps also need to be considered to perform a complete and
appropriate analysis. Multivariate HFA associated literature as papers and reports is
not easily accessible to practitioners and students. Hence, this has led to an
increasing gap between research and practice in this field. Therefore, the desperate
need of a reference book where the reader can find all the relevant material
covering the different steps and situations of a multivariate HFA in a simplified and
accessible presentation, the connections between them as well as a complete
overview of all steps of the analysis has been felt.
Palaeoflood Hydrology
Gerardo Benito, Andrés Díez-Herrero, in
Hydro-Meteorological Hazards, Risks and Disasters, 2015
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 5/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
able to describe the temporal variation of statistical parameters (mean, variance) in
probability distribution functions (Gumbel, Lognormal, Weibull, Gamma). The
statistical parameters may show increasing/decreasing trends that can be modelled
using time as covariate (characterizing the trend or as a smooth function via cubic
splines; Villarini et al., 2009), or they can be related to hydroclimatic covariates such
as climatic indices that reflect low-frequency climatic variability (e.g., Pacific
Decadal Oscillation, North Atlantic Oscillation, Arctic Oscillation; López and
Francés, 2013). The application of these nonstationary models to palaeoflood
hydrology requires a characterization of the occurrence rate (covariate) during the
recorded period.
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 6/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
descriptive and informative physiographic and climatic catchment descriptors to be
used as proxies for the flood frequency regime (see, e.g., Castellarin et al., 2001;
Merz and Blöschl, 2005) and (2) the development of pooling procedures as
objective and nonsupervised as possible. Several objective approaches have been
proposed by the scientific literature, such as cluster analysis (Burn, 1989) or
unsupervised artificial neural networks (ANNs) (see, e.g., Hall and Minns, 1999;
Toth, 2009). Furthermore, the scientific community is dedicating an increasing
attention to the possibilities offered by the application of geostatistical techniques
to the problem of statistical regionalisation. These techniques have been showed to
have a significant potential for regionalization and, for this reason, will be briefly
discussed and presented in this section.
Geostatistical procedures were originally developed for the spatial interpolation of
point data (see, e.g., kriging: Kitanidis, 1997). The literature proposes two different
ways to apply geostatistics to the problem of regionalisation of hydrological
information. The first technique is called physiographic space-based interpolation
(PSBI) and performs the spatial interpolation of the desired hydrometric variable
(e.g., T-year flood, but also annual streamflow, peak flow with a certain return
period, low flows, etc.) in the bidimensional space of geomorphoclimatic
descriptors (Chokmani and Ouarda, 2004; Castiglioni et al., 2009). The x and y
orthogonal coordinates of the bidimensional space are derived from an adequate
set of n>1 geomorphologic and climatic descriptors of the river basin (such as
drainage area, main channel length, mean annual precipitation, and indicators of
seasonality; see Castellarin et al., 2001) through the application of multivariate
techniques, such as the principal components or canonical correlation analysis (Shu
and Ouarda, 2007). The second technique, named Topological kriging or
Topkriging, is a spatial estimation method for streamflow-related variables. It
interpolates the streamflow value of interest (i.e., T-year flood, low-flow indices,
etc.) along the stream network by taking the area and the nested nature of
catchments into account (Skøien et al., 2006; Skøien and Blöschl, 2007).
The philosophy behind these innovative approaches to regionalization is rather
interesting because they enable one to regionalize hydrometric variables
dispensing with the definition or identification of homogeneous regions or
pooling groups of sites (see Figure 13). The approaches are particularly appealing
for predictions in ungauged basins as they provide a continuous representation of
the quantity of interest (e.g., T-year flood) along the stream network (Topkriging) or
in the physiographic space (PSBI), providing the user with an estimate of the
uncertainty associated with the interpolated value. In particular, the final output for
Topkriging is the estimation of the measure of interest (with uncertainty) along the
stream network (see Figure 16). A little less intuitive is the output of PSBI. With this
technique any given basin (gauged or ungauged) can be represented as a point in
the x–y space described above; in the same way the set of gauged basins of the
study area can be represented by a cloud of points in this space. The empirical
values of the quantity of interest (e.g., T-year flood) can be represented along the
third dimension z for each gauged catchment, and can then be spatially
interpolated (with uncertainty) by applying a standard interpolation algorithm (e.g.,
ordinary or universal kriging). The spatial interpolation enables one to represent
the quantity of interest over the entire portion of the x–y space containing
empirical data, and therefore to estimate it at ungauged sites lying within the same
portion of the space (see Figure 16).
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 7/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
Figure 16. (a) Topkriging: 100-year flood per unit area (color codes in m3 s−1 km−2)
for a portion of the Mur region (Austria): Topkriging estimates along the stream
network and empirical values as circles. From Skøien JO, Merz R and Bloschl G
(2006) Top-kriging - geostatistics on stream networks. Hydrology and Earth System
Sciences 10(2): 277–287, Fig. 7. (b) PSBI: 3D representation of standardised value of
100-year flood over the physiographic space identified for a set of basins in
northern central Italy (gauged basins are represented as dots).
30.1 Introduction
Flood frequency analysis is a technique commonly used to relate the magnitude of
extreme runoff or river flow events to their frequency of occurrence through the
use of probability distribution functions. Historical records provide essential
information to predict the recurrence interval of hydrological extremes. Such
information on high flow events is crucial in many cases; dike design for flood
protection; reservoir design and management, river pollution control, and in
ecology management (Engeland et al., 2005). Data observed over an extended
period of time at river gaging stations are main input in flood frequency analysis.
Assumption is made that flood data are space and time independent, identically
distributed, and uninfluenced by natural or man-made changes in the hydrological
system (Rao and Hamed, 2000). This analysis estimates the recurrence interval of
extreme events by extrapolating beyond the length of the record. Obviously, the
longer the available data the more accurate the estimations are.
It is likely that sufficient information can be extracted from the “at-site” analysis for
sites with available data. Nevertheless, it is a challenge to estimate floods for
ungaged sites or sites with short time series, which influences the reliability of
estimation. In such cases, regionalization techniques assist in estimating flows of
required return periods at ungaged locations of interest. The primary goal of
regional frequency analysis is to investigate typical observations of the same
variable at numerous measuring sites within a suitably defined “region” (Hosking
and Wallis, 1997). This leads to more accurate conclusions when analyzing all data
samples rather than using only a single sample. The final result of the regional
analysis is a regional curve, which permits flood frequency and quantile
estimations at any location along the basin’s networks. This principle is applied to
the Baro-Akobo River basin to support the basin’s water management decisions.
The regionalization concept applicable to all drainage basins within a
homogeneous region was first introduced by Dalrymple (1960), which “trades
space for time,” by using data from nearby or similar sites to estimate quantiles of
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 8/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
the underlying variable at each site in the homogenous region of consideration
(Stedinger et al., 1993). The concept was continuously developed since, and new
approaches were regularly developed by researchers (Benson, 1962; Matalas and
Gilroy, 1968; Vicens et al., 1975; NERC, 1975; Greiss and Wood, 1981; Rossi et al.,
1984; Hosking et al., 1985; Lettenmaier et al., 1987; Burn, 1990; Stedinger and Lu,
1995; Hosking and Wallis, 1997; Reed et al., 1999; Sveinsson et al., 2001).
One of the most important and challenging steps in regional flood frequency
analysis is the delineation of homogeneous regions. There are no universally
accepted objective methods for such delineation. This is due to the complexity of
factors that affect the generation of floods. To mention some examples, in the UK
flood studies report the British Isles were divided into 11 regions on the basis of
general knowledge of hydrological regime (NERC, 1975). Matalas et al. (1975)
divided the United States into 14 geographical regions, based on the variability of
skewness of instantaneous flood peak data. In New Zealand, Mosley (1981) used
cluster analysis. In Pennsylvania, White (1975) used factor analysis of a collection of
physically similar basins. Acreman and Sinclair (1986) used cluster analysis on a
matrix of basin characteristics to identify similar basins in Scotland. Wiltshire
(1986) used an iterative search procedure to optimally divide the basin
characteristic data space for flood regionalization and also used multivariate
discriminant analysis to form groups for British catchments. A number of other
regionalization techniques were also developed for objective determination of
homogeneous regions. In this study, regionalization of the basin is based on Q-Q
plot method for identification of groups of stations with similar distribution tail
behavior and developed regression equations between the mean annual maximum
flow (MAF) of the catchments and their corresponding physical characteristics.
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 9/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 10/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 11/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
Table 1. Classification of flooded area according to land use for 5, 10, 50, and
100 years return period.
Return
period 5 Years 10 Years
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 12/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
Figures depict the generated river flood hazard maps based on water depth, and
the intersection of land use with flood area boundaries for each flood event
simulation. As shown in Figs. 11–14.
Fig. 11. Floodplain coverage extent on the land use for 5-year return period.
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 13/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
Fig. 12. Floodplain coverage extent on the land use for 10-year return period.
Fig. 13. Floodplain coverage extent on the land use for 50-year return period.
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 14/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
Fig. 14. Floodplain coverage extent on the land use for 100-year return period.
The flood inundation maps indicates a high risk to the land use and Wadi with
considerable water depth.
Wadi flood hazard mapping and management water depth distribution is more
essential in order to determine flooded areas. Floodplain inundation map for
various return periods (5, 10, 50, and 100-year) of the study area can be mapped
and overlay on Land use at Bou Saâda Wadi—Subbasin to analyze of damages
caused to Built-Up area, vegetation.
The approximation of a flood-prone area on a map is shown in Figs. 11–14:
As we have seen in the results, gradually varied steady flow is characterized by
minor changes in water depth from one cross-section to another.
The flooded areas along the Bou Saâda Wadi—Subbasin are 230.76 km2,
239.31 km2, 254.52 km2, and 265.77 km2 for 5, 10, 50, and 100 year return
periods, respectively. The roads affected by floods have considerably affected the
trade exchanges.
On the other hand, the inundation depth of the 100 years return period ranged
from 0.000061035 to 5.40 m, the Built-Up areas, and the Vegetation (forest, grass)
were exposed to overflows, and Whereas the impacts on Built Up areas are highly.
So that the flood inundation area within 100 years return period covers 66.54% of
the Built-up area, 33.45% of Vegetation (forest, grass). Similarly, the land use
affected by the 5, 10, 25, and 50 years of flood frequency has different inundation
depths.
People who are living near the river banks affected by catastrophic flooding are
directly exposed to inundation depths of over 5.40 m. And need further
considerations to flood protection.
Flood is usually a result of natural causes, it may also be caused by man-made
factors. The reason lies in urbanization, density can increase risk, with the
development of considered unplanned construction on flood plains, especially in
high-risk areas, without proper infrastructural.
In Bou Saâda Wadi—Subbasin, flood causes can be attributed to:
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 15/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
1. Characteristics of the Bou Saâda Wadi, that are characterized by a gradual
slope, where the floodplain is relatively lowland the hazard of the river flood is
higher than the other locations.
2. Increase in Bou Saâda Subbasin land-use development causes an increase of
imperviousness which leads to an increase in runoff volume.
3. Improper drainage system throughout the Bou Saâda Subbasin shares.
Based on the findings of this study, a variety of mitigation measures can be
identified which will minimize the impact of flooding in Bou Saâda Wadi—
Subbasin:
• Should give solutions:
• Structural measures.
• Nonstructural measures: include land use regulations, e.g., relocate the
population residing along the wadi banks, and Preventing any new residential
structures in the areas with high risk of flood.
• Early warning networks should be constructed in the city for the people to take
the necessary precaution.
• Filed data related to channel roughness coefficient need to be collected which
can help to get better results of the hydraulic model.
Mass-Movement Geomorphology
Simon Loew, ... Werner Gerber, in Treatise on Geomorphology (Second Edition),
2022
𝑉 −𝐵 (1)
𝐹 = 𝐴𝑉0 𝑉
0
Where A is the frequency of rockfalls with a volume bigger than V0 (an activity
parameter) and B is a uniformity coefficient, which reflects the decrease of the
frequency when the volume increases. V0 is the minimal value of the considered
volume range or a minimal volume of interest, which depends on the context of
the analysis. Examples of spatio-temporal frequency distributions are given in
Fig. 9. The retreat rate of the cliff can be estimated by integration of the volume-
frequency relation, which can be extrapolated in order to account for rare large
events (Hantz et al., 2003a, b).
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 16/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 17/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
As initially suggested by Gumbel (1954), AMS is the data set most commonly used
in probabilistic analyses. POT is sometimes preferred when the objective is to
increase the time series length, especially while dealing with heavy-tailed
distributions (Madsen et al., 1997). However, it must be considered that the POT
method introduces discretional terms in the choice of the threshold value, as
clarified by Claps and Laio (2003). CDS advantages are described by Marani and
Zanetti (2015) and can be taken into account when critical high-return period
estimation is the objective. However, when using CDS, more evidence is still
required on the selection of the minimum set of necessary data, given a target
return period (Requena et al., 2019).
To apply analyses on long records, however, one must consider that for older
records only AMS may be available, and this issue may limit the operational
options.
Whatever the available observations are, the definition of the variables to use for
the derivation of the regional dimensionless frequency curve K(T) is not a trivial
task, similarly to what happens in the case of flood peaks. The main alternatives in
literature are to proceed towards regionalization of rainfall quantiles (Svensson and
Jones, 2010) for one duration within the simple scaling assumption (Soltani et al.,
2017) or for different durations under the multiscaling approach (i.e., when Kd(T)
changes with the duration d, Burlando and Rosso, 1996). However, many authors
assume that it is preferable to regionalize the probability distribution parameters
(Svensson and Jones, 2010 and reference therein) or the sample ordinary moments,
or the L-moments (Modarres and Sarhadi, 2012; Ngongondo et al., 2012; Smithers
and Schulze, 2001). In all cases, a thorough understanding of the dependence of
the variance of the regionalized parameters on the record length is necessary,
especially for the estimation of high-return period quantiles.
Our evaluation is that the average practitioner would not easily distinguish the
differences in the outcomes when different variables have been chosen for the
regionalization. Differences can emerge in “critical applications,” where the aim is
the estimation of high-return period design rainfall, or in the evaluation of rainfall
spatial variability in regions with few data and high hydro-climatic variability (e.g.,
mountainous regions). In both cases, however, marked differences can arise from
the intervention of individual (significant) records, even if included in short time
series (Libertino et al., 2018)
12.2.2 Data preparation and data scarcity
As mentioned by CSAGroup (2019), practitioners are now more and more
concerned with data and methods that can emphasize the evolution of rainfall
hazard due to climate change. For this reason, the management of large databases
of rainfall measurements requires to consider and manage the presence of gaps in
the time series, the use of recent but short records, and the possible role of rain
gauges relocation.
In the literature of rainfall extremes two main approaches are applied to deal with
the data fragmentation issue. The first one is draconian and is based on the
definition of a minimum acceptable threshold of record length: in this approach, to
achieve the data homogeneity required to proceed with the RFA, only the time
series longer than a given threshold will be considered in the analysis. Although
this approach can be considered as precautionary, one can end up discarding
important information included in some short records, and this can affect the
reliability of the RFA (Ouali et al., 2016; Libertino et al., 2018).
The second approach tries to preserve the available information in relation to its
influence on each parameter to be estimated (Laio et al., 2011). In essence, even if
short time series cannot provide great influence on higher-order moments, their
data can be confidently used to improve the estimation of the first order
parameters. Considering that this second approach may lead to errors if based on
non-robust assumptions (Teegavarapu and Nayak, 2017) significant research efforts
are being recently addressed on data augmentation techniques, using also ancillary
information. A brief review is reported below.
Most of the methodologies proposed to retain information from short records are
based on interpolation or data reconstruction. For instance, Pappas et al. (2014) use
time series autocorrelation to deal with sporadic time gaps. However, low
autocorrelation or frequent and systematic missing data can make the method not
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 18/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
applicable. Ouali et al. (2016) propose a conditional quantile regression model that,
even if not address a data reconstruction, can be ascribed to the “data
augmentation” category. The “patched kriging” (PK) approach, proposed by
Libertino et al. (2018), reconstructs missing values using contemporary
observations at nearby sites through a sequential application of the ordinary
kriging. Homogeneous annual maxima series at each location can also be obtained
with a bootstrap procedure, that accounts for all the measurements in nearby
stations (Uboldi et al. 2014). In this case, data relevance decreases as the distance
between the rain gauge and the estimation point increases. However, the results of
this application end up being highly sensitive to the presence of outliers in the
observed series, and can produce quantiles that can deviate substantially from
those computed from local samples even of consistent length.
Data augmentation in rainfall frequency analysis from ancillary data are the last
frontier, both for applications to areas affected by data scarcity, and for the increase
in the spatial detail of the statistical analyses. In the first case, some studies
investigate the possibility of evaluating statistics using hydrometeorology attributes
rather than inadequately observed precipitation measurements (Satyanarayana and
Srinivas, 2008). This approach, preliminarily tested in India over the summer
monsoon affected regions, suggests using both large-scale atmospheric variables
that affect the precipitation in the study region, and location attributes as
additional features in the regionalization. Over large areas, satellite data can
certainly provide additional insights in RFA (Qamar et al., 2017) since global-scale
rainfall datasets like those acquired by TRMM and GPM constellations have
homogeneous and reasonably long time series. However, additional work is
required to take enough advantage of this information, considering its low
resolution in space and some inherent inaccuracies (Zorzetto and Marani, 2020).
Data scarcity is, however, a relative concept, as it can depend on the required
details of the spatial rainfall analysis. Urban hydrological applications, for instance,
tend to be data-hungry for the need to reproducing rainfall estimates at high
resolution. In this context, Weather Radar data can provide interesting support to
the regional frequency analyses, as witnessed by recent literature (Goudenhoofdt
et al., 2017; Ochoa-Rodriguez et al., 2019; Kašpar et al., 2021). As seen from the
above applications, new and interesting developments in rainfall frequency analysis
can therefore come from the inclusion of additional information from other
sources, providing advances already seen with the use of data assimilation in
rainfall nowcasting applications (Li et al., 2018).
𝜎
for 𝑥 ≥ ( 𝜇 − ) if 𝑘 > 0 ( Fréchet ) , 𝑥 ∈ ℝ if 𝑘 = 0 ( Gumbel ) and
𝑘
𝜎
𝑥≤ (𝜇 − 𝑘 ) if 𝑘 < 0 ( Weibull ) where 𝜇 ∈ ℝ, 𝜎 > 0 and 𝑘 ∈ ℝ are the
location, scale, and shape parameters, respectively. However, for LN2 and LN3
distributions, it is preferable to present their density functions, respectively, as
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 19/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
( log 𝑥 − 𝜇 )
2 (7.6)
1 -
𝑓𝐿𝑁2 (𝑥; 𝜇, 𝜎) = 𝑥𝜎 𝑒 2𝜎2 for𝑥 > 0
√2𝜋
where 𝜇 ∈ ℝ and 𝜎 > 0 are, respectively, the mean and standard deviation of log𝑥,
whereas 𝑚 ∈ ℝ is a threshold parameter (lower bound).
The considered nonstationary GEV version 𝐺𝐸𝑉𝜇𝜐, 𝜎𝜐, 𝑘 incorporates
nonstationarity in its location and scale parameters by linking them to the covariate
𝜐. As mentioned above, for simplicity and for practical reasons, the shape
parameter k is considered constant. According to the shape of the trend, we
consider the notations GEV00: no trend; GEV10: linear trend in the location
parameter 𝜇𝜐 = 𝜇0 + 𝜇1 𝜐; GEV20: quadratic trend in the location parameter
𝜇𝜐 = 𝜇0 + 𝜇1 𝜐 + 𝜇2 𝜐2 ; and GEV11: linear trend in both the location and scale
parameters 𝜇𝜐 = 𝜇0 + 𝜇1 𝜐, 𝜎𝜐 = exp ( 𝜎0 + 𝜎1 𝜐 ) . As mentioned earlier, to keep
the scale parameter 𝜎 ( 𝜐 ) positive, we use the transformation
𝜑𝜎 ( 𝜐 ) = log ( 𝜎 ( 𝜐 ) ) . Using similar notations, we obtain the models LN200,
LN210, and LN220 in analogy with the location parameter 𝜇, respectively, as in
GEV00, GEV10, and GEV20. For the LN3 case, the models LN300, LN310, LN320, and
LN311 can be considered as the possible trend in threshold parameter m, that is,
𝐿𝑁3𝜇𝜐, 𝜎, 𝑚𝜐.
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 20/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
lack of proper Precambrian apparent polar wander paths (APWPs) hampers the use
of seventh criterion, which states that poles must not have a resemblance to
younger paleomagnetic poles, and therefore only the first six criteria (Q1–6) are
employed here as done in Veikkolainen et al. (2014a).
The methodological proof of inclination frequency analysis rests on the
assumption of random sampling of continents over globe. Rolf and Pesonen (2018)
applied 3D spherical mantle convection models to investigate whether latitudes
have been sampled uniformly by the Earth’s continents in the far past, finding out
that this requirement is not met for a period shorter than 1.3 Gyr. This strengthens
the conclusion of Meert et al. (2003) who used random walk models to prove that
in sampling periods over 600 Myr, inclination distribution resembles that of GAD
only for ~30% of time even if GAD is assumed valid for the entire geologic past.
Originally, Evans (1976) argued that 500–600 Myr is adequate for random
sampling, an obviously outdated hypothesis. A major weakness of the inclination
frequency analysis is also the fact that distributions similar in appearance may
result from different combinations of multipolar fields, that is, various solutions
are not unique. For example, distribution of GAD is similar to the distribution of a
multipolar field with GAD, G2=0.17 and G3=0.03 (Veikkolainen et al., 2014a) This
emphasizes the need for using other independent methods to test the GAD
hypothesis.
Our analysis of PALEOMAGIA was restricted to igneous rocks due to the proven
shallowing of sedimentary magnetization and the cooling-related problems
associated with metamorphic rocks. This does not rule out using sedimentary data
in other analyses, such as studying variations between N and R polarity in certain
period by using magnetostratigraphy and analyzing transitional geomagnetic field
data (see Fig. 3.9A). Our data handling followed a few steps that included removal
of combined polarity (N+R) entries in cases of dual-polarity paleomagnetic data,
removal of certain subentries if they were part of a combined entry with similar or
higher Q1–6 value, and removal of all data with Q1–6 < 3. We also removed entries
that were obviously superseded by new entries. However, to ensure that an
adequate number of data are available for statistical tests, we did not carry out
spatiotemporal binning which, if applied, must be done within each terrane
separately without regard to present-day geography (Veikkolainen et al., 2014c).
Apart from the actual polarity timescale, the significantly larger proportion of
positive inclinations (62%; N=695) compared to negative ones (38%; N=443) may
be a sign of insufficient spatiotemporal sampling. The distinction is partly a result
of the large number (c.200 entries) of data with low positive inclinations from
Baltica and Laurentia in the Nuna supercontinent. The potential removal of steep-
inclination data due to their resemblance to the present-day field (PEF) in Europe
and North America (Lapointe et al., 1978) does not explain the difference, because
PEF in those regions is represented by positive inclination data. The PALEOMAGIA
subset of 635 data entries (Q1–6 ≥ 3) had a mean |I|=42.3 degrees, slightly smaller
than the value of |I|=49.1 degrees expected from GAD, although variations in
different stages of Precambrian are substantial (Fig. 3.10).
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 21/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 22/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 23/24
6/29/23, 7:08 PM Frequency Analysis - an overview | ScienceDirect Topics
https://www.sciencedirect.com/topics/earth-and-planetary-sciences/frequency-analysis 24/24