 Cumulative frequency analysis

Cumulative frequency analysis is the applcation of estimation theory to exceedance probability (or equivalently to its complement). The complement, the nonexceedance probability concerns the frequency of occurrence of values of a phenomenon staying below a reference value. The phenomenon may be time or space dependent. Cumulative frequency is also called frequency of non−exceedance.
Cumulative frequency analysis is done to obtain insight into how often a certain phenomenon (feature) is below a certain value. This may help in describing or explaining a situation in which the phenomenon is involved, or in planning interventions, for example in flood protection.^{[1]}
Contents
Principles
Definition
Frequency analysis ^{[2]} applies to a record of observed data on a variable phenomenon. The record may be time dependent (e.g. rainfall measured in one spot) or space dependent (e.g. crop yields in an area) or otherwise.
The variable phenomenon is modelled by a random variable X and the parameter p_{C} of interest is the nonexceeding probability:
in which X_{r} is a given reference value.
The record of observed data consists of a random sample taken from X.
In this context the number of times the reference value X_{r} is not exceeded in the sample is called the cumulative frequency M_{Xr}. The relative cumulative frequency is this number divided by the sample size N and is sometimes denoted as Fc(X≤Xr), or briefly as Fc(Xr). It is calculatd as;
Briefly the expression can be noted as: F_{C} = M / N.
When Xr=Xmax, where Xmax is the maximum value observed, it is found that Fc=1, because M=N.
Estimation
The parameter p_{C} is estimated on the basis of the relative cumulative frequency Fc. The estimate is:
In brief:
 Fc = M / (N+1)
The denominator N+1 is introduced instead of N to create the possibility for X to be greater than Xmax as now Pc(Xmax) is less than 1. There exist also other proposals for the denominator (see plotting positions), but they are claimed to be incorrect.^{[3]}
Fitting of probability distributions
Continuous distributions
To present the cumulative frequency distribution as a continuous mathematical equation instead of a discrete set of data, one may try to fit the cumulative frequency distribution to a known cumulative probability distribution,.^{[2]}^{[4]}
If successful, the known equation is enough to report the frequency distribution and a table of data will not be required. Further, the equation helps interpolation and extrapolation.However, care should be taken with extrapolating a cumulative frequency distribution, because this may be a source of errors. One possible error is that the frequency distribution does not follow the selected probability distribution any more beyond the range of the observed data.Any equation that gives the value 1 when integrated from a lower limit to an upper limit agreeing well with the data range, can be used as a probability distribution for fitting. A sample of probability distributions that may be used can be found in probability distributions.
Probability distributions can be fitted by several methods,^{[2]} for example:
 the parametric method, determining the parameters like mean and standard deviation from the X data using the method of moments, the maximum likelihood method and the method of probability weighted moments.
 the regression method, linearizing the probability distribution through transformation and determining the parameters from a linear regression of the transformed Pc (obtained from ranking) on the transformed X data.
Application of both types of methods using for example the
 the normal distribution, the lognormal distribution, the logistic distribution, the loglogistic distribution, the exponential distribution, the Fréchet distribution, the Gumbel distribution, the Pareto distribution, the Weibull distribution and other
often shows that a number of distributions fit the data well and do not yield significantly different results, while the differences between them may be small compared to the width of the confidence interval.^{[2]} This illustrates that it may be difficult to determine which distribution gives better results.
Discontinuous distributions
Sometimes it is possible to fit one type of probability distribution to the lower part of the data range and another type to the higher part, separated by a breakpoint, whereby the overall fit is improved.
The figure gives an example of a useful introduction of a such a discontinuous distribution for rainfall data in northern Peru, where the climate is subject to the behavior Pacific ocean current El Niño. When the Niño extends to the south of Ecuador and enters the ocean along the coast of Peru, the climate in Northern Peru becomes tropical and wet. When the Niño does not reach Peru, the climate is semi−arid. For this reason, the higher rainfalls follow a different frequency distribution than the lower rainfalls.^{[5]}
Prediction
Uncertainty
When a cumulative frequency distribution is derived from a record of data, it can be questioned if it can be used for predictions. For example, given a distribution of river discharges for the years 1950 to 2000, can this distribution be used to predict how often a certain river discharge will be exceeded in the years 2000 to 2050? The answer is yes, provided that the environmental conditions do not change. If the environmental conditions do change, such as alterations in the infrastructure of the river's watershed or in the rainfall pattern due to climatic changes, the prediction on the basis of the historical record is subject to a systematic error. Even when there is no systematic error, there may be a random error, because by chance the observed discharges during 1950 − 2000 may have been higher or lower than normal, while on the other hand the discharges from 2000 to 2050 may by chance be lower or higher than normal.
Confidence intervals
Probability theory can help to estimate the range in which the random error may be. In the case of cumulative frequency there are only two possibilities: a certain reference value X is exceeded or it is not exceeded. The sum of exceedance frequency and cumulative frequency is 1 or 100%. Therefore the binomial distribution can be used in estimating the range of the random error.
According to the normal theory, the binomial distribution can be approximated and for large N standard deviation Sd can be calculated as follows:
 Sd =√{Pc(1−Pc)/N}
where Pc is the cumulative probability and N is the number of data. It is seen that the standard deviation Sd reduces at an increasing number of observations N.
The determination of the confidence interval of Pc makes use of Student's ttest (t). The value of t depends on the number of data and the confidence level of the estimate of the confidence interval. Then, the lower (L) and upper (U) confidence limits of Pc in a symmetrical distribution are found from:
 L = Pc − t.Sd.
 U = Pc + t.Sd
This is known as Wald interval^{[6]} . However, the binomial distribution is only symmetrical around the mean when Pc = 0.5, but it becomes asymmetrical and more and more skew when Pc approaches 0 or 1. Therefore, by approximation, Pc and 1−Pc can be used as weight factors in the assignation of t.Sd to L and U :
 L = Pc − 2Pc.t.Sd
 U = Pc + 2(1−Pc).t.Sd
where it can be seen that these expressions for Pc = 0.5 are the same as the previous ones.
Example
N = 25, Pc = 0.8, Sd = 0.08, confidence level is 90%, t = 1.71, L = 0.70, U = 0.85
Thus, with 90% confidence, it is found that 0.70 < Pc < 0.85
Still, there is 10% chance that Pc < 0.70, or Pc > 0.85Wald interval is known to perform poorly ^{[7]} ^{[8]}^{[9]}.
Wilson score interval^{[10]} provides confidence interval for binomial distributions based on score tests and has better sample coverage, see ^{[11]} and binomial proportion confidence interval for a more detailed overview.
Return period
The cumulative probability Pc can also be called probability of non−exceedance. The probability of exceedance Pe is found from:
 Pe = 1 − Pc.
The return period T defined as:
 T = 1/Pe
and indicates the expected number of observations that have to be done again to find the value of the variable in study greater than the value used for T.
The upper (T_{U}) and lower (T_{L}) confidence limits of return periods can be found respectively as: T_{U} = 1/(1−U)
 T_{L} = 1/(1−L)
For extreme values of the variable in study, U is close to 1 and small changes in U originate large changes in T_{U}. Hence, the estimated return period of extreme values is subject to a large random error. Moreover, the confidence intervals found hold for a long term prediction. For predictions at a shorter run, the confidence intervals U−L and T_{U}−T_{L} may actually be wider. Together with the limited certainty (less than 100%) used in the t−test, this explains why, for example, a 100−year rainfall might occur twice in 10 years.
The strict notion of return period actually has a meaning only when it concerns a time−dependent phenomenon, like point rainfall. The return period then corresponds to the expected waiting time until the exceedance occurs again. The return period has the same dimension as the length of time for which each observation is representative. For example, when the observations concern daily rainfalls, the return period is expressed in days, and for yearly rainfalls it is in years.
Need of confidence belts
The figure shows the variation that may occur when obtaining samples of a variate that follows a certain probability distribution. The data were provided by Benson ^{[1]}.
The confidence belt around an experimental cumulative frequency or return period curve gives an impression of the region in which the true distribution may be found.
Also, it clarifies that the experimentally found best fitting probability distribution may deviate from the true distribution.
Histogram
The observed data can be arranged in classes or groups with serial number k. Each group has a lower limit (L_{k}) and an upper limit (U_{k}). When the class (k) contains m_{k} data and the total number of data is N, then the relative class or group frequency is found from:
 Fg(L_{k}<X≤U_{k}) = m_{k} / N
or briefly:
 Fg_{k} = m/N
or in percentage:
 Fg(%) = 100m/N
The presentation of all class frequencies gives a frequency distribution or histogram. Histograms, even when made from the same record, are different for different class limits.
The histogram can also be derived from the fitted cumulative probability distribution:
 Pg_{k} = Pc(U_{k}) − Pc(L_{k})
There may be a difference between Fg_{k} and Pg_{k} due to the deviations of the observed data from the fitted distribution (see figure).
See also
References
 ^ ^{a} ^{b} Benson, M.A. 1960. Characteristics of frequency curves based on a theoretical 1000 year record. In: T.Dalrymple (ed.), Flood frequency analysis. U.S. Geological Survey Water Supply paper 1543−A, pp. 51−71
 ^ ^{a} ^{b} ^{c} ^{d} Frequency and Regression Analysis. Chapter 6 in: H.P.Ritzema (ed., 1994), Drainage Principles and Applications, Publ. 16, pp. 175−224, International Institute for Land Reclamation and Improvement (ILRI), Wageningen, The Netherlands. ISBN 90 70754 3 39 . Free download from the from webpage [1] under nr. 12, or directly as PDF : [2]
 ^ Makkonen, L. 2008. Communications in Statistics − Theory and Methods, 37: 460−467
 ^ David Vose, Fitting distributions to data
 ^ CumFreq, a program for cumulative frequency analysis with confidence bands, return periods, and a discontinuity option. Free download from : [3]
 ^ Wald, A.; J. Wolfowitz (1939). "Confidence limits for continuous distribution functions". The Annals of Mathematical Statistics: 105–118.
 ^ Ghosh, B.K (1979). "A comparison of some approximate confidence intervals for the binomial parameter". Journal of the American Statistical Association: 894–900.
 ^ Blyth, C.R.; H.A. Still (1983). "Binomial confidence intervals". Journal of the American Statistical Association: 108–116.
 ^ Agresti, A.; B. Caffo (2000). "Simple and effective confidence intervals for pro portions and differences of proportions result from adding two successes and two failures.". The American Statistician: 280–288.
 ^ Wilson, E.B. (1927). "Probable inference, the law of succession, and statistical inference". Journal of the American Statistical Association: 209–212.
 ^ Hogg, R.V. (2001). Probability and statistical inference (6th ed.). Prentice Hall, NJ: Upper Saddle River.
Categories: Data analysis
 Statistical models
 Probability distributions
Wikimedia Foundation. 2010.
Look at other dictionaries:
Cumulative frequency — may refer to: Cumulative distribution function Cumulative frequency analysis This disambiguation page lists articles associated with the same title. If an internal link led you here, you may … Wikipedia
Cumulative distribution function — for the normal distributions in the image below … Wikipedia
Frequency — For other uses, see Frequency (disambiguation). Three cyclically flashing lights, from lowest frequency (top) to highest frequency (bottom). f is the frequency in hertz (Hz), meaning the number of cycles per second. T is the period in seconds (s) … Wikipedia
Frequency distribution — In statistics, a frequency distribution is an arrangement of the values that one or more variables take in a sample. Each entry in the table contains the frequency or count of the occurrences of values within a particular group or interval, and… … Wikipedia
Cumulative incidence — or incidence proportion is a measure of frequency, as in epidemiology, where it is a measure of disease frequency during a period of time. Where the period of time considered is an entire lifetime, the incidence proportion is called lifetime risk … Wikipedia
Survival analysis — is a branch of statistics which deals with death in biological organisms and failure in mechanical systems. This topic is called reliability theory or reliability analysis in engineering, and duration analysis or duration modeling in economics or … Wikipedia
Principal component analysis — PCA of a multivariate Gaussian distribution centered at (1,3) with a standard deviation of 3 in roughly the (0.878, 0.478) direction and of 1 in the orthogonal direction. The vectors shown are the eigenvectors of the covariance matrix scaled by… … Wikipedia
Medical analysis of circumcision — Numerous medical studies have examined the effects of male circumcision with mixed opinions regarding the benefits and risks of the procedure. Opponents of circumcision say it is medically unnecessary, is unethical when performed on newborns, is… … Wikipedia
Pareto analysis — is a statistical technique in decision making that is used for selection of a limited number of tasks that produce significant overall effect. It uses the Pareto principle the idea that by doing 20% of work you can generate 80% of the advantage… … Wikipedia
Intelligence analysis — This article deals with the intellectual process of analysis itself, as opposed to intelligence analysis management, which, in turn, is a subcomponent of intelligence cycle management. For a complete hierarchical list of articles in this series,… … Wikipedia