Loading AI tools
Extensions of the concept of randomness From Wikipedia, the free encyclopedia
The seven states of randomness in probability theory, fractals and risk analysis are extensions of the concept of randomness as modeled by the normal distribution. These seven states were first introduced by Benoît Mandelbrot in his 1997 book Fractals and Scaling in Finance, which applied fractal analysis to the study of risk and randomness.[1] This classification builds upon the three main states of randomness: mild, slow, and wild.
The importance of seven states of randomness classification for mathematical finance is that methods such as Markowitz mean variance portfolio and Black–Scholes model may be invalidated as the tails of the distribution of returns are fattened: the former relies on finite standard deviation (volatility) and stability of correlation, while the latter is constructed upon Brownian motion.
These seven states build on earlier work of Mandelbrot in 1963: "The variations of certain speculative prices"[2] and "New methods in statistical economics"[3] in which he argued that most statistical models approached only a first stage of dealing with indeterminism in science, and that they ignored many aspects of real world turbulence, in particular, most cases of financial modeling.[4][5] This was then presented by Mandelbrot in the International Congress for Logic (1964) in an address titled "The Epistemology of Chance in Certain Newer Sciences"[6]
Intuitively speaking, Mandelbrot argued[6] that the traditional normal distribution does not properly capture empirical and "real world" distributions and there are other forms of randomness that can be used to model extreme changes in risk and randomness. He observed that randomness can become quite "wild" if the requirements regarding finite mean and variance are abandoned. Wild randomness corresponds to situations in which a single observation, or a particular outcome can impact the total in a very disproportionate way.
The classification was formally introduced in his 1997 book Fractals and Scaling in Finance,[1] as a way to bring insight into the three main states of randomness: mild, slow, and wild . Given N addends, portioning concerns the relative contribution of the addends to their sum. By even portioning, Mandelbrot meant that the addends were of same order of magnitude, otherwise he considered the portioning to be concentrated. Given the moment of order q of a random variable, Mandelbrot called the root of degree q of such moment the scale factor (of order q).
The seven states are:
Wild randomness has applications outside financial markets, e.g. it has been used in the analysis of turbulent situations such as wild forest fires.[7]
Using elements of this distinction, in March 2006, a year before the Financial crisis of 2007–2010, and four years before the Flash crash of May 2010, during which the Dow Jones Industrial Average had a 1,000 point intraday swing within minutes,[8] Mandelbrot and Nassim Taleb published an article in the Financial Times arguing that the traditional "bell curves" that have been in use for over a century are inadequate for measuring risk in financial markets, given that such curves disregard the possibility of sharp jumps or discontinuities. Contrasting this approach with the traditional approaches based on random walks, they stated:[9]
We live in a world primarily driven by random jumps, and tools designed for random walks address the wrong problem.
Mandelbrot and Taleb pointed out that although one can assume that the odds of finding a person who is several miles tall are extremely low, similar excessive observations can not be excluded in other areas of application. They argued that while traditional bell curves may provide a satisfactory representation of height and weight in the population, they do not provide a suitable modeling mechanism for market risks or returns, where just ten trading days represent 63 per cent of the returns between 1956 and 2006.[dubious – discuss]
If the probability density of is denoted , then it can be obtained by the double convolution .
When u is known, the conditional probability density of u′ is given by the portioning ratio:
In many important cases, the maximum of occurs near , or near and . Take the logarithm of and write:
Splitting the doubling convolution into three parts gives:
p(u) is short-run concentrated in probability if it is possible to select so that the middle interval of () has the following two properties as u→∞:
Consider the formula , if p(u) is the scaling distribution the integrand is maximum at 0 and ∞, on other cases the integrand may have a sharp global maximum for some value defined by the following equation:
One must also know in the neighborhood of . The function often admits a "Gaussian" approximation given by:
When is well-approximated by a Gaussian density, the bulk of originates in the "q-interval" defined as . The Gaussian q-intervals greatly overlap for all values of . The Gaussian moments are called delocalized. The lognormal's q-intervals are uniformly spaced and their width is independent of q; therefore if the log-normal is sufficiently skew, the q-interval and (q + 1)-interval do not overlap. The lognormal moments are called uniformly localized. In other cases, neighboring q-intervals cease to overlap for sufficiently high q, such moments are called asymptotically localized.
Seamless Wikipedia browsing. On steroids.
Every time you click a link to Wikipedia, Wiktionary or Wikiquote in your browser's search results, it will show the modern Wikiwand interface.
Wikiwand extension is a five stars, simple, with minimum permission required to keep your browsing private, safe and transparent.