1000/1000
Hot
Most Recent
Dragon king (DK) is a double metaphor for an event that is both extremely large in size or impact (a "king") and born of unique origins (a "dragon") relative to its peers (other events from the same system). DK events are generated by / correspond to mechanisms such as positive feedback, tipping points, bifurcations, and phase transitions, that tend to occur in nonlinear and complex systems, and serve to amplify DK events to extreme levels. By understanding and monitoring these dynamics, some predictability of such events may be obtained. The theory has been developed by Didier Sornette, who hypothesizes that many of the crises that we face are in fact DK rather than black swans—i.e., they may be predictable to some degree. Given the importance of crises to the long-term organization of a variety of systems, the DK theory urges that special attention be given to the study and monitoring of extremes, and that a dynamic view be taken. From a scientific viewpoint, such extremes are interesting because they may reveal underlying, often hidden, organizing principles. Practically speaking, one should study extreme risks, but not forget that significant uncertainty will almost always be present, and should be rigorously considered in decisions regarding risk management and design. The theory of DK is related to concepts such as black swan theory, outliers, complex systems, nonlinear dynamics, power laws, extreme value theory, prediction, extreme risks, and risk management.
A black swan can be considered a metaphor for an event that is surprising (to the observer), has a major effect, and, after being observed, is rationalized in hindsight. The theory of black swans is epistemological, relating to the limited knowledge and understanding of the observer. The term was introduced and popularized by Nassim Taleb and has been associated with concepts such as heavy tails, non-linear payoffs, model error, and even Knightian uncertainty, whose “unknowable unknown” event terminology was popularized by former United States Secretary of Defense Donald Rumsfeld. Taleb claims that black swan events are not predictable, and in practice, the theory encourages one to “prepare rather than predict”, and limit one's exposure to extreme fluctuations.
The black swan concept is important and poses a valid criticism of people, firms, and societies that are irresponsible in the sense that they are overly confident in their ability to anticipate and manage risk. However, claiming that extreme events are – in general – unpredictable may also lead to a lack of accountability in risk management roles. In fact, it is known that in a wide range of physical systems that extreme events are predictable to some degree.[1][2][3][4] One simply needs to have a sufficiently deep understanding of the structure and dynamics of the focal system, and the ability to monitor it. This is the domain of the dragon kings. Such events have been referred to as Grey Swans by Taleb. A more rigorous distinction between black swans, grey swans, and dragon kings is difficult as black swans are not precisely defined in physical and mathematical terms. However technical elaboration of concepts in the Black Swan book are elaborated in the Silent Risk document. An analysis of the precise definition of a black swan in a risk management context was written by Prof. Terje Aven.[5]
It is well known that many phenomena in both the natural and social sciences have power law statistics (Pareto distribution) [7] [8] [9] . Furthermore, from extreme value theory, it is known that a broad range of distributions (the Frechet class) have tails that are asymptotically power law. The result of this is that, when dealing with crises and extremes, power law tails are the “normal” case. The unique property of power laws is that they are scale-invariant / self-similar / fractal. This property implies that all events – both large and small – are generated by the same mechanism, and thus there will be no distinct precursors by which the largest events may be predicted. A well known conceptual framework for events of this type is self-organized criticality. Such concepts are compatible with the theory of the black swan. However Taleb has also stated that considering the power law as a model instead of a model with lighter tails (e.g., a Gaussian) “converts black swans into gray ones”, in the sense that the power law model gives non-negligible probability to large events.
In a variety of studies it has been found that, despite the fact that a power law models the tail of the empirical distribution well, the largest events are significantly outlying (i.e., much larger than what would be expected under the model)[6] [10] [11] . Such events are interpreted as dragon kings as they indicate a departure from the generic process underlying the power law. Examples of this include the largest radiation release events occurring in nuclear power plant accidents, the largest city (agglomeration) within the sample of cities in a country, the largest crashes in financial markets, and intraday wholesale electricity prices.[6][12]
Physically speaking, dragon kings may be associated with the regime changes, bifurcations, and tipping points of complex out-of-equilibrium systems.[14] For instance, the catastrophe (fold bifurcation) of the global ecology illustrated in the figure could be considered to be a dragon king: Many observers would be surprised by such a dramatic change of state. However, it is well known that in dynamic systems, there are many precursors as the system approaches the catastrophe.
Positive feedback is also a mechanism that can spawn dragon kings. For instance, in a stampede the number of cattle running increases the level of panic which causes more cattle to run, and so on. In human dynamics such herding and mob behavior has also been observed in crowds, stock markets, and so on (see herd behavior).
Dragon kings are also caused by attractor bubbling in coupled oscillator systems.[15] Attractor bubbling is a generic behavior appearing in networks of coupled oscillators where the system typically orbits in an invariant manifold with a chaotic attractor (where the peak trajectories are low), but is intermittently pushed (by noise) into a region where orbits are locally repelled from the invariant manifold (where the peak trajectories are large). These excursions form the dragon kings, as illustrated in the figure. It is claimed that such models can describe many real phenomena such as earthquakes, brain activity, etc.[15] A block and spring mechanical model, considered as a model of geological faults and their earthquake dynamics, produced a similar distribution.[16]
It could also be the case that dragon kings are created as a result of system control or intervention. That is, trying to suppress the release of stress or death in dynamic complex systems may lead to an accumulation of stress or a maturation towards instability. For instance, brush/forest fires are a natural occurrence in many areas. Such fires are inconvenient and thus we may wish that they are diligently extinguished. This leads to long periods without inconvenient fires, however, in the absence of fires, dead wood accumulates. Once this accumulation reaches a critical point, and a fire starts, the fire becomes so large that it cannot be controlled – a singular event that could be considered to be a dragon king. Other policies, such as doing nothing (allowing for small fires to occur naturally), or performing strategic controlled burning, would avoid enormous fires by allowing for frequent small ones. Another example is monetary policy. Quantitative easing programs and low interest rate policies are common, with the intention of avoiding recessions, promoting growth, etc. However, such programs build instability by increasing income inequality, keeping weak firms alive, and inflating asset bubbles.[17][18] Ultimately such policies, aimed at smoothing out economic fluctuations, will enable an enormous correction – a dragon king.
DK are outliers by definition. However, when calling DK outliers there is an important proviso: In standard statistics outliers are typically erroneous values and are discarded, or statistical methods are chosen that are somehow insensitive to outliers. Contrarily, DK are outliers that are highly informative, and should be the focus of much statistical attention. Thus a first step is identifying DK in historical data. Existing tests are either based on the asymptotic properties of the empirical distribution function (EDF)[12] or on an assumption about the underlying cumulative distribution function (CDF) of the data.[6]
It turns out that testing for outliers relative to an exponential distribution is very general. The latter follows from the Pickands–Balkema–de Haan theorem of Extreme value theory which states that a wide range of distributions asymptotically (above high thresholds) have exponential or power law tails. As an aside, this is one explanation why power law tails are so common when studying extremes. To finish the point, since the natural logarithm of a power law tail is exponential, one can take the logarithm of power law data and then test for outliers relative to an exponential tail. There are many test statistics and techniques for testing for outliers in an exponential sample. An inward test sequentially tests the largest point, then the second largest, and so on, until the first test that is not rejected (i.e., the null hypothesis that the point is not an outlier is not rejected). The number of rejected tests identifies the number of outliers. For instance, where [math]\displaystyle{ x_{(1)}\gt x_{(2)}\gt \dots\gt x_{(n)} }[/math] is the sorted sample, the inward robust test uses the test statistic [math]\displaystyle{ T_{r,m}=x_{(r)}/(x_{(m)}+...+x_{(n)}) }[/math] where r is the point being tested [math]\displaystyle{ (r=1,2,3...) }[/math], and [math]\displaystyle{ r\lt m\lt n }[/math] where m is the pre-specified maximum number of outliers. At each step the p-value for the test statistic must be computed and, if lower than some level, the test rejected. This test has many desirable properties: It does not require that the number of outliers be specified, it is not prone to under (masking) and over (swamping) estimation of the number outliers, it is easy to implement, and the test is independent of the value of the parameter of the exponential tail.[6]
thumbnail|(II) Empirical [[CCDF of the log of radiation released (grey dashed) and damage (black) caused by accidents at nuclear power plants, with solid lines giving the fitted power law tail. (III) Empirical CCDF of populations in urban agglomerations within a country, scaled so that the second largest has size 1. Outliers are labeled.[6]]] Some examples of where dragon kings have been detected as outliers include:[6][12]
(I) financial crashes as measured by drawdowns, where the outliers correspond to terrorist attacks (e.g., the 2005 London bombing), and the Flash crash of 2010;
(II) the radiation released and loss ($) caused by accidents at nuclear power plants, where outliers correspond to runaway disasters where safety mechanisms were overwhelmed;
(III) the largest city (measured by the population in its agglomeration) in the population of cities within a country, where the largest city plays a disproportionately important role in the dynamics of the country, and benefits from unique growth; and,
(IV) intraday wholesale electricity prices.
(V) Three-wave nonlinear interaction. It is possible to suppress the emergence of Dragon Kings [19]
How one models and predicts dragon kings depends on the underlying mechanism. However, the common approach will require continuous monitoring of the focal system and comparing measurements with a (non-linear or complex) dynamic model. It has been proposed that the more homogeneous the system, and the stronger its interactions, the more predictable it will be.[20]
For instance, in non-linear systems with phase transitions at a critical point, it is well known that a window of predictability occurs in the neighborhood of the critical point due to precursory signs: the system recovers more slowly from perturbations, autocorrelation changes, variance increases, spatial coherence increases, etc.[22][23] These properties have been used for prediction in many applications ranging from changes in the bio-sphere[13] to rupture of pressure tanks on the Ariane rocket.[24]
For the phenomena of unsustainable growth (e.g., of populations or stock prices), one can consider a growth model that features a finite time singularity, which is a critical point where the growth regime changes. In systems that are discrete scale invariant such a model is power law growth, decorated with a log-periodic function.[26][27] Fitting this model on the growth data (non-linear regression) allows for the prediction of the singularity, i.e., the end of unsustainable growth. This has been applied to many problems,[4] for instance: rupture in materials,[24][28] earthquakes,[29] and the growth and burst of bubbles in financial markets[11][30][31][32] (see http://www.er.ethz.ch/financial-crisis-observatory.html for bubble indicators based on such techniques).
An interesting dynamic to consider, that may reveal the development of a block-buster success, is Epidemic phenomena: e.g., the spread of plague, viral phenomena in media, the spread of panic and volatility in stock markets, etc. In such a case, a powerful approach is to decompose activity/fluctuations into exogeneous and endogeneous parts, and learn about the endogenous dynamics that may lead to high impact bursts in activity.[25][33][34]
Given a model and data, one can obtain a statistical model estimate. This model estimate can then be used to compute interesting quantities such as the conditional probability of the occurrence of a dragon king event in a future time interval, and the most probable occurrence time. When doing statistical modeling of extremes, and using complex or nonlinear dynamic models, there is bound to be substantial uncertainty. Thus, one should be diligent in uncertainty quantification: not only considering the randomness present in the fitted stochastic model, but also the uncertainty of its estimated parameters (e.g., with Bayesian techniques or by first simulating parameters and then simulating from the model with those parameters), and the uncertainty in model selection (e.g., by considering an ensemble of different models).
One can then use the estimated probabilities and their associated uncertainties to inform decisions. In the simplest case, one performs a binary classification: predicting that a dragon king will occur in a future interval if its probability of occurrence is high enough, with sufficient certainty. For instance, one may take a specific action if a dragon king is predicted to occur. An optimal decision will then balance the cost of false negatives/false positives and misses/false alarms according to a specified loss function. For instance, if the cost of a miss is very large relative to the cost of a false alarm, the optimal decision will detect dragon kings more frequently than they occur. One should also study the true positive rate of the prediction. The smaller this value is, the weaker the test, and the closer one is to black swan territory. In practice the selection of the optimal decision, and the computation of its properties must be done by cross validation with historical data (if available), or on simulated data (if one knows how to simulate the dragon kings).
In a dynamic setting the dataset will grow over time, and the model estimate, and its estimated probabilities will evolve. One may then consider combining the sequence of estimates/probabilities when performing prediction. In this dynamic setting, the test will likely be weak most of the time (e.g., when the system is around equilibrium), but as one approaches a dragon king, and precursors become visible, the true positive rate should increase.
Dragon kings form special kinds of events leading to extreme risks (which can also be opportunities). That extreme risks are important should be self-evident. Natural disasters provide many examples (e.g., asteroid impacts leading to extinction). Some statistical examples of the impact of extremes are that: the largest nuclear power plant accident (2011 Fukushima disaster) caused more damage than all (>200) other historical accidents together,[35] the largest 10 percent of private data breaches from organizations accounts for 99 percent of the total breached private information,[36] the largest 5 epidemics since 1900 caused 20 times the fatalities of the remaining 1363,[6][37] etc. In general such statistics arrive in the presence of heavy tailed distributions, and the presence of dragon kings will augment the already oversized impact of extreme events.
Despite the importance of extreme events, due to ignorance, misaligned incentives, and cognitive biases, we often fail to adequately anticipate them. Technically speaking, this leads to poorly specified models where distributions that are not heavy tailed enough, and under-appreciating both serial and multivariate dependence of extreme events. Some examples of such failures in risk assessment include the use of Gaussian models in finance (Black-Scholes, the Gaussian copula, LTCM), the use of Gaussian processes and linear wave theory failing to predict the occurrence of rogue waves, the failure of economic models in general to predict the financial crisis of 2007–2008, and the under-appreciation of external events, cascades, and nonlinear effects in probabilistic risk assessment, leading to not anticipating the Fukushima Daiichi nuclear disaster in 2011. Such high impact failures emphasize the importance of the study of extremes.
The dragon king concept raises many questions about how one can deal with risk. Of course, if possible, one should avoid exposure to large risks (often referred to as the black swan approach). However, in many developments, exposure to risk is a necessity, and a trade-off between risk and return needs to be navigated. This nontrivial case is the one that must be considered.
In an adaptive system, where prediction of dragon kings is successful, one can act to defend the system or even profit. How to design such resilient systems, as well as their real time risk monitoring systems[38] , is an important and interdisciplinary problem where dragon kings must be considered.
On another note, when it comes to the quantification of risk in a given system (whether it be a bank, an insurance company, a dike, a bridge, or a socio-economic system), risk needs to be accounted for over a period, such as annually. Typically one is interested in statistics such as the annual probability of loss or damage in excess of some value (Value at risk), other tail risk measures, and return periods. To provide such risk characterizations, the dynamic dragon kings must be reasoned about in terms of annual frequency and severity statistics. These frequency and severity statistics can then be brought together in a model such as a Compound Poisson Process.
Provided that the statistical properties of the system are consistent over time (stationary), frequency and severity statistics may be constructed based on past observations, simulations, and/or assumptions. If not, one may only construct scenarios. However, in any case, given the uncertainty present, a range of scenarios should be considered. Due to the shortage of data for extreme events, the principle of parsimony, and theoretical results from extreme value theory about universal tail models, one typically relies on a GPD tail model. However such a model excludes DK. Thus, when one has sufficient reason to believe that DK are present, or if one simply wants to consider a scenario, one may e.g., consider a density mixture of a GPD and a density for the DK regime.