Review Article
Austin J Bus Adm Manage. 2023; 7(2): 1059.
Why the Basic Axioms of Risk Assessment Can Be Problematic to Identify Risk? – The Development of the VUCA Meter
Thordur Vikingur Fridgeirsson¹*; Helgi Thor Ingason²
¹Assistant Professor, Reykjavik University, Iceland
²Professor at Reykjavik University Reykjavík, Capital Region, Iceland
*Corresponding author: Thordur Vikingur Fridgeirsson Assistant Professor, Reykjavik University, Iceland. Email: [email protected]
Received: June 08, 2023 Accepted: June 30, 2023 Published: July 07, 2023
Abstract
The VUCA world is the epitome of the challenges linked with some of the undercurrents currently shaping businesses in the projectified society. The terms Volatility, Uncertainty, Complexity and Ambiguity (VUCA) are used as descriptive for the continuous flux project managers and project planners are forced to keep at bay. In this article, the authors present an idea to connect the VUCA concept as a risk identification platform to access, identify and isolate low-probability but high-impact events often called black swans or fat tail events. It is argued that the assumptions of conventional risk assessment in the domain of project management can lead to skewed general outcomes due to limitations of the intellect to assess probability. The outline of the VUCA meter is drafted and argued that the meter can augment the conventional risk assessment approach.
Keywords: VUCA meter; Probability; Black swans; Risk management
Introduction
In his much-applauded book “Against the Gods – The Remarkable Story of Risk” Peter L. Bernstein claims that risk management has existed for more than 2000 years as part of decision-making [7]. However, modern risk management only started after World War II 1[12]. The first academic book on risk management was published by Mehr and Hedges as late as 1963 [31]. Risk management commenced by being a financial instrument to hedge companies against fluctuations related to interest rates, stock market returns, exchange rates, and the prices of raw materials or commodities. Risk management has over the years evolved to be a corporate framework to handle risk and uncertainty. It entered the realm of project management in 1987 when PMI (Project Management Institute) added risk management section to the PMBOK guide [37]. Georges Dionne states that “In general, a pure risk is a combination of the probability or frequency of an event and its consequences, which is usually negative [12].” This definition of risk is complying with well-known classifications of what is a risk. Oxford English Dictionary defines risk as: “a chance or possibility of danger, loss, injury or other adverse consequences”. The ISO 31000 standard refers to risk as the “effect of uncertainty on objectives.” Institute of Risk Management (IRM) forwards this definition: “Risk is the combination of the probability of an event and its consequence.” HM Treasury in the UK refers to risk as: “uncertainty of outcome, within a range of exposure, arising from a combination of the impact and the probability of potential events.” Finally, from the Institute of Internal Auditors we have “risk is measured in terms of consequences and likelihood.”. [21] Brought together we can assume that a common formulation of risk is:
Risk is the combination of the probability of an event and its consequence. Risk = Likelihood of event × Impact of event.
Here we have an issue. If risk is a product of likelihood and impact, e.g. on a scale from 1-5, a risk matrix for some hypothetical project could look like shown in Figure 1.
Figure 1: A conventional risk matrix based on Risk = Likelihood x Impact [26].
Figure 2: The unknown unknowns lie in the tails.
The problem here is that a high-impact and low likelihood risk event might be assumed to be of limited significance in the general project risk profile.
Black swans and the Vadlaheidi tunnel project a paradigm called the black swan risk factor has recently emerged and is currently widely used and the work presented in this paper is inspired by this. The term has become widely recognised because of the book The Black Swan by Nassim Taleb [39]. The financial crash in 2008 is one of the most recent and well-known black swan events. The effect of the crash was catastrophic and global, and only a few outliers were able to predict it happen ing. And although somewhat disputed, the COVID-19 pandemic [34] and the war in Ukraine may be considered black swans (Financial Times, 2022). It should be noted that the vocabulary of this type of risk management has been expanded in the last years by using colorful naming from the animal kingdom. Some risk analysts would name COVID-19 and the war in Ukraine as a Grey Rhino and some stock crash as Dragon Kings [13]. However, in the present work, risk events that are governed by power or Poisson distributions, rather than the normal distribution are called black swans or fat tail events, for the sake of simplifying.
To prepare for the postulation of this paper the authors like to present the Vadlaheidi tunnel project (here after we use the nam Vadlaheidi project for simplicity). This project is a 7.5km mountain tunnel at the north coast of Iceland connecting the city of Akureyri with Fnjoskadalur. The case presents some interesting shortcomings of the traditional definition of risk.
• The initial business model for the project was presented in 2002. It was assumed that the construction and the operation of the tunnel would be a private-public enterprise with high feasibility and limited technical difficulties. Road tolls would recover all investment costs within 20 years plus a macroeconomic gain of 8%. Then came the financial crisis in 2008. Market financing folded as a consequence. The arrangement was modified and the Icelandic government was forced to guarantee a loan to make the construction possible.
• When the construction commenced the project soon hit some serious unforeseen problems. In the beginning of 2014, a major hot water leak was detected, due to unexpected geothermal activity in the mountain. As a consequence, drilling was impossible due to heat and steam. To be able to proceed with the project, the contractor had to move the equipment to the other side of the mountain and continue drilling from there.
• In April 2015, a major unexpected leak of cold water was discovered at the new drilling site. The water completely floated the tunnel causing serious problems. A famous news clip from this period shows a TV reporter rowing a boat inside the tunnel to investigate and describe the conditions.
• The tunnel was scheduled to be ready for traffic in 2016 (the initial plan assumed 2011). However, it was only operative in December 2018. The cost overrun in 2017 was estimated at 44%. However, it should be noted that in the presented cost overrun number, the cost of finance was not included and the real total cost overrun is thus much higher. In July 2019, it was reported that the income from the tolls was 35–40% less than expected. Moreover, the COVID-19 pandemic resulted in a major breach in traffic volume in 2020, as the Icelandic tourist industry collapsed and local people were encouraged to limit their mobility as much as possible.
In a nutshell, the drastic events that troubled the Vadlaheidi project are:
1. An unexpected international financial crisis ruined the initial business model transferring the financial risk to the public.
2. Unexpected geothermal activity inside the tunnel prevented drilling the tunnel with negative consequences for the schedule and the budget.
3. An unexpected cold-water leak inside the tunnel delayed significantly excavating the tunnel with negative consequences for the schedule and the budget.
4. The unexpected global pandemic reduced the tunnel traffic and the revenue stream was severely reduced. [40].
In short, the project was hit by four highly low-probability but high-impact events with serious consequences. It can be argued that the hypothetical risk matrix in figure 1 would not capture any of these risk events due to how improbable they are. Some might recall the famous press conference held on February 12th, 2002 by then the US Secretary of Defence, Donald Rumsfeld. During this conference, Mr. Rumsfeld stated: But there are also unknown unknowns—the ones we don't know we don't know. And if one looks throughout the history of our country and other free countries, it is the latter category that tends to be the difficult one [35]. This comment was actually forwarded to justify the ill-conceived claims that Iraq had “weapons of mass destruction”. However, these words have become legendary as they exemplify risks that come from scenarios that are so unexpected that they would not be considered.
The VUCA world
The unknown-unknowns and black swans in fact embody a situation referred to as the VUCA world. The VUCA world refers to the volatile, uncertain, complex and ambiguous situation that challenges businesses and societies. We have for instance several dubious political leaders that do not hesitate to break with rules and protocols, shaking the norms of the diplomatic and economic world. Former president Trump is accused of being indifferent - even encouraging - when the United States Capitol was mobbed by his supporters in January 2021 [20]. Hungary, the country led by Viktor Orbán, has been declared by the EU not being a full democracy [3]. Boris Johnson was forced to resign in 2022 when proven to tell outright lies repeatedly [22]. Vladimir Putin wages an ill-conceived war in Ukraine in 2022 with terrible humanitarian consequences and with destabilizing effects on worldwide economic balances in magnitudes not fully known at the time of this writing [42].
Even in less political contexts, the manageability that was once associated with planning and forecasting and the continuity of established actors can no longer be re-lied upon. When the giant container ship, Ever Given, blocked passage through the Suez Canal in March 2021 the global supply chain was thrown into chaos [27]. Much hyped financial ideas like the Bitcoin cryptocurrency have turned out to be highly volatile. From November 2021 to November 2022 Bitcoin lost more than 70% of its value (Khalfaoui et al., 2023). And what impact do new and disruptive technologies like automation and quantum computing driven by artificial intelligence have on business and daily lives? Global demographics, migration, trade strategies, disruptive technologies, unconventional politicians, cultural uprisings (e.g. MeToo) and the en-counter of climate change (Friðgeirsson & Steindórsdóttir, 2018); all of this create an acute challenge for management in foreseeing the future and orchestrating preventive measures in the attempt to gain control.
Bennett and Lemoine (2014b) call this “the VUCA world” employing an acronym for the volatility, uncertainty, complexity, and ambiguity that exemplifies the world we inhabit.
Another noteworthy paradigm in the context of this study is the term “projectification”. Projectification is a well-known concept to describe the diffusion of projects as a form of organizing in all sectors of the economy [32,36]. Concluded that more than one third of an economy in a developed country can be traced to projects and this progression is escalating. Arguably, the management of risk in projects is more current than ever due to the significance of this management form and the uncertainty of the VUCA world.
The authors theory is that the conventional definition of risk as being the likelihood of event multiplied by the impact of the event - can be problematic when assessing project risk in the VUCA world. The reason is on one hand the escalation of unlikely events that cause big problems and on the other hand the limitations of the human mind to assess probabilities.
The Problem of Subjective Probability Assessment
In project management, likelihood of event (probability) is acquired either by a subjective estimate or it is based on empirical evidences. Firstly, let’s discuss what is called the subjective probability of an event. People make such judgments all the time, not only to estimate risk but to cope with their daily lives. We look out of the window prior to walking the dog or heading for the golf course to estimate if it will rain, we speculate on sport event results, who will win the election and so on. Both these approaches, using subjective estimates or empirical evidences, are valid and current per se. However, they come with shortcomings. Group of experts assessing risk may be prone to cognitive biases often referred to as “planning fallacies”. Planning fallacies are described in the seminal work of Daniel Kahneman and Amos Tversky [24] and have been proven to affect judgments and decisions. The planning fallacy, see for instance [29] and [9], stems from agents taking an inside view focusing on the constituents of the specific planned action rather than on the outcomes of similar actions already completed. Thus, for example, the estimated costs put forward by cities competing to hold the Olympic Games have consistently been underestimated; yet, every four years these errors are repeated [14]. Some interesting psychological issues may affect how the estimating and planning plays out. The backbone of the conventional approach in risk assessment is the attempt to predict the likelihood of a risk event to happen. The estimate of subjective probabilities is in particular interesting in context of this paper and the VUCA world. The reason is that subjective probability judgments are based on data of limited validity which are processed through mental heuristic rules. Daniel Kahneman calls these heuristics rules “mental shotguns” to answer complex questions [23]. The application of these mental rules is often very useful but the drawback is that heuristics can lead to biased probability assessments. In particular, three types of heuristic rules are applied by people to estimate subjective probabilities and they are well worth mentioning in the context of risk identification. These rules are called Representativeness, Availability and Anchoring and adjustment.
Representativeness Heuristic
The representativeness heuristic may work in such a way that when trying to assess how likely a certain event is, we ignore base rate frequencies and sample sizes and opt instead for what we find a fit to our question. A well-known example of this is when a group of people were asked how likely it is those two hospitals, a large one and a small one, report that more than 60% of children born on any given day are boys. Most people estimate that the odds of such a report is even for both hospitals, thus ignoring the sample size (the small hospital has larger variation in gender birth ratios). Another example is when a description of an individual fits a certain occupation, for instance a librarian. If the description fits the stereotypic description of a librarian (shy, withdrawn, helpful, little interest in people, detailed) people assume the person belonging to this rather rare occupation, ignoring the prior probability of the outcome (the number of librarians in US is about 160.000 but the number of lawyers is about 1,3 million (www.ala.org and www.clio.com)). This can lead to serious errors as similarities do not necessary represent probabilities of occurrence. People also expect that random sequences are likely to have a certain pattern. When tossing a coin, people consider the sequence H-T-H-T-T-H more likely to happen than H-H-H-T-T-T although statistically, both sequences are evenly likely to happen. This is sometimes referred to as the gambler’s fallacy (luck must be on my side next game as I have been so unlucky until now). People also tend to ignore how outliers work. For instance that a record sales month will be followed by equally successful month, or even better. In other words, that contemporary change actually represent future conditions. But outliers usually do not work like that and the most likely outcome for the sales is that the they will regress to the mean (the normal condition). The representativeness heuristic and the cognitive biases incurred can deprive people from assessing correctly the probabilities of a risk event.
Availability Heuristics
Another interesting heuristic leading to flawed probability assessment is he availability heuristics. The availability heuristic mechanism leads people to confuse the probability of an event with how easily it can be brought to mind. This can lead to biased judgment of the probability of occurrence. A risk assessor that hears about a terrorist attack in the news the morning prior to a risk assessment seminar might be affected by the images that occupies his/her mind. However, an event that can be easily imagined is not necessarily an event that is particularly likely to happen. A well-known example of the availability heuristic is the fear of flying. Airplane accidents are very rare but when they happen they are extensively covered in the news. Pictures of people stuck in an air cabin waiting helplessly to meet their destiny are easily brought to mind. The probability of dying in an air crash are extremely low and in fact less than dying in a car crash (flyfright.com, 2022).
Anchoring and Adjustment Heuristic
The last heuristic presented here to portray the challenges of assigning probabilities to risk events is the anchoring and adjustment heuristic. This heuristic is well known from negotiation techniques. A well-known trick when trying to get a low price is to make the first move and name a low number, in the hope that the proposed number will serve as an anchor that will be adjusted in the negotiation process. Anchoring also plays a role when the decision is based on incomplete computations. The human mind is actually not very good at computations. It is not easy to multiply 1x2x3x4x5x6x7x8 without a calculator. Interestingly, unaided multiplication of 8x7x6x5x4x3x2x1 gives a totally different estimate as studies show. The reason is the different arrangement of the numbers. People note the first numbers in the row and attempt to calculate in their minds the product, but wind-up in guessing. As a rule, the row with the descending sequence produces higher estimates than the ascending sequence. In the original experiment the median of the ascending estimates was 512 while the median of the descending was 2.250 (by the way, the correct answer is 40.320). An example of other biases under this category is the typical overestimation of the probability of conjunctive events as people tend to anchor the probability of a single event and make insufficient adjustment consequently. For instance, a system that has seven dependent elements, meaning that they all have to be functional so the system can operate. Each element has p=0,9 to be operational at any given moment. Studies show that many would consider the probability of functional system to be 0,9. The correct probability is only p=0,48 (0,9 in the power of seven).
The text above is based on the inspiring work of Daniel Kahneman and Amos Tversky [23]. Daniel Kahneman received the Nobel Memorial Prize in Economic Sciences 2002 for his ground-breaking work in applying psychological insights to economic theory, particularly in the areas of judgment and decision-making under uncertainty. So, in spite of the many wonderful attributes of the human mind, estimating probabilities is not one of them.
Assessing Probability with Empirical Evidences
Empirical evidence also come with some well-known downsides in risk management. It is often assumed that the likelihood of outcomes is normally distributed, pointing the risk focus towards the highest frequency outcomes - the most probable outcomes (the mean). However, as mentioned before, prolific scholars and thinkers accentuate that the most impressive risks are not found in the averages but in the tails of the normal distribution. That is to say, the biggest challenges with the direst consequences are low probability high impact outliers, often called black swans. The term became famous because of the book, The Black Swan, by Nassim Taleb [39]. World War I, the rise of the Internet and 9/11 have been identified as examples of black swan events. The financial crash in 2008 is one of the most recent and well-known black swan events. The effect of the crash was catastrophic and global, and only a few outliers were able to predict it happening. And although somewhat disputed, the COVID-19 pandemic [34] and the war in Ukraine may be considered black swans (Financial Times, 2022). It should be noted that the vocabulary of risk management has been augmented the last year by using colorful naming’s from the animal kingdom. For instance some risk analysts would name COVID-19 and the war in Ukraine as a Grey Rhino´s and some stock crash’s as Dragon Kings. [13] However, in this work risk events that are governed by power or Poisson distributions, rather than the normal distribution, are called black swans or fat tail events for the sake of simplifying.
Taleb argued that because black swan events are impossible to predict due to their rarity, yet they have catastrophic consequences, it is important for people to always assume a black swan event is a possibility, whatever it may be, and to try to plan accordingly. [10] Moreover, in the VUCA world, the speedy and complex environment we live in, empirical evidence may be difficult to enquire and may even be rapidly redundant. Fresh management concepts like Agile and Beyond Budgeting are specifically designed to overcome the challenges of rigid up-front planning in our turbulent world. [11] Lastly, finding a suitable reference class can also be a challenge. An example of this is deciding the probability of a programmer getting a job within a certain period. Should the reference class include only unemployed programmers or should it contain other attributes like gender, age, health, address, etc. [18]?
The authors therefore like to introduce the VUCA concept as an additional platform to assess project risks with an emphasis on isolating low-frequency but high-impact risk events by a stepwise procedure.
The VUCA Meter Development
The attempt to augment the risk assessment procedure is not unpresented. A case study from Ackermann et al., (2007) discussed the shortcomings of systemic risk assessment. The authors argued that the most attention in the systemic risk assessment is devoted to the technical risk in projects, not other risk categories such as political risk, customer risk, partner and supplier risk, human risk, reputation risk, market, and financial risk. In 2011, Geraldi, Maylor and Williams published an article where they systematically reviewed the complexities of projects and pointed out that this is a key variable that impacts decisions in project management. [17] The type of complexity that is most frequently mentioned is structural complexity. Still, uncertainty is a relevant type of complexity and is one of the four concepts that constitute VUCA. The internal connection between complexity and risk was mentioned in the literature as early as 1920 and has thus long been recognized [2]. In 2006 Linehan and Kavanagh defined projects as a confusing phenomenon that contain a lot of complexity and ambiguity, and the idea of a single clear goal is not realistic [28].
The VUCA approach originates from Warren Bennis and Burt Nanus´s book - Leaders: Strategies for Taking Charge where the acronym VUCA stands for volatility, uncertainty, complexity, and ambiguity [6]. According to the U.S. Army Heritage and Education Center, the VUCA acronym has been used within the U.S. Army War College for leadership training. It is often cited in US Army War College documents and has become popular in strategic and leadership research throughout the 1990s [41]. The VUCA term has made its way into the business lexicon, with popular explanatory articles appearing in publications, such as the Harvard Business Review [4] and Forbes [25].
The VUCA concept has also entered the project management realm. In 2016 Szpitter and Sadkowska proposed VUCA as a useful matrix for understanding and assessing project environment complexity and risk [38]. Another paper argues that why projects fail is due to project complexity of the modern VUCA world and through a survey the authors find out that stakeholders do often not treat project complexity with due respect but rather with ignorance or “stupidity” [33]. In 2016 the book “Managing in a VUCA World” took a close look at Volatility, Uncertainty, Complexity, and Ambiguity (VUCA) as concepts and frameworks to understand and manage unpredictability and rapid change [30]. The VUCA concept has also been used in an attempt to rethink project management in an era of digital transformation and the development of the project management profession until 2030 [8].
The authors therefore have worked on the idea to compile a methodology that would enhance the risk assessment doctrine by using the VUCA concept as the platform. Early in the process the methodology was coined as “VUCA meter”. [15] and [16] This study bases its definition of the VUCA components on Bennett and Lemoine, who use VUCA to describe the rapidly changing environment in which modern businesses must navigate. They warn against conflating the distinct terms of the VUCA framework when one is faced with the unpredictability of VUCA situations. Despite the myriad of popular articles, they claim, “there is a lack of information regarding just what it is that leaders should do in order to confront […] these conditions” [5]. Properly identifying them, they claim, is crucial to take appropriate action, as they each “require their own separate and unique responses. […] Failure to use the right label will lead to a misallocation of what could be considerable corporate resources” [5].
All things considered, Bennett and Lemoine provided the authors with the semantics needed to form a textual assessment approach. Based on this, a definition of each of the VUCA terms is summarized in Table 1 and the relevant questions in the context of project management formed [15,16].
- Volatility: Unstable and unpredictable resource cost and/or availability at unpredictable times and durations, expected fluctuations on resources with unknown timing, and magnitude. Questions aim to capture key reasons why resources might be suddenly unavailable or expensive, and the challenges in resolving resourcing for unforeseen new needs.
- Uncertainty: Lack of knowledge, the unclear impact of change, but causes and effects are known. Questions aim to capture key risks for lack of comprehensive and reliable knowledge in projects, difficulty in communicating or accessing relevant information, or effectively resolving questions on unforeseen new issues.
- Complexity: Many interconnected parts, complex regulatory/political environments, multiple components and parts. Questions aim to capture key risks for the structural or systems-related complexity in the environment, internal and external, of a project, as well as whether appropriate governance to effectively deal with these complexities is in place.
- Ambiguity: Doubt about the nature of cause and effect, little to no historical information available as to predict an outcome, hence difficult to forecast or plan for. Questions aim to capture whether there are risks due to lack of experience and predictability that might affect the project and whether the project delivers unpredictable new changes, either internal or external to the project and/or the organization.
Table 1: VUCA semantics and relevant questions [5,15].
The purpose of the VUCA meter is to detect and expose fat tail risk events referred to as black swans. Nancy Green provided a passable description of the characteristics of a risk event that might surpass the conventional risk assessment procedure based on the work of Nassim Taleb, see Table 2 [19].
- Emergency response to the problem and a fixing the problem are different aspects.
- A solution to the problem is unknown and must be created under dismal circumstances.
- Public relations issues can be massive putting pressure on reputation, credibility and perception of the public.
- Governmental and regulatory agencies may demand response.
- Productivity and cash flow may be affected negatively, liquidity could become uncertain and asset prices disturbed.
- Despite the problem the day-to-day operation must continue.
Table 2: Criteria to isolate black swan risk events [19].
The semantics of the Bennet and Lemoine study on the characteristics of VUCA provided the authors with means to develop the sections of the VUCA risk identification meter. Each section of the meter must be addressed individually since they require a unique response. Based on this, seven sequential steps were identified as shown in Table 3.
1. Pre-preparation
· Primarily to prepare the expert participating in the workshop.
2. Workshop kick-off
· Primarily to educate and motivate the experts to be creative and productive in the four VUCA rounds.
3. Four VUCA rounds
· Each round is centered on a particular focus question. The object in each round is based on table 1. The risk matrices for each round can be seen in figure 3.
4. Categorizing
· Primarily to sort and bring together similar events and identify risk patterns.
5. Risk event formulation
· Primarily to generate clear textual description of risk events that can be assessed and valued individually.
6. Risk event impact
· The experts estimate individually the impact of the risk events from step 5 on a Likert scale.
7. Black swan identification
· The criteria´s from table 2 are used to isolate black swan risk events.
Table 3: The VUCA meter procedure in seven steps.
Of particular interest are the four VUCA rounds, see Figure 3. The VUCA rounds are structured brain storm sessions. The rounds are carefully facilitated to enhance creative thinking.
Figure 3: Each VUCA round is supported by a matrix each depicting a particular round.
Figure 4: A schematic progression of the VUCA meter identification of high impact but low probability risk events.
It is noteworthy that the VUCA procedure assumes that steps 2-5 are physical collective workshops, carefully planned and facilitated. Steps 6 and 7 are on the other hand individual assessments were the experts evaluate the outcomes from the previous steps on pre-designed scales. These steps can be carried out online if applicable.
Discussions and Conclusion
Risk was inserted into the domain of project management as late as 1987 when PMI introduced a risk chapter in its main doctrine, the PMBOK. Since then, risk and uncertainty have gradually been accepted as a major dogma in the discipline of project management. The aspiration to hedge projects against the unexpected by assessing risk is currently on pair with scheduling and budgeting. The significance of risk management is further enveloped by the projectification of the society. Schoper et al. (2018) argues that a third of the economy can be traced to project work. In Germany the share of project work was 29% in 2009 but was projected to be over 41% in 2019. It can be argued that the escalation of projects and project management is a response to the VUCA world. The volatile, uncertain, complex and ambiguous business environment managers and decision makers must cope with.
Arguably, the most sensitive step of the risk management procedure is the first step, the assessment of the root causes to later troubles. If a risk analyst fails to identify events that can put the project objectives into jeopardize at the initial stages, the realization of such an event later in the project lifecycle can be difficult and costly to rectify. In this study the authors point out that the basic axioms of risk analysis are the function of probability and impact of a risk event. The authors do not oppose to this approach per se. However, they call attention to interesting psychological theories forwarded initially by Daniel Kahneman and Amos Tversky in the seventies, but have now become a mainstream concept to understand for instance cost overruns, late budgets, scope creeps, etc. The Prospect Theory [24] for instance claims that due to cognitive biases people have difficulties assessing probabilities, chance, likelihood or whatever terms we use to describe a certain frequency of a possible outcome. This is not a statement to undermine the traditional approach. The authors are on the contrary a firm believer on using for instance empirical data, reference class forecasting, simulations, risk models, etc. However, the problem described is this conceptual article can be instrumental in assessing project risk as it is mathematically the most important gradient for quantifying the qualitative assessment. Moreover, the most probable risk events are not equal to severity of the risk event that further underpins the theory that risk analysts should be on the hunt for fat tail events.
The authors theory is that the conventional definition of risk;
Risk = Likelihood of event × Impact of event can be problematic assessing project risk in the VUCA world. In this work we have presented firstly; why it is problematic and, secondly how the accepted process can be, not be replaced, but augmented by the VUCA meter. The VUCA meter is a normative approach aimed at isolating low frequency but high-impact events that could bring dilemma to the project.
The VUCA meter has hitherto been tested rigorously on nine projects. The prototype was tested on five large technical installations. The first version was used to supplement risk assessment on a large infrastructure project in the planning phase. This was followed by using the VUCA meter to identify risk in three very different projects; one humanitarian program due to the war in Ukraine, one a Waste-to-energy plant and lastly a geo-thermal power plant. All assessments were carried out and facilitated by professional project managers. Furthermore, in two of the assessments the VUCA meter was benchmarked against the conventional risk assessment procedure. An accurate description of the VUCA meter and the application of it await another publication. It can though be stated that there are clear evidences that the meter is a valuable new instrument that can be placed with other risk identification means in the project manager’s toolbox.
References
- Ackermann F, Eden C, Williams T, Howick S. Systemic risk assessment: A case study. J Oper Res Soc. 2007; 58: 39-51.
- Baccarini D. The concept of project complexity: a review. Int J Manag Projects Bus. 1996; 14: 201-4.
- Batory A. More power, less support: the Fidesz government and the coronavirus pandemic in Hungary. Gov Opposition. 2022: 1-17.
- Bennett N, Lemoine G. What VUCA really means. Harv Bus Rev. 2014a.
- Bennett N, Lemoine GJ. What a difference a word makes: understanding threats to performance in a VUCA world. Bus Horiz. 2014b; 57: 311-7.
- Bennis WG, Nanus B, Bennis S. Leaders: strategies for taking charge. New York: Harper & Row; 1985; 200.
- Bernstein P. Against the Gods—the remarkable story of risk. Wiley; 1996.
- Bierwolf R, Romero D, Pelk H, Stettina CJ. On the future of project management innovation: A call for discussion towards project management. Technol Innov (ICE/ITMC). International Conference on Engineering. IEEE Publications. 2017; 2017: 689-98.
- Buehler R, Griffin D, Peetz J. The planning fallacy. Adv Exp Soc Psychol. 2010; 43: 1-62.
- Cirillo P, Taleb NN. Tail risk of contagious diseases. Nat Phys. 2020; 16: 606-13.
- Denning S. Reinventing Management-How Agile an Beyond Budgeting have converged Management. Forbes. Available from: https://www.forbes.com/sites/stevedenning/2018/03/27/reinventing-management-how-agile-and-beyond-budgeting-have-converged/?sh=19e6967b4c58; 2018.
- Dionne G. Risk management: history, definition, and critique. Risk Manag Ins Rev. 2013; 16: 147-16.
- Ferguson N. Doom -The politics of catastrophe. Penguin; 2022.
- Flyvbjerg B, Stewart A, Budzier A. The Oxford Olympics Study 2016: cost and cost overrun at the games; 2016. Available from: https://doi.org/arXiv preprint arXiv:1607.04484.
- Fridgeirsson TV, Ingason HT, Björnsdottir SH, Gunnarsdottir AY. Can the ’VUCA meter’ augment the traditional project risk identification process? A case study. Sustainability. 2021; 13: 12769.
- Fridgeirsson TV, Ingason HT, Jonasson HI, Kristjansdottir BH. The VUCAlity of projects: A new approach to assess a project risk in a complex world. Sustainability. 2021; 13: 3808.
- Geraldi J, Maylor H, Williams T. Now, let’s Make it Really Complex (Complicated): A Systematic Review of the Complexities of Projects. Int J Oper Prod Manag. 2011; 31: 966-90.
- Goodwin P, Wright G. Decision analysis for management judgment. 4th ed. John Wiley & Sons; 2014.
- Green N. Keys to success in managing a black swan event. AON Corporation; 2015. Available from: http://www.aon.com/attachments/risk-services/Manage_Black_Swan_Even_Whitepaper_31811.pdf.
- Holtzman E. The January 6th Report. Simon & Schuster; 2022.
- Hopkin P. Fundamentals of risk management: understanding, evaluating and implementing effective risk management. Kogan Page Publishers; 2018.
- Judge D. ”Would I lie to you?”: Boris Johnson and Lying in the House of Commons. Pol Q. 2022; 93: 77-86.
- Kahneman D. Thinking, fast and slow; 2011.
- Kahneman D, Tversky A. Prospect theory: an analysis of decision under risk. Econometrica. 1979; 47: 263-91.
- Kraaijenbrink J. What does VUCA really mean? [Forbes, desember]. (35) Forbes; 2018. What Does VUCA Really Mean? Available from: https://www.forbes.com/sites/jeroenkraaijenbrink/2018/12/19/what-does-vuca-really-mean/?sh=7fbd6f2c17d6.
- Larson EW, Gray CF. Project management: the managerial process. McGraw-Hill Education; 2014.
- Lee JMY, Wong EYC. Suez Canal blockage: an analysis of legal impact, risks and liabilities to the global supply chain. MATEC Web Conf. 2021: 339.
- Linehan C, Kavanagh D. From project ontologies to communities of virtue. Making projects critical. 2006; 51-67.
- Lovallo D, Kahneman D. Delusions of success. How optimism undermines executives’ decisions. Harv Bus Rev. 2003; 81: 56-63.
- Mack O, Jungen M. Program management in VUCA environments: theoretical and pragmatical thoughts on a systemic management of projects and programs. In: Mack O, Khare A, Krämer A, Burgartz T, editors. Managing in a VUCA world. 2016; 41-57.
- Mehr RI, H, RA. Risk management in the business enterprise. RD Irwin. 1963.
- Midler C. Projectification of the firm: the Renault case. Scandinavian Journal of Management. 1995; 11: 363-75.
- Mikkelsen MF, Marnewick C, Klein K. On Stupidity in Project Management—A critical reflection of PM in a VUCA world. Mod Proj Manag. 2020; 8.
- Mohammed FS, Qadri H, Mohammed S. COVID-19 in the era of artificial intelligence: A black swan event? J Med Artif Intell. 2022; 5: 6.
- Pozen DE. Deep secrecy. Stanford Law Rev. 2010; 257-339.
- Schoper YG, Wald A, Ingason HT, Fridgeirsson TV. Projectification in Western economies: A comparative study of Germany, Norway and Iceland. Int J Proj Manag. 2018; 36: 71-82.
- Stretton A. A short history of modern project management. PM World Today. 2007; 9: 1-18.
- Szpitter A, Sadkowska J. Using VUCA matrix for the assessment of project environment risk. Zarzadzanie i finanse, 14. 2016; 4013: 401.
- Taleb N. The impact of the highly improbable. 2nd ed. Swan: Black. Random House; 2007.
- Vadlaheidi. Sagan; n.d. Available from: https://www.vadlaheidi.is/is/sagan.
- Van Tulder R, Jankowska B, Verbeke A. Introduction: progress in international business research in an increasingly VUCA world. In: International business in a VUCA world: the changing role of states and firms. Emerald Publishing Limited; 2019.
- Welfens PJ. Beginnings of the russo-Ukrainian war. In: Russia’s invasion of Ukraine: economic challenges, embargo issues and a new global economic order. Springer. 2023; 3-43.