Risk

Risk is the possibility of something bad happening.[1] Risk involves uncertainty about the effects/implications of an activity with respect to something that humans value (such as health, well-being, wealth, property or the environment), often focusing on negative, undesirable consequences.[2] Different practice areas of risk management have used many different definitions.

Definitions

Firefighters at work

Oxford English Dictionary

The Oxford English Dictionary (OED) cites the earliest use of the word in English (in the spelling of risque from its French original, 'risque') as of 1621, and the spelling as risk from 1655. While including several other definitions, the OED 3rd edition defines risk as:

(Exposure to) the possibility of loss, injury, or other adverse or unwelcome circumstance; a chance or situation involving such a possibility.[3]

International Organization for Standardization

The International Organization for Standardization (ISO) Guide 73 provides basic vocabulary to develop common understanding on risk management concepts and terms across different applications. ISO Guide 73:2009 defines risk as:

effect of uncertainty on objectives

Note 1: An effect is a deviation from the expected – positive or negative.

Note 2: Objectives can have different aspects (such as financial, health and safety, and environmental goals) and can apply at different levels (such as strategic, organization-wide, project, product and process).

Note 3: Risk is often characterized by reference to potential events and consequences or a combination of these.

Note 4: Risk is often expressed in terms of a combination of the consequences of an event (including changes in circumstances) and the associated likelihood of occurrence.

Note 5: Uncertainty is the state, even partial, of deficiency of information related to, understanding or knowledge of, an event, its consequence, or likelihood.[4]

This definition was developed by an international committee representing over 30 countries and is based on the input of several thousand subject matter experts. It was first adopted in 2002. Its complexity reflects the difficulty of satisfying fields that use the term risk in different ways. Some restrict the term to negative impacts (“downside risks”), while others include positive impacts (“upside risks”).

ISO 31000:2018 “Risk management — Principles” uses the same definition with a simpler set of notes.[5]

Other

Many other definitions of risk have been influential:

“Source of harm”. The earliest use of the word “risk” was as a synonym for the much older word “hazard”, meaning a potential source of harm. This definition comes from Blount’s “Glossographia” (1661)[6] and was the main definition in the OED 1st (1914) and 2nd (1989) editions. Modern equivalents refer to “unwanted events” [7] or “something bad that might happen”.[1]
“Chance of harm”. This definition comes from Johnson’s “Dictionary of the English Language” (1755), and has been widely paraphrased, including “possibility of loss” [8] or “probability of unwanted events”.[7]
“Uncertainty about loss”. This definition comes from Willett’s “Economic Theory of Risk and Insurance” (1901).[9] This links “risk” to “uncertainty”, which is a broader term than chance or probability.
“Measurable uncertainty”. This definition comes from Knight’s “Risk, Uncertainty and Profit” (1921).[10] It allows “risk” to be used equally for positive and negative outcomes. In insurance, risk involves situations with unknown outcomes but known probability distributions.[11]
“Variance of return”. This definition comes from Markovitz’s “Portfolio Selection” (1952).[12] In practice, risk was measured using the standard deviation of returns.[7]
“Statistically expected loss”. The probability of an event multiplied by its magnitude was proposed as a definition of risk by the US Nuclear Regulatory Commission (1975),[13] and remains widely used.[7] It is properly known as the ‘expectation value’.
“Likelihood and severity of events”. The “triplet” definition of risk as “scenarios, probabilities and consequences” was proposed by Kaplan & Garrick (1981).[14] Many definitions refer to the likelihood/probability of events/effects/losses of different severity/consequence, e.g. ISO Guide 73 Note 4.[4]
“Consequences and associated uncertainty”. This was proposed by Kaplan & Garrick (1981).[14] This definition is preferred in Bayesian analysis, which sees risk as the combination of events and uncertainties about them.[15]
“Uncertain events affecting objectives”. This definition was adopted by the Association for Project Management (1997).[16][17] With slight rewording it became the definition in ISO Guide 73.[4]
“Uncertainty of outcome”. This definition was adopted by the UK Cabinet Office (2002)[18] to encourage innovation to improve public services. It allowed “risk” to describe either “positive opportunity or negative threat of actions and events”.
“Human interaction with uncertainty”. This definition comes from Cline (2015)[19] in the context of adventure education.

Some resolve these differences by arguing that the definition of risk is subjective. For example:

No definition is advanced as the correct one, because there is no one definition that is suitable for all problems. Rather, the choice of definition is a political one, expressing someone’s views regarding the importance of different adverse effects in a particular situation.[20]

The Society for Risk Analysis concludes that “experience has shown that to agree on one unified set of definitions is not realistic”. The solution is “to allow for different perspectives on fundamental concepts and make a distinction between overall qualitative definitions and their associated measurements.”[2]

Practice areas

Risk is ubiquitous in all areas of life and risk management is something that we all must do, whether we are managing a major organisation or simply crossing the road. When describing risk however, it is convenient to consider that risk practitioners operate in some specific practice areas.

Economic risk

Economic risks can be manifested in lower incomes or higher expenditures than expected. The causes can be many, for instance, the hike in the price for raw materials, the lapsing of deadlines for construction of a new operating facility, disruptions in a production process, emergence of a serious competitor on the market, the loss of key personnel, the change of a political regime, or natural disasters.

Health

Risks in personal health may be reduced by primary prevention actions that decrease early causes of illness or by secondary prevention actions after a person has clearly measured clinical signs or symptoms recognised as risk factors. Tertiary prevention reduces the negative impact of an already established disease by restoring function and reducing disease-related complications. Ethical medical practice requires careful discussion of risk factors with individual patients to obtain informed consent for secondary and tertiary prevention efforts, whereas public health efforts in primary prevention require education of the entire population at risk. In each case, careful communication about risk factors, likely outcomes and certainty must distinguish between causal events that must be decreased and associated events that may be merely consequences rather than causes.

In epidemiology, the lifetime risk of an effect is the cumulative incidence, also called incidence proportion over an entire lifetime.[21]

Health, safety, and environment

In terms of occupational health & safety management, the term 'risk' may be defined as the most likely consequence of a hazard, combined with the likelihood or probability of it occurring.

Health, safety, and environment (HSE) are separate practice areas; however, they are often linked. The reason for this is typically to do with organizational management structures; however, there are strong links among these disciplines. One of the strongest links between these is that a single risk event may have impacts in all three areas, albeit over differing timescales. For example, the uncontrolled release of radiation or a toxic chemical may have immediate short-term safety consequences, more protracted health impacts, and much longer-term environmental impacts. Events such as Chernobyl, for example, caused immediate deaths, and in the longer term, deaths from cancers, and left a lasting environmental impact leading to birth defects, impacts on wildlife, etc.

Over time, a form of risk analysis called environmental risk analysis has developed. Environmental risk analysis is a field of study that attempts to understand events and activities that bring risk to human health or the environment.[22]

Human health and environmental risk is the likelihood of an adverse outcome (See adverse outcome pathway). As such, risk is a function of hazard and exposure. Hazard is the intrinsic danger or harm that is posed, e.g. the toxicity of a chemical compound. Exposure is the likely contact with that hazard. Therefore, the risk of even a very hazardous substance approaches zero as the exposure nears zero, given a person's (or other organism's) biological makeup, activities and location (See exposome).[23] Another example of health risks are when certain behaviours, such as risky sexual behaviours, increase the likelihood of contracting HIV.[24]

Information technology and information security

Information technology risk, or IT risk, IT-related risk, is a risk related to information technology. This relatively new term was developed as a result of an increasing awareness that information security is simply one facet of a multitude of risks that are relevant to IT and the real world processes it supports.

The increasing dependencies of modern society on information and computers networks (both in private and public sectors, including military)[25][26][27] has led to new terms like IT risk and Cyberwarfare.

Information security means protecting information and information systems from unauthorised access, use, disclosure, disruption, modification, perusal, inspection, recording or destruction.[28] Information security grew out of practices and procedures of computer security.
Information security has grown to information assurance (IA) i.e. is the practice of managing risks related to the use, processing, storage, and transmission of information or data and the systems and processes used for those purposes.
While focused dominantly on information in digital form, the full range of IA encompasses not only digital but also analogue or physical form.
Information assurance is interdisciplinary and draws from multiple fields, including accounting, fraud examination, forensic science, management science, systems engineering, security engineering, and criminology, in addition to computer science.

So, IT risk is narrowly focused on computer security, while information security extends to risks related to other forms of information (paper, microfilm). Information assurance risks include the ones related to the consistency of the business information stored in IT systems and the information stored by other means and the relevant business consequences.

Insurance

Insurance is a risk treatment option which involves risk sharing. It can be considered as a form of contingent capital and is akin to purchasing an option in which the buyer pays a small premium to be protected from a potential large loss.

Insurance risk is often taken by insurance companies, who then bear a pool of risks including market risk, credit risk, operational risk, interest rate risk, mortality risk, longevity risks, etc.[29]

Business and management

Means of assessing risk vary widely between professions. Indeed, they may define these professions; for example, a doctor manages medical risk, while a civil engineer manages risk of structural failure. A professional code of ethics is usually focused on risk assessment and mitigation (by the professional on behalf of client, public, society or life in general).

In the workplace, incidental and inherent risks exist. Incidental risks are those that occur naturally in the business but are not part of the core of the business. Inherent risks have a negative effect on the operating profit of the business.

In human services

The experience of many people who rely on human services for support is that 'risk' is often used as a reason to prevent them from gaining further independence or fully accessing the community, and that these services are often unnecessarily risk averse.[30] "People's autonomy used to be compromised by institution walls, now it's too often our risk management practices", according to John O'Brien.[31] Michael Fischer and Ewan Ferlie (2013) find that contradictions between formal risk controls and the role of subjective factors in human services (such as the role of emotions and ideology) can undermine service values, so producing tensions and even intractable and 'heated' conflict.[32]

High reliability organisations (HROs)

A high reliability organisation (HRO) is an organisation that has succeeded in avoiding catastrophes in an environment where normal accidents can be expected due to risk factors and complexity. Most studies of HROs involve areas such as nuclear aircraft carriers, air traffic control, aerospace and nuclear power stations. Organizations such as these share in common the ability to consistently operate safely in complex, interconnected environments where a single failure in one component could lead to catastrophe. Essentially, they are organisations which appear to operate 'in spite' of an enormous range of risks.

Some of these industries manage risk in a highly quantified and enumerated way. These include the nuclear power and aircraft industries, where the possible failure of a complex series of engineered systems could result in highly undesirable outcomes. The usual measure of risk for a class of events is then: R = probability of the event × the severity of the consequence.

The total risk is then the sum of the individual class-risks; see below.[33]

In the nuclear industry, consequence is often measured in terms of off-site radiological release, and this is often banded into five or six-decade-wide bands.

The risks are evaluated using fault tree/event tree techniques (see safety engineering). Where these risks are low, they are normally considered to be "broadly acceptable". A higher level of risk (typically up to 10 to 100 times what is considered broadly acceptable) has to be justified against the costs of reducing it further and the possible benefits that make it tolerable—these risks are described as "Tolerable if ALARP", where ALARP stands for "as low as reasonably practicable". Risks beyond this level are classified as "intolerable".

The level of risk deemed broadly acceptable has been considered by regulatory bodies in various countries—an early attempt by UK government regulator and academic F. R. Farmer used the example of hill-walking and similar activities, which have definable risks that people appear to find acceptable. This resulted in the so-called Farmer Curve of acceptable probability of an event versus its consequence.

The technique as a whole is usually referred to as probabilistic risk assessment (PRA) (or probabilistic safety assessment, PSA). See WASH-1400 for an example of this approach.

Finance

In finance, risk is the chance that the return achieved on an investment will be different from that expected, and also takes into account the size of the difference. This includes the possibility of losing some or all of the original investment. In a view advocated by Damodaran, risk includes not only "downside risk" but also "upside risk" (returns that exceed expectations).[34] Some regard the standard deviation of the historical returns or average returns of a specific investment as providing some historical measure of risk; see modern portfolio theory. Financial risk may be market-dependent, determined by numerous market factors, or operational, resulting from fraudulent behaviour (e.g. Bernard Madoff).

A fundamental idea in finance is the relationship between risk and return (see modern portfolio theory). The greater the potential return one might seek, the greater the risk that one generally assumes. A free market reflects this principle in the pricing of an instrument: strong demand for a safer instrument drives its price higher (and its return correspondingly lower) while weak demand for a riskier instrument drives its price lower (and its potential return thereby higher). For example, a US Treasury bond is considered to be one of the safest investments. In comparison to an investment or speculative grade corporate bond, US Treasury notes and bonds yield lower rates of return. The reason for this is that a corporation is more likely to default on debt than the US government. Because the risk of investing in a corporate bond is higher, investors are offered a correspondingly higher rate of return.

A popular risk measure is Value-at-Risk (VaR).

There are different types of VaR: long term VaR, marginal VaR, factor VaR and shock VaR. The latter is used in measuring risk during the extreme market stress conditions.

In finance, risk has no single definition.

Artzner et al.[35] write "we call risk the investor's future net worth". In Novak[36] "risk is a possibility of an undesirable event".

In financial markets, one may need to measure credit risk, information timing and source risk, probability model risk, operational risk and legal risk if there are regulatory or civil actions taken as a result of "investor's regret".

With the advent of automation in financial markets, the concept of "real-time risk" has gained a lot of attention. Aldridge and Krawciw[37] define real-time risk as the probability of instantaneous or near-instantaneous loss, and can be due to flash crashes, other market crises, malicious activity by selected market participants and other events. A well-cited example[38] of real-time risk was a US$440 million loss incurred within 30 minutes by Knight Capital Group (KCG) on 1 August 2012; the culprit was a poorly-tested runaway algorithm deployed by the firm. Regulators have taken notice of real-time risk as well. Basel III[39] requires real-time risk management framework for bank stability.

It is not always obvious if financial instruments are "hedging" (purchasing/selling a financial instrument specifically to reduce or cancel out the risk in another investment) or "speculation" (increasing measurable risk and exposing the investor to catastrophic loss in pursuit of very high windfalls that increase expected value).

Some people may be "risk seeking", i.e. their utility function's second derivative is positive. Such an individual willingly pays a premium to assume risk (e.g. buys a lottery ticket).

Financial auditing

The financial audit risk model expresses the risk of an auditor providing an inappropriate opinion (or material misstatement) of a commercial entity's financial statements. It can be analytically expressed as

where AR is audit risk, IR is inherent risk, CR is control risk and DR is detection risk.

Note: As defined, audit risk does not consider the impact of an auditor misstatement and so is stated as a simple probability. The impact of misstatement must be considered when determining an acceptable audit risk.[40]

Security

AT YOUR OWN RISK
Popular labelling

Security risk management involves protection of assets from harm caused by deliberate acts. A more detailed definition is: "A security risk is any event that could result in the compromise of organizational assets i.e. the unauthorized use, loss, damage, disclosure or modification of organizational assets for the profit, personal interest or political interests of individuals, groups or other entities constitutes a compromise of the asset, and includes the risk of harm to people. Compromise of organizational assets may adversely affect the enterprise, its business units and their clients. As such, consideration of security risk is a vital component of risk management."[41]

Human factors

One of the growing areas of focus in risk management is the field of human factors where behavioural and organizational psychology underpin our understanding of risk based decision making. This field considers questions such as "how do we make risk based decisions?", "why are we irrationally more scared of sharks and terrorists than we are of motor vehicles and medications?"

In decision theory, regret (and anticipation of regret) can play a significant part in decision-making, distinct from risk aversion[42][43](preferring the status quo in case one becomes worse off).

Framing[44] is a fundamental problem with all forms of risk assessment. In particular, because of bounded rationality (our brains get overloaded, so we take mental shortcuts), the risk of extreme events is discounted because the probability is too low to evaluate intuitively. As an example, one of the leading causes of death is road accidents caused by drunk driving – partly because any given driver frames the problem by largely or totally ignoring the risk of a serious or fatal accident.

For instance, an extremely disturbing event (an attack by hijacking, or moral hazards) may be ignored in analysis despite the fact it has occurred and has a nonzero probability. Or, an event that everyone agrees is inevitable may be ruled out of analysis due to greed or an unwillingness to admit that it is believed to be inevitable. These human tendencies for error and wishful thinking often affect even the most rigorous applications of the scientific method and are a major concern of the philosophy of science.

All decision-making under uncertainty must consider cognitive bias, cultural bias, and notational bias: No group of people assessing risk is immune to "groupthink": acceptance of obviously wrong answers simply because it is socially painful to disagree, where there are conflicts of interest.

Framing involves other information that affects the outcome of a risky decision. The right prefrontal cortex has been shown to take a more global perspective[45] while greater left prefrontal activity relates to local or focal processing.[46]

From the Theory of Leaky Modules[47] McElroy and Seta proposed that they could predictably alter the framing effect by the selective manipulation of regional prefrontal activity with finger tapping or monaural listening.[48] The result was as expected. Rightward tapping or listening had the effect of narrowing attention such that the frame was ignored. This is a practical way of manipulating regional cortical activation to affect risky decisions, especially because directed tapping or listening is easily done.

Psychology of risk taking

A growing area of research has been to examine various psychological aspects of risk taking. Researchers typically run randomised experiments with a treatment and control group to ascertain the effect of different psychological factors that may be associated with risk taking. Thus, positive and negative feedback about past risk taking can affect future risk taking. In an experiment, people who were led to believe they are very competent at decision making saw more opportunities in a risky choice and took more risks, while those led to believe they were not very competent saw more threats and took fewer risks.[49]

Maintenance

The concept of risk-based maintenance is an advanced form of Reliability centred maintenance. In case of chemical industries, apart from probability of failure, consequences of failure is also very important. Therefore, the selection of maintenance policies should be based on risk, instead of reliability. Risk-based maintenance methodology acts as a tool for maintenance planning and decision making to reduce the probability of failure and its consequences. In risk-based maintenance decision making, the maintenance resources can be used optimally based on the risk class (high, medium, or low) of equipment or machines, to achieve tolerable risk criteria.[50]

Cybersecurity[51]

Closely related to information assurance and security risk, cybersecurity is the application of system security engineering[52] in order to address the compromise of company cyber-assets required for business or mission purposes. In order to address cyber-risk, cybersecurity applies security to the supply chain, the design and production environment for a product or service, and the product itself in order to provide efficient and appropriate security commensurate with the value of the asset to the mission or business process.

Risk assessment and analysis

Since risk assessment and management is essential in security management, both are tightly related. Security assessment methodologies like CRAMM contain risk assessment modules as an important part of the first steps of the methodology. On the other hand, risk assessment methodologies like Mehari evolved to become security assessment methodologies. An ISO standard on risk management (Principles and guidelines on implementation) was published under code ISO 31000 on 13 November 2009.

Quantitative analysis

There are many formal methods used to "measure" risk.

Often the probability of a negative event is estimated by using the frequency of past similar events. Probabilities for rare failures may be difficult to estimate. This makes risk assessment difficult in hazardous industries, for example nuclear energy, where the frequency of failures is rare, while harmful consequences of failure are severe.

Statistical methods may also require the use of a cost function, which in turn may require the calculation of the cost of loss of a human life. This is a difficult problem. One approach is to ask what people are willing to pay to insure against death[53] or radiological release (e.g. GBq of radio-iodine), but as the answers depend very strongly on the circumstances it is not clear that this approach is effective.

Risk is often measured as the expected value of an undesirable outcome. This combines the probabilities of various possible events and some assessment of the corresponding harm into a single value. See also Expected utility. The simplest case is a binary possibility of Accident or No accident. The associated formula for calculating risk is then:

For example, if performing activity X has a probability of 0.01 of suffering an accident of A, with a loss of 1000, then total risk is a loss of 10, the product of 0.01 and 1000.

Situations are sometimes more complex than the simple binary possibility case. In a situation with several possible accidents, total risk is the sum of the risks for each different accident, provided that the outcomes are comparable:

For example, if performing activity X has a probability of 0.01 of suffering an accident of A, with a loss of 1000, and a probability of 0.000001 of suffering an accident of type B, with a loss of 2,000,000, then total loss expectancy is 12, which is equal to a loss of 10 from an accident of type A and 2 from an accident of type B.

One of the first major uses of this concept was for the planning of the Delta Works in 1953, a flood protection program in the Netherlands, with the aid of the mathematician David van Dantzig.[54] The kind of risk analysis pioneered there has become common today in fields like nuclear power, aerospace and the chemical industry.

In statistical decision theory, the risk function is defined as the expected value of a given loss function as a function of the decision rule used to make decisions in the face of uncertainty.

Fear as intuitive risk assessment

People may rely on their fear and hesitation to keep them out of the most profoundly unknown circumstances. Fear is a response to perceived danger. Risk could be said to be the way we collectively measure and share this "true fear"—a fusion of rational doubt, irrational fear, and a set of unquantified biases from our own experience.

The field of behavioural finance focuses on human risk-aversion, asymmetric regret, and other ways that human financial behaviour varies from what analysts call "rational". Risk in that case is the degree of uncertainty associated with a return on an asset. Recognizing and respecting the irrational influences on human decision making may do much to reduce disasters caused by naive risk assessments that presume rationality but in fact merely fuse many shared biases.

Anxiety, risk and decision making

Fear, anxiety and risk

According to one set of definitions, fear is a fleeting emotion ascribed to a particular object, while anxiety is a trait of fear (this is referring to "trait anxiety", as distinct from how the term "anxiety" is generally used) that lasts longer and is not attributed to a specific stimulus (these particular definitions are not used by all authors cited on this page).[55] Some studies show a link between anxious behaviour and risk (the chance that an outcome will have an unfavorable result).[56] Joseph Forgas introduced valence based research where emotions are grouped as either positive or negative (Lerner and Keltner, 2000). Positive emotions, such as happiness, are believed to have more optimistic risk assessments and negative emotions, such as anger, have pessimistic risk assessments. As an emotion with a negative valence, fear, and therefore anxiety, has long been associated with negative risk perceptions. Under the more recent appraisal tendency framework of Jennifer Lerner et al., which refutes Forgas' notion of valence and promotes the idea that specific emotions have distinctive influences on judgments, fear is still related to pessimistic expectations.[57]

Psychologists have demonstrated that increases in anxiety and increases in risk perception are related and people who are habituated to anxiety experience this awareness of risk more intensely than normal individuals.[58] In decision-making, anxiety promotes the use of biases and quick thinking to evaluate risk. This is referred to as affect-as-information according to Clore, 1983. However, the accuracy of these risk perceptions when making choices is not known.[59]

Consequences of anxiety

Experimental studies show that brief surges in anxiety are correlated with surges in general risk perception.[59] Anxiety exists when the presence of threat is perceived (Maner and Schmidt, 2006).[58] As risk perception increases, it stays related to the particular source impacting the mood change as opposed to spreading to unrelated risk factors.[59] This increased awareness of a threat is significantly more emphasised in people who are conditioned to anxiety.[60] For example, anxious individuals who are predisposed to generating reasons for negative results tend to exhibit pessimism.[60] Also, findings suggest that the perception of a lack of control and a lower inclination to participate in risky decision-making (across various behavioural circumstances) is associated with individuals experiencing relatively high levels of trait anxiety.[58] In the previous instance, there is supporting clinical research that links emotional evaluation (of control), the anxiety that is felt and the option of risk avoidance.[58]

There are various views presented that anxious/fearful emotions cause people to access involuntary responses and judgments when making decisions that involve risk. Joshua A. Hemmerich et al. probes deeper into anxiety and its impact on choices by exploring "risk-as-feelings" which are quick, automatic, and natural reactions to danger that are based on emotions. This notion is supported by an experiment that engages physicians in a simulated perilous surgical procedure. It was demonstrated that a measurable amount of the participants' anxiety about patient outcomes was related to previous (experimentally created) regret and worry and ultimately caused the physicians to be led by their feelings over any information or guidelines provided during the mock surgery. Additionally, their emotional levels, adjusted along with the simulated patient status, suggest that anxiety level and the respective decision made are correlated with the type of bad outcome that was experienced in the earlier part of the experiment.[61] Similarly, another view of anxiety and decision-making is dispositional anxiety where emotional states, or moods, are cognitive and provide information about future pitfalls and rewards (Maner and Schmidt, 2006). When experiencing anxiety, individuals draw from personal judgments referred to as pessimistic outcome appraisals. These emotions promote biases for risk avoidance and promote risk tolerance in decision-making.[60]

Dread risk

It is common for people to dread some risks but not others: They tend to be very afraid of epidemic diseases, nuclear power plant failures, and plane accidents but are relatively unconcerned about some highly frequent and deadly events, such as traffic crashes, household accidents, and medical errors. One key distinction of dreadful risks seems to be their potential for catastrophic consequences,[62] threatening to kill a large number of people within a short period of time.[63] For example, immediately after the 11 September attacks, many Americans were afraid to fly and took their car instead, a decision that led to a significant increase in the number of fatal crashes in the time period following the 9/11 event compared with the same time period before the attacks.[64][65]

Different hypotheses have been proposed to explain why people fear dread risks. First, the psychometric paradigm[62] suggests that high lack of control, high catastrophic potential, and severe consequences account for the increased risk perception and anxiety associated with dread risks. Second, because people estimate the frequency of a risk by recalling instances of its occurrence from their social circle or the media, they may overvalue relatively rare but dramatic risks because of their overpresence and undervalue frequent, less dramatic risks.[65] Third, according to the preparedness hypothesis, people are prone to fear events that have been particularly threatening to survival in human evolutionary history.[66] Given that in most of human evolutionary history people lived in relatively small groups, rarely exceeding 100 people,[67] a dread risk, which kills many people at once, could potentially wipe out one's whole group. Indeed, research found[68] that people's fear peaks for risks killing around 100 people but does not increase if larger groups are killed. Fourth, fearing dread risks can be an ecologically rational strategy.[69] Besides killing a large number of people at a single point in time, dread risks reduce the number of children and young adults who would have potentially produced offspring. Accordingly, people are more concerned about risks killing younger, and hence more fertile, groups.[70]

Anxiety and judgmental accuracy

The relationship between higher levels of risk perception and "judgmental accuracy" in anxious individuals remains unclear (Joseph I. Constans, 2001). There is a chance that "judgmental accuracy" is correlated with heightened anxiety. Constans conducted a study to examine how worry propensity (and current mood and trait anxiety) might influence college student's estimation of their performance on an upcoming exam, and the study found that worry propensity predicted subjective risk bias (errors in their risk assessments), even after variance attributable to current mood and trait anxiety had been removed.[59] Another experiment suggests that trait anxiety is associated with pessimistic risk appraisals (heightened perceptions of the probability and degree of suffering associated with a negative experience), while controlling for depression.[58]

Other considerations

Risk and uncertainty

In his seminal work Risk, Uncertainty, and Profit, Frank Knight (1921) established the distinction between risk and uncertainty.

... Uncertainty must be taken in a sense radically distinct from the familiar notion of Risk, from which it has never been properly separated. The term "risk," as loosely used in everyday speech and in economic discussion, really covers two things which, functionally at least, in their causal relations to the phenomena of economic organization, are categorically different. ... The essential fact is that "risk" means in some cases a quantity susceptible of measurement, while at other times it is something distinctly not of this character; and there are far-reaching and crucial differences in the bearings of the phenomenon depending on which of the two is really present and operating. ... It will appear that a measurable uncertainty, or "risk" proper, as we shall use the term, is so far different from an unmeasurable one that it is not in effect an uncertainty at all. We ... accordingly restrict the term "uncertainty" to cases of the non-quantitive type.:[71]

Thus, Knightian uncertainty is immeasurable, not possible to calculate, while in the Knightian sense risk is measurable.

Another distinction between risk and uncertainty is proposed by Douglas Hubbard:[72][73]

Uncertainty: The lack of complete certainty, that is, the existence of more than one possibility. The "true" outcome/state/result/value is not known.
Measurement of uncertainty: A set of probabilities assigned to a set of possibilities. Example: "There is a 60% chance this market will double in five years"
Risk: A state of uncertainty where some of the possibilities involve a loss, catastrophe, or other undesirable outcome.
Measurement of risk: A set of possibilities each with quantified probabilities and quantified losses. Example: "There is a 40% chance the proposed oil well will be dry with a loss of $12 million in exploratory drilling costs".

In this sense, one may have uncertainty without risk but not risk without uncertainty. We can be uncertain about the winner of a contest, but unless we have some personal stake in it, we have no risk. If we bet money on the outcome of the contest, then we have a risk. In both cases there are more than one outcome. The measure of uncertainty refers only to the probabilities assigned to outcomes, while the measure of risk requires both probabilities for outcomes and losses quantified for outcomes.

Risk attitude, appetite and tolerance

The terms risk attitude, appetite, and tolerance are often used similarly to describe an organisation's or individual's attitude towards risk-taking. One's attitude may be described as risk-averse, risk-neutral, or risk-seeking. Risk tolerance looks at acceptable/unacceptable deviations from what is expected. Risk appetite looks at how much risk one is willing to accept. There can still be deviations that are within a risk appetite. For example, recent research finds that insured individuals are significantly likely to divest from risky asset holdings in response to a decline in health, controlling for variables such as income, age, and out-of-pocket medical expenses.[74]

Gambling is a risk-increasing investment, wherein money on hand is risked for a possible large return, but with the possibility of losing it all. Purchasing a lottery ticket is a very risky investment with a high chance of no return and a small chance of a very high return. In contrast, putting money in a bank at a defined rate of interest is a risk-averse action that gives a guaranteed return of a small gain and precludes other investments with possibly higher gain. The possibility of getting no return on an investment is also known as the rate of ruin.

Risk as a vector quantity

Hubbard also argues that defining risk as the product of impact and probability presumes, unrealistically, that decision-makers are risk-neutral.[73] A risk-neutral person's utility is proportional to the expected value of the payoff. For example, a risk-neutral person would consider 20% chance of winning $1 million exactly as desirable as getting a certain $200,000. However, most decision-makers are not actually risk-neutral and would not consider these equivalent choices. This gave rise to prospect theory and cumulative prospect theory. Hubbard proposes to instead describe risk as a vector quantity that distinguishes the probability and magnitude of a risk. Risks are simply described as a set or function of possible payoffs (gains or losses) with their associated probabilities. This array is collapsed into a scalar value according to a decision-maker's risk tolerance.

This is a list of books about risk issues.

Title Author(s) Year
Acceptable RiskBaruch Fischhoff, Sarah Lichtenstein, Paul Slovic, Steven L. Derby, and Ralph Keeney1984
Against the Gods: The Remarkable Story of RiskPeter L. Bernstein1996
At risk: Natural hazards, people's vulnerability and disastersPiers Blaikie, Terry Cannon, Ian Davis, and Ben Wisner1994
Building Safer Communities. Risk Governance, Spatial Planning and Responses to Natural HazardsUrbano Fra Paleo2009
Dangerous Earth: An introduction to geologic hazardsBarbara W. Murck, Brian J. Skinner, Stephen C. Porter1998
Disasters and DemocracyRutherford H. Platt1999
Earth Shock: Hurricanes, volcanoes, earthquakes, tornadoes and other forces of natureW. Andrew Robinson1993
Human System Response to Disaster: An Inventory of Sociological FindingsThomas E. Drabek1986
Judgment Under Uncertainty: heuristics and biasesDaniel Kahneman, Paul Slovic, and Amos Tversky1982
Mapping Vulnerability: disasters, development, and peopleGreg Bankoff, Georg Frerks, and Dorothea Hilhorst2004
Man and Society in Calamity: The Effects of War, Revolution, Famine, Pestilence upon Human Mind, Behavior, Social Organization and Cultural LifePitirim Sorokin1942
Mitigation of Hazardous Comets and AsteroidsMichael J.S. Belton, Thomas H. Morgan, Nalin H. Samarasinha, Donald K. Yeomans2005
Natural Disaster Hotspots: a global risk analysisMaxx Dilley2005
Natural Hazard Mitigation: Recasting disaster policy and planningDavid Godschalk, Timothy Beatley, Philip Berke, David Brower, and Edward J. Kaiser1999
Natural Hazards: Earth’s processes as hazards, disasters, and catastrophesEdward A. Keller, and Robert H. Blodgett2006
Normal Accidents. Living with high-risk technologiesCharles Perrow1984
Paying the Price: The status and role of insurance against natural disasters in the United StatesHoward Kunreuther, and Richard J. Roth1998
Planning for Earthquakes: Risks, politics, and policyPhilip R. Berke, and Timothy Beatley1992
Practical Project Risk Management: The ATOM MethodologyDavid Hillson and Peter Simon2012
Reduction and Predictability of Natural DisastersJohn B. Rundle, William Klein, Don L. Turcotte1996
Regions of Risk: A geographical introduction to disastersKenneth Hewitt1997
Risk Analysis: a quantitative guideDavid Vose2008
Risk: An introduction (ISBN 978-0-415-49089-4)Bernardus Ale2009
Risk and Culture: An essay on the selection of technical and environmental dangersMary Douglas, and Aaron Wildavsky1982
Socially Responsible Engineering: Justice in Risk Management (ISBN 978-0-471-78707-5)Daniel A. Vallero, and P. Aarne Vesilind2006
Swimming with Crocodiles: The Culture of Extreme Drinking Marjana Martinic and Fiona Measham (eds.) 2008
The Challenger Launch Decision: Risky Technology, Culture and Deviance at NASADiane Vaughan1997
The Environment as Hazard Ian Burton, Robert Kates, and Gilbert F. White1978
The Social Amplification of RiskNick Pidgeon, Roger E. Kasperson, and Paul Slovic2003
What is a Disaster? New answers to old questionsRonald W. Perry, and Enrico Quarantelli2005
Floods: From Risk to Opportunity (IAHS Red Book Series) Ali Chavoshian, and Kuniyoshi Takeuchi2013
The Risk Factor: Why Every Organization Needs Big Bets, Bold Characters, and the Occasional Spectacular FailureDeborah Perry Piscione2014

See also

  • Ambiguity aversion
  • Audit risk
  • Benefit shortfall
  • Civil defence
  • Countermeasure
  • Early case assessment
  • External risk
  • Enterprise risk
  • Event chain methodology
  • Financial risk
  • Fuel price risk management
  • Global catastrophic risk
  • Hazard (risk)
  • Identity resolution
  • Information assurance
  • Inherent risk
  • Inherent risk (accounting)
  • International Risk Governance Council
  • ISO/PAS 28000
  • IT risk
  • Legal risk
  • Life-critical system
  • Liquidity risk
  • Loss aversion
  • Moral hazard
  • Operational risk
  • Preventive maintenance
  • Probabilistic risk assessment
  • Process risk
  • Reputational risk
  • Reliability engineering
  • Risk analysis
  • Risk assessment
  • Risk compensation
    • Peltzman effect
  • Risk management
  • Risk-neutral measure
  • Risk perception
  • Risk register
  • Sampling risk
  • Systemic risk
  • Systematic risk
  • Uncertainty
  • Vulnerability

References

  1. "Risk". Cambridge Dictionary.
  2. "Glossary" (PDF). Society for Risk Analysis. Retrieved 13 April 2020.
  3. "risk". Oxford English Dictionary (3rd ed.). Oxford University Press. September 2005. (Subscription or UK public library membership required.)
  4. "Guide 73:2009 Risk Management - Vocabulary". ISO.
  5. "ISO 31000:2018 Risk Management - Guidelines". ISO.
  6. Blount, Thomas (1661). Glossographia, or, A dictionary interpreting all such hard words of whatsoever language now used in our refined English tongue. London.
  7. Hansson, Sven Ove, "Risk", The Stanford Encyclopedia of Philosophy (Fall 2018 Edition), Edward N. Zalta (ed.)
  8. "risk". Oxford English Dictionary (3rd ed.). Oxford University Press. September 2005. (Subscription or UK public library membership required.)
  9. Willett, Allan (1901). Economic Theory of Risk and Insurance. Columbia University Press. p. 6.
  10. Knight, Frank (1921). Risk, Uncertainty and Profit.
  11. Masci, Pietro (Spring 2011). "The History of Insurance: Risk, Uncertainty and Entrepreneurship". Journal of the Washington Institute of China Studies. 5 (3): 25-68. Retrieved 13 April 2020.
  12. Markovitz, H. (March 1952). "Portfolio Selection". The Journal of Finance. 7 (1): 77–91.
  13. Rasmussen (1975). An Assessment of Accident Risks in U.S. Commercial Nuclear Power Plants. US Nuclear Regulatory Commission.
  14. Kaplan, S.; Garrick, B.J. (1981). "On the Quantitative Definition of Risk". Risk Analysis. 1 (1).
  15. Aven, Terje (2011). Quantitative Risk Assessment – The Scientific Platform. Cambridge University Press.
  16. Project Risk Analysis and Management Guide. Association of Project Management. 1997.
  17. A Guide to the Project Management Body of Knowledge (4th Edition) ANSI/PMI 99-001-2008
  18. Risk: Improving government’s capability to handle risk and uncertainty (PDF). Cabinet Office Strategy Unit. 2002.
  19. Cline, Preston B. (3 March 2015). "The Merging of Risk Analysis and Adventure Education" (PDF). Wilderness Risk Management. 5 (1): 43–45. Retrieved 12 December 2016.
  20. Fischhoff, B; Watson, S.R.; Hope, C. (1984). "Defining Risk". Policy Sciences (17): 123–139.
  21. Rychetnik L, Hawe P, Waters E, Barratt A, Frommer M (July 2004). "A glossary for evidence based public health". J Epidemiol Community Health. 58 (7): 538–45. doi:10.1136/jech.2003.011585. PMC 1732833. PMID 15194712.
  22. Gurjar, Bhola Ram; Mohan, Manju (2002). "Environmental Risk Analysis: Problems and Perspectives in Different Countries". Risk: Health, Safety & Environment. 13: 3. Retrieved 23 March 2013.
  23. Vallero, Daniel A. (2016). "Environmental Biotechnology: A Biosystems Approach." Amsterdam: Academic Press. ISBN 978-0-12-407776-8.
  24. Potter, Patricia (2013). Fundamentals of nursing. St. Louis, Mo: Mosby Elsevier. p. 386. ISBN 9780323079334.
  25. Cortada, James W. (4 December 2003). The Digital Hand: How Computers Changed the Work of American Manufacturing, Transportation, and Retail Industries. USA: Oxford University Press. p. 512. ISBN 978-0-19-516588-3.
  26. Cortada, James W. (3 November 2005). The Digital Hand: Volume II: How Computers Changed the Work of American Financial, Telecommunications, Media, and Entertainment Industries. USA: Oxford University Press. ISBN 978-0-19-516587-6.
  27. Cortada, James W. (6 November 2007). The Digital Hand, Vol 3: How Computers Changed the Work of American Public Sector Industries. USA: Oxford University Press. p. 496. ISBN 978-0-19-516586-9.
  28. 44 U.S.C. § 3542(b)(1).
  29. Carson, James M.; Elyasiani, Elyas; Mansur, Iqbal (2008). "Market Risk, Interest Rate Risk, and Interdependencies in Insurer Stock Returns: A System-GARCH Model". The Journal of Risk and Insurance. 75 (4): 873–891. CiteSeerX 10.1.1.568.4087. doi:10.1111/j.1539-6975.2008.00289.x.
  30. A Positive Approach To Risk Requires Person Centred Thinking, Neill et al., Tizard Learning Disability Review http://pierprofessional.metapress.com/content/vr700311x66j0125/%5B%5D
  31. John O'Brien cited in Sanderson, H. Lewis, J. A Practical Guide to Delivering Personalisation; Person Centred Practice in Health and Social Care p211
  32. Fischer, Michael Daniel; Ferlie, Ewan (1 January 2013). "Resisting hybridisation between modes of clinical risk management: Contradiction, contest, and the production of intractable conflict" (PDF). Accounting, Organizations and Society. 38 (1): 30–49. doi:10.1016/j.aos.2012.11.002.
  33. Sebastián Martorell, Carlos Guedes Soares, Julie Barnett (2014). Safety, Reliability and Risk Analysis: Theory, Methods and Applications. CRC Press. p. 671. ISBN 9781482266481.CS1 maint: multiple names: authors list (link)
  34. Damodaran, Aswath (2003). Investment Philosophies: Successful Investment Philosophies and the Greatest Investors Who Made Them Work. Wiley. p. 15. ISBN 978-0-471-34503-9.
  35. Artzner, P.; Delbaen, F.; Eber, J.-M.; Heath, D. (1999). "Coherent measures of risk". Math. Finance. 9 (3): 203–228. doi:10.1111/1467-9965.00068.
  36. Novak S.Y. 2011. Extreme value methods with applications to finance. London: CRC. ISBN 978-1-43983-574-6.
  37. Aldridge, I., Krawciw, S., 2017. Real-Time Risk: What Investors Should Know About Fintech, High-Frequency Trading and Flash Crashes. Hoboken: Wiley. ISBN 978-1119318965.
  38. McCrank, John (17 October 2012). "Knight Capital posts $389.9 million loss on trading glitch". Reuters.
  39. "Basel III: international regulatory framework for banks". www.bis.org. 7 December 2017.
  40. Arco van de Ven. Marijn van Daelen, Christoph van der Elst (eds.). Risk Management and Corporate Governance: Interconnections in Law: Chapter: Risk Management from an accounting perspective. pp. 16–17.CS1 maint: uses editors parameter (link)
  41. Julian Talbot and Miles Jakeman Security Risk Management Body of Knowledge, John Wiley & Sons, 2009.
  42. Virine, L., & Trumper, M. ProjectThink. Gower. 2013
  43. Virine, L., & Trumper, M. Project Risk Analysis Made Ridiculously Simple. World Scientific Publishing. 2017
  44. Amos Tversky / Daniel Kahneman, 1981. "The Framing of Decisions and the Psychology of Choice."
  45. Schatz, J.; Craft, S.; Koby, M.; DeBaun, M. R. (2004). "Asymmetries in visual-spatial processing following childhood stroke". Neuropsychology. 18 (2): 340–352. doi:10.1037/0894-4105.18.2.340. PMID 15099156.
  46. Volberg, G.; Hubner, R. (2004). "On the role of response conflicts and stimulus position for hemispheric differences in global/local processing: An ERP study". Neuropsychologia (Submitted manuscript). 42 (13): 1805–1813. doi:10.1016/j.neuropsychologia.2004.04.017. PMID 15351629.
  47. Drake, R. A. (2004). Selective potentiation of proximal processes: Neurobiological mechanisms for spread of activation. Medical Science Monitor, 10, 231–234.
  48. McElroy, T.; Seta, J. J. (2004). "On the other hand, am I rational? Hemisphere activation and the framing effect" (PDF). Brain and Cognition. 55 (3): 572–580. doi:10.1016/j.bandc.2004.04.002. PMID 15223204.
  49. Krueger, Norris, and Peter R. Dickson. "How believing in ourselves increases risk taking: perceived self-efficacy and opportunity recognition." Decision Sciences 25, no. 3 (1994): 385–400.
  50. Arunraj, N.S.; Maiti, J. (2007). "Risk-based maintenance—Techniques and applications". Journal of Hazardous Materials. 142 (3): 653–661. doi:10.1016/j.jhazmat.2006.06.069. PMID 16887261.
  51. noun. The state of being protected against the criminal or unauthorized use of electronic data, or the measures taken to achieve this. 'But despite the number of agencies involved, cybersecurity generally seems to have slipped in importance for the administration
  52. Commerce, Computer Security Division, Information Technology Laboratory, National Institute of Standards and Technology, U.S. Department of (15 November 2016). "NIST Releases SP 800-160, Systems Security Engineering". csrc.nist.gov.
  53. Landsburg, Steven (3 March 2003). "Is your life worth $10 million?". Everyday Economics. Slate. Retrieved 17 March 2008.
  54. Wired Magazine, Before the levees break, page 3.
  55. Hartley, Catherine A.; Phelps, Elizabeth A. (2012). "Anxiety and Decision-Making". Biological Psychiatry. 72 (2): 113–118. doi:10.1016/j.biopsych.2011.12.027. PMC 3864559. PMID 22325982.
  56. Jon Gertner. What Are We Afraid Of, Money 32.5 (2003): 80.
  57. Lerner, Jennifer S.; Keltner, Dacher (2000). "Beyond Valence: Toward A Model of Emotion-Specific Influences on Judgment and Choice". Cognition & Emotion. 14 (4): 473–493. CiteSeerX 10.1.1.318.6023. doi:10.1080/026999300402763.
  58. Jon K. Maner, Norman B. Schmidt, The Role of Risk Avoidance in Anxiety, Behavior Therapy, Volume 37, Issue 2, June 2006, pp. 181–189, ISSN 0005-7894, 10.1016/j.beth.2005.11.003.
  59. Constans, Joseph I. (2001). "Worry propensity and the perception of risk". Behaviour Research and Therapy. 39 (6): 721–729. doi:10.1016/S0005-7967(00)00037-1. PMID 11400715.
  60. Jon K. Maner, J. Anthony Richey, Kiara Cromer, Mike Mallott, Carl W. Lejuez, Thomas E. Joiner, Norman B. Schmidt, Dispositional anxiety and risk-avoidant decision-making, Personality and Individual Differences, Volume 42, Issue 4, March 2007, pp. 665–675, ISSN 0191-8869, 10.1016/j.paid.2006.08.016.
  61. Joshua A. Hemmerich, Arthur S. Elstein, Margaret L. Schwarze, Elizabeth Ghini Moliski, William Dale, Risk as feelings in the effect of patient outcomes on physicians' future treatment decisions: A randomized trial and manipulation validation, Social Science & Medicine, Volume 75, Issue 2, July 2012, pp. 367–376, ISSN 0277-9536, 10.1016/j.socscimed.2012.03.020.
  62. Slovic, P (1987). "Perception of risk". Science. 236 (4799): 280–285. Bibcode:1987Sci...236..280S. doi:10.1126/science.3563507. PMID 3563507.
  63. Gigerenzer G (2004) Dread risk, 11 September, and fatal traffic accidents. Psych Sci 15:286−287.
  64. Gaissmaier, W.; Gigerenzer, G. (2012). "9/11, Act II: A fine-grained analysis of regional variations in traffic fatalities in the aftermath of the terrorist attacks". Psychological Science. 23 (12): 1449–1454. doi:10.1177/0956797612447804. PMID 23160203.
  65. Lichtenstein, S; Slovic, P; Fischhoff, B; Layman, M; Combs, B (1978). "Judged frequency of lethal events". Journal of Experimental Psychology: Human Learning and Memory. 4 (6): 551–578. doi:10.1037/0278-7393.4.6.551. hdl:1794/22549.
  66. Öhman, A; Mineka, S (2001). "Fears, phobias, and preparedness: Toward an evolved module of fear and fear learning". Psychol Rev. 108 (3): 483–522. doi:10.1037/0033-295x.108.3.483. PMID 11488376.
  67. Hill, KR; Walker, RS; Bozicevic, M; Eder, J; Headland, T; et al. (2011). "Co-residence patterns in hunter-gatherer societies show unique human social structure". Science. 331 (6022): 1286–1289. Bibcode:2011Sci...331.1286H. doi:10.1126/science.1199071. PMID 21393537.
  68. Galesic, M; Garcia-Retamero, R (2012). "The risks we dread: A social circle account". PLOS ONE. 7 (4): e32837. Bibcode:2012PLoSO...732837G. doi:10.1371/journal.pone.0032837. PMC 3324481. PMID 22509250.
  69. Bodemer, N.; Ruggeri, A.; Galesic, M. (2013). "When dread risks are more dreadful than continuous risks: Comparing cumulative population losses over time". PLOS One. 8 (6): e66544. Bibcode:2013PLoSO...866544B. doi:10.1371/journal.pone.0066544. PMC 3694073. PMID 23840503.
  70. Wang, XT (1996). "Evolutionary hypotheses of risk-sensitive choice: Age differences and perspective change". Ethol Sociobiol. 17: 1–15. CiteSeerX 10.1.1.201.816. doi:10.1016/0162-3095(95)00103-4.
  71. Frank Hyneman Knight "Risk, uncertainty and profit" pg. 19, Hart, Schaffner, and Marx Prize Essays, no. 31. Boston and New York: Houghton Mifflin. 1921.
  72. Douglas Hubbard "How to Measure Anything: Finding the Value of Intangibles in Business" pg. 46, John Wiley & Sons, 2007.
  73. Douglas Hubbard "The Failure of Risk Management: Why It's Broken and How to Fix It, John Wiley & Sons, 2009. Page 22 of https://canvas.uw.edu/courses/1066599/files/37549842/download?verifier=ar2VjVOxCU8sEQr23I5LEBpr89B6fnwmoJgBinqj&wrap=1
  74. Federal Reserve Bank of Chicago, Health and the Savings of Insured versus Uninsured, Working-Age Households in the U.S., November 2009

Bibliography

Referred literature

  • James Franklin, 2001: The Science of Conjecture: Evidence and Probability Before Pascal, Baltimore: Johns Hopkins University Press.
  • John Handmer and Paul James (2005). "Trust Us and Be Scared: The Changing Nature of Risk". Global Society. 21 (1): 119–30.CS1 maint: uses authors parameter (link)
  • Niklas Luhmann, 1996: Modern Society Shocked by its Risks (= University of Hong Kong, Department of Sociology Occasional Papers 17), Hong Kong, available via HKU Scholars HUB

Books

  • Historian David A. Moss' book When All Else Fails explains the US government's historical role as risk manager of last resort.
  • Bernstein P. L. Against the Gods ISBN 0-471-29563-9. Risk explained and its appreciation by man traced from earliest times through all the major figures of their ages in mathematical circles.
  • Rescher, Nicholas (1983). A Philosophical Introduction to the Theory of Risk Evaluation and Measurement. University Press of America.
  • Porteous, Bruce T.; Pradip Tapadar (December 2005). Economic Capital and Financial Risk Management for Financial Services Firms and Conglomerates. Palgrave Macmillan. ISBN 978-1-4039-3608-0.
  • Tom Kendrick (2003). Identifying and Managing Project Risk: Essential Tools for Failure-Proofing Your Project. AMACOM/American Management Association. ISBN 978-0-8144-0761-5.
  • Hillson D. (2007). Practical Project Risk Management: The Atom Methodology. Management Concepts. ISBN 978-1-56726-202-5.
  • Kim Heldman (2005). Project Manager's Spotlight on Risk Management. Jossey-Bass. ISBN 978-0-7821-4411-6.
  • Dirk Proske (2008). Catalogue of risks – Natural, Technical, Social and Health Risks. Eos Transactions. 90. Springer. p. 18. Bibcode:2009EOSTr..90...18E. doi:10.1029/2009EO020009. ISBN 978-3-540-79554-4.
  • Gardner D. Risk: The Science and Politics of Fear, Random House Inc. (2008) ISBN 0-7710-3299-4.
  • Novak S.Y. Extreme value methods with applications to finance. London: CRC. (2011) ISBN 978-1-43983-574-6.
  • Hopkin P. Fundamentals of Risk Management. 2nd Edition. Kogan-Page (2012) ISBN 978-0-7494-6539-1

Articles and papers

  • Cevolini, A (2015). ""Tempo e decisione. Perché Aristotele non-ha un concetto di rischio?" PDF". Divus Thomas. 118 (1): 221–249.
  • Clark, L.; Manes, F.; Antoun, N.; Sahakian, B. J.; Robbins, T. W. (2003). "The contributions of lesion laterality and lesion volume to decision-making impairment following frontal lobe damage". Neuropsychologia. 41 (11): 1474–1483. doi:10.1016/s0028-3932(03)00081-2. PMID 12849765.
  • Cokely, E. T.; Galesic, M.; Schulz, E.; Ghazal, S.; Garcia-Retamero, R. (2012). "Measuring risk literacy: The Berlin Numeracy Test" (PDF). Judgment and Decision Making. 7: 25–47.
  • Drake, R. A. (1985). "Decision making and risk taking: Neurological manipulation with a proposed consistency mediation". Contemporary Social Psychology. 11: 149–152.
  • Drake, R. A. (1985). "Lateral asymmetry of risky recommendations". Personality and Social Psychology Bulletin. 11 (4): 409–417. doi:10.1177/0146167285114007.
  • Gregory, Kent J.; Bibbo, Giovanni; Pattison, John E. (2005). "A Standard Approach to Measurement Uncertainties for Scientists and Engineers in Medicine". Australasian Physical and Engineering Sciences in Medicine. 28 (2): 131–139. doi:10.1007/bf03178705. PMID 16060321.
  • Hansson, Sven Ove. (2007). "Risk", The Stanford Encyclopedia of Philosophy (Summer 2007 Edition), Edward N. Zalta (ed.), forthcoming .
  • Holton, Glyn A. (2004). "Defining Risk", Financial Analysts Journal, 60 (6), 19–25. A paper exploring the foundations of risk. (PDF file).
  • Knight, F. H. (1921) Risk, Uncertainty and Profit, Chicago: Houghton Mifflin Company. (Cited at: , § I.I.26.).
  • Kruger, Daniel J., Wang, X.T., & Wilke, Andreas (2007) "Towards the development of an evolutionarily valid domain-specific risk-taking scale" Evolutionary Psychology (PDF file).
  • Metzner-Szigeth, Andreas (2009). "Contradictory approaches? On realism and constructivism in the social sciences research on risk, technology and the environment" (PDF). Futures. 41 (3): 156–170. doi:10.1016/j.futures.2008.09.017.
  • Miller, L (1985). "Cognitive risk taking after frontal or temporal lobectomy I. The synthesis of fragmented visual information". Neuropsychologia. 23 (3): 359–369. doi:10.1016/0028-3932(85)90022-3. PMID 4022303.
  • Miller, L.; Milner, B. (1985). "Cognitive risk taking after frontal or temporal lobectomy II. The synthesis of phonemic and semantic information". Neuropsychologia. 23 (3): 371–379. doi:10.1016/0028-3932(85)90023-5. PMID 4022304.
  • Neill, M. Allen, J. Woodhead, N. Reid, S. Irwin, L. Sanderson, H. 2008 "A Positive Approach to Risk Requires Person Centred Thinking" London, CSIP Personalisation Network, Department of Health. Available from: https://web.archive.org/web/20090218231745/http://networks.csip.org.uk/Personalisation/Topics/Browse/Risk/ [Accessed 21 July 2008].
  • Wildavsky, Aaron; Wildavsky, Adam (2008). "Risk and Safety". In David R. Henderson (ed.). Concise Encyclopedia of Economics (2nd ed.). Indianapolis: Library of Economics and Liberty. ISBN 978-0865976658. OCLC 237794267.
  • Risk – The entry of the Stanford Encyclopedia of Philosophy

This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.