Collarspace Discussion Forums


Home  Login  Search 

Risk Analysis - Stats and Math Help Sorely Needed.


View related threads: (in this forum | in all forums)

Logged in as: Guest
 
All Forums >> [Casual Banter] >> Off the Grid >> Risk Analysis - Stats and Math Help Sorely Needed. Page: [1]
Login
Message << Older Topic   Newer Topic >>
Risk Analysis - Stats and Math Help Sorely Needed. - 3/13/2007 8:02:13 PM   
FirmhandKY


Posts: 8948
Joined: 9/21/2004
Status: offline
Several years ago, I was trained in Risk Analysis, but have since lost most of my training material, although I do have some experience to fall back on, as well as a fair amount of training and education in statistical analysis.

But, I don't consider myself an expert at statistical analysis, and would like some opinions and interpretations about some interesting RA (Risk Analysis) material I happened upon.

Here is the ranking scale for a specific project I'm currently researching.  I have done some looking at their source material, but haven't been able to locate every reference in their supporting source material, but it appears not to be anything other than general RA material such as I have used before.

Here are the important information that they use in order to quantify the industrial risks for the project.  I'll place my specific questions afterwards:

13. A level of confidence, as defined in Table 3, can be used to characterize uncertainty that is based on expert judgment as to the correctness of a model, an analysis or a statement. The last two terms in this scale should be reserved for areas of major concern that need to be considered from a risk or opportunity perspective, and the reason for their use should be carefully explained.

Table 3. Quantitatively calibrated levels of confidence.

Terminology                     Degree of confidence in being correct
Very High confidence      At least 9 out of 10 chance of being correct
High confidence               About 8 out of 10 chance
Medium confidence          About 5 out of 10 chance
Low confidence                About 2 out of 10 chance
Very low confidence         Less than 1 out of 10 chance


14. Likelihood, as defined in Table 4, refers to a probabilistic assessment of some well defined outcome having occurred or occurring in the future. The categories defined in this table should be considered as having ‘fuzzy’ boundaries. Use other probability ranges where more appropriate but do not then use the terminology in table 4. Likelihood may be based on quantitative analysis or an elicitation of expert views. The central range of this scale should not be used to express a lack of knowledge – see paragraph 12 and Table 2 for that situation.

Table 4.                Likelihood Scale.

Terminology                Likelihood of the occurrence/ outcome
Virtually certain            > 99% probability of occurrence
Very likely                    > 90% probability
Likely                            > 66% probability
About as likely as not        33 to 66% probability
Unlikely                        < 33% probability
Very unlikely                < 10% probability
Exceptionally unlikely    < 1% probability

Questions.

1.  If, according to their analysis, they have a "high level of confidence" (8 out of 10 chance) of something being "likely" (>66% probability) ... what is the statistical probability of that event happening?

2.  In the "likely" category .... what is the statistical difference between a 66% chance of something happening, and an 89% chance of something happening?

3.  Without a detailed breakdown of the statistical background, and the math that was used to arrive at their "level of confidences" and their "Likelihood Scale", how confident are you in their conclusions?

Any input and thoughts would be appreciated.

FirmKY


_____________________________

Some people are just idiots.
Profile   Post #: 1
RE: Risk Analysis - Stats and Math Help Sorely Needed. - 3/13/2007 10:13:52 PM   
Termyn8or


Posts: 18681
Joined: 11/12/2005
Status: offline
Firm, where did you get your head screwed on, it seems straight. You know alignment is a PITA.

Anyway, from what I am reading, this looks like a test or study material for an insurance actuary. It is exactly what an actuary does.

As such, some of the evidence is heuristic and some is not. For example, you are an actuary analysing data for the constant rate adjustments. This is necessary for a company to do so their rates can remain competitive. The better job the actuary does in his portent, the higher the profit margin of the company.

Say you are analysing theft coverage, all the other components are handled by someone else. With your data the spreadsheets and all that will draw you a map practically automatically. Your job as an actuary is to determine what of these data are valid, and affective on actual probablility. You are also to determine their affective weight, as used to determine premium rates. Ideally you can attract customers with lower rates, if they are in the contingent of customers who are not likely to have a claim.

As far as the probablilities go you can almost use raw data. When they want predictions you need a bit of help. Computer or otherwise.

More later.

T

(in reply to FirmhandKY)
Profile   Post #: 2
RE: Risk Analysis - Stats and Math Help Sorely Needed. - 3/14/2007 12:25:08 AM   
FirmhandKY


Posts: 8948
Joined: 9/21/2004
Status: offline


PITA indeed, T.

It is a complex study, and a number of people are not comfortable with some of the math, or assumptions. 

I'm specifically concerned with the way that both probability and level of confidence can be viewed in a statistical sense i.e.  if something has an 8 in 10 chance of happening at the >66% likelihood of happening, then your actual probability of occurance would be reduced by at least 20%.  I think this would mean that your lower range limit would be reduced to 52.5% or thereabouts [66 - (.20 x 66) = 52.5].  But ... I'm not sure that it shouldn't be even lower, since I'm only reducing it by 20% of the remaining figure, not the entire base amount.

The second thing that concerns me is the "spread" between their "likelihood".  The "Likely" range is 89% to 66%, but I can see no real way to determine a difference between the two ends of the range.  In effect, they appear to be a difference without a distinction.  If so, then I feel that the lower limit should be the more defensible figure, and the more prudent to adopt.

The fact that they admit that any boundaries are "fuzzy" also concerns me.  What the hell does "fuzzy" mean, in this context?

As well, the method used to come up with the base percentages for likelihood AND levels of confidence is "expert judgment" or "expert views", especially as there seems to be differing "experts" dueling over the issue. 

There is some supporting statistical material, but there seems to be a disconnect between much of that material, and their later conclusions of "likelihood" and "confidence", with no clear linkage between the facts presented, and the conclusions drawn.


I know that when I did Risk Analysis professionally, I had to document each step between my data, and my probabilities and then explain my level of confidence in a statistical manner as well.

In the explanation I posted above, they seem to have much of the form, but not all the substance.

I thought there might be an insurance, risk analyst or mathematician that could lend me a hand.

I realize that this may all seem a little esoteric, but I hope to get the chance to explain in more detail.  I know that there are a lot of people on the OTF forums who have a wide variety of scientic and mathematical abilities (or so it seems ), so I was hoping for some input.

I know my "Missing Two Dollars" thread in "Random Stupidity" forum got some good answers when I posted a couple of weeks ago.  Maybe that's where I should have posted this one?

FirmKY

_____________________________

Some people are just idiots.

(in reply to Termyn8or)
Profile   Post #: 3
RE: Risk Analysis - Stats and Math Help Sorely Needed. - 3/14/2007 12:47:34 AM   
Wildfleurs


Posts: 1650
Joined: 9/24/2004
From: Connecticut
Status: offline
quote:

ORIGINAL: FirmhandKY

Questions.

1.  If, according to their analysis, they have a "high level of confidence" (8 out of 10 chance) of something being "likely" (>66% probability) ... what is the statistical probability of that event happening?

2.  In the "likely" category .... what is the statistical difference between a 66% chance of something happening, and an 89% chance of something happening?

3.  Without a detailed breakdown of the statistical background, and the math that was used to arrive at their "level of confidences" and their "Likelihood Scale", how confident are you in their conclusions?

Any input and thoughts would be appreciated.

FirmKY



I have a statistical background, and its 3:30 in the morning so take my comments for what they are:

1) The thing that is throwing me with their parameters of confidence is that there literally is a statistical test that you can do for confidence, which is distinct from a self-defined determination of confidence that they seem to be (setting) arbitrarily .

2) It looks like they are using Likert scales to set their parameters, which is why you'll never be able to fully tease out the difference between a 89% and 66% chance of something happening.  Thats just the bad that happens with doing groupings like that, which is inevitable for risk analysis (that I've done a bit of and seems to be largely a function of plugging in numbers in spreadsheets).

3) Not very confident - the devil is always in the details, especially details as general as what you provided would never be enough for me.

C~

< Message edited by Wildfleurs -- 3/14/2007 12:48:34 AM >


_____________________________

"Just because you've always done it that way doesn't mean it's not incredibly stupid." -despair.com

~~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~
The heart of it all - http://www.wildfleurs.com
~~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~*~

(in reply to FirmhandKY)
Profile   Post #: 4
RE: Risk Analysis - Stats and Math Help Sorely Needed. - 3/14/2007 10:09:08 AM   
FirmhandKY


Posts: 8948
Joined: 9/21/2004
Status: offline
quote:

ORIGINAL: Wildfleurs

quote:

ORIGINAL: FirmhandKY

Questions.

1.  If, according to their analysis, they have a "high level of confidence" (8 out of 10 chance) of something being "likely" (>66% probability) ... what is the statistical probability of that event happening?

2.  In the "likely" category .... what is the statistical difference between a 66% chance of something happening, and an 89% chance of something happening?

3.  Without a detailed breakdown of the statistical background, and the math that was used to arrive at their "level of confidences" and their "Likelihood Scale", how confident are you in their conclusions?

Any input and thoughts would be appreciated.

FirmKY



I have a statistical background, and its 3:30 in the morning so take my comments for what they are:

1) The thing that is throwing me with their parameters of confidence is that there literally is a statistical test that you can do for confidence, which is distinct from a self-defined determination of confidence that they seem to be (setting) arbitrarily .

2) It looks like they are using Likert scales to set their parameters, which is why you'll never be able to fully tease out the difference between a 89% and 66% chance of something happening.  Thats just the bad that happens with doing groupings like that, which is inevitable for risk analysis (that I've done a bit of and seems to be largely a function of plugging in numbers in spreadsheets).

3) Not very confident - the devil is always in the details, especially details as general as what you provided would never be enough for me.

C~


Wildfleurs,

Thank you very much. 

1. I had forgotten the details of both the mathematical confidence interval and it's associated measurement criteria, but I knew what I was seeing didn't sit well with me.

I agree that they seem to be having a problem with it.  I did go back and google "margin of error" and "confidence interval" and found at least three good sources:

a. Wikipedia confidence interval is pretty complete.  I found some interesting thoughts which I've placed in italics:

Confidence levels are typically given alongside statistics resulting from sampling.

In a statement: we are 90% confident that between 35% and 45% of voters favor Candidate A, 90% is our confidence level and 35%-45% is our confidence interval.

It is very tempting to misunderstand this statement in the following way.

[math deleted]

This conclusion does not follow from the laws of probability because θ is not a "random variable"; i.e., no probability distribution has been assigned to it. Confidence intervals are generally a frequentist method, i.e., employed by those who interpret "90% probability" as "occurring in 90% of all cases". Suppose, for example, that θ is the mass of the planet Neptune, and the randomness in our measurement error means that 90% of the time our statement that the mass is between this number and that number will be correct. The mass is not what is random. Therefore, given that we have measured it to be 82 units, we cannot say that in 90% of all cases, the mass is between 82 − 1.645 and 82 + 1.645. There are no such cases; there is, after all, only one planet Neptune.

But if probabilities are construed as degrees of belief rather than as relative frequencies of occurrence of random events, i.e., if we are Bayesians rather than frequentists, can we then say we are 90% sure that the mass is between 82 − 1.645 and 82 + 1.645? Many answers to this question have been proposed, and are philosophically controversial. The answer will not be a mathematical theorem, but a philosophical tenet. Less controversial are Bayesian credible intervals, in which one starts with a prior probability distribution of θ, and finds a posterior probability distribution, which is the conditional probability distribution of θ given the data.

For users of frequentist methods, the explanation of a confidence interval can amount to something like: "The confidence interval represents values for the population parameter for which the difference between the parameter and the observed estimate is not statistically significant at the 10% level". Critics of frequentist methods suggest that this hides the real and, to the critics, incomprehensible frequentist interpretation which might be expressed as: "If the population parameter in fact lies within the confidence interval, then the probability that the estimator either will be the estimate actually observed, or will be closer to the parameter, is less than or equal to 90%". Users of Bayesian methods, if they produced a confidence interval, might by contrast say "My degree of belief that the parameter is in fact in the confidence interval is 90%". Disagreements about these issues are not disagreements about solutions to mathematical problems. Rather they are disagreements about the ways in which mathematics is to be applied.

From this, it appears that they are using a Bayesian method of evaluation, and it's more about a philosophical ideal, than a mathematical formula.


b.  Bayesian Probability;

The Bayesian interpretation of probability allows probabilities to be assigned to all propositions (or, in some formulations, to the events signified by those propositions) in any reference class of events, independent of whether the events can be interpreted as having a relative frequency in repeated trials. Controversially, Bayesian probability assignments are relative to a (possibly hypothetical) rational subject: it is therefore not inconsistent to be able to assign different probabilities to the same proposition by Bayesian methods based on the same observed data.

Differences arise either because different models of the data generating process are used or because of different prior probability assignments to model parameters. Differences due to model differences are possible in both Frequentist and Bayesian analyses, but differences due to prior probabilities assignments are distinctive to Bayesian analyses. Such probabilities are sometimes called 'personal probabilities', although there may be no particular individual to whom they belong. Frequentists argue that this makes Bayesian analyses subjective, in the negative sense of 'not determined from the data'. Bayesians typically reply that differences due to alternative models of the data generating process are equally subjective, and that such model choices are also (ideally) chosen prior to analysis of the data, by the analyst. Since the analyst is also a natural person to assign prior probabilities, the two subjective inputs can be seen as deriving from the same source.

Although there is no reason why different interpretations (senses) of a word cannot be used in different contexts, there is a history of antagonism between Bayesians and frequentists, with the latter often rejecting the Bayesian interpretation as ill-grounded. The groups have also disagreed about which of the two senses reflects what is commonly meant by the term 'probable'. More importantly, the groups have agreed that Bayesian and Frequentist analyses answer genuinely different questions, but disagreed about which class of question it is more important to answer in scientific and engineering contexts.

A third source that I found seems to be the way I was taught about probability and confidence intervals in my stats classes, lo these many years ago, and is a "frequencists" interpretation:


b.  Confidence level calculator: (specifically used in context of mailing campaigns and responses):

Confidence Level is the likelihood - expressed as a percentage - that the results of a test are real and repeatable, and not just random. The idea is based on the concept of the "normal distribution curve," which shows that variation in almost any data (such as the heights of all fourth-graders, or the amount of rainfall in January) tends to be clustered around an average value, with relatively few individual measurements at the extremes.

So if your confidence level is, say, 92%, that means, according to probability theory, there's a 92% chance that you'd see similar results in a repeat of the test. (It does not mean you'd receive the same number of gifts, or that the difference between the packages would be the same. It only means that the package that received more gifts in the first test would be likely to receive more gifts in the second as well - unless, of course, some significant other factors have changed.)

A confidence level of 50% would mean the difference is truly random, with only a 50-50 chance that you'd see the same results in a repeat of the test. Even at 75% the odds are not good - there's a one in four chance that your results are meaningless. Some statisticians consider 90% to be the minimum confidence level for statistically significant results, and that's reportedly the standard used in many election polls. Others insist on a minimum of 95% to be considered significant. And in medical research, for obvious reasons, there's a strong preference for even higher levels of confidence.

It's important to remember that we're talking about probabilities, and there's no magic number that guarantees your results will be repeatable. While it's always best to have a confidence level of 95% or higher, you shouldn't ignore results in the 80% to 90% range. Those results may indicate trends and provide clues about how to improve your mailings; at the very least, they're worth re-testing, preferably in larger quantities. (In any test, a larger sample size will generally give more reliable results.)

Finally, a few words of caution: These confidence levels are only valid when you're comparing test panels that can be thought of as a single event. Don't try to adapt them to a situation that changes over time (such as your total number of active donors), or use it to compare appeals that mailed at different times (there are far too many uncontrolled variables in that case).

This source combines the use of probability and likelihood in the way that I was trained and understood, and makes sense to me without the reference to the Bayesian method of subjective confidence.

I think it points out the use of Likert scales (thanks WF, I couldn't remember the name before, so couldn't google it) - which are ordinal measures - as a percentage is an attempt to overcome the lack of a valid mathematical system to establish probability.

Overall, I am like you.  I have a low level of confidence in the conclusions of the report, based on their playing with "subjective" data as if it were "objective" and then attempting to overcome the lack by arbitarily assigning percentages of "confidence" based on a system which inherently doesn't function that way.

In other words, they seem to be using Bayesian logic disguised as Frequentist math in order to give it a higher degree of credibility. 

Or am I off base in my thoughts?

Anyone else have any thoughts?

FirmKY


_____________________________

Some people are just idiots.

(in reply to Wildfleurs)
Profile   Post #: 5
Page:   [1]
All Forums >> [Casual Banter] >> Off the Grid >> Risk Analysis - Stats and Math Help Sorely Needed. Page: [1]
Jump to:





New Messages No New Messages
Hot Topic w/ New Messages Hot Topic w/o New Messages
Locked w/ New Messages Locked w/o New Messages
 Post New Thread
 Reply to Message
 Post New Poll
 Submit Vote
 Delete My Own Post
 Delete My Own Thread
 Rate Posts




Collarchat.com © 2025
Terms of Service Privacy Policy Spam Policy

0.063