## Probability: Normal Distribution

### Introduction

The normal distribution is defined by a mathematical formula, namely, where we are given values for the mean, μ, and the standard deviation, σ. With such a definition we get a different value for each different combination of μ and σ. Fortunately, we can look at a single example, called the standard normal distribution, where we have μ=0 and σ=1. With those values the general formula becomes A graph of that function appears in Figure 1. Remember that as a probability density function, the area under the curve and above the axis is equal to 1 square unit. Each small rectangle in Figure 1 represents 0.5*0.05 or 0.025 square units. It takes 40 such areas to equal 1 square unit.

Figure 1 We must note here that the curve never gets to the x-axis. That is, the curve is always above the x-axis even for values more extreme than -4 or 4.

The normal distribution is often called a bell-shaped curve for the obvious reason that the graph of the function can be made to look somewhat like a bell. A careful "read" of the graph in Figure 1 reveals that we are calling values along the traditional x-axis z values. This is not a typo. It is quite important. We refer to the values along the traditional x-axis as z values; we will learn, later, how to change values from a non-standard normal distribution into z values.

The "can be made to look somewhat like" is added to point out that part of the visual effect of the curve depends upon not having equal x and y scales. In Figure 1 we go from -4 to 4 on the x scale, but we magnify the y scale by stretching it so that the vertical spread shows 0 to 0.40. If we had equal scales the curve would appear as in Figure 1a. This figure is no more or less accurate in displaying the normal curve, but Figure 1a no longer gives that impression of a "bell" shape.

Figure 1a Looking back at Figure 1 (or at Figure 1a if you like) we see that the normal distribution is a symmetric distribution. Since the probability that a random variable, X, is less than a given z value is the area under the curve and to the left of the vertical line at that value, this symmetry means that P(X<0)=0.5.

Because the standard normal distribution has mean zero, μ=0, and standard deviation one, σ=1, the region of the graph between -1 and 1 on the graph represents values that are within 1 standard deviation of the mean. The area under the curve between -1 and 1 is about 0.68. (A better value would be 0.68269, but we will just use 0.68 for now.) Therefore, again using symmetry, the red area in Figure 2 represents two pieces, each accounting for about 34% of the total area.

Figure 2 Knowing the values shown in Figure 2 we can say that P(X < -1) ≈ 0.16, or P(X > 1) ≈ 0.16, or P(X < 1) ≈ 0.84, or P(X < -1 or X > 1) ≈ 0.32. Quite a bit of knowledge from just one graph!

If we move out another standard deviation, we find that the region between -2 and 2 accounts for about 95% of the total area (1 square unit) under the curve. (A better value would be 0.954499876, but we will just use 0.95 for now.)

Figure 3 Using symmetry and the earlier information we have assigned probabilities to the various regions in Figure 3. Now we can say that P(X < -2) ≈ 0.025, or P(X < 2) ≈ 0.975, or P(-2 < X < 1) ≈ 0.815, and so on.

Extending our range, the area under the curve between -3 and 3 is about 0.997 square units.

Figure 4 Thus, 99.7% of all the area under the curve is within 3 standard deviations of the mean.

### The Table

As we have seen in the general discussion of continuous probability functions, having these graphs is nice, but having a table would be even nicer. We do have access to a Normal Distribution Table. In addition, there is a different web page that is meant to help you with Reading Tables. As noted on that page, there are many different styles of tables in general use for the normal distribution. If you have a powerful calculator or a computer, it is much less likely that you will need to use tables. However, reading such tables is not a bad skill to acquire.

There are just four kinds of probabilities that we want to evaluate for z values in a standard normal distribution. They are
1. P(X < z)
2. P(X > z)
3. P(z1 < X < z2)
4. P(X < z1 or X > z2)
Approaches to solving all of these were presented on the Probability: Continuous Cases web page. Here we will just run through an example of each, but with minimal discussion.

An example of the first problem might be to find P(X < -2.14). Figure 5 displays the portion of the table, from the Normal Distribution Table, that we need for this.

Figure 5 Having looked at the web page that holds this table, we know that this table gives the area under the curve to the left of the z value. Therefore, we can just read the correct answer as 0.0162.

An example of the second problem might be to find P(X > 0.97). Figure 6 displays the portion of the same table that we need for this.

Figure 6 In this case we want to use the fact that P(X > 0.97) = 1 - P(X < 0.97). Therefore, our answer will be 1 - 0.8340 = 0.1660.

An example of the third style of problem might be to find P(-2.14 < X < 0.97). We know that this will be equal to P(X < 0.97) - P(X < -2.14). Reading the required values from the table gives this as 0.8340 - 0.0162 = 0.8178.

Using the same z values, an example of the fourth style might be to find P(X < -2.14 or X > 0.97). We could approach this as just finding the sum P(X < -2.14) + P(X > 0.97) which is 0.0162 + 0.1660 = 0.1822. Alternatively we could compute this as 1 - P(-2.14 < X < 0.97), or 1 - 0.8178 = 0.1822.

### pnorm() in R

Of course, our goal is to see how we can do all of this in R. The design and use of the functions papfelton() and pblumenkopf was presented on the Probability: Continuous Cases web page. Those functions were modeled upon the R function pnorm(), a function that, in its most simple form, will give the area under the standard normal distribution curve and left of a specified z value. Thus, we can find P(X < -2.14) by using the command pnorm(-2.14), and we can find P(X > 0.97) by using the command 1 - pnorm(0.97). These are shown in Figure 7.

Figure 7 These are the same results we found above using the table, although these have more significant digits.

We can find P(-2.14 < X < 0.97) by finding P(X < 0.97) - P(X < -2.14) which we write as pnorm(0.97) - pnorm(-2.14). We can find P(X < -2.14 or X > 0.97) by finding P(X < -2.14) + P(X > 0.97) which we rewrite as P(X < -2.14) + (1 - P(X < 0.97)) which we then code as pnorm(-2.14)+(1 - pnorm(0.97)). These are shown in Figure 8.

Figure 8 When we use pnorm() we may specify an argument that changes its behavior from the default of finding the area under the curve to the left of the z value to finding the area under the curve to the right of the z value. The former is called the left tail or lower tail area. The latter is called the right tail or upper tail area. The new argument for pnorm() has the name lower.tail and, if given, it must be assigned the value TRUE or FALSE. By default, it is understood to be TRUE and, as a result, pnorm() by default produces the lower tail area. If, however, we specify lower.tail=FALSE then the function produces the upper tail area.

We have already seen that we can get the upper tail area by subtracting the lower tail area from 1. That is how we found P(X > 0.97). We calculated 1 - P(X < 0.97). Using the parameter lower.tail=FALSE we can get the same value directly from pnorm(0.97,lower.tail=FALSE) as shown at the start of Figure 9.

Figure 9 Figure 9 also shows an alternative to coding our fourth example above. The use of the lower.tail=FALSE argument saved us from having to construct the expression (1 - some left-tail probability). We get the same answer either way. Both are correct. Use the form that best suits you.

### Reading "backward" from the table

In our earlier discussion of general continuous probability functions our next step was to read the table backwards. That is, we want to find a z value such that P(X <z)=0.05. We learned how we could find bracketing values in the table and then interpolate between them to find a good approximation to the desired z value. Figure 10 shows the portion of the table that we want to use for this problem. The two bracketing cells are highlighted.

Figure 10 The z values associated with those two cells are -1.64 and -1.65. We want to be halfway between the two cell values, 0.0505 and 0.0495. Therefore, our approximation will be halfway between the two z values, (-1.64 +-1.65)/2 = -1.645.

### qnorm() in R

The R function that does this "backwards" work is qnorm(). In R we could get the answer to finding a z value such that P(X <z)=0.05 by using the command qnorm(0.05) as is shown in Figure 11.

Figure 11 Again, we get a better answer from R than we could get from the table because the table values are all rounded to 4 decimal places. As we have seen in other pages, even the value shown in Figure 11 is a rounded value. We could get R to give us more digits in the answer, as shown in Figure 12, but there is really no need to do this. For our work there is really no need to go beyond four significant digits (the table result).

Figure 12 The work above used the standard normal distribution, the distribution that has mean zero (μ = 0) and standard deviation one (σ = 1). That is an ideal distribution. We have a table for P(x < z) for that distribution. We have the functions pnorm() and qnorm() for that distribution. But what do we do when we have a non-standard normal distribution? That is the BIG QUESTION. The next section leads us to the BIG ANSWER.

### Normalizing a non-standard normal distribution

Here is a separate script that can be used to demonstrate the same concepts shown below. To use the script download it to an appropriate directory or open it and then copy it to an appropriate script file.

Or, istead of using the script, consider the data in Table 1.
Figure 13 shows the generation of that same data along with a computation of the mean and, seemingly, the standard deviation.

Figure 13 Although it is tempting to just move ahead at this point, we will pause long enough to remember that sd() computes the standard deviation of a sample. For this discussion we want to use the standard deviation of a population.

Recall in the much earlier discussion of computing standard deviations we developed a small function, pop_sd() to compute the standard deviation of a population. The code for that function was
```pop_sd<-function( input_list )
{ n <- length( input_list)
sd( input_list )*sqrt((n-1)/n)
}
```
Figure 14 shows this function entered into R along with a new computation for the value of σ.

Figure 14 At this point we have assigned the mean to a variable called mu and we know that mu holds the value 346.5833, and the standard deviation, 21.19191 is stored in sigma.

Before we go further it might be worth looking at the distribution of values in L1. We start by looking at a summary(L1) just so that we can find the minimum and maximum values, and then move to a histogram command. (It is a bit disingenuous to throw out this rather long histogram command as if I just created it by looking at the values from the summary. The truth is that I slowly developed the command, adding directions along the way, changing those directions as I reviewed the resulting graph. The command that you find here is the result of that development.)

Figure 15 The graph produced is in Figure 16.

Figure 16 The distribution shown in Figure 16 is clearly not that shown in Figure 1. But that is not a surprise. Figure 1 shows an "ideal" distribution defined by a mathematics. Figure 16 shows a real population with just 96 values in it. Still, the distribution of L1 shown in Figure 16 is at least "approximately" normal. (It is also good to remember from our earlier presentation on histograms that the decision that we made on the number of intervals and the placement of the intervals does influence the appearance of that histogram.)

We know, from Figure 13, that we have stored the mean in the variable mu. We will create a new population by subtracting the mean from each value in L1. The statement `L2 <- L1 - mu` does just that. If we follow that with the statement L2 we get to see all 96 of the values in the new variable L2. This is shown in Figure 17.

Figure 17 We did not have to look at all those values, but such a listing provides a nice confirmation of the action that we have taken.

Now we can look at the new population in L2. Figure 18 shows the commands and the results for finding the mean, population standard deviation and the Min and Max values of L2. Figure 18 ends with a new hist() statement to display the distribution of the new population.

Figure 18 Please note that 1.894762e-14 means 1.894762*10-14 which we could expand to 0.00000000000001894762, a value that is essentially zero. When that value is rounded to 3 decimal places by the summary(L2) command it is displayed as 0.000.

Another point to note is that the standard deviation of L2 is exactly the same as the standard deviation of L1. This is an important fact: Adding the same value (or subtracting the same value) from every element in a population produces a new population but this new population will have the same standard deviation as did the original population. We could even use our R function to express this. If k is any particular single value, then for a population Ln it will always be true that
pop_sd(Ln) == pop_sd(Ln + k)
where we understand == to mean is equal to. This should make sense because the standard deviation is a measure of dispersion. It is a measure of the deviations from the mean. All we accomplished by subtracting the mean from every element of L1 was to shift all the values from being centered around approximately 346.5833 to being centered around approximately 0.00. We did not alter the deviations of the values from the mean value. The shape of the distribution in Figure 18 is like the shape of the distribution in Figure 16. Again, changing the intervals would alter the appearance of the graph, but not by all that much. The graph would still generally resemble that of a normal distribution.

Figure 19 Now take the values in L2 and divide each of them by sigma, the standard deviation of L1, which we know is also the standard deviation of L2. Put the results in the variable L3 and then display L3. This is shown in Figure 20.

Figure 20 Now that we have a new population we will examine its values as shown in Figure 21.

Figure 21 The mean of L3 is even closer to 0, as we would expect since we divided each value in L2 by a number around 21.19191. The standard deviation of L3 is essentially 1.0, where we need to say "essentially" because all we can see is that R has rounded the value to 1. This is another important observation: Multiplying (or dividing) every element in a population by the same value produces a new population but this new population will have a new standard deviation that is equal to the standard deviation of the original population multiplied (or divided) by that same value. We could even use our R function to express this. If k is any particular single value then for a population Ln it will always be true that
pop_sd(Ln * k) == k * pop_sd(Ln)
where we understand == to mean is equal to.

Figure 21 ends with a new hist() statement to display the distribution of the new population. That histogram appears in Figure 22.

Figure 22 Again, this resembles the shape of Figure 1. In addition, this new population, L3, has mean=0 and standard deviation=1.

If we have a population Lorig that has mean=μ and standard deviation=σ then we can get a new population Lnew by subtracting μ from each value in Lorig and then dividing that difference by σ. When we do this the new population Lnew will have mean=0 and standard deviation=1.

Here is a link to a script demonstrating this.

If x is a value in Lorig then using that transformation produces a new value, we will call it z in Lnew where Furthermore, z will have mean=0 and standard deviation=1. This is the BIG ANSWER for the question above: How do we deal with a normal distribution that has a mean different from 0 and/or a standard deviation different from 1? The answer is that we use this z = (x - μ) / σ to convert any value in this non-standard population to a z-score and then we can use our table or our pnorm() function to answer questions about that value.

For example, IQ scores have a normal distribution with mean equal to 100 and standard deviation equal to 15. If this is the case then what is P(X ≤ 88)? We find the z-score equivalent to 88 by doing the computation (88-100)/15=-12/15≈-0.8. Then, if we are using the table, we look up that z-score in the table. Figure 23 shows that part of the table.

Figure 23 From the table we see that P(z≤-0.8)=0.2119, therefore P(X≤88)=0.2119.

Alternatively, we could just use the pnorm(-0.8) command and let R do the work. Figure 24 shows exactly that approach.

Figure 24 In Figure 24 we see the slightly more accurate answer.
 This is a great time to point out just how sloppy we have gotten. We talk about IQ scores as being normally distributed with mean 100 and standard deviation of 15. Yet, if we think about it, this is absurdly wrong. IQ scores are whole numbers. You do not get an IQ score of 104.24! IQ scores are not continuous. Normal distributions are continuous. It is impossible to have discrete distributions be normal distributions. And yet, we refer to IQ scores as a normally distribution. We will return to this complication in the box below the following examples, after the answers following Figure 25.

We use the same approach to answer questions such as
1. Selecting a random person, how strange is it for that person to have an IQ greater than or equal to 138?
2. What portion of the population has an IQ between 85 and 105, inclusive?
3. What is the likelihood of selecting a random person whose IQ is outside the range from 94 to 119?
Those questions become
1. P(X ≥ 138)
2. P(85 ≤ X ≤ 105)
3. P( X < 94 or X > 119 )
We convert the values in these problems to z-scores
1. 138 becomes z=(138-100)/15=38/15≈2.53333
2. 85 becomes (85-100)/15=-15/15 = -1.000
105 becomes (105-100)/15 = 5/15 = 0.33333
3. 94 becomes (94-100)/15 = -6/15 =-0.4
119 becomes (119-100)/15 = 19/15 ≈ 1.26666
Then we can transform the original problems into problems using those z-scores.
1. P(X ≥ 138) = P(z ≥ 2.5333) = 1 - P(z ≤ 2.5333)
2. P(85 ≤ X ≤ 105) = P(-1 ≤ z ≤ 0.3333) = P( z &le:0.3333) - P( z ≤ -1 )
3. P( X < 94 or X > 119 ) = P( z < -0.4 or z > 1.2666) = 1 - (P(z < 1.2666) - P(z < -0.4) )
If we are using the table then we need to look up all these values in the table. A value of 2.5333 poses a problem in that the table gives us values for 2.53, namely 0.9943, and 2.54, namely 0.9945, but that is as close as the table comes. Since we want the value that is 1/3 the way between 2.53 and 2.54, and since the change in the probability is 0.9945-0.9943=0.0002, we can interpolate a value that is 0.0002/3 beyond 0.9943 or 0.9943+0.0000666≈0.99437. You see why using the tables is a pain!

Rather than get bogged down with the tables, we could just get the answers using R. The problems restated in R become
1. 1-pnorm(2.5333) or pnorm(2.5333,lower.tail=FALSE)
2. pnorm(0.3333)-pnorm(-1)
3. 1-(pnorm(1.2666)-pnorm(-0.4) )
Figure 25 shows all of these:

Figure 25 1. P(X ≥ 138) ≈ 0.00564971
2. P(85 ≤ X ≤ 105) ≈ 0.4718908
3. P( X < 94 or X > 119 ) ≈ 0.4472274
 Here we return to the question of calling the IQ distribution, which we know to be a discrete distribution, a normal distribution. We do this because the distribution of the IQ scores approximates a normal distribution. It is just easier for us the deal with the distribution as a normal one rather than as a discrete one. However, we should recall that in a continuous distribution the probability of getting an exact value is zero (remember that there is a height of the curve in Figure 1 above every point, but there is no width at a point and area is length times width which means the probability is zero). A consequence of this is that when we find, as we did above, P(X≥138) we really are not sure just what that means because there really should be a P(X=138). If we want to include 138 and all the values above it, then we should really have looked at P(X>137.5). Using the halfway point between our discrete IQ values makes a lot of sense. We would say, then, that P(X=138) = P(137.5

As you have seen, z-scores and the relation have been essential for converting values from non-standard normal distributions to our standard normal distribution. In the same way, solving the relation for x to get x = z*σ + μ, we can go backwards from the standard normal distribution to a non-standard one.

For example, for a distribution that is normal with mean 75 and standard deviation 14, we want to know the value in that distribution that has 12% of the values less than it. In a standard normal distribution we would find the z-score by looking up the area in a cell of the table, or by using the qnrom(.12) function. Figure 26 shows the portion of the table that we could use.

Figure 26 The two cells that bracket 0.12 hold 0.1210 and 0.1190, and they hold the values for -1.17 and -1.18, respectively. We need to be halfway between so our table generated z-score will be -1.175.

The R command is shown in Figure 27.

Figure 27 Here we get a z-score of -1.174987. The R-generated value is almost identical, so we will just use -1.175 as the z-score. Then, using x = z*σ + μ, we can find the value in the non-standard distribution by calculating -1.175*14+75 = -16.45+75 = 58.55. Of course, using the R value we could have done this as qnorm(.12)*14+75 as is shown in Figure 28.

Figure 28 All done in one step!

### The N(μ,σ) Notation

Let us take a minute here to introduce a shorthand method for saying that a distribution is normal with mean=a and standard deviation=b. That shorthand is to say that the distribution is N(a,b). The standard normal distribution is N(0,1). We say that a normal distribution with mean 75 and standard deviation 14 is N(75,14).

### The Expanded Notation for pnorm() and qnorm()

It should not come as a surprise that most of our real work will be with non-standard normal distributions. If you understand that for ages we have had but the one table of the Standard Normal Distribution, then you must appreciate the importance of the z-scores, the relation , and the relation x = z*σ + μ. Furthermore, unlike the discrete binomial distribution, you must appreciate that it would be impossible to produce tables for all of the different combinations of non-standard values for the mean and standard deviation of these non-standard normal distributions.

However, our calculators and computers do not rely on tables to do their work. Therefore, it is possible for our technology to deal directly with non-standard situations as long as we supply the non-standard values for the mean and standard deviation. In particular, in R the pnorm() function actually allows us to specify one or both of those values. For a distribution that is N(75,14), to find the P(X < 83.4) we just need to give the command pnorm(83.4, mean=75, sd=14) as is shown in Figure 29. Of course, we could do the extra work the way have done it above, also shown in Figure 29, but using the arguments seems a bit cleaner and easier to understand. We get the same answer either way.

Figure 29 The four answers, and then corrected answers, to the three example problems from earlier (see Figure 25 and the corrections in the boxed area that followed it) are shown, using the expanded form of pnorm() in Figure 30.

Figure 30 And, of course, there is a similar expanded notation for qnorm(). Thus, to find the value, X, for a population that is N(75,14) we just need to use the command qnorm(0.12,mean=75, sd=14) as shown in Figure 31.

Figure 31 As we have seen, having the named arguments mean= and sd= in both pnorm() and qnorm() can be quite convenient. There does seem to be a little "overhead" in using them in the sense that we seem to need to have to type the mean= and sd= as well as the values assigned to them. In most cases this can be avoided because the formal definition of the two functions gives these named parameters in the second and third parameter position, respectively. Figure 32 holds an excerpt from that formal definition.

Figure 32 In R, once named arguments have been assigned, all unnamed arguments are assigned to the remaining parameters in the order specified in the formal definition. Therefore if we were to give the command qnorm(0.12,75,14) the 0.12 is assigned to `q`, the 75 is assigned to `mean`, and the 14 is assigned to `sd`. This more compact statement is shown in Figure 33.

Figure 33 Although this more compact form is available, it is my recommendation that you stay with the longer named argument form of the commands. If nothing else, it makes it much easier to check and confirm your work.
Review for pnorm and qnorm

```x <- seq(-4, 4, length=200)
hx <- dnorm(x)
plot(x, hx, type="l", lty=1,
xlim=c(-4,4), xaxp=c(-4,4,16),
ylim=c(0,0.4), yaxp=c(0,0.4,8),
las=2, cex.axis=0.75,
xlab="z value",
ylab="Density", main="Normal Distribution")
abline( v=0, lty=1, lwd=2)
abline( h=0, lty=1, lwd=2)
abline( h=seq(0.05,4,0.05), lty=3, col="darkgray")
abline( v=seq(-4,4,0.5), lty=3, col="darkgray")

plot(x, hx, type="l", lty=1,
xlim=c(-4,4), xaxp=c(-4,4,16),
ylim=c(0,8), yaxp=c(0,8,16),
las=2, cex.axis=0.75,
xlab="z value",
ylab="Density", main="Normal Distribution")
abline( v=0, lty=1, lwd=2)
abline( h=0, lty=1, lwd=2)
abline( h=seq(0.5,16,0.5), lty=3, col="darkgray")
abline( v=seq(-4,4,0.5), lty=3, col="darkgray")

pnorm(-2.14)
pnorm(0.97)
1 - pnorm(0.97)

pnorm(0.97) - pnorm(-2.14)
pnorm(-2.14)+(1 - pnorm(0.97))

pnorm(0.97, lower.tail=FALSE)
# re-do P(X<-2.14) or P(X>0.97)
pnorm(-2.14)+pnorm(0.97,lower.tail=FALSE)

qnorm(0.05)

options(digits=13)
qnorm(0.05)
options(digits=7)

source("http://courses.wccnet.edu/~palay/math160r/gnrnd4.R")
gnrnd4( key1=121859504, key2=0002100345 )
L1
mu<-mean(L1)
mu
sd(L1)
pop_sd<-function( input_list )
{ n <- length( input_list)
sd( input_list )*sqrt((n-1)/n)
}
pop_sd( L1)
sigma<-pop_sd( L1)
sigma

summary(L1)

hist(L1,xlim=c(280,400),
breaks=seq(280,400,5),
xaxp=c(280,400,24),
cex.axis=0.75, las=2, right=FALSE)

L2<-L1-mu
L2
mean(L2)
pop_sd( L2)
summary( L2)
hist(L2,xlim=c(-60,60),
breaks=seq(-60,60,5),
xaxp=c(-60,60,24),
cex.axis=0.75, las=2)

L3<-L2/sigma
L3
mean(L3)
pop_sd(L3)
summary( L3)
hist(L3,xlim=c(-3,3),
breaks=seq(-3,3,0.25),
xaxp=c(-3,3,24),
cex.axis=0.75, las=2)

pnorm(-0.8)

1-pnorm(2.5333)
pnorm(2.5333,lower.tail=FALSE)
pnorm(0.3333)-pnorm(-1)
1-(pnorm(1.2666)-pnorm(-0.4))

pnorm((138.5-100)/15)-pnorm((137.5-100)/15)

1-pnorm((137.5-100)/15)
pnorm((137.5-100)/15,lower.tail=FALSE)
pnorm((105.5-100)/15)-pnorm((84.5-100)/15)
1-(pnorm((118.5-100)/15)-pnorm((93.5-100)/15))

qnorm(.12)
qnorm(.12)*14+75

pnorm(83.4, mean=75, sd=14)
pnorm((83.4-75)/14)

1-pnorm(138,mean=100, sd=15)
pnorm(138,mean=100, sd=15,
lower.tail=FALSE)
pnorm(105,mean=100, sd=15) -
pnorm(85,mean=100, sd=15)
1-(pnorm(119,mean=100, sd=15) -
pnorm(94,mean=100, sd=15))
#
#corrected to
#
1-pnorm(137.5,mean=100, sd=15)
pnorm(137.5,mean=100, sd=15,
lower.tail=FALSE)
pnorm(105.5,mean=100, sd=15) -
pnorm(84.5,mean=100, sd=15)
1-(pnorm(118.5,mean=100, sd=15) -
pnorm(93.5,mean=100, sd=15))

qnorm(0.12,mean=75, sd=14)
qnorm(0.12,75,14)
```