<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://www.micahveilleux.com/statistics/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Statadmin</id>
	<title>Statistics - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://www.micahveilleux.com/statistics/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Statadmin"/>
	<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Special:Contributions/Statadmin"/>
	<updated>2026-05-06T14:12:34Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.45.1</generator>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_2-1,_Comparing_a_single_mean_to_a_specified_value&amp;diff=43</id>
		<title>Problem 2-1, Comparing a single mean to a specified value</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_2-1,_Comparing_a_single_mean_to_a_specified_value&amp;diff=43"/>
		<updated>2011-11-04T15:40:13Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Problem statement==&lt;br /&gt;
&#039;&#039;The breaking strength of a fiber is required to be at least 150 psi. Past experience has indicated that the standard deviation of breaking strength is σ = 3 psi. A random sample of four specimens is tested, and the results are y&amp;lt;sub&amp;gt;1&amp;lt;/sub&amp;gt; = 145, y&amp;lt;sub&amp;gt;2&amp;lt;/sub&amp;gt; = 153, y&amp;lt;sub&amp;gt;3&amp;lt;/sub&amp;gt; = 150, and y&amp;lt;sub&amp;gt;4&amp;lt;/sub&amp;gt; = 147.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;State the hypotheses that you think should be tested in this experiment.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Test these hypothesis using α = 0.05. What are your conclusions?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Find the &#039;&#039;P&#039;&#039;-value for the test in part (b).&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Construct a 95 percent confidence interval on the mean breaking strength.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
[[Image:Gaussian1.png|thumb|left|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; Our data compared to a theoretical Gaussian distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Section A: Choosing hypotheses===&lt;br /&gt;
In this problem we are given a set of four data points. These data points all come from a distribution of breaking strengths which has an unknown mean μ. We will call this the &#039;&#039;true distribution&#039;&#039;. Previous experience indicates that breaking strengths follow a Gaussian &#039;&#039;theoretical distribution&#039;&#039; with a standard deviation of 3 psi, so we assume this for our distribution also. Our task is to determine whether or not the true mean, which is impossible to know exactly, is greater than or equal to 150 psi. We plot this data and the distribution in figure 1.&lt;br /&gt;
&lt;br /&gt;
Since the sample mean is an approximation of the true mean, we define the &#039;&#039;standard error of the mean&#039;&#039; (SEM) to be &amp;lt;math&amp;gt;\sigma/\sqrt{n}=1.5&amp;lt;/math&amp;gt;, where n=4 is the number of data points.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Distribution type&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Mean&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Standard deviation&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;True distribution&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;Normal&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\mu\approx\overline{y}=148.25&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\sigma=3&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Theoretical distribution&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;Normal&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\mu_0=150&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\sigma_0=3&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We first state two hypotheses. The null hypothesis is that our data does come from the theoretical distribution: the true mean μ = μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;. Our alternative hypothesis states that the data comes from a distribution centered around a different mean.&lt;br /&gt;
&lt;br /&gt;
There are three choices for the alternative hypothesis: μ &amp;lt; 150, μ &amp;gt; 150, and μ ≠ 150. We adopt the convention that the alternative hypothesis will be true if the data does not meet the requirements. In this case, the breaking strength of the fiber is required to be at least 150 psi, so we choose μ &amp;lt; 150 as our alternative hypothesis.&lt;br /&gt;
&lt;br /&gt;
Formally, we state our hypotheses as:&amp;lt;br/&amp;gt;&lt;br /&gt;
&amp;lt;center&amp;gt;H&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;: μ = 150&amp;lt;br/&amp;gt;&lt;br /&gt;
H&amp;lt;sub&amp;gt;1&amp;lt;/sub&amp;gt;: μ &amp;lt; 150&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br clear=&#039;all&#039;&amp;gt;&lt;br /&gt;
[[Image:Gaussian2.png|thumb|left|&#039;&#039;&#039;Figure 2:&#039;&#039;&#039; Our plot after normalizing.]]&lt;br /&gt;
&lt;br /&gt;
===Section B: Z-values===&lt;br /&gt;
For convenience, we start by standardizing our theoretical distribution to have a mean of zero and a standard deviation of one. To do this, we first center the distribution around zero by subtracting the theoretical mean (150) from each point in the distribution. We then divide each point by the standard deviation (3). The sample mean can be standardized in the same manner. We plot the normalized distribution and sample mean in figure 2.&lt;br /&gt;
&lt;br /&gt;
We now assume that the null hypothesis is true and ask whether or not this assumption makes sense. Given that this assumption is true, the sample mean is most likely to be close to zero. To test this, we define a range over which we consider our sample mean to be unacceptable, the &#039;&#039;rejection region&#039;&#039;. If the sample mean is in the rejection region, it is too far from zero and we reject the null hypothesis.&lt;br /&gt;
&lt;br /&gt;
We will define the lower limit to be z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;, where α=0.05. Graphically, given a standard Gaussian distribution, the area under the curve left of z&amp;lt;sub&amp;gt;0.05&amp;lt;/sub&amp;gt; is equal to 5% of the total area. You can either look up z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt; in a table or calculate it using a software package. Using Excel, the appropriate function is &amp;lt;tt&amp;gt;=NORMSINV(alpha)&amp;lt;/tt&amp;gt;. The corresponding function in R is &amp;lt;tt&amp;gt;qnorm(alpha)&amp;lt;/tt&amp;gt;. Using one of these methods, we find that z&amp;lt;sub&amp;gt;0.05&amp;lt;/sub&amp;gt;=−1.64.&lt;br /&gt;
&lt;br /&gt;
We then find the z-value of our data and compare the z-value to z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. The formula for the z-value is as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;z=(\overline{y}-\mu_0)(\frac{1}{\sigma})(\sqrt{n})&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Because we have already standardized our data, μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;=0 and σ=1, so this formula simplifies to &amp;lt;math&amp;gt;\overline{y} \sqrt{n}=-0.417\cdot2=-0.833&amp;lt;/math&amp;gt;. Note that the formula above normalizes the data, if it has not already been normalized. The z-value can be interpreted as the distance between the sample mean and μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;, scaled by a factor which makes the z-value more extreme with large sample sizes. If we take many samples, our z-value is more likely to fall in the rejection region, because we are more certain of the accuracy of our sample mean.&lt;br /&gt;
&lt;br /&gt;
The rejection region for our z-value is from negative infinity to z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. We see that our z-value is greater than z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. Therefore, we cannot reject the null hypothesis.&lt;br /&gt;
&lt;br /&gt;
[[Image:Gaussian3.png|thumb|left|&#039;&#039;&#039;Figure 3:&#039;&#039;&#039; Illustrating the P-value.]]&lt;br /&gt;
&lt;br /&gt;
===Section C: P-values===&lt;br /&gt;
Another way to judge how likely it is that our null hypothesis is true is to calculate the P-value. If we were to redo the experiment, taking four new data points, the P-value gives us the probability of our new sample mean being at least as extreme as our original sample mean. Graphically, if we extend the critical region until it reaches our z-value, the P-value is equal to the area of the shaded region (see figure 3). &lt;br /&gt;
&lt;br /&gt;
To calculate the P-value in Excel, use &amp;lt;tt&amp;gt;=NORMSDIST(-ABS(z))&amp;lt;/tt&amp;gt;. In R, use &amp;lt;tt&amp;gt;pnorm(-abs(z))&amp;lt;/tt&amp;gt;. (We use the negative absolute value because &amp;lt;tt&amp;gt;NORMSDIST&amp;lt;/tt&amp;gt; and &amp;lt;tt&amp;gt;pnorm&amp;lt;/tt&amp;gt; integrate from negative infinity to the z-value. If the z-value is positive, we instead want to integrate from the z-value to positive infinity, which is mathematically equivalent to integrating from negative infinity to the negative of the z-value.)&lt;br /&gt;
&lt;br /&gt;
For this problem, we find that the P-value is 0.202. Note that a P-value of 0.5 indicates that the sample mean is equal to the mean of the theoretical distribution. You can see this graphically by noting that the z-value will be zero in this case, and integrating the theoretical distribution to zero covers half of the area. (Recall that the total area under a standard Gaussian curve is one.) The further the P-value is from 0.5, the greater the distance between the two means.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div style=&amp;quot;float:left; vertical-align: top; padding-right: 20px; padding-bottom: 20px;&amp;quot;&amp;gt;[[Image:Gaussian4.png|thumb|none|&#039;&#039;&#039;Figure 4:&#039;&#039;&#039; The confidence interval about the sample mean.]]&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Gaussian5.png|thumb|none|&#039;&#039;&#039;Figure 5:&#039;&#039;&#039; The confidence interval about the theoretical mean.]]&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section D: Confidence intervals===&lt;br /&gt;
We now return to our original data set and theoretical distribution with the mean of 150 psi; that is, we will no longer use our normalized space.&lt;br /&gt;
&lt;br /&gt;
We will now calculate the range of sample means that would lead us to conclude that the breaking strength of our fiber is at least 150 psi, given an α of 0.05. This range is known as the confidence interval about the sample mean.&lt;br /&gt;
&lt;br /&gt;
To calculate this interval, we ask what sample mean would give us a z-value equal to z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. We can determine this by substituting z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt; for z into the formula for z, and solving for &amp;lt;math&amp;gt;\overline{y}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;z_\alpha=(\overline{y}-\mu_0)(\frac{1}{\sigma})(\sqrt{n}) \Rightarrow \overline{y} = \mu_0+\frac{z_\alpha \sigma}{\sqrt{n}}=147.53&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is the lower limit of our confidence interval. Because any sample mean greater than 150 is acceptable, the upper limit of the confidence interval is infinity. We plot this interval in figure 4. Formally, our confidence interval about the sample mean is&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;147.53 &amp;lt; \overline{y} &amp;lt; \infty&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We next calculate a confidence interval about the mean of the theoretical distribution, μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;. This will give us the range of minimum breaking strengths we could have specified and still found our data acceptable. We can calculate this in much the same way as the previous confidence interval: substitute z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt; for z in the formula for z, but this time solve for μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mu_0=\overline{y}-\frac{z_\alpha \sigma}{\sqrt{n}}=151.22&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is the upper limit of our confidence interval. The lower limit is zero, because we simply require the theoretical mean to be less than this number. Formally, our confidence interval about the theoretical mean is&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;0 \le \mu_0 &amp;lt; 151.22&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_5-1,_Factorial_designs&amp;diff=42</id>
		<title>Problem 5-1, Factorial designs</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_5-1,_Factorial_designs&amp;diff=42"/>
		<updated>2010-08-30T23:25:53Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: /* Solution */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Problem Statement==&lt;br /&gt;
&#039;&#039;The yield of a chemical process is being studied. The two most important variables are thought to be the pressure and the temperature. Three levels of each factor are selected, and a factorial experiment with two replicates is performed. The yield data follow:&#039;&#039;&lt;br /&gt;
&amp;lt;table cellpadding=2 cellspacing=0 style=&amp;quot;border-top: 1px solid black; border-bottom: 1px solid black&amp;quot;&amp;gt;&amp;lt;tr&amp;gt;&amp;lt;th rowspan=2 valign=bottom style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;Temperature (°C)&amp;lt;/th&amp;gt;&amp;lt;th colspan=3 style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;Pressure (psig)&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;200&amp;lt;/td&amp;gt;&amp;lt;td style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;215&amp;lt;/td&amp;gt;&amp;lt;td style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;230&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td rowspan=2 valign=top&amp;gt;150&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.4&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.7&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.2&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td&amp;gt;90.2&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.6&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.4&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td rowspan=2 valign=top&amp;gt;160&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.1&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.5&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;89.9&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td&amp;gt;90.3&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.6&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.1&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td rowspan=2 valign=top&amp;gt;170&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.5&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.8&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.4&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td&amp;gt;90.7&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.9&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.1&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Analyze the data and draw conclusions. Use α=0.05.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Prepare appropriate residual plots and comment on the model&#039;s adequacy.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Under what conditions would you operate this process?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
&amp;lt;div style=&amp;quot;float:left; vertical-align: top; padding-right: 20px; padding-bottom: 20px;&amp;quot;&amp;gt;&lt;br /&gt;
[[Image:5-1yield.png|thumb|180px|none|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; This plot of yield vs. pressure, at the different levels of temperature, helps to visualize the different effects in our problem.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1dataplot.png|thumb|180px|none|&#039;&#039;&#039;Figure 2:&#039;&#039;&#039; This plot illustrates the various effects on the data in problem 5-1. The hats (^) indicate that these are approximate effects calculated from our sample data.]]&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We are given a set of data that describes the yield of a chemical process, but the yield is affected by two factors: temperature and pressure.  At each combination of &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; levels of factor A (temperature) and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; levels of factor B (pressure), &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; data points have been collected (see chart in problem statement).  Not only do these two factors influence the yield data directly, but they might also interact to affect the yield in an unexpected manner.&lt;br /&gt;
&lt;br /&gt;
The average of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; data points at a specific temperature and pressure in our problem is &amp;lt;math&amp;gt;\bar{y}_{ij}.&amp;lt;/math&amp;gt;. An interaction effect would cause the difference between two &amp;lt;math&amp;gt;\bar{y}_{ij}.&amp;lt;/math&amp;gt; values in a column of our chart to vary as pressure is changed.  For example, say we set the pressure to 200 psig and observe the average yield increase from 90 to 95 as temperature is increased from 150 °C to 160 °C.  But when we set the pressure to 215 psig the yield drops from 90 to 85 as temperature is again increased from 150 °C to 160 °C. This would be an interaction effect, since the effect on the yield cannot be separated into independent temperature and pressure effects. Of course, what appears to be an interaction effect may just be large random error, so we&#039;ll have to check for that.&lt;br /&gt;
&lt;br /&gt;
The plot shown in Figure 1 is very useful for visualizing the factor and interaction effects. We can see an apparent temperature effect that indicates 170 °C produces the highest yield. 215 psig also appears to be the best pressure if we want high yield, as the green line is higher than the others.  Interaction effects in this type of plot are indicated by non-parallel lines, so there is not much apparent interaction here.  Before we draw any conclusions, we must test that this plot provides a true image of the chemical process, and is not just a product of high random error.&lt;br /&gt;
&lt;br /&gt;
Each data point in our problem can be thought of as the sum of several effects:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;y_{ijk} = \mu + \tau_i + \beta_j + (\tau \beta)_{ij} + \epsilon_{ijk}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; i = 1,2,...,a &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; j = 1,2,...,b &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; k = 1,2,...,n &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
and:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; y_{ijk} = &amp;lt;/math&amp;gt; data point from the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of factor A and the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th level of factor B&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \mu = &amp;lt;/math&amp;gt; the grand mean&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \tau_i = &amp;lt;/math&amp;gt; the effect of the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of factor A&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \beta_j = &amp;lt;/math&amp;gt; the effect of the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th level of factor B&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; (\tau \beta)_{ij} = &amp;lt;/math&amp;gt; the effect of interaction between factors A and B&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \epsilon_{ijk} = &amp;lt;/math&amp;gt; a random error affecting the data point from the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of factor A and the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th level of factor B&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 2 illustrates how these effects sum up to each data point.  The blue line indicates the grand mean &amp;lt;math&amp;gt;\bar{y}.. \approx \mu &amp;lt;/math&amp;gt;, the red X indicates the sample mean of the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of temperature &amp;lt;math&amp;gt;\bar{y}_i.&amp;lt;/math&amp;gt;, the green # indicates the #th pressure level effect indicator, the brown triangle indicates the interaction effect indicator, and the black # indicates the data point from the #th level of pressure. The projection of each line onto the x-axis tells you the quantitatively what its estimated effect is, and each point can be thought of as where the data would be if effects closer to the real data point were not present.&lt;br /&gt;
&lt;br /&gt;
=== Hypothesis Testing ===&lt;br /&gt;
We would like to know if any of our estimated effects are significant.  Therefore, we have three tests to do, one for each type of effect.  For factor A, our null hypothesis will be that all the treatment means are equal (they all come from the same distribution). Our alternative hypothesis is that they are not all equal:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; H_0\!:~\tau_1 = \tau_2 = ... = \tau_a = 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; H_1\!:~\tau_i \ne 0 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similarly for factor B and the interaction effect:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; H_0\!:~\beta_1 = \beta_2 = ... = \beta_b = 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; H_1\!:~\beta_i \ne 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; H_0\!:~(\tau \beta)_{ij} = 0 &amp;lt;/math&amp;gt;  (for all &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;)&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; H_1\!:~(\tau \beta)_{ij} \ne 0 &amp;lt;/math&amp;gt;  (for at least one combination of &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;)&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To test these hypotheses, we are interested in the following sums of squares. Each of the following equations estimates the effect on our data due to the subscripted factor (&amp;lt;math&amp;gt;A &amp;lt;/math&amp;gt; → factor A, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; → factor B, &amp;lt;math&amp;gt;AB&amp;lt;/math&amp;gt; → interaction, and &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; → random error).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;SS_A = bn \Sigma^a_{i=1} (\bar{y}_i.. - \bar{y}...)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;SS_B = an \Sigma^b_{j=1} (\bar{y}._j. - \bar{y}...)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;SS_{AB} = n \Sigma^a_{i=1} \Sigma^b_{j=1} (\bar{y}_{ij}. - \bar{y}_{i}.. - \bar{y}._j. + \bar{y}...)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; SS_E = \Sigma^a_{i=1} \Sigma^b_{j=1} \Sigma^n_{k=1} (y_{ijk} - \bar{y}_{ij}.)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before we can compare these quantities to each other, we must divide by the degrees of freedom to normalize. We calculate the normalized &#039;&#039;mean squares&#039;&#039; with:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; MS_A = \frac{SS_A}{a-1} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; MS_B = \frac{SS_B}{b-1} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; MS_{AB} = \frac{SS_{AB}}{(a-1)(b-1)} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; MS_E = \frac{SS_E}{ab(n-1)} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We test whether our factor and interaction effects are significant by dividing their mean squares by the mean square of our random error to get &amp;lt;math&amp;gt;F_0&amp;lt;/math&amp;gt;.  This value must be large for us to conclude that the effect we&#039;re testing is significant (the effect must be large relative to the random error). We pick a value from the F distribution based on &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt; and our degrees of freedom – this determines what &amp;quot;large&amp;quot; is. Do this in Excel with &amp;lt;tt&amp;gt;FINV(α,numerator&#039;s DOF,denominator&#039;s DOF)&amp;lt;/tt&amp;gt; and in R with &amp;lt;tt&amp;gt;qf(1-alpha, numerator&#039;s DOF, denominator&#039;s DOF)&amp;lt;/tt&amp;gt; (DOF = degrees of freedom).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table cellspacing=0 cellpadding=5 style=&amp;quot;border-top: 1px solid black; border-bottom: 1px solid black&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Source of Variation &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Sum of Squares &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Degrees of Freedom &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Mean Square &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_0&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_{\alpha}&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Factor A (Temperature) &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.301 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 2     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.151 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 8.469 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4.256 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Factor B (Pressure) &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.768  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 2      &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.384  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 21.594 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4.256  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Interaction &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.069 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.017 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.969 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 3.633 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Error &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.160 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 9     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.018 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; Total &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 1.298 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 17    &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We conclude from these tests that the effects from factors A and B (temperature and pressure) are significant, and there is no significant interaction effect. However, these tests are only accurate if the model is correct.&lt;br /&gt;
&lt;br /&gt;
===Residual Plots and Model Adequacy Testing===&lt;br /&gt;
[[Image:5-1norm_prob.png|thumb|360px|left|&#039;&#039;&#039;Figure 3:&#039;&#039;&#039; These plots indicate that our data is sufficiently gaussian and we can believe the results of our hypothesis testing.]]&lt;br /&gt;
&lt;br /&gt;
To test whether our data is sufficiently gaussian that our hypothesis testing is valid, we create a normal probability plot.  As before, we calculate the residuals (by subtracting the expected values from the datapoints, &amp;lt;math&amp;gt;y_{ijk} - \bar{y}_{ij}.&amp;lt;/math&amp;gt;) and sorting them in an array.  Then we plot the sorted residuals against z-values, where the z-values are calculated in Excel by doing &amp;lt;tt&amp;gt;NORMSINV(percent)&amp;lt;/tt&amp;gt; or in R with &amp;lt;tt&amp;gt;qnorm(percent)&amp;lt;/tt&amp;gt;.  &amp;lt;tt&amp;gt;Percent&amp;lt;/tt&amp;gt; is equal to index_of_array/(DOF + 1). If the plot does not roughly fall along a straight line, the data is not from a Gaussian distribution.&lt;br /&gt;
&lt;br /&gt;
There are several other useful plots we can make with residuals.  Plotting &amp;lt;math&amp;gt;\bar{y}_{ij}.&amp;lt;/math&amp;gt; against residuals produces a plot that should have data randomly scattered throughout its entire area, if not the data may not be gaussian. Plotting the residuals against either of the factors should indicate that the data is more or less equally random in each level of the factors.  Note that because &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt;, there is some symmetry present in each of these plots.&lt;br /&gt;
&lt;br /&gt;
Given that our data fits the model reasonably well, and that each of the factors are significant, the optimum conditions for producing high yield are when temperature is 170 °C and pressure is 200 psig.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=MediaWiki:Mainpage&amp;diff=41</id>
		<title>MediaWiki:Mainpage</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=MediaWiki:Mainpage&amp;diff=41"/>
		<updated>2010-08-06T01:41:55Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Statistics_Tutorial:_Design_and_Analysis_of_Experiments&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=MediaWiki:Mainpage&amp;diff=40</id>
		<title>MediaWiki:Mainpage</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=MediaWiki:Mainpage&amp;diff=40"/>
		<updated>2010-08-06T01:39:35Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;quot;Statistics Tutorial: Design and Analysis of Experiments&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Statistics Tutorial: Design and Analysis of Experiments&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Main_Page&amp;diff=39</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Main_Page&amp;diff=39"/>
		<updated>2010-08-06T01:28:56Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: moved Main Page to Statistics Tutorial: Design and Analysis of Experiments: Replace &amp;quot;Main Page&amp;quot; with meaningful title.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Statistics Tutorial: Design and Analysis of Experiments]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=38</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=38"/>
		<updated>2010-08-06T01:28:56Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: moved Main Page to Statistics Tutorial: Design and Analysis of Experiments: Replace &amp;quot;Main Page&amp;quot; with meaningful title.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Edward Brash&#039;s statistics class, at Christopher Newport University. Detailed solutions are presented for several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions were prepared using the &amp;lt;tt&amp;gt;R&amp;lt;/tt&amp;gt; statistical programming language, which is much more powerful than &amp;lt;tt&amp;gt;MS Excel&amp;lt;/tt&amp;gt;, but has a significantly steeper learning curve.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1, Comparing a single mean to a specified value]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2, Comparing a single mean to a specified value (second example)|Problem 2-2, Comparing a single mean to a specified value (second example)]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-4, Determining required sample size|Problem 2-4, Determining required sample size]]&lt;br /&gt;
&lt;br /&gt;
[[Problems 3-1 through 3-3, Analysis of variance|Problems 3-1 through 3-3, Analysis of variance]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 4-1, Randomized blocks|Problem 4-1, Randomized blocks]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 5-1, Factorial designs|Problem 5-1, Factorial designs]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=37</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=37"/>
		<updated>2010-07-31T19:15:34Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Edward Brash&#039;s statistics class, at Christopher Newport University. Detailed solutions are presented for several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions were prepared using the &amp;lt;tt&amp;gt;R&amp;lt;/tt&amp;gt; statistical programming language, which is much more powerful than &amp;lt;tt&amp;gt;MS Excel&amp;lt;/tt&amp;gt;, but has a significantly steeper learning curve.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1, Comparing a single mean to a specified value]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2, Comparing a single mean to a specified value (second example)|Problem 2-2, Comparing a single mean to a specified value (second example)]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-4, Determining required sample size|Problem 2-4, Determining required sample size]]&lt;br /&gt;
&lt;br /&gt;
[[Problems 3-1 through 3-3, Analysis of variance|Problems 3-1 through 3-3, Analysis of variance]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 4-1, Randomized blocks|Problem 4-1, Randomized blocks]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 5-1, Factorial designs|Problem 5-1, Factorial designs]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:5-1norm_prob.png&amp;diff=36</id>
		<title>File:5-1norm prob.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:5-1norm_prob.png&amp;diff=36"/>
		<updated>2010-07-31T19:03:06Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: These plots indicate that our data is sufficiently gaussian and we can believe the results of our hypothesis testing.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;These plots indicate that our data is sufficiently gaussian and we can believe the results of our hypothesis testing.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:5-1dataplot.png&amp;diff=35</id>
		<title>File:5-1dataplot.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:5-1dataplot.png&amp;diff=35"/>
		<updated>2010-07-31T19:02:16Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: This plot illustrates the various effects on the data in problem 5-1. The hats (^) indicate that these are approximate effects calculated from our sample data.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This plot illustrates the various effects on the data in problem 5-1. The hats (^) indicate that these are approximate effects calculated from our sample data.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:5-1yield.png&amp;diff=34</id>
		<title>File:5-1yield.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:5-1yield.png&amp;diff=34"/>
		<updated>2010-07-31T19:01:46Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: This plot of yield vs. pressure, at the different levels of temperature, helps to visualize the different effects in our problem.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This plot of yield vs. pressure, at the different levels of temperature, helps to visualize the different effects in our problem.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_5-1,_Factorial_designs&amp;diff=33</id>
		<title>Problem 5-1, Factorial designs</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_5-1,_Factorial_designs&amp;diff=33"/>
		<updated>2010-07-31T19:01:07Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;#039;==Problem Statement== &amp;#039;&amp;#039;The yield of a chemical process is being studied. The two most important variables are thought to be the pressure and the temperature. Three levels of eac…&amp;#039;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Problem Statement==&lt;br /&gt;
&#039;&#039;The yield of a chemical process is being studied. The two most important variables are thought to be the pressure and the temperature. Three levels of each factor are selected, and a factorial experiment with two replicates is performed. The yield data follow:&#039;&#039;&lt;br /&gt;
&amp;lt;table cellpadding=2 cellspacing=0 style=&amp;quot;border-top: 1px solid black; border-bottom: 1px solid black&amp;quot;&amp;gt;&amp;lt;tr&amp;gt;&amp;lt;th rowspan=2 valign=bottom style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;Temperature (°C)&amp;lt;/th&amp;gt;&amp;lt;th colspan=3 style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;Pressure (psig)&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;200&amp;lt;/td&amp;gt;&amp;lt;td style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;215&amp;lt;/td&amp;gt;&amp;lt;td style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt;230&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td rowspan=2 valign=top&amp;gt;150&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.4&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.7&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.2&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td&amp;gt;90.2&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.6&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.4&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td rowspan=2 valign=top&amp;gt;160&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.1&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.5&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;89.9&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td&amp;gt;90.3&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.6&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.1&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td rowspan=2 valign=top&amp;gt;170&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.5&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.8&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.4&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td&amp;gt;90.7&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.9&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;90.1&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Analyze the data and draw conclusions. Use α=0.05.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Prepare appropriate residual plots and comment on the model&#039;s adequacy.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Under what conditions would you operate this process?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
&amp;lt;div style=&amp;quot;float:left; vertical-align: top; padding-right: 20px; padding-bottom: 20px;&amp;quot;&amp;gt;&lt;br /&gt;
[[Image:5-1yield.png|thumb|180px|none|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; This plot of yield vs. pressure, at the different levels of temperature, helps to visualize the different effects in our problem.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:5-1dataplot.png|thumb|180px|none|&#039;&#039;&#039;Figure 2:&#039;&#039;&#039; This plot illustrates the various effects on the data in problem 5-1. The hats (^) indicate that these are approximate effects calculated from our sample data.]]&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We are given a set of data that describes the yield of a chemical process, but the yield is affected by two factors: temperature and pressure.  At each combination of &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; levels of factor A (temperature) and &amp;lt;math&amp;gt;b&amp;lt;/math&amp;gt; levels of factor B (pressure), &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; data points have been collected (see chart in problem statement).  Not only do these two factors influence the yield data directly, but they might also interact to affect the yield in an unexpected manner.&lt;br /&gt;
&lt;br /&gt;
The average of the &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; data points at a specific temperature and pressure in our problem is &amp;lt;math&amp;gt;\bar{y}_{ij}.&amp;lt;/math&amp;gt;. An interaction effect would cause the difference between two &amp;lt;math&amp;gt;\bar{y}_{ij}.&amp;lt;/math&amp;gt; values in a column of our chart to vary as pressure is changed.  For example, say we set the pressure to 200 psig and observe the average yield increase from 90 to 95 as temperature is increased from 150 °C to 160 °C.  But when we set the pressure to 215 psig the yield drops from 90 to 85 as temperature is again increased from 150 °C to 160 °C. This would be an interaction effect, since the effect on the yield cannot be separated into independent temperature and pressure effects. Of course, what appears to be an interaction effect may just be large random error, so we&#039;ll have to check for that.&lt;br /&gt;
&lt;br /&gt;
The plots shown in Figure 1 are very useful for visualizing the factor and interaction effects. We can see an apparent temperature effect that indicates 170 °C produces the highest yield. 215 psig also appears to be the best pressure if we want high yield, as the green line is higher than the others.  Interaction effects in this type of plot are indicated by non-parallel lines, so there is not much apparent interaction here.  Before we draw any conclusions, we must test that this plot provides a true image of the chemical process, and is not just a product of high random error.&lt;br /&gt;
&lt;br /&gt;
Each data point in our problem can be thought of as the sum of several effects:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;y_{ijk} = \mu + \tau_i + \beta_j + (\tau \beta)_{ij} + \epsilon_{ijk}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; i = 1,2,...,a &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; j = 1,2,...,b &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; k = 1,2,...,n &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
and:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; y_{ijk} = &amp;lt;/math&amp;gt; data point from the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of factor A and the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th level of factor B&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \mu = &amp;lt;/math&amp;gt; the grand mean&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \tau_i = &amp;lt;/math&amp;gt; the effect of the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of factor A&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \beta_j = &amp;lt;/math&amp;gt; the effect of the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th level of factor B&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; (\tau \beta)_{ij} = &amp;lt;/math&amp;gt; the effect of interaction between factors A and B&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \epsilon_{ijk} = &amp;lt;/math&amp;gt; a random error affecting the data point from the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of factor A and the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th level of factor B&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Figure 2 illustrates how these effects sum up to each data point.  The blue line indicates the grand mean &amp;lt;math&amp;gt;\bar{y}.. \approx \mu &amp;lt;/math&amp;gt;, the red X indicates the sample mean of the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th level of temperature &amp;lt;math&amp;gt;\bar{y}_i.&amp;lt;/math&amp;gt;, the green # indicates the #th pressure level effect indicator, the brown triangle indicates the interaction effect indicator, and the black # indicates the data point from the #th level of pressure. The projection of each line onto the x-axis tells you the quantitatively what its estimated effect is, and each point can be thought of as where the data would be if effects closer to the real data point were not present.&lt;br /&gt;
&lt;br /&gt;
=== Hypothesis Testing ===&lt;br /&gt;
We would like to know if any of our estimated effects are significant.  Therefore, we have three tests to do, one for each type of effect.  For factor A, our null hypothesis will be that all the treatment means are equal (they all come from the same distribution). Our alternative hypothesis is that they are not all equal:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; H_0\!:~\tau_1 = \tau_2 = ... = \tau_a = 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; H_1\!:~\tau_i \ne 0 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similarly for factor B and the interaction effect:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; H_0\!:~\beta_1 = \beta_2 = ... = \beta_b = 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; H_1\!:~\beta_i \ne 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; H_0\!:~(\tau \beta)_{ij} = 0 &amp;lt;/math&amp;gt;  (for all &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;)&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; H_1\!:~(\tau \beta)_{ij} \ne 0 &amp;lt;/math&amp;gt;  (for at least one combination of &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;)&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To test these hypotheses, we are interested in the following sums of squares. Each of the following equations estimates the effect on our data due to the subscripted factor (&amp;lt;math&amp;gt;A &amp;lt;/math&amp;gt; → factor A, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; → factor B, &amp;lt;math&amp;gt;AB&amp;lt;/math&amp;gt; → interaction, and &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; → random error).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;SS_A = bn \Sigma^a_{i=1} (\bar{y}_i.. - \bar{y}...)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;SS_B = an \Sigma^b_{j=1} (\bar{y}._j. - \bar{y}...)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;SS_{AB} = n \Sigma^a_{i=1} \Sigma^b_{j=1} (\bar{y}_{ij}. - \bar{y}_{i}.. - \bar{y}._j. + \bar{y}...)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; SS_E = \Sigma^a_{i=1} \Sigma^b_{j=1} \Sigma^n_{k=1} (y_{ijk} - \bar{y}_{ij}.)^2 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Before we can compare these quantities to each other, we must divide by the degrees of freedom to normalize. We calculate the normalized &#039;&#039;mean squares&#039;&#039; with:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; MS_A = \frac{SS_A}{a-1} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; MS_B = \frac{SS_B}{b-1} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; MS_{AB} = \frac{SS_{AB}}{(a-1)(b-1)} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; MS_E = \frac{SS_E}{ab(n-1)} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We test whether our factor and interaction effects are significant by dividing their mean squares by the mean square of our random error to get &amp;lt;math&amp;gt;F_0&amp;lt;/math&amp;gt;.  This value must be large for us to conclude that the effect we&#039;re testing is significant (the effect must be large relative to the random error). We pick a value from the F distribution based on &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt; and our degrees of freedom – this determines what &amp;quot;large&amp;quot; is. Do this in Excel with &amp;lt;tt&amp;gt;FINV(α,numerator&#039;s DOF,denominator&#039;s DOF)&amp;lt;/tt&amp;gt; and in R with &amp;lt;tt&amp;gt;qf(1-alpha, numerator&#039;s DOF, denominator&#039;s DOF)&amp;lt;/tt&amp;gt; (DOF = degrees of freedom).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table cellspacing=0 cellpadding=5 style=&amp;quot;border-top: 1px solid black; border-bottom: 1px solid black&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Source of Variation &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Sum of Squares &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Degrees of Freedom &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Mean Square &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_0&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_{\alpha}&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Factor A (Temperature) &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.301 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 2     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.151 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 8.469 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4.256 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Factor B (Pressure) &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.768  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 2      &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.384  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 21.594 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4.256  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Interaction &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.069 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.017 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.969 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 3.633 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Error &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.160 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 9     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.018 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; Total &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 1.298 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 17    &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We conclude from these tests that the effects from factors A and B (temperature and pressure) are significant, and there is no significant interaction effect. However, these tests are only accurate if the model is correct.&lt;br /&gt;
&lt;br /&gt;
===Residual Plots and Model Adequacy Testing===&lt;br /&gt;
[[Image:5-1norm_prob.png|thumb|360px|left|&#039;&#039;&#039;Figure 3:&#039;&#039;&#039; These plots indicate that our data is sufficiently gaussian and we can believe the results of our hypothesis testing.]]&lt;br /&gt;
&lt;br /&gt;
To test whether our data is sufficiently gaussian that our hypothesis testing is valid, we create a normal probability plot.  As before, we calculate the residuals (by subtracting the expected values from the datapoints, &amp;lt;math&amp;gt;y_{ijk} - \bar{y}_{ij}.&amp;lt;/math&amp;gt;) and sorting them in an array.  Then we plot the sorted residuals against z-values, where the z-values are calculated in Excel by doing &amp;lt;tt&amp;gt;NORMSINV(percent)&amp;lt;/tt&amp;gt; or in R with &amp;lt;tt&amp;gt;qnorm(percent)&amp;lt;/tt&amp;gt;.  &amp;lt;tt&amp;gt;Percent&amp;lt;/tt&amp;gt; is equal to index_of_array/(DOF + 1). If the plot does not roughly fall along a straight line, the data is not from a Gaussian distribution.&lt;br /&gt;
&lt;br /&gt;
There are several other useful plots we can make with residuals.  Plotting &amp;lt;math&amp;gt;\bar{y}_{ij}.&amp;lt;/math&amp;gt; against residuals produces a plot that should have data randomly scattered throughout its entire area, if not the data may not be gaussian. Plotting the residuals against either of the factors should indicate that the data is more or less equally random in each level of the factors.  Note that because &amp;lt;math&amp;gt;n = 2&amp;lt;/math&amp;gt;, there is some symmetry present in each of these plots.&lt;br /&gt;
&lt;br /&gt;
Given that our data fits the model reasonably well, and that each of the factors are significant, the optimum conditions for producing high yield are when temperature is 170 °C and pressure is 200 psig.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:4-1dataplot.png&amp;diff=32</id>
		<title>File:4-1dataplot.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:4-1dataplot.png&amp;diff=32"/>
		<updated>2010-07-31T18:59:19Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_4-1,_Randomized_blocks&amp;diff=31</id>
		<title>Problem 4-1, Randomized blocks</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_4-1,_Randomized_blocks&amp;diff=31"/>
		<updated>2010-07-31T18:58:16Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;#039;==Problem Statement== &amp;#039;&amp;#039;A chemist wishes to test the effect of four chemical agents on the strength of a particular type of cloth.  Because there might be variability from one bo…&amp;#039;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Problem Statement==&lt;br /&gt;
&#039;&#039;A chemist wishes to test the effect of four chemical agents on the strength of a particular type of cloth.  Because there might be variability from one bolt to another, the chemist decides to use a randomized block design, with the bolts of cloth considered as blocks.  She selects five bolts and applies all four chemicals in random order to each bolt.  The resulting tensile strengths follow.  Analyze data from this experiment (use &amp;lt;math&amp;gt;\alpha = 0.05&amp;lt;/math&amp;gt;) and draw appropriate conclusions.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th colspan=5 bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Bolt&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Chemical&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;1&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;4&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;5&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;1&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;73&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;68&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;74&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;71&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;67&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;73&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;67&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;75&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;72&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;70&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;75&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;68&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;78&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;73&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;68&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;4&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;73&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;71&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;75&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;75&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;69&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
[[Image:4-1dataplot.png|thumb|360px|left|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; Illustration of data.]]&lt;br /&gt;
&lt;br /&gt;
In this problem we need to compare &amp;lt;math&amp;gt;a = 4&amp;lt;/math&amp;gt; treatments (chemical agents), but the data is being influenced by &amp;lt;math&amp;gt;b = 5&amp;lt;/math&amp;gt; blocks (different bolts of fabric).  Each datapoint can be thought of as the sum of several factors:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; y_{ij} = \mu + \tau_i + \beta_j + \epsilon_{ij} &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
Where:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; i = 1,2,...,a &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; j = 1,2,...,b &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
and:&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; y_{ij} = &amp;lt;/math&amp;gt; datapoint from the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th treatment and the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th block&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \mu = &amp;lt;/math&amp;gt; the grand mean&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \tau_i = &amp;lt;/math&amp;gt; the effect of the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th treatment&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \beta_j = &amp;lt;/math&amp;gt; the effect of the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th block&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; \epsilon_{ij} = &amp;lt;/math&amp;gt; a random error affecting the datapoint from the &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt;th treatment and the &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt;th block&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In the left plot of Figure 1, this equation is illustrated.  The blue line marks the grand sample mean &amp;lt;math&amp;gt;\bar{y}.. \approx \mu&amp;lt;/math&amp;gt;, and the red X marks the treatment mean &amp;lt;math&amp;gt;\bar{y}_i.&amp;lt;/math&amp;gt; (mean of all data treated by the same chemical).  The green number marks where the datapoint would be if the treatment effect and block effect were present, but there was no random error.  Finally, the black number marks the location of the datapoint – the sum of the grand mean and each effect.  The value of the numbers indicates which block that point was affected by.&lt;br /&gt;
&lt;br /&gt;
In order to examine only the effect of the chemical agent, each bolt was divided into four pieces, each of which was treated by a different chemical agent.  This serves to average out the effect of the different bolts. This is displayed in the left plot in Figure 1.  The green lines deviate from each treatment mean identically, illustrating that the block effect is constant across chemical treatments.&lt;br /&gt;
&lt;br /&gt;
The plot on the right in Figure 1 illustrates the data similarly, but the squares indicate the block means, and the treatment effects are shown deviating from them.  This is merely an alternate visualization of the sum of the grand mean and each effect.&lt;br /&gt;
&lt;br /&gt;
===Hypothesis Testing===&lt;br /&gt;
&lt;br /&gt;
As usual, our null hypothesis will be that all the treatment means are equal (and come from the same distribution).  Our alternative hypothesis is that they are not all equal:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;H_0\!:~\tau_1 = \tau_2 = ... = \tau_a = 0 &amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;H_1\!:~\tau_i \ne 0 &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to test these hypotheses, we are interested in the sums of squares of all the error effects.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt; SS_{Total} = \sum_{i=1}^a \sum_{j=1}^b y^2_{ij} - \frac{y^2\!\!..}{N} = 191.75&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; SS_{Treatments} = \frac{1}{b}\sum_{i=1}^a y^2_i\!. - \frac{y^2\!\!..}{N} = 12.95&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; SS_{Blocks} = \frac{1}{a} \sum_{j=1}^b y^2\!\!._j - \frac{y^2\!\!..}{N} = 157.00&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt; SS_{Error} = SS_{Total} - SS_{Treatments} - SS_{Blocks} = 21.80&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Each of these is a measure of its respective type of error.  Graphically, in the left plot of Figure 1 &amp;lt;math&amp;gt;SS_{Total}&amp;lt;/math&amp;gt; is the sum of the squares of the distance along the x-axis from the grand mean to each of the black numbers.  &amp;lt;math&amp;gt;SS_{Treatments}&amp;lt;/math&amp;gt; is similar, but for the x-components of the red lines. For &amp;lt;math&amp;gt;SS_{Blocks}&amp;lt;/math&amp;gt; we are summing the squares of the x-component of the green lines.  And for &amp;lt;math&amp;gt;SS_{Error}&amp;lt;/math&amp;gt; it is the x-component of the black lines that are squared and summed (it is easier to calculate this by arithmetic as in the equation above).&lt;br /&gt;
&lt;br /&gt;
By dividing the statistics above by their respective degrees of freedom, we get a normalized measure of the error effects, or a &#039;&#039;mean sum of squares&#039;&#039;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;MS_{Treatments} = \frac{SS_{Treatments}}{a-1} = 4.32&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;MS_{Blocks} = \frac{SS_{Blocks}}{b-1} = 39.25&amp;lt;/math&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;MS_{Error} = \frac{SS_{Error}}{(a-1)(b-1)} = 1.81&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In order to test our hypothesis, we must compare the treatment error to random error.  We will do this with the F statistic, &amp;lt;math&amp;gt;F_0 = \frac{MS_{Treatments}}{MS_{Error}} = 2.38&amp;lt;/math&amp;gt;.  We compare this quantity to a &amp;lt;math&amp;gt;F_{\alpha,~a-1,~(a-1)(b-1)} = 3.49&amp;lt;/math&amp;gt;, which comes from the F-distribution (this can be calculated in Excel with &amp;lt;tt&amp;gt;FINV(α,a−1,N−a)&amp;lt;/tt&amp;gt;; in R use &amp;lt;tt&amp;gt;qf(1-alpha, a-1, (a-1)*(b-1))&amp;lt;/tt&amp;gt;.  If &amp;lt;math&amp;gt;F_{0} &amp;gt; F_{\alpha,~a-1,~(a-1)(b-1)}&amp;lt;/math&amp;gt;, we reject the null hypothesis.  As this is not the case, we cannot reject our null hypothesis.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table cellspacing=0 cellpadding=5 style=&amp;quot;border-top: 1px solid black; border-bottom: 1px solid black&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Source of Variation &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Sum of Squares &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Degrees of Freedom &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Mean Square &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_0&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_{\alpha}&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; P-Value &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Treatments (Chemical agent) &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 12.95 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 3     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4.317 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 2.376 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 3.490 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 0.121 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Blocks (Bolts) &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 157.00 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 4      &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 39.250 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Error &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 21.80 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 12    &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 1.817 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; Total &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 191.75 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 19     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=30</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=30"/>
		<updated>2010-07-31T18:57:51Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1, Comparing a single mean to a specified value]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2, Comparing a single mean to a specified value (second example)|Problem 2-2, Comparing a single mean to a specified value (second example)]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-4, Determining required sample size|Problem 2-4, Determining required sample size]]&lt;br /&gt;
&lt;br /&gt;
[[Problems 3-1 through 3-3, Analysis of variance|Problems 3-1 through 3-3, Analysis of variance]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 4-1, Randomized blocks|Problem 4-1, Randomized blocks]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 5-1, Factorial designs|Problem 5-1, Factorial designs]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:3-1fake4.png&amp;diff=29</id>
		<title>File:3-1fake4.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:3-1fake4.png&amp;diff=29"/>
		<updated>2010-07-31T18:39:51Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Confidence interval on the differences in means.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Confidence interval on the differences in means.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:3-1fake3.png&amp;diff=28</id>
		<title>File:3-1fake3.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:3-1fake3.png&amp;diff=28"/>
		<updated>2010-07-31T18:39:14Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Confidence interval on the mean tensile strength for each mixing technique.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Confidence interval on the mean tensile strength for each mixing technique.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:3-1e.png&amp;diff=27</id>
		<title>File:3-1e.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:3-1e.png&amp;diff=27"/>
		<updated>2010-07-31T18:23:51Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Residuals vs. predicted tensile strength.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Residuals vs. predicted tensile strength.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:Npp.png&amp;diff=26</id>
		<title>File:Npp.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:Npp.png&amp;diff=26"/>
		<updated>2010-07-31T18:20:51Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Normal probability plot.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Normal probability plot.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:3-1b.png&amp;diff=25</id>
		<title>File:3-1b.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:3-1b.png&amp;diff=25"/>
		<updated>2010-07-31T18:19:58Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Comparison of data to a T distribution.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Comparison of data to a T distribution.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:3-1error_comparison.png&amp;diff=24</id>
		<title>File:3-1error comparison.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:3-1error_comparison.png&amp;diff=24"/>
		<updated>2010-07-31T18:18:05Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Comparison of hypothetical data sets for which most of the error is between the treatment means (left) and within the treatment means (right).&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Comparison of hypothetical data sets for which most of the error is between the treatment means (left) and within the treatment means (right).&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:3-1dataplot.png&amp;diff=23</id>
		<title>File:3-1dataplot.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:3-1dataplot.png&amp;diff=23"/>
		<updated>2010-07-31T18:17:19Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Illustration of data.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Illustration of data.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problems_3-1_through_3-3,_Analysis_of_variance&amp;diff=22</id>
		<title>Problems 3-1 through 3-3, Analysis of variance</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problems_3-1_through_3-3,_Analysis_of_variance&amp;diff=22"/>
		<updated>2010-07-31T18:11:59Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;#039;We will provide solutions for problems 3-1, 3-2 and 3-3 in this tutorial.  ==Problem statement== ===Problem 3-1=== &amp;#039;&amp;#039;The tensile strength of portland cement is being studied. Fou…&amp;#039;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;We will provide solutions for problems 3-1, 3-2 and 3-3 in this tutorial.&lt;br /&gt;
&lt;br /&gt;
==Problem statement==&lt;br /&gt;
===Problem 3-1===&lt;br /&gt;
&#039;&#039;The tensile strength of portland cement is being studied. Four different mixing techniques can be used economically. A completely randomized experiment was conducted and the following data collected.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Mixing&amp;lt;br&amp;gt;Technique&amp;lt;/th&amp;gt;&amp;lt;th colspan=4 bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Tensile Strength (lb/in&amp;lt;sup&amp;gt;2&amp;lt;/sup&amp;gt;)&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;1&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;3129&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;3000&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2865&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2890&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;3200&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;3300&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2975&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;3150&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2800&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2900&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2985&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;3050&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;td bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;4&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2600&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2700&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2600&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;2765&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Test the hypothesis that mixing techniques affect the strength of the cement. Use α = 0.05.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Construct a graphical display as described in section 3-5.3 to compare the mean tensile strengths for the four mixing techniques. What are your conclusions?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Use the Fisher LSD method with α = 0.05 to make comparisons between pairs of means.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Construct a normal probability plot of the residuals. What conclusion would you draw about the validity of the normality assumption?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Plot the residuals vs. the predicted tensile strength. Comment on the plot.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Prepare a scatter plot of the results to aid the interpretation of the results of this experiment.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Problem 3-2===&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Rework part (b) of problem 3-1 using Tukey&#039;s test with α = 0.05. Do you get the same conclusions from Tukey&#039;s test that you did from the graphical procedure and/or the Fisher LSD method?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Explain the difference between the Tukey and Fisher procedures.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Problem 3-3===&lt;br /&gt;
&#039;&#039;Reconsider the experiment in problem 3-1. Find a 95% confidence interval on the mean tensile strength of the portland cement produced by each of the four mixing techniques. Also find a 95% confidence interval on the difference in means for techniques 1 and 3. Does this aid you in interpreting the results of the experiment?&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
[[Image:3-1dataplot.png|thumb|left|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; Illustration of data.]]&lt;br /&gt;
&lt;br /&gt;
In these problems, we are given a data set with &amp;lt;math&amp;gt;a=4&amp;lt;/math&amp;gt; subsets, each containing &amp;lt;math&amp;gt;n=4&amp;lt;/math&amp;gt; values for a total of &amp;lt;math&amp;gt;N=a n=16&amp;lt;/math&amp;gt; data points. Each subset contains measurements of tensile strength of cement samples that were produced with a different mixing technique, or &#039;&#039;&#039;treatment&#039;&#039;&#039;. We define the mean over all data points as the &#039;&#039;&#039;grand mean&#039;&#039;&#039;, and the mean of each point within a given treatment as the &#039;&#039;&#039;treatment mean&#039;&#039;&#039;. To compare these data subsets, it is useful to think of each data point as the sum of the grand mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt;, the ith treatment mean &amp;lt;math&amp;gt;\tau_i&amp;lt;/math&amp;gt;, and a random error ϵ&amp;lt;sub&amp;gt;ij&amp;lt;/sub&amp;gt; specific to the jth data point in the ith treatment (refer to figure 1).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;y_{ij}=\mu+\tau_i+\epsilon_{ij}\begin{cases}i=1,2,\ldots,a\\j=1,2,\ldots,n\end{cases}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since we are given a finite set of data, we must approximate these means by calculating sample means. The grand sample mean is given by:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\bar{y}..=\frac{1}{N}\sum_{i=1}^a \sum_{j=1}^n y_{ij} = 2932&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The sample treatment means are given by:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\bar{y}_i.=\frac{1}{n}\sum_{j=1}^n y_{ij}&amp;lt;/math&amp;gt; where &amp;lt;math&amp;gt;i=1, 2, \ldots, a&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The dot indicates that you are summing over the variable it replaces.&lt;br /&gt;
&lt;br /&gt;
===Section 3-1 (A): Hypothesis testing===&lt;br /&gt;
[[Image:3-1error_comparison.png|thumb|360px|left|&#039;&#039;&#039;Figure 2:&#039;&#039;&#039; Comparison of hypothetical data sets for which most of the error is between the treatment means (left) and within the treatment means (right).]]&lt;br /&gt;
We would like to know if one of our data subsets is significantly different from the others, as this may indicate that one of our manufacturing techniques is superior (or inferior) to the others.  To compare our data subsets we are interested in whether most of the error is within the treatments (ϵ) or between the treatments (&amp;lt;math&amp;gt;\tau&amp;lt;/math&amp;gt;).  If most of the error is between the treatment means, then we can claim there are significant differences between them.  If there is too much error within the treatment means we cannot claim that they are significantly different (see figure x).  Mathematically, we can approximate the error between the treatment means as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mathrm{MS_{Treatments}}=\frac{n \sum_{i=1}^a (\bar{y}_i.-\bar{y}..)^2}{a-1}=163,247&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To approximate the error within the treatment means, it is easiest to subtract the error between the means from the total error:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mathrm{MS_{Error}}=\frac{\sum_{i=1}^a \sum_{j=1}^n (y_{ij}-\bar{y}..)^2-n \sum_{i=1}^a (\bar{y}_i.-\bar{y}..)^2}{N-a}&amp;lt;/math&amp;gt;&amp;lt;math&amp;gt;=\frac{\sum_{i=1}^a \sum_{j=1}^n y_{ij}^2-\frac{1}{n}\sum_{i=1}^{a} y_i.^2}{N-a}=12,826&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We are interested in the ratio&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;F_0=\frac{\mathrm{MS_{Treatments}}}{\mathrm{MS_{Error}}}=12.7&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To determine whether or not there are significant differences between our treatments, we will compare F&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt; to &amp;lt;math&amp;gt;F_{\alpha,\,a-1,\,N-a}=3.5&amp;lt;/math&amp;gt; from the F distribution. In Excel this value can be found using the function &amp;lt;tt&amp;gt;FINV(α,a−1,N−a)&amp;lt;/tt&amp;gt;; in R it can be found using &amp;lt;tt&amp;gt;qf(1−α,a−1,N−a)&amp;lt;/tt&amp;gt;. If &amp;lt;math&amp;gt;F_0&amp;gt;F_{\alpha,\,a-1,\,N-a}&amp;lt;/math&amp;gt;, which is the case here, then the error between treatment means is large enough compared to the error within treatment means to conclude that there is a significant difference between at least one treatment and the others.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table cellspacing=0 cellpadding=5 style=&amp;quot;border-top: 1px solid black; border-bottom: 1px solid black&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Source of Variation &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Sum of Squares &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Degrees of Freedom &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; Mean Square &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_0&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-bottom: 1px solid black;&amp;quot;&amp;gt; &amp;lt;math&amp;gt;F_{\alpha}&amp;lt;/math&amp;gt; &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Mixing Technique &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 489740 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 3      &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 163247 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 12.728 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 3.490  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt; Error &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 153908 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 12     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td&amp;gt; 12826  &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr align=&amp;quot;center&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot; style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; Total &amp;lt;/th&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 643648 &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt; 15     &amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;td style=&amp;quot;border-top: 1px solid black;&amp;quot;&amp;gt;&amp;lt;/td&amp;gt;&lt;br /&gt;
&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section 3-1 (B): Graphical display to compare mean tensile strengths===&lt;br /&gt;
[[Image:3-1b.png|thumb|left|Comparison of data to a T distribution.]]&lt;br /&gt;
A relatively simple way to visualize the treatment means, and see whether or not they are statistically equal (qualitatively), is to simply plot the four averages on the same graph as a T distribution.  We need to know what mean and standard deviation to use for our T distribution.  For the mean, we will simply use the grand mean, and we will approximate the standard deviation with &amp;lt;math&amp;gt;\sqrt{\mathrm{MS_{Error}}/n}=\sqrt{12826/4}=56.6&amp;lt;/math&amp;gt;. This approximation for the standard deviation relies on &amp;lt;math&amp;gt;\mathrm{MS_{Error}}&amp;lt;/math&amp;gt;, which does not take into account the differences between the treatment means.  It assumes that the treatment means are all equal – if they are not statistically equivalent, it will be obvious when we plot the treatment means on the same plot as the distribution.&lt;br /&gt;
&lt;br /&gt;
Looking at the plot on the left, we see that for our data it is unlikely that all of the treatment means come from the plotted distribution.  The two treatment means under the tails of the T distribution appear to be significantly different from those under the center.&lt;br /&gt;
&lt;br /&gt;
===Section 3-1 (C): Fisher LSD comparisons===&lt;br /&gt;
Fisher LSD comparisons allow each pair of treatment means to be compared.  This is done using a t-test as we did in problem 2.11 B (solution: [http://www.jlab.org/~pcarter/stats/2-11.xlsx Excel], [http://www.jlab.org/~pcarter/stats/2-11.R R]), but replacing &amp;lt;math&amp;gt;S_p&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;\sqrt{\mathrm{MS_E}}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;t_0=\frac{\bar{y}_i.-\bar{y}_j.}{\sqrt{\frac{2\mathrm{MS_E}}{n}}}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving for &amp;lt;math&amp;gt;\bar{y}_{i.} - \bar{y}_{j.}&amp;lt;/math&amp;gt; yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\bar{y}_i.-\bar{y}_j.=t_0\sqrt{\frac{2\mathrm{MS_E}}{n}}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We will compare this to a theoretical value called the least significant difference:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mathrm{LSD}=t_{\alpha/2,~N-a} \sqrt{\frac{2\mathrm{MS_E}}{n}}=174.5&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If &amp;lt;math&amp;gt;|\bar{y}_i-\bar{y}_j| &amp;gt; \mathrm{LSD}&amp;lt;/math&amp;gt; then the treatment means &amp;lt;math&amp;gt;i&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;j&amp;lt;/math&amp;gt; are significantly different.&lt;br /&gt;
&lt;br /&gt;
In Excel, you can calculate LSD using &amp;lt;tt&amp;gt;TINV(α,N-a)*sqrt(2*MSe/n)&amp;lt;/tt&amp;gt;, where &amp;lt;tt&amp;gt;MSe&amp;lt;/tt&amp;gt; is &amp;lt;math&amp;gt;\mathrm{MS_{Error}}&amp;lt;/math&amp;gt;. In R, the equivalent command is &amp;lt;tt&amp;gt;qt(1-α/2,N-a)*sqrt(2*MSe/n)&amp;lt;/tt&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
The following table shows the differences between each pair of treatment means.  Differences highlighted in &amp;lt;span style=&amp;quot;background-color: #ddddff&amp;quot;&amp;gt;blue&amp;lt;/span&amp;gt; are large enough for that pair to be considered significantly different.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th colspan=4 align=center&amp;gt;&amp;lt;math&amp;gt;|\bar{y}_i-\bar{y}_j|&amp;lt;/math&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;4&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;1&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;185.25&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;37.25&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;304.75&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;222.50&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;490.00&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td                 align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;267.50&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
To apply the Fisher LSD method to our data, we will compare LSD to each of the numbers in the chart to the left. For example, we see that 185.25 &amp;gt; 174.5, so there is a statistically significant difference between treatment 1 and 2.&lt;br /&gt;
&lt;br /&gt;
===Section 3-1 (D): Normal probability plot===&lt;br /&gt;
[[Image:npp.png|thumb|left|Normal probability plot.]]&lt;br /&gt;
&lt;br /&gt;
We have been assuming that our data is distributed normally (on a Gaussian), and that it is therefore valid to do t-tests. To be sure, we should check our normality assumption by creating a normal probability plot. This is done by plotting the residuals against values from a z-distribution.  Residuals are calculated by subtracting the corresponding treatment mean from each data point, and must be sorted before using them to make the plot.  The values we seek from a z-distribution are obtained by doing &amp;lt;tt&amp;gt;NORMSINV(percent)&amp;lt;/tt&amp;gt; in Excel, or &amp;lt;tt&amp;gt;qnorm(percent)&amp;lt;/tt&amp;gt; in R. In these commands, &amp;lt;tt&amp;gt;percent&amp;lt;/tt&amp;gt; is a number from 1/(dof+1) to dof/(dof+1) where &amp;lt;tt&amp;gt;dof&amp;lt;/tt&amp;gt; is the degrees of freedom. These commands return z-distribution values that represent ideal residual values.  If the resulting plot is roughly linear, then the normality assumption is valid.&lt;br /&gt;
&amp;lt;br clear=&amp;quot;all&amp;quot; /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section 3-1 (E): Plot of residuals vs. predicted tensile strength===&lt;br /&gt;
[[Image:3-1e.png|thumb|left|Residuals vs. predicted tensile strength.]]&lt;br /&gt;
As an estimate of the tensile strength for each treatment, we use the treatment mean. The plot of residuals vs. their treatment means gives an indication of the relative sizes of errors between (x-axis) and within (y-axis) treatments.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br clear=&amp;quot;all&amp;quot; /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section 3-1 (F): Plot of all data===&lt;br /&gt;
[[Image:3-1dataplot.png|Plot of all data.]]&lt;br /&gt;
&lt;br /&gt;
===Section 3-2 (A): Tukey test===&lt;br /&gt;
Tukey&#039;s test is similar to Fisher LSD comparisons in that they allow pairs of treatment means to be compared. However, instead of using the t-statistic, Tukey&#039;s test uses the Studentized range statistic q:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;q=\frac{\bar{y}_{max}-\bar{y}_{min}}{\sqrt{\mathrm{MS_{Error}/n}}}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Solving for &amp;lt;math&amp;gt;\bar{y}_{max}-\bar{y}_{min}&amp;lt;/math&amp;gt; yields:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\bar{y}_{max}-\bar{y}_{min} = q\sqrt{\mathrm{MS_{Error}}/n}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We will compare this with the theoretical value:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;T_\alpha = q_\alpha(a,~f) \sqrt{\mathrm{MS_{Error}}/n}=4.2 \sqrt{12826/4}=237.75&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This can be calculated in R using &amp;lt;tt&amp;gt;qtukey(1-alpha,a,N-a)*sqrt(MSe/n)&amp;lt;/tt&amp;gt; where &amp;lt;tt&amp;gt;MSe&amp;lt;/tt&amp;gt; is &amp;lt;math&amp;gt;\mathrm{MS_{Error}}&amp;lt;/math&amp;gt;. If &amp;lt;math&amp;gt;|\bar{y}_i - \bar{y}_j| &amp;gt; T_\alpha&amp;lt;/math&amp;gt;, there is a significant difference between the two treatments.&lt;br /&gt;
&lt;br /&gt;
We now compare this statistic to the differences between the treatment means. Differences highlighted in &amp;lt;span style=&amp;quot;background-color: #ddddff&amp;quot;&amp;gt;blue&amp;lt;/span&amp;gt; are large enough for that pair to be considered significantly different. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th colspan=4 align=center&amp;gt;&amp;lt;math&amp;gt;|\bar{y}_i-\bar{y}_j|&amp;lt;/math&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;4&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;1&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;185.25&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;37.25&amp;lt;/td&amp;gt;&amp;lt;td      align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;304.75&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;2&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;222.50&amp;lt;/td&amp;gt;&amp;lt;td           align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;490.00&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;3&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td                 align=&#039;right&#039; bgcolor=&amp;quot;#ddddff&amp;quot;&amp;gt;267.50&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section 3-2 (B): Difference between Tukey and Fisher procedures===&lt;br /&gt;
The Fisher procedure uses the T-statistic to compare pairs of treatment means, while the Tukey test uses the Studentized range statistic.  One consequence of this is that the Fisher procedure controls the error rate &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt; for each individual pairwise comparison, whereas the Tukey test controls the overall error rate.&lt;br /&gt;
&lt;br /&gt;
===Section 3-3: Confidence intervals===&lt;br /&gt;
&amp;lt;div style=&amp;quot;float:left; vertical-align: top; padding-right: 20px; padding-bottom: 20px;&amp;quot;&amp;gt;&lt;br /&gt;
[[Image:3-1fake3.png|thumb|none|&#039;&#039;&#039;Figure 7:&#039;&#039;&#039; Confidence interval on the mean tensile strength for each mixing technique.]]&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:3-1fake4.png|thumb|none|&#039;&#039;&#039;Figure 8:&#039;&#039;&#039; Confidence interval on the differences in means.]]&amp;lt;/div&amp;gt;&lt;br /&gt;
We want to find a 95% confidence interval on the mean tensile strength for each mixing technique. The upper bound of each confidence interval is the treatment mean plus the least significant difference, &amp;lt;math&amp;gt;\bar{y}_i.+\mathrm{LSD}&amp;lt;/math&amp;gt;. The lower bound is &amp;lt;math&amp;gt;\bar{y}_i.-\mathrm{LSD}&amp;lt;/math&amp;gt;. LSD was calculated for &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt;=0.95, so this gives us a 95% confidence interval (see figure 7).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mathrm{LSD}=t_{\alpha/2,~N-a} \sqrt{\frac{2\mathrm{MS_E}}{n}}=174.5&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;lower bound&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;treatment mean&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;upper bound&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 1&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2848&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2971&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;3094&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 2&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;3033&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;3156&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;3280&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 3&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2810&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2933&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;3057&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 4&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2543&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2666&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;2790&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
To find the confidence interval on the differences in means, we simply subtract to get the difference between our treatment means, and then use the formula above to calculate the confidence interval (see figure 8).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;lower bound&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;math&amp;gt;\bar{y}_{i.} - \bar{y}_{j.}&amp;lt;/math&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;upper bound&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 1 - 2&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;-359&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;-185&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;-10&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 1 - 3&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;-137&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;37&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;211&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 1 - 4&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;130&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;304&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;479&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 2 - 3&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;48&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;222&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;396&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 2 - 4&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;315&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;490&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;664&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Treatment 3 - 4&amp;lt;/th&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;93&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;267&amp;lt;/td&amp;gt;&amp;lt;td align=&#039;right&#039;&amp;gt;441&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=21</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=21"/>
		<updated>2010-07-31T18:10:51Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1, Comparing a single mean to a specified value]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2, Comparing a single mean to a specified value (second example)|Problem 2-2, Comparing a single mean to a specified value (second example)]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-4, Determining required sample size|Problem 2-4, Determining required sample size]]&lt;br /&gt;
&lt;br /&gt;
[[Problems 3-1 through 3-3, Analysis of variance|Problems 3-1 through 3-3, Analysis of variance]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/4-1|Problem 4-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/5-1|Problem 5-1]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:2-4fig1.png&amp;diff=20</id>
		<title>File:2-4fig1.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:2-4fig1.png&amp;diff=20"/>
		<updated>2010-07-31T16:52:45Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Confidence interval of the sample mean.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Confidence interval of the sample mean.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_2-4,_Determining_required_sample_size&amp;diff=19</id>
		<title>Problem 2-4, Determining required sample size</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_2-4,_Determining_required_sample_size&amp;diff=19"/>
		<updated>2010-07-31T16:52:17Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;#039;===Problem Statement===  &amp;#039;&amp;#039;A normally distributed random variable has an unknown mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and a known variance &amp;lt;math&amp;gt;\sigma^2=9&amp;lt;/math&amp;gt;. Find the sample size required…&amp;#039;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;===Problem Statement===&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;A normally distributed random variable has an unknown mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; and a known variance &amp;lt;math&amp;gt;\sigma^2=9&amp;lt;/math&amp;gt;. Find the sample size required to contruct a 95 percent confidence interval on the mean that has a total length of 1.0.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
===Solution===&lt;br /&gt;
[[Image:2-4fig1.png|thumb|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; Confidence interval of the sample mean.]]&lt;br /&gt;
&lt;br /&gt;
The problem is asking us how much data we need in order to say with 95% certainty that the mean &amp;lt;math&amp;gt;\mu&amp;lt;/math&amp;gt; is within a range of length one.  We know from previous problems how to calculate the boundaries of a confidence interval for the sample mean, and we are told the total length of the confidence interval should be one.  So we begin by subtracting the equation for the left edge of the confidence interval from the equation for the right edge, and setting this equal to one:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;(\mu + \frac{z_{\alpha/2}\sigma}{\sqrt{N}} ) - (\mu - \frac{z_{\alpha/2}\sigma}{\sqrt{N}}) = 1&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Simplifying and solving for N, we get:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;N = (2 z_{\alpha/2} \sigma)^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can plug in for &amp;lt;math&amp;gt;\sigma&amp;lt;/math&amp;gt; with the value given in the problem statement, and we can solve for &amp;lt;math&amp;gt;z_{alpha/2}&amp;lt;/math&amp;gt; with &amp;lt;tt&amp;gt;NORMSINV(1-0.05)&amp;lt;/tt&amp;gt; in Excel, or &amp;lt;tt&amp;gt;qnorm(1-0.05/2)&amp;lt;/tt&amp;gt; in R (we assumed an &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt; of 0.05).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;N = (2 \cdot 1.96 \cdot 3)^2 = 138.29&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can&#039;t take a .29th of a datapoint, so we round up to &amp;lt;math&amp;gt;N = 139&amp;lt;/math&amp;gt;.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=18</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=18"/>
		<updated>2010-07-31T16:51:50Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1, Comparing a single mean to a specified value]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2, Comparing a single mean to a specified value (second example)|Problem 2-2, Comparing a single mean to a specified value (second example)]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-4, Determining required sample size|Problem 2-4, Determining required sample size]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/3-1 through 3-3|3-1 through 3-3]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/4-1|Problem 4-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/5-1|Problem 5-1]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig5.png&amp;diff=17</id>
		<title>File:2-2fig5.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig5.png&amp;diff=17"/>
		<updated>2010-07-31T16:46:30Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: The confidence interval about the theoretical mean.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The confidence interval about the theoretical mean.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig4.png&amp;diff=16</id>
		<title>File:2-2fig4.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig4.png&amp;diff=16"/>
		<updated>2010-07-31T16:45:22Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: The confidence interval about the sample mean.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The confidence interval about the sample mean.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig2.png&amp;diff=15</id>
		<title>File:2-2fig2.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig2.png&amp;diff=15"/>
		<updated>2010-07-31T16:44:08Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Our plot after normalizing.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Our plot after normalizing.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig1.png&amp;diff=14</id>
		<title>File:2-2fig1.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:2-2fig1.png&amp;diff=14"/>
		<updated>2010-07-31T16:43:33Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Our data compared to a theoretical Gaussian distribution.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Our data compared to a theoretical Gaussian distribution.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_2-2,_Comparing_a_single_mean_to_a_specified_value_(second_example)&amp;diff=13</id>
		<title>Problem 2-2, Comparing a single mean to a specified value (second example)</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_2-2,_Comparing_a_single_mean_to_a_specified_value_(second_example)&amp;diff=13"/>
		<updated>2010-07-31T16:43:07Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;#039;==Problem Statement== &amp;#039;&amp;#039;The viscosity of a liquid detergent is supposed to average 800 centistokes at 25 °C. A random sample of 16 batches of detergent is collected, and the ave…&amp;#039;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Problem Statement==&lt;br /&gt;
&#039;&#039;The viscosity of a liquid detergent is supposed to average 800 centistokes at 25 °C. A random sample of 16 batches of detergent is collected, and the average viscosity is 812. Suppose we know that the standard deviation of viscosity is σ = 25 centistokes.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;State the hypotheses that should be tested.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Test these hypotheses using α = 0.05. What are your conclusions?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;What is the P-value for the test?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Find a 95 percent confidence interval on the mean.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
[[Image:2-2fig1.png|thumb|left|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; Our data compared to a theoretical Gaussian distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Section A: Choosing hypotheses===&lt;br /&gt;
&lt;br /&gt;
In this problem we are told we would like our liquid detergent to have a mean of &amp;lt;math&amp;gt;\mu_0 = 800&amp;lt;/math&amp;gt; and a standard deviation of &amp;lt;math&amp;gt;\sigma = 25&amp;lt;/math&amp;gt;: this is our theoretical distribution.  We are also told that a sample of &amp;lt;math&amp;gt;N = 16&amp;lt;/math&amp;gt; batches of detergent have an average viscosity of &amp;lt;math&amp;gt;\overline{y} = 812&amp;lt;/math&amp;gt;, which estimates the mean of our true distribution.  We would like to know if the means of our true and theoretical distributions are likely to be the same.  There are two hypotheses to consider here. Our null hypothesis is that the means of the two distributions are equal, and our alternative hypothesis is that they are not equal.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;H&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;: μ = μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;&amp;lt;br /&amp;gt;&lt;br /&gt;
H&amp;lt;sub&amp;gt;1&amp;lt;/sub&amp;gt;: μ ≠ μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The alternate hypothesis is called &#039;&#039;two-tailed&#039;&#039; because it is true if &amp;lt;math&amp;gt;\mu &amp;lt; \mu_0&amp;lt;/math&amp;gt; and if &amp;lt;math&amp;gt;\mu &amp;gt; \mu_0&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br clear=&#039;all&#039;&amp;gt;&lt;br /&gt;
[[Image:2-2fig2.png|thumb|left|&#039;&#039;&#039;Figure 2:&#039;&#039;&#039; Our plot after normalizing.]]&lt;br /&gt;
&lt;br /&gt;
===Section B: Z-values===&lt;br /&gt;
&lt;br /&gt;
To compare the mean of the true distribution to that of the theoretical distribution, we test the null hypothesis with a z-test.  The z-value is calculated as in problem 2-1:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;z=\frac{\bar{y}-\mu_0}{\sigma/\sqrt{n}}=1.92&amp;lt;/math&amp;gt;&amp;lt;br/&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Since we have a two-tailed alternative hypothesis, we must define a rejection region at both extremes of our theoretical distribution.  Our value for &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt; determines the total size of the rejection region, so we simply declare that 2.5% (since &amp;lt;math&amp;gt;\alpha&amp;lt;/math&amp;gt; = .05, or 5%) of the area on the left of our theoretical distribution is a rejection region, and 2.5% of the area on the right is also a rejection region (see Figure 2). We calculate &amp;lt;math&amp;gt;z_{\alpha/2}&amp;lt;/math&amp;gt; to determine the x-value that corresponds to the rightmost edge of the rejection region on the left, and &amp;lt;math&amp;gt;z_{1-\alpha/2}&amp;lt;/math&amp;gt; to find the leftmost edge of the rejection region on the right (see Figure 2 again).&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;z_{\alpha/2}=-1.96&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;z_{1-\alpha/2} = 1.96&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If z is between &amp;lt;math&amp;gt;{z_\alpha}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;z_{1-\alpha/2}&amp;lt;/math&amp;gt;, it is not in the rejection region and we claim the null hypothesis to be true (with a confidence of 95%).  Otherwise we claim the alternative hypothesis to be true (with the same confidence).&lt;br /&gt;
&lt;br /&gt;
Note that &amp;lt;math&amp;gt;|z_{\alpha/2}| = z_{1-\alpha/2}&amp;lt;/math&amp;gt;.  This is always true.  You can simply compare your z-value to &amp;lt;math&amp;gt;|z_{\alpha/2}|&amp;lt;/math&amp;gt; to perform a z-test.  If &amp;lt;math&amp;gt;z &amp;lt; |z_{\alpha/2}|&amp;lt;/math&amp;gt;, claim your null hypothesis to be true, otherwise claim that your alternative hypothesis is true.&lt;br /&gt;
&lt;br /&gt;
===Section C: P-values===&lt;br /&gt;
We now calculate a P-value the same way we did in problem 2-1.  Graphically, we extend the rejection region inwards from both tails until we run into our z-value.  The P-value is the area of the shaded area, calculated in Excel with &amp;lt;tt&amp;gt;=2*NORMSDIST(-ABS(z))&amp;lt;/tt&amp;gt; or in R with &amp;lt;tt&amp;gt;2*pnorm(-abs(z))&amp;lt;/tt&amp;gt;.  These functions integrate a normal distribution from negative infinity to the number we give it (in this case the negative absolute value of z), so we give it our negative z-value and multiply by two in order to get the total area of the shaded regions on the graph.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;P-value &amp;lt;math&amp;gt;= 0.0549&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Our z-value is very close to the rejection region, so a plot illustrating the extended rejection regions will look nearly identical to Figure 2.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div style=&amp;quot;float:left; vertical-align: top&amp;quot;&amp;gt;[[Image:2-2fig4.png|thumb|left|&#039;&#039;&#039;Figure 4:&#039;&#039;&#039; The confidence interval about the sample mean.]]&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:2-2fig5.png|thumb|left|&#039;&#039;&#039;Figure 5:&#039;&#039;&#039; The confidence interval about the theoretical mean.]]&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section D: Confidence intervals===&lt;br /&gt;
To calculate the limits of the confidence interval for the sample mean, we use the following formula:&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\bar{y}&amp;lt;/math&amp;gt; confidence interval limits = &amp;lt;math&amp;gt;\mu_0\pm z_{\alpha/2} \sigma/\sqrt{n}&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This tells us the range in which a sample mean could lie in order for us to accept our null hypothesis:&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;787.75&amp;lt;\bar{y}&amp;lt;812.25&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can calculate the confidence interval for &amp;lt;math&amp;gt;\mu_0&amp;lt;/math&amp;gt; in a similar way, which tells us the range in which the mean of our theoretical distribution (given a sample mean of 812) could lie in order for us to accept our null hypothesis:&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mu_0&amp;lt;/math&amp;gt; confidence interval limits = &amp;lt;math&amp;gt;\bar{y}\pm z_{\alpha/2} \sigma/\sqrt{n}&amp;lt;/math&amp;gt;&amp;lt;br /&amp;gt;&lt;br /&gt;
&amp;lt;math&amp;gt;799.75&amp;lt;\mu_0&amp;lt;824.25&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=12</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=12"/>
		<updated>2010-07-31T16:42:21Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1, Comparing a single mean to a specified value]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2, Comparing a single mean to a specified value (second example)|Problem 2-2, Comparing a single mean to a specified value (second example)]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/2-4|Problem 2-4]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/3-1 through 3-3|3-1 through 3-3]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/4-1|Problem 4-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/5-1|Problem 5-1]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=11</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=11"/>
		<updated>2010-07-31T16:40:59Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1]]&lt;br /&gt;
&lt;br /&gt;
[[Problem 2-2], Comparing a single mean to a specified value (second example)|Problem 2-2]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/2-4|Problem 2-4]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/3-1 through 3-3|3-1 through 3-3]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/4-1|Problem 4-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/5-1|Problem 5-1]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian5.png&amp;diff=10</id>
		<title>File:Gaussian5.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian5.png&amp;diff=10"/>
		<updated>2010-07-31T16:09:16Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: The confidence interval about the theoretical mean.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The confidence interval about the theoretical mean.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian4.png&amp;diff=9</id>
		<title>File:Gaussian4.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian4.png&amp;diff=9"/>
		<updated>2010-07-31T16:08:40Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: The confidence interval about the sample mean.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The confidence interval about the sample mean.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian3.png&amp;diff=8</id>
		<title>File:Gaussian3.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian3.png&amp;diff=8"/>
		<updated>2010-07-31T16:07:40Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Illustrating the P-value.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Illustrating the P-value.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian2.png&amp;diff=7</id>
		<title>File:Gaussian2.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian2.png&amp;diff=7"/>
		<updated>2010-07-31T16:06:19Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Our plot after normalizing.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Our plot after normalizing.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian1.png&amp;diff=6</id>
		<title>File:Gaussian1.png</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=File:Gaussian1.png&amp;diff=6"/>
		<updated>2010-07-31T16:04:59Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Our data compared to a theoretical Gaussian distribution.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Our data compared to a theoretical Gaussian distribution.&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Problem_2-1,_Comparing_a_single_mean_to_a_specified_value&amp;diff=5</id>
		<title>Problem 2-1, Comparing a single mean to a specified value</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Problem_2-1,_Comparing_a_single_mean_to_a_specified_value&amp;diff=5"/>
		<updated>2010-07-31T15:57:26Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: Created page with &amp;#039;==Problem statement== &amp;#039;&amp;#039;The breaking strength of a fiber is required to be at least 150 psi. Past experience has indicated that the standard deviation of breaking strength is σ …&amp;#039;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Problem statement==&lt;br /&gt;
&#039;&#039;The breaking strength of a fiber is required to be at least 150 psi. Past experience has indicated that the standard deviation of breaking strength is σ = 3 psi. A random sample of four speciments is tested, and the results are y&amp;lt;sub&amp;gt;1&amp;lt;/sub&amp;gt; = 145, y&amp;lt;sub&amp;gt;2&amp;lt;/sub&amp;gt; = 153, y&amp;lt;sub&amp;gt;3&amp;lt;/sub&amp;gt; = 150, and y&amp;lt;sub&amp;gt;4&amp;lt;/sub&amp;gt; = 147.&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol style=&amp;quot;list-style-type:lower-latin&amp;quot;&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;State the hypotheses that you think should be tested in this experiment.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Test these hypothesis using α = 0.05. What are your conclusions?&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Find the &#039;&#039;P&#039;&#039;-value for the test in part (b).&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&#039;&#039;Construct a 95 percent confidence interval on the mean breaking strength.&#039;&#039;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==Solution==&lt;br /&gt;
[[Image:Gaussian1.png|thumb|left|&#039;&#039;&#039;Figure 1:&#039;&#039;&#039; Our data compared to a theoretical Gaussian distribution.]]&lt;br /&gt;
&lt;br /&gt;
===Section A: Choosing hypotheses===&lt;br /&gt;
In this problem we are given a set of four data points. These data points all come from a distribution of breaking strengths which has an unknown mean μ. We will call this the &#039;&#039;true distribution&#039;&#039;. Previous experience indicates that breaking strengths follow a Gaussian &#039;&#039;theoretical distribution&#039;&#039; with a standard deviation of 3 psi, so we assume this for our distribution also. Our task is to determine whether or not the true mean, which is impossible to know exactly, is greater than or equal to 150 psi. We plot this data and the distribution in figure 1.&lt;br /&gt;
&lt;br /&gt;
Since the sample mean is an approximation of the true mean, we define the &#039;&#039;standard error of the mean&#039;&#039; (SEM) to be &amp;lt;math&amp;gt;\sigma/\sqrt{n}=1.5&amp;lt;/math&amp;gt;, where n=4 is the number of data points.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;table border=1 cellspacing=0 cellpadding=4&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Distribution type&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Mean&amp;lt;/th&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Standard deviation&amp;lt;/th&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;True distribution&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;Normal&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\mu\approx\overline{y}=148.25&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\sigma=3&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;tr&amp;gt;&amp;lt;th bgcolor=&amp;quot;#eeeeee&amp;quot;&amp;gt;Theoretical distribution&amp;lt;/th&amp;gt;&amp;lt;td&amp;gt;Normal&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\mu_0=150&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;td&amp;gt;&amp;lt;math&amp;gt;\sigma_0=3&amp;lt;/math&amp;gt;&amp;lt;/td&amp;gt;&amp;lt;/tr&amp;gt;&lt;br /&gt;
&amp;lt;/table&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We first state two hypotheses. The null hypothesis is that our data does come from the theoretical distribution: the true mean μ = μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;. Our alternative hypothesis states that the data comes from a distribution centered around a different mean.&lt;br /&gt;
&lt;br /&gt;
There are three choices for the alternative hypothesis: μ &amp;lt; 150, μ &amp;gt; 150, and μ ≠ 150. We adopt the convention that the alternative hypothesis will be true if the data does not meet the requirements. In this case, the breaking strength of the fiber is required to be at least 150 psi, so we choose μ &amp;lt; 150 as our alternative hypothesis.&lt;br /&gt;
&lt;br /&gt;
Formally, we state our hypotheses as:&amp;lt;br/&amp;gt;&lt;br /&gt;
&amp;lt;center&amp;gt;H&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;: μ = 150&amp;lt;br/&amp;gt;&lt;br /&gt;
H&amp;lt;sub&amp;gt;1&amp;lt;/sub&amp;gt;: μ &amp;lt; 150&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;br clear=&#039;all&#039;&amp;gt;&lt;br /&gt;
[[Image:Gaussian2.png|thumb|left|&#039;&#039;&#039;Figure 2:&#039;&#039;&#039; Our plot after normalizing.]]&lt;br /&gt;
&lt;br /&gt;
===Section B: Z-values===&lt;br /&gt;
For convenience, we start by standardizing our theoretical distribution to have a mean of zero and a standard deviation of one. To do this, we first center the distribution around zero by subtracting the theoretical mean (150) from each point in the distribution. We then divide each point by the standard deviation (3). The sample mean can be standardized in the same manner. We plot the normalized distribution and sample mean in figure 2.&lt;br /&gt;
&lt;br /&gt;
We now assume that the null hypothesis is true and ask whether or not this assumption makes sense. Given that this assumption is true, the sample mean is most likely to be close to zero. To test this, we define a range over which we consider our sample mean to be unacceptable, the &#039;&#039;rejection region&#039;&#039;. If the sample mean is in the rejection region, it is too far from zero and we reject the null hypothesis.&lt;br /&gt;
&lt;br /&gt;
We will define the lower limit to be z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;, where α=0.05. Graphically, given a standard Gaussian distribution, the area under the curve left of z&amp;lt;sub&amp;gt;0.05&amp;lt;/sub&amp;gt; is equal to 5% of the total area. You can either look up z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt; in a table or calculate it using a software package. Using Excel, the appropriate function is &amp;lt;tt&amp;gt;=NORMSINV(alpha)&amp;lt;/tt&amp;gt;. The corresponding function in R is &amp;lt;tt&amp;gt;qnorm(alpha)&amp;lt;/tt&amp;gt;. Using one of these methods, we find that z&amp;lt;sub&amp;gt;0.05&amp;lt;/sub&amp;gt;=−1.64.&lt;br /&gt;
&lt;br /&gt;
We then find the z-value of our data and compare the z-value to z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. The formula for the z-value is as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;z=(\overline{y}-\mu_0)(\frac{1}{\sigma})(\sqrt{n})&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Because we have already standardized our data, μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;=0 and σ=1, so this formula simplifies to &amp;lt;math&amp;gt;\overline{y} \sqrt{n}=-0.417\cdot2=-0.833&amp;lt;/math&amp;gt;. Note that the formula above normalizes the data, if it has not already been normalized. The z-value can be interpreted as the distance between the sample mean and μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;, scaled by a factor which makes the z-value more extreme with large sample sizes. If we take many samples, our z-value is more likely to fall in the rejection region, because we are more certain of the accuracy of our sample mean.&lt;br /&gt;
&lt;br /&gt;
The rejection region for our z-value is from negative infinity to z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. We see that our z-value is greater than z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. Therefore, we cannot reject the null hypothesis.&lt;br /&gt;
&lt;br /&gt;
[[Image:Gaussian3.png|thumb|left|&#039;&#039;&#039;Figure 3:&#039;&#039;&#039; Illustrating the P-value.]]&lt;br /&gt;
&lt;br /&gt;
===Section C: P-values===&lt;br /&gt;
Another way to judge how likely it is that our null hypothesis is true is to calculate the P-value. If we were to redo the experiment, taking four new data points, the P-value gives us the probability of our new sample mean being at least as extreme as our original sample mean. Graphically, if we extend the critical region until it reaches our z-value, the P-value is equal to the area of the shaded region (see figure 3). &lt;br /&gt;
&lt;br /&gt;
To calculate the P-value in Excel, use &amp;lt;tt&amp;gt;=NORMSDIST(-ABS(z))&amp;lt;/tt&amp;gt;. In R, use &amp;lt;tt&amp;gt;pnorm(-abs(z))&amp;lt;/tt&amp;gt;. (We use the negative absolute value because &amp;lt;tt&amp;gt;NORMSDIST&amp;lt;/tt&amp;gt; and &amp;lt;tt&amp;gt;pnorm&amp;lt;/tt&amp;gt; integrate from negative infinity to the z-value. If the z-value is positive, we instead want to integrate from the z-value to positive infinity, which is mathematically equivalent to integrating from negative infinity to the negative of the z-value.)&lt;br /&gt;
&lt;br /&gt;
For this problem, we find that the P-value is 0.202. Note that a P-value of 0.5 indicates that the sample mean is equal to the mean of the theoretical distribution. You can see this graphically by noting that the z-value will be zero in this case, and integrating the theoretical distribution to zero covers half of the area. (Recall that the total area under a standard Gaussian curve is one.) The further the P-value is from 0.5, the greater the distance between the two means.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;div style=&amp;quot;float:left; vertical-align: top; padding-right: 20px; padding-bottom: 20px;&amp;quot;&amp;gt;[[Image:Gaussian4.png|thumb|none|&#039;&#039;&#039;Figure 4:&#039;&#039;&#039; The confidence interval about the sample mean.]]&amp;lt;br&amp;gt;&lt;br /&gt;
[[Image:Gaussian5.png|thumb|none|&#039;&#039;&#039;Figure 5:&#039;&#039;&#039; The confidence interval about the theoretical mean.]]&amp;lt;/div&amp;gt;&lt;br /&gt;
&lt;br /&gt;
===Section D: Confidence intervals===&lt;br /&gt;
We now return to our original data set and theoretical distribution with the mean of 150 psi; that is, we will no longer use our normalized space.&lt;br /&gt;
&lt;br /&gt;
We will now calculate the range of sample means that would lead us to conclude that the breaking strength of our fiber is at least 150 psi, given an α of 0.05. This range is known as the confidence interval about the sample mean.&lt;br /&gt;
&lt;br /&gt;
To calculate this interval, we ask what sample mean would give us a z-value equal to z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt;. We can determine this by substituting z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt; for z into the formula for z, and solving for &amp;lt;math&amp;gt;\overline{y}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;z_\alpha=(\overline{y}-\mu_0)(\frac{1}{\sigma})(\sqrt{n}) \Rightarrow \overline{y} = \mu_0+\frac{z_\alpha \sigma}{\sqrt{n}}=147.53&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is the lower limit of our confidence interval. Because any sample mean greater than 150 is acceptable, the upper limit of the confidence interval is infinity. We plot this interval in figure 4. Formally, our confidence interval about the sample mean is&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;147.53 &amp;lt; \overline{y} &amp;lt; \infty&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We next calculate a confidence interval about the mean of the theoretical distribution, μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;. This will give us the range of minimum breaking strengths we could have specified and still found our data acceptable. We can calculate this in much the same way as the previous confidence interval: substitute z&amp;lt;sub&amp;gt;α&amp;lt;/sub&amp;gt; for z in the formula for z, but this time solve for μ&amp;lt;sub&amp;gt;0&amp;lt;/sub&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;\mu_0=\overline{y}-\frac{z_\alpha \sigma}{\sqrt{n}}=151.22&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is the upper limit of our confidence interval. The lower limit is zero, because we simply require the theoretical mean to be less than this number. Formally, our confidence interval about the theoretical mean is&lt;br /&gt;
&lt;br /&gt;
&amp;lt;center&amp;gt;&amp;lt;math&amp;gt;0 \le \mu_0 &amp;lt; 151.22&amp;lt;/math&amp;gt;&amp;lt;/center&amp;gt;&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=4</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=4"/>
		<updated>2010-07-31T15:56:38Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[Problem 2-1, Comparing a single mean to a specified value|Problem 2-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/2-2|Problem 2-2]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/2-4|Problem 2-4]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/3-1 through 3-3|3-1 through 3-3]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/4-1|Problem 4-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/5-1|Problem 5-1]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
	<entry>
		<id>https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=2</id>
		<title>Statistics Tutorial: Design and Analysis of Experiments</title>
		<link rel="alternate" type="text/html" href="https://www.micahveilleux.com/statistics/index.php?title=Statistics_Tutorial:_Design_and_Analysis_of_Experiments&amp;diff=2"/>
		<updated>2010-07-31T15:37:25Z</updated>

		<summary type="html">&lt;p&gt;Statadmin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This is a tutorial for Dr. Brash&#039;s statistics class, at Christopher Newport University. Here, we will present detailed solutions to several problems from the first few chapters of &#039;&#039;Design and Analysis of Experiments&#039;&#039;, 6th ed. by Douglas Montgomery. Solutions are available in Microsoft Excel format and [http://www.r-project.org R] format. Although beginners typically find Microsoft Excel easier to use, it proves to be very limiting for more advanced statistical analysis. R is a much more flexible and powerful software package for statistical analysis which is freely available, but has a somewhat steeper learning curve than Excel.&lt;br /&gt;
&lt;br /&gt;
==Solutions==&lt;br /&gt;
[[User:Pcarter/2-1|Problem 2-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/2-2|Problem 2-2]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/2-4|Problem 2-4]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/3-1 through 3-3|3-1 through 3-3]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/4-1|Problem 4-1]]&lt;br /&gt;
&lt;br /&gt;
[[User:Pcarter/5-1|Problem 5-1]]&lt;/div&gt;</summary>
		<author><name>Statadmin</name></author>
	</entry>
</feed>