Applied statistics and probability for engineers / Douglas C. Montgomery, George C. enhancements in this edition, including reorganizing and rewriting major. By Montgomery, Runger and Hubele Engineers, Fifth Edition Introduction to engineering statistics, By Montgomery and Runger appropriate for a one- semester course Introduction .. The pdf is variable and shown in Fig. engineering statistics montgomery 5th pdf. Engineering Statistics IES Office Hours: By Douglas C. Montgomery and George C. Runger. 5th Edition PDF.

Engineering Statistics Montgomery 5th Edition Pdf

Language:English, Japanese, German
Published (Last):03.10.2015
ePub File Size:30.57 MB
PDF File Size:19.60 MB
Distribution:Free* [*Register to download]
Uploaded by: LEROY

engineering statistics solution manual in pdf format, then you've come to the right site. we furnish the complete engineering statistics 5th edition montgomery. Engineering Statistics, 5th Edition - site edition by Douglas C. Description Montgomery, Runger, and Hubele's Engineering Statistics, 5th Edition provides modern . Introduction To Mathematical Programming 4th Edition Solutions Pdf. Editorial Reviews. About the Author. Douglas C. Montgomery, Regents' Professor of Industrial Engineering Statistics 5e + WileyPLUS Registration Card.

A comprehensive treatment of statistical methodology for quality control and improvement. Includes some aspects of quality management, such as Six Sigma. Montgomery An introduction to design and analysis of experiments, for senior and graduate students, and practitioners.

Managing, Controlling, and Improving Quality By Montgomery, Jennings, and Pfund For a first course in quality management or total quality, an organized approach to quality management, control, and improvement, focusing on both management structure and statistical and analytical tools. Introduction to Linear Regression Analysis, Fourth Edition By Montgomery, Peck, and Vining A comprehensive and thoroughly up-todate look at regression analysis, still the most widely used technique in statistics today.

Response Surface Methodology: Process and Product Optimization Using Designed Experiments, Third Edition By Myers, Montgomery and AndersonCook The exploration and optimization of response surfaces, for graduate courses in experimental design, and for applied statisticians, engineers, and chemical and physical scientists.

The range of theoretical topics and applications appeals both to students and practicing professionals. An in-depth understanding of probability is not nec- essary to understand how to use statistics for effective engineering problem solving. Other topics in this chapter include expected values, variances, probability plotting, and the central limit theorem. Techniques for a single sample are in Chapter 4, and two-sample inference techniques are in Chapter 5.

Our presentation is distinctly applications oriented and stresses the simple comparative-experiment nature of these procedures. We want engineering students to become interested in how these methods can be used to solve real- world problems and to learn some aspects of the concepts behind them so that they can see how to apply them in other settings.

We give a logical, heuristic development of the tech- niques, rather than a mathematically rigorous one. In this edition, we have focused more extensively on the P-value approach to hypothesis testing because it is relatively easy to un- derstand and is consistent with how modern computer software presents the concepts.

Empirical model building is introduced in Chapter 6. Both simple and multiple linear re- gression models are presented, and the use of these models as approximations to mechanistic models is discussed.

Chapter 7 formally introduces the design of engineering experiments, although much of Chapters 4 and 5 was the foundation for this topic. We emphasize the factorial design and, in particular, the case in which all of the experimental factors are at two levels.

Our practical ex- perience indicates that if engineers know how to set up a factorial experiment with all factors at two levels, conduct the experiment properly, and correctly analyze the resulting data, they can successfully attack most of the engineering experiments that they will encounter in the real world. Consequently, we have written this chapter to accomplish these objectives. We also introduce fractional factorial designs and response surface methods.

Statistical quality control is introduced in Chapter 8. The important topic of Shewhart control charts is emphasized. The X and R charts are presented, along with some simple control charting techniques for individuals and attribute data. We also discuss some aspects of estimating the capability of a process. The students should be encouraged to work problems to master the subject matter. The end-of-section exercises are intended to reinforce the concepts and techniques introduced in that section.

These exercises xii PREFACE are more structured than the end-of-chapter supplemental exercises, which generally require more formulation or conceptual thinking. We use the supplemental exercises as integrating problems to reinforce mastery of concepts as opposed to analytical technique. The team exercises challenge the student to apply chapter methods and concepts to problems requiring data collection. As noted later, the use of statistics software in problem solution should be an integral part of the course.

There is a tendency in teaching these courses to spend a great deal of time on probability and random variables and, indeed, some engineers, such as industrial and electrical engineers, do need to know more about these subjects than stu- dents in other disciplines and to emphasize the mathematically oriented aspects of the subject.

This type of course can be fun to teach and much easier on the instructor because it is almost always easier to teach theory than application, but it does not prepare the student for professional practice. In our course taught at Arizona State University, students meet twice weekly, once in a large classroom and once in a small computer laboratory.

Students are responsible for reading assignments, individual homework problems, and team projects. In-class team activities in- clude designing experiments, generating data, and performing analyses. The supplemental problems and team exercises in this text are a good source for these activities.

The intent is to provide an active learning environment with challenging problems that foster the development of skills for analysis and synthesis.

Therefore, we strongly recommend that the computer be integrated into the course. Throughout the book, we have presented output from Minitab as typical examples of what can be done with modern computer software.

In teaching, we have used Statgraphics, Minitab, Excel, and several other statistics packages or spreadsheets. We did not clutter the book with examples from many different packages because how the instructor integrates the software into the class is ultimately more important than which package is used.

All text data and the instructor manual are available in electronic form. WileyPLUS provides an online environment that integrates rele- vant resources, including the entire digital textbook, in an easy-to-navigate framework that helps students study more effectively. One-on-one Engagement. Students engage with related examples in various media and sample practice items, including: Craig Downing, including: Throughout each study session, students can assess their progress and gain immediate feedback.

WileyPLUS provides reliable, customizable resources that reinforce course goals inside and outside of the classroom as well as visibility into individual student progress. Pre-created materials and activities help instructors optimize their time: Customizable Course Plan: Pre-created Activity Types Include: WileyPLUS provides instant access to reports on trends in class performance, student use of course materials and progress towards learning objectives, helping inform deci- sions and drive classroom discussions.

Learn More. Powered by proven technology and built on a foundation of cognitive research, WileyPLUS has enriched the education of millions of students, in over 20 countries around the world. We are grateful to Dr.

Dale Kennedy and Dr. Mary Anderson—Rowland for their generous feedback and suggestions in teaching our course at Arizona State University. We also thank Dr. Lora Zimmer, and Dr. Sharon Lewis for their work in the development of the course based on this text when they were graduate assistants.

Engineering Statistics, 5th Edition

We are very thankful to Ms. Busaba Laungrungrong, Dr. Nuttha Lurponglukana, and Dr.

Sarah Street, Dr. James C. Ford, Dr. Craig Downing, and Mr. Patrick Egbunonu for their assistance in checking the accuracy and completeness of the text, solutions manual, and WileyPLUS.

We appreciate the staff support and resources provided by the Industrial Engineering program at Arizona State University, and our director Dr. Ronald Askin. Several reviewers provided many helpful suggestions, including Dr. Thomas Willemain, Rensselaer, Dr. John D. David Mathiason.

We are also indebted to Dr. Smiley Cheng of the University of Manitoba for permission to adapt many of the statistical tables from his excellent book with Dr. This project was supported, in part, by the National Science Foundation Opinions expressed are those of the authors and not necessarily those of the Foundation Douglas C. Runger Norma Faris Hubele. The Ironto Wayside Footbridge was built in and is the oldest standing metal bridge in Virginia.

Although it has now been restored as a footbridge, in its former life it routinely carried heavy wagonloads, three tons or more, of goods and materials. Huffman conducted a historical survey of the bridge and found that a load-bearing analysis had never been done. After gathering the available structural data on the bridge, she created a computer model stress analysis based on typical loads that it would have carried. After analyzing her results, she tested them on the bridge itself to verify her model.

She set up dial gauges under the cen- ter of each truss. She then had a 3.

Introduction to Time Series Analysis and Forecasting

Her results and conclusions will be helpful in maintaining the bridge and in helping others to restore and study historic bridges. Her adviser Cris Moen points out that her computer model can be used to create structural models to test other bridges.

It is an excellent example of using sample data to verify an engineering model. The steps in the engineering method are as follows: Develop a clear and concise description of the problem. Identify, at least tentatively, the important factors that affect this problem or that may play a role in its solution. State any limitations or assumptions of the model. Conduct appropriate experiments and collect data to test or validate the tentative model or conclusions made in steps 2 and 3.

Manipulate the model to assist in developing a solution to the problem. Draw conclusions or make recommendations based on the problem solution.

Develop Identify Propose or Manipulate Confirm Draw a clear the important refine the model the solution conclusions description a model and make of the problem factors recommendations Collect data Figure The engineering problem-solving method.

Steps 2—4 in Fig. Many of the engineering Many aspects of engineering practice involve collecting, working with, and using data in the sciences are employed in solution of a problem, so knowledge of statistics is just as important to the engineer as the engineering problem- knowledge of any of the other engineering sciences. Statistical methods are a powerful aid solving method: For example, consider the gasoline mileage performance of your car.

Do you always get as thermodynamics and exactly the same mileage performance on every tank of fuel? Of course not—in fact, sometimes heat transfer the mileage performance varies considerably. These factors represent potential sources of variability in the system. Statistics gives us a framework for describing this variability and for learning about which potential sources of variability are the most important or have the greatest impact on the gasoline mileage performance. We also encounter variability in most types of engineering problems.

For example, suppose that an engineer is developing a rubber compound for use in O-rings. The O-rings are to be employed as seals in plasma etching tools used in the semiconductor industry, so their resistance to acids and other corrosive substances is an important characteristic. The tensile strengths in psi of the eight O-rings are , , , , , , , and As we should have anticipated, not all the O-ring specimens exhibit the same measurement of tensile strength.

There is variability in the tensile strength measurements. Because the measurements exhibit variability, we say. The constant remains uncover patterns in data. However, this never happens in engineering practice, so the actual measurements we observe exhibit variability.

Figure is a dot diagram of the O-ring tensile strength data. The dot diagram is a very useful plot for displaying a small body of data, say, up to about 20 observations. This plot allows us to easily see two important features of the data: The need for statistical thinking arises often in the solution of engineering problems.

Applied Statistics And Probability For Engineers Books

Consider the engineer developing the rubber O-ring material. From testing the initial specimens, he knows that the average tensile strength is Eight O-ring specimens are made from this modified rubber compound and subjected to the nitric acid emersion test described earlier. The tensile test results are , , , , , , , and The tensile test data for both groups of O-rings are plotted as a dot diagram in Fig.

However, there are some obvious questions to ask. For instance, how do we know that another set of O-ring specimens will not give different results?

In other words, are these results due entirely to chance? Is a sample of eight O-rings adequate to give reliable results? Statistical inference is has no effect on tensile strength? Statistical thinking and methodology can help answer the process of deciding if these questions. This reasoning is from a sample such as the eight rubber O-rings to a population such as the O-rings that will be sold to customers and is re- ferred to as statistical inference. See Fig.

Clearly, reasoning based on measurements from some objects to measurements on all objects can result in errors called sampling errors. We can think of each sample of eight O-rings as a random and representative sample of all parts that will ultimately be manufactured. The order in which each O-ring was tested was also randomly determined. This is an example of a completely randomized designed experiment. Sometimes the objects to be used in the comparison are not assigned at random to the treatments.

For example, the September issue of Circulation a medical journal pub- lished by the American Heart Association reports a study linking high iron levels in the body with increased risk of heart attack.

The researchers just tracked the subjects over time. This type of study is called an observational study. Designed experiments and observational studies are discussed in more detail in the next section. For example, the difference in heart attack risk could be attributable to the dif- ference in iron levels or to other underlying factors that form a reasonable explanation for the observed results—such as cholesterol levels or hypertension.

In the engineering environment, the data are almost always a sample that has been se- lected from some population. In the previous section we introduced some simple methods for summarizing and visualizing data. In the engineering environment, the data are almost always a sample that has been selected from some population.

A sample is a subset of the population containing the observed objects or the outcomes and the resulting data. Generally, engineering data are collected in one of three ways: When little thought is put into the data collection procedure, serious problems for both the statistical analysis and the practical inter- pretation of results can occur. Montgomery, Peck, and Vining describe an acetone-butyl alcohol distillation col- umn. A schematic of this binary column is shown in Fig.

For this column, production person- nel maintain and archive the following records: The production personnel very infrequently change this rate. In most such studies, the engineer is interested in using the data to construct a model relating the variables of interest. These types of models are called empirical models, and they are illustrated in more detail in Section A retrospective study takes advantage of previously collected, or historical, data.

It has the advantage of minimizing the cost of collecting the data for the study. However, there are several potential problems: The historical data on the two temperatures and the acetone concentration do not correspond directly.

Constructing an approximate correspondence would probably require making several assumptions and a great deal of effort, and it might be impos- sible to do reliably. Because the two temperatures do not vary. Within the narrow ranges that they do vary, the condensate temperature tends to in- crease with the reboil temperature. Retrospective studies, although often the quickest and easiest way to collect engineering process data, often provide limited useful information for controlling and analyzing a process.

In general, their primary disadvantages are as follows: Some of the important process data often are missing. The reliability and validity of the process data are often questionable. The nature of the process data often may not allow us to address the problem at hand. The engineer often wants to use the process data in ways that they were never in- tended to be used.

Using historical data always involves the risk that, for whatever reason, some of the important data were not collected or were lost or were inaccurately transcribed or recorded. Consequently, historical data often suffer from problems with data quality.

These errors also make historical data prone to outliers. Just because data are convenient to collect does not mean that these data are useful. Historical data cannot provide this information if information on some important variables was never collected. For example, the ambient temperature may affect the heat losses from the distillation column.

On cold days, the column loses more heat to the environment than during very warm days. The production logs for this acetone-butyl alcohol column do not routinely record the ambient temperature. Also, the concentration of acetone in the input feed stream has an effect on the acetone concentra- tion in the output product stream.

However, this variable is not easy to measure routinely, so it is not recorded either. Consequently, the historical data do not allow the engineer to include either of these factors in the analysis even though potentially they may be important.

The purpose of many engineering data analysis efforts is to isolate the root causes underly- ing interesting phenomena. With historical data, these interesting phenomena may have occurred months, weeks, or even years earlier. Analyses based on historical data often identify interesting phenomena that go unexplained.

Finally, retrospective studies often involve very large indeed, even massive data sets. As the name implies, an observational study simply observes the process or population during a period of routine operation. Usually, the engineer interacts or disturbs the process only as much as is required to obtain data on the system, and often a special effort is made to collect data on variables that are not routinely recorded, if it is thought that such data might be useful.

With proper planning, observational studies can ensure accurate, complete, and reliable data. The data col- lection form should provide the ability to add comments to record any other interesting phenomena that may occur, such as changes in ambient temperature.

It may even be possible to arrange for the input feed stream acetone concentration to be measured along with the other variables during this relatively short-term study. An observational study conducted in this manner would help ensure accurate and reliable data collection and would take care of problem 2 and possibly some aspects of problem 1 associated with the retrospective study. This approach also minimizes the chances of observing an outlier related to some error in the data.

Unfortunately, an observational study cannot address problems 3 and 4. Observational studies can also involve very large data sets.

In a designed experiment, the engineer makes deliberate or purposeful changes in controllable variables called factors of the system, observes the resulting system output, and then makes a decision or an inference about which variables are responsible for the changes that he or she observes in the output performance. An important distinction between a designed experiment and either an observational or retrospective study is that the different combinations of the factors of interest are applied randomly to a set of experimental units.

This allows cause-and-effect relationships to be established, something that cannot be done with observational or retrospective studies.

The O-ring example is a simple illustration of a designed experiment. That is, a deliber- ate change was introduced into the formulation of the rubber compound with the objective of discovering whether or not an increase in the tensile strength could be obtained. This is an experiment with a single factor. We can view the two groups of O-rings as having the two formu- lations applied randomly to the individual O-rings in each group.

This establishes the desired cause-and-effect relationship. These techniques are introduced and illustrated extensively in Chapters 4 and 5. A designed experiment can also be used in the distillation column problem. Suppose that we have three factors: The experimental design must ensure that we can separate out the effects of these three factors on the response variable, the concentration of acetone in the output product stream.

In a designed experiment, often only two or three levels of each factor are employed. The best experimental strategy to use when there are several factors of interest is to conduct a factorial experiment. In a factorial experiment, the factors are varied together in an arrangement that tests all possible combinations of factor levels. Figure illustrates a factorial experiment for the distillation column. Because all three fac- tors have two levels, there are eight possible combinations of factor levels, shown geometrically as the eight corners of the cube in Fig.

The tabular representation in Fig. The actual experimental runs would be conducted in random order, thus establishing the random assignment of factor-level combinations to experi- mental units that is the key principle of a designed experiment. Two trials, or replicates, of the experiment have been performed in random order , resulting in 16 runs also called observations.

Some very interesting tentative conclusions can be drawn from this experiment. First, com- pare the average acetone concentration for the eight runs with condenser temperature at the high level with the average concentration for the eight runs with condenser temperature at the low level these are the averages of the eight runs on the left and right faces of the cube in Fig. Thus, increasing the condenser temperature from the low to the high level increases the average concentration by 0.

The reboil temperature effect can be evaluated by comparing the average of the eight runs in the top of the cube with the average of the eight runs in the bottom, or The effect of increasing the reboil temperature is to increase the average concentration by 0. This graph was constructed by calculating the Rate Temp. Figure A four-factor factorial experiment for the distillation column. This is an example of an interaction between two factors.

Interactions occur often in physical and chemical systems, and factorial experiments are the only way to investigate their effects.

In fact, if interactions are present and the factorial experimental strategy is not used, incorrect or misleading results may be obtained. We can easily extend the factorial strategy to more factors.

Suppose that the engineer wants to consider a fourth factor, the concentration of acetone in the input feed stream. Figure illustrates how all four factors could be investigated in a factorial design. Note that as in any factorial design, all possible combinations of the four factors are tested. The experiment requires 16 trials. If each combi- nation of factor levels in Fig. Generally, if there are k factors and they each have two levels, a factorial experimental design will require 2k runs.

Clearly, as the number of factors increases, the number of trials required in a factorial experiment increases rapidly; for instance, eight factors each at two levels would require trials.

This amount of testing quickly becomes unfeasible from the viewpoint of time and other resources. A fractional factorial experiment is a variation of the basic factorial arrangement in which only a subset of the factor combinations are actually tested.

Figure shows a fractional factorial experimental design for the four- factor version of the distillation column experiment.

This experimental design requires only 8 runs instead of the original 16; consequently, it would be called a one-half fraction. This is an excellent experimental design in which to study all four factors. It will provide good information about the individual effects of the four factors and some information about how these factors interact. Factorial and fractional factorial experiments are used extensively by engineers and scientists in industrial research and development, where new technology, products, and processes are.

Reboil temp. Figure A fractional factorial experiment for the distillation column. Chapter 7 focuses on these principles, concentrating on the factorial and fractional factorials that we have introduced here. The objective is to use the sample data to make decisions or learn something about the population. Recall that the population is the complete collection of items or objects from which the sample is taken.

A sample is just a subset of the items in the population. For example, suppose that we are manufacturing semiconductor wafers, and we want to learn about the resistivity of the wafers in a particular lot. In this case, the lot is the population. Data are often collected as a result of an engineering experiment.The important topic of Shewhart control charts is emphasized.

David Mathiason. The best experimental strategy to use when there are several factors of interest is to conduct a factorial experiment. Figure also displays the data obtained when one adjustment a decrease of two units is applied to the mean after the shift is detected at observation number Column Solution.

Upon completion of the review period, please return the evaluation copy to Wiley. Students are responsible for reading assignments, individual homework problems, and team projects.