Industrial Waste
Management
-------
This Guide provides state-of-the-art tools and
practices to enable you to tailor hands-on
solutions to the industrial waste management
challenges you face.
WHAT'S AVAILABLE
• Quick reference to multimedia methods for handling and disposing of wastes
from all types of industries
• Answers to your technical questions about siting, design, monitoring, operation.
and closure of waste facilities
• Interactive, educational tools, including air and ground water risk assessment
models, fact sheets, and a facility siting tool.
• Best management practices, from risk assessment and public participation to
waste reduction, pollution prevention, and recycling
-------
;NOWLEDGEMENTS
The rdowing members of the Industrial Waste Focus Group and the Industrial Waste Steering Commiw are grateUy
acknowledged far al of their time and assistance in the development of this guidance document
Current Industrial Waste Focus
Group Members
Paul Bar*, The Dow Chemical
Company
Walter Carey. Nestle USA Inc and
New Miltord Farms
Rama Chaturvedi Bethlehem Steel
Corporation
H.C. Clark. Rice University
Barbara Dodds, League of Women
voters
Chuck Feerick. Exxon Mobil
Corporation
Stacey Ford. Exxon Mobil
Corporation
Robert Giraud OuPont Company
John Harney Citizens Round
Tabte/PURE
Kyle Isakower. American Petroleum
Institute
Richard Jarman, National Food
Processors Association
James Meiers, Cinergy Power
Generation Services
Scott Murto. General Motors and
American Foundry Society
James Roewer, Edison Electric
Institute
Edward Repa. Environmental
Industry Association
Tim Savior, International Paper
Amy Schaffer. Weyerhaeuser
Ed Skemofc, WMX Technologies. Inc
Michael Wach Western
Environmental Law Center
David Wens, University of South
Wabnms Medical Center
Pat Gwn Cherokee Nation of
Oklahoma
Past industrial Waste Focus
Group Members
Dora Cetofius. Sierra Club
Brian Forrestal. Laidlaw Waste
Systems
Jonathan Greenberg. Browning-
Ferris Industries
Michael Gregory, Arizona Toxics
Information and Sierra Club
Andrew Mites The Dexter
Corporation
Gary Robbins, Exxon Company
Kevin Sail. National Paint & Coatings
Association
Bruce SteJne. American Iron & Steel
Lisa Williams, Aluminum Association
Cuircnt Industrial Waste Steering
Committee Members
Keiiy Catalan Aaaocauon oi Slate
and Territorial Solid Waste
Management Officials
Marc Crooks, Washington State
Department ot Ecology
Cyndi Darling. Maine Department of
Environmental Protection
Jon DilDard Montana Department of
Environmental Qualty
Anne Dobbs. Texas Natural
Resources Conservation
Commission
Richard Hammond New York State
Department of Environmental
Conservation
Elizabeth Haven California State
Waste Resources Control Board
Jim Hul Missouri Department of
Natural Resources
Jim Knudson, Washington State
Department of Ecology
Chris McGuire, Florida Department
of Environmental Protection
Gene Mitchell Wisconsin
Department of Natural Resources
William Pounds, Pennsylvania
Department of Environmental
Protection
Bijan Sharafkhani Louisiana
Department of Environmental
Qualty
James Warner, Minnesota Pollution
Control Agency
ittustrial Waste Steering
Pamela um*. nianie
Environmental Protection
NormGumenik Arizona Department
of Environmental Qualty
Steve Jenkins, Alabama Department
of Environmental Management
Jim North Arizona Department of
Environmental Quality
-------
Industrial waste is generated by the production
of commercial goods, products, or services.
Examples include wastes from the production
of chemicals, iron and steel, and food goods.
-------
United States Office of Environmental EPA/600/R-96/084
Environmental Protection Information July, 2000
Agency Washington, DC 20460
<&EPA Guidance for
Data Quality Assessment
Practical Methods for
Data Analysis
EPA QA/G-9
QAOO UPDATE
-------
-------
FOREWORD
This document is the 2000 (QAOO) version of the Guidance for Data Quality Assessment
which provides general guidance to organizations on assessing data quality criteria and
performance specifications for decision making. The Environmental Protection Agency (EPA) has
developed a process for performing Data Quality Assessment (DQA) Process for project
managers and planners to determine whether the type, quantity, and quality of data needed to
support Agency decisions has been achieved. This guidance is the culmination of experiences in
the design and statistical analyses of environmental data in different Program Offices at the EPA.
Many elements of prior guidance, statistics, and scientific planning have been incorporated into
this document.
This document is distinctly different from other guidance documents; it is not intended to
be read in a linear or continuous fashion. The intent of the document is for it to be used as a
"tool-box" of useful techniques in assessing the quality of data. The overall structure of the
document will enable the analyst to investigate many different problems using a systematic
methodology.
This document is one of a series of quality management guidance documents that the EPA
Quality Staff has prepared to assist users in implementing the Agency-wide Quality System. Other
related documents include:
EPA QA/G-4 Guidance for the Data Quality Objectives Process
EPA QA/G-4D DEFT Software for the Data Quality Objectives Process
EPA QA/G-4HW Guidance for the Data Quality Objectives Process for Hazardous
Waste Site Investigations
EPA QA/G-9D Data Quality Evaluation Statistical Toolbox (DataQUEST)
This document is intended to be a "living document" that will be updated periodically to
incorporate new topics and revisions or refinements to existing procedures. Comments received
on this 2000 version will be considered for inclusion in subsequent versions. Please send your
written comments on Guidance for Data Quality Assessment to:
Quality Staff (2811R)
Office of Environmental Information
U.S. Environmental Protection Agency
1200 Pennsylvania Avenue, NW
Washington, DC 20460
Phone: (202) 564-6830
Fax: (202)565-2441
E-mail: quality@epa.gov
EPA QA/G-9 Final
QAOO Version i July 2000
-------
This page is intentionally blank.
EPA QA/G-9 Final
QAOO Version ii July 2000
-------
TABLE OF CONTENTS
Page
INTRODUCTION 0-1
0.1 PURPOSE AND OVERVIEW 0-1
0.2 DQA AND THE DATA LIFE CYCLE 0-2
0.3 THE 5 STEPS OF DQA 0-2
0.4 INTENDED AUDIENCE 0-4
0.5 ORGANIZATION 0-4
0.6 SUPPLEMENTAL SOURCES 0-4
STEP 1: REVIEW DQOs AND THE SAMPLING DESIGN 1-1
1.1 OVERVIEW AND ACTIVITIES 1-3
1.1.1 Review Study Objectives 1-4
1.1.2 Translate Objectives into Statistical Hypotheses 1-4
1.1.3 Develop Limits on Decision Errors 1-5
1.1.4 Review Sampling Design 1-7
1.2 DEVELOPING THE STATEMENT OF HYPOTHESES 1-9
1.3 DESIGNS FOR SAMPLING ENVIRONMENTAL MEDIA 1-11
1.3.1 Authoritative Sampling 1-11
1.3.2 Probability Sampling 1-13
1.3.2.1 Simple Random Sampling 1-13
1.3.2.2 Sequential Random Sampling 1-13
1.3.2.3 Systematic Samples 1-14
1.3.2.4 Stratified Samples 1-14
1.3.2.5 Compositing Physical Samples 1-15
1.3.2.6 Other Sampling Designs 1-15
STEP 2: CONDUCT A PRELIMINARY DATA REVIEW 2-1
2.1 OVERVIEW AND ACTIVITIES 2-3
2.1.1 Review Quality Assurance Reports 2-3
2.1.2 Calculate Basic Statistical Quantities 2-4
2.1.3 Graph the Data 2-4
2.2 STATISTICAL QUANTITIES 2-5
2.2.1 Measures of Relative Standing 2-5
2.2.2 Measures of Central Tendency 2-6
2.2.3 Measures of Dispersion 2-8
2.2.4 Measures of Association 2-8
2.2.4.1 Pearson's Correlation Coefficient 2-8
2.2.4.2 Spearman's Rank Correlation Coefficient 2-11
2.2.4.3 Serial Correlation Coefficient 2-11
EPA QA/G-9 Final
QAOO Version iii July 2000
-------
Page
2.3 GRAPHICAL REPRESENTATIONS 2-13
2.3.1 Histogram/Frequency Plots 2-13
2.3.2 Stem-and-Leaf Plot 2-15
2.3.3 Box and Whisker Plot 2-17
2.3.4 Ranked Data Plot 2-17
2.3.5 Quantile Plot 2-21
2.3.6 Normal Probability Plot (Quantile-Quantile Plot) 2-22
2.3.7 Plots for Two or More Variables 2-26
2.3.7.1 Plots for Individual Data Points 2-26
2.3.7.2 Scatter Plot 2-27
2.3.7.3 Extensions of the Scatter Plot 2-27
2.3.7.4 Empirical Quantile-Quantile Plot 2-30
2.3.8 Plots for Temporal Data 2-30
2.3.8.1 Time Plot 2-32
2.3.8.2 Plot of the Autocorrelation Function (Correlogram) 2-33
2.3.8.3 Multiple Observations Per Time Period 2-35
2.3.9 Plots for Spatial Data 2-36
2.3.9.1 Posting Plots 2-37
2.3.9.2 Symbol Plots 2-37
2.3.9.3 Other Spatial Graphical Representations 2-39
2.4 Probability Distributions 2-39
2.4.1 The Normal Distribution 2-39
2.4.2 The t-Distribution 2-40
2.4.3 The Lognormal Distribution 2-40
2.4.4 Central Limit Theorem 2-41
STEP 3: SELECT THE STATISTICAL TEST 3-1
3.1 OVERVIEW AND ACTIVITIES 3-3
3.1.1 Select Statistical Hypothesis Test 3-3
3.1.2 Identify Assumptions Underlying the Statistical Test 3-3
3.2 TESTS OF HYPOTHESES ABOUT A SINGLE POPULATION 3-4
3.2.1 Tests for a Mean 3-4
3.2.1.1 The One-Sample t-Test 3-5
3.2.1.2 The Wilcoxon Signed Rank (One-Sample) Test 3-11
3.2.1.3 The Chen Test 3-15
3.2.2 Tests for a Proportion or Percentile 3-16
3.2.2.1 The One-Sample Proportion Test 3-18
3.2.3 Tests for a Median 3-18
3.2.4 Confidence Intervals 3-20
3.3 TESTS FOR COMPARING TWO POPULATIONS 3-21
EPA QA/G-9 Final
QAOO Version iv July 2000
-------
Page
3.3.1 Comparing Two Means 3-22
3.3.1.1 Student's Two-Sample t-Test (Equal Variances) 3-23
3.3.1.2 Satterthwaite's Two-Sample t-Test (Unequal Variances) . . 3-23
3.3.2 Comparing Two Proportions or Percentiles 3-27
3.3.2.1 Two-Sample Test for Proportions 3-28
3.3.3 Nonparametric Comparisons of Two Population 3-31
3.3.3.1 The Wilcoxon Rank Sum Test 3-31
3.3.3.2 The Quantile Test 3-35
3.3.4 Comparing Two Medians 3-36
3.4 Tests for Comparing Several Populations 3-37
3.4.1 Tests for Comparing Several Means 3-37
3.4.1.1 Dunnett's Test 3-38
STEP 4: VERIFY THE ASSUMPTIONS OF THE STATISTICAL TEST 4-1
4.1 OVERVIEW AND ACTIVITIES 4-3
4.1.1 Determine Approach for Verifying Assumptions 4-3
4.1.2 Perform Tests of Assumptions 4-4
4.1.3 Determine Corrective Actions 4-5
4.2 TESTS FOR DISTRIBUTIONAL ASSUMPTIONS 4-5
4.2.1 Graphical Methods 4-7
4.2.2 Shapiro-Wilk Test for Normality (the W test) 4-8
4.2.3 Extensions of the Shapiro-Wilk Test (Filliben's Statistic) 4-8
4.2.4 Coefficient of Variation 4-8
4.2.5 Coefficient of Skewness/Coefficient of Kurtosis Tests 4-9
4.2.6 Range Tests 4-10
4.2.7 Goodness-of-Fit Tests 4-12
4.2.8 Recommendations 4-13
4.3 TESTS FOR TRENDS 4-13
4.3.1 Introduction 4-13
4.3.2 Regression-Based Methods for Estimating and Testing for Trends .4-14
4.3.2.1 Estimating a Trend Using the Slope of the Regression Line 4-14
4.3.2.2 Testing for Trends Using Regression Methods 4-15
4.3.3 General Trend Estimation Methods 4-16
4.3.3.1 Sen's Slope Estimator 4-16
4.3.3.2 Seasonal Kendall Slope Estimator 4-16
4.3.4 Hypothesis Tests for Detecting Trends 4-16
4.3.4.1 One Observation per Time Period for
One Sampling Location 4-16
4.3.4.2 Multiple Observations per Time Period
for One Sampling Location 4-19
EPA QA/G-9 Final
QAOO Version v July 2000
-------
Page
4.3.4.3 Multiple Sampling Locations with Multiple Observations .4-20
4.3.4.4 One Observation for One Station with Multiple Seasons .4-22
4.3.5 A Discussion on Tests for Trends 4-23
4.3.6 Testing for Trends in Sequences of Data 4-24
4.4 OUTLIERS 4-24
4.4.1 Background 4-24
4.4.2 Selection of a Statistical Test 4-27
4.4.3 Extreme Value Test (Dixon's Test) 4-27
4.4.4 Discordance Test 4-29
4.4.5 Rosner's Test 4-30
4.4.6 Walsh's Test 4-32
4.4.7 Multivariate Outliers 4-32
4.5 TESTS FOR DISPERSIONS 4-33
4.5.1 Confidence Intervals for a Single Variance 4-33
4.5.2 The F-Test for the Equality of Two Variances 4-33
4.5.3 Bartlett's Test for the Equality of Two or More Variances 4-33
4.5.4 Levene's Test for the Equality of Two or More Variances 4-35
4.6 TRANSFORMATIONS 4-39
4.6.1 Types of Data Transformations 4-39
4.6.2 Reasons for Transforming Data 4-41
4.7 VALUES BELOW DETECTION LIMITS 4-42
4.7.1 Less than 15% Nondetects - Substitution Methods 4-43
4.7.2 Between 15-50% Nondetects 4-43
4.7.2.1 Cohen's Method 4-43
4.7.2.2 Trimmed Mean 4-45
4.7.2.3 Winsorized Mean and Standard Deviation 4-45
4.7.2.4 Atchison's Method 4-46
4.7.2.5 Selecting Between Atchison's Method or Cohen's Method .4-49
4.7.3 Greater than 5-% Nondetects - Test of Proportions 4-50
4.7.4 Recommendations 4-50
4.8 INDEPENDENCE 4-51
STEP 5: DRAW CONCLUSIONS FROM THE DATA 5-1
5.1 OVERVIEW AND ACTIVITIES 5-3
5.1.1 Perform the Statistical Hypothesis Test 5-3
5.1.2 Draw Study Conclusions 5-3
5.1.3 Evaluate Performance of the Sampling Design 5-5
5.2 INTERPRETING AND COMMUNICATING THE TEST RESULTS 5-6
5.2.1 Interpretation of p-Values 5-7
5.2.2 "Accepting" vs. "Failing to Reject" the Null Hypothesis 5-7
EPA QA/G-9 Final
QAOO Version vi July 2000
-------
5.2.3 Statistical Significance vs. Practical Significance
5.2.4 Impact of Bias on Test Results
5.2.5 Quantity vs. Quality of Data
5.2.6 "Proof of Safety" vs. "Proof of Hazard"
APPENDIX A: STATISTICAL TABLES A - 1
APPENDIX B: REFERENCES B - 2
EPA QA/G-9 Final
QAOO Version vii July 2000
-------
This page is intentionally blank.
EPA QA/G-9 Final
QAOO Version viii July 2000
-------
INTRODUCTION
0.1 PURPOSE AND OVERVIEW
Data Quality Assessment (DQA) is the scientific and statistical evaluation of data to
determine if data obtained from environmental data operations are of the right type, quality, and
quantity to support their intended use. This guidance demonstrates how to use DQA in
evaluating environmental data sets and illustrates how to apply some graphical and statistical tools
for performing DQA. The guidance focuses primarily on using DQA in environmental decision
making; however, the tools presented for preliminary data review and verifying statistical
assumptions are useful whenever environmental data are used, regardless of whether the data are
used for decision making.
DQA is built on a fundamental premise: data quality, as a concept, is meaningful only
when it relates to the intended use of the data. Data quality does not exist in a vacuum; one must
know in what context a data set is to be used in order to establish a relevant yardstick for judging
whether or not the data set is adequate. By using the DQA, one can answer two fundamental
questions:
1. Can the decision (or estimate) be made with the desired confidence, given the quality of
the data set?
2. How well can the sampling design be expected to perform over a wide range of possible
outcomes? If the same sampling design strategy is used again for a similar study, would
the data be expected to support the same intended use with the desired level of
confidence, particularly if the measurement results turned out to be higher or lower than
those observed in the current study?
The first question addresses the data user's immediate needs. For example, if the data
provide evidence strongly in favor of one course of action over another, then the decision maker
can proceed knowing that the decision will be supported by unambiguous data. If, however, the
data do not show sufficiently strong evidence to favor one alternative, then the data analysis alerts
the decision maker to this uncertainly. The decision maker now is in a position to make an
informed choice about how to proceed (such as collect more or different data before making the
decision, or proceed with the decision despite the relatively high, but acceptable, probability of
drawing an erroneous conclusion).
The second question addresses the data user's potential future needs. For example, if
investigators decide to use a certain sampling design at a different location from where the design
was first used, they should determine how well the design can be expected to perform given that
the outcomes and environmental conditions of this sampling event will be different from those of
the original event. Because environmental conditions will vary from one location or time to
another, the adequacy of the sampling design approach should be evaluated over a broad range of
possible outcomes and conditions.
EPA QA/G-9 Final
QAOO Version 0 - 1 July 2000
-------
0.2 DQA AND THE DATA LIFE CYCLE
The data life cycle (depicted in Figure 0-1) comprises three steps: planning,
implementation, and assessment. During the planning phase, the Data Quality Objectives (DQO)
Process (or some other systematic planning procedure) is used to define quantitative and
qualitative criteria for determining when, where, and how many samples (measurements) to
collect and a desired level of confidence. This information, along with the sampling methods,
analytical procedures, and appropriate quality assurance (QA) and quality control (QC)
procedures, are documented in the QA Project Plan. Data are then collected following the QA
Project Plan specifications. DQA completes the data life cycle by providing the assessment
needed to determine if the planning objectives were achieved. During the assessment phase, the
data are validated and verified to ensure that the sampling and analysis protocols specified in the
QA Project Plan were followed, and that the measurement systems performed in accordance with
the criteria specified in the QA Project Plan. DQA then proceeds using the validated data set to
determine if the quality of the data is satisfactory.
PLANNING
Data Quality Objectives Process
Quality Assurance Project Plan Development
,
IMPLEMENTATION
Field Data Collection and Associated
Quality Assurance / Quality Control Activities
1
1
ASSESSMENT
Data Validation/Verification
Data Quality Assessment
/
1
1
I
I
/
t
t
QUALITY ASSURANCE ASSESSMENT
/_,.„, / /QC/Performance /
' Routine Data //,_,.. /
/ / Evaluation Data /
• INPUTS •
DATA VALIDATION/VERIFICATION
• Verify measurement performance
• Verify measurement procedures and
reporting requirements
^ OUTPUT
/ VALIDATED/VERIFIED DATA /
| INPUT
DATA QUALITY ASSESSMENT
• Review DQOs and design
• Select statistical test
• Verify assumptions
• Draw conclusions
• OUTPUT
1 CONCLUSIONS DRAWN FROM DATA /
/
Figure 0-1. DQA in the Context of the Data Life Cycle
0.3 THE 5 STEPS OF THE DQA
The DQA involves five steps that begin with a review of the planning documentation and
end with an answer to the question posed during the planning phase of the study. These steps
roughly parallel the actions of an environmental statistician when analyzing a set of data. The five
EPA QA/G-9
QAOO Version
0-2
Final
July 2000
-------
steps, which are described in detail in the remaining chapters of this guidance, are briefly
summarized as follows:
1. Review the Data Quality Objectives (DQOs) and Sampling Design: Review the DQO
outputs to assure that they are still applicable. If DQOs have not been developed, specify
DQOs before evaluating the data (e.g., for environmental decisions, define the statistical
hypothesis and specify tolerable limits on decision errors; for estimation problems, define
an acceptable confidence or probability interval width). Review the sampling design and
data collection documentation for consistency with the DQOs.
2. Conduct a Preliminary Data Review: Review QA reports, calculate basic statistics, and
generate graphs of the data. Use this information to learn about the structure of the data
and identify patterns, relationships, or potential anomalies.
3. Select the Statistical Test: Select the most appropriate procedure for summarizing and
analyzing the data, based on the review of the DQOs, the sampling design, and the
preliminary data review. Identify the key underlying assumptions that must hold for the
statistical procedures to be valid.
4. Verify the Assumptions of the Statistical Test: Evaluate whether the underlying
assumptions hold, or whether departures are acceptable, given the actual data and other
information about the study.
5. Draw Conclusions from the Data: Perform the calculations required for the statistical
test and document the inferences drawn as a result of these calculations. If the design is to
be used again, evaluate the performance of the sampling design.
These five steps are presented in a linear sequence, but the DQA is by its very nature iterative.
For example, if the preliminary data review reveals patterns or anomalies in the data set that are
inconsistent with the DQOs, then some aspects of the study planning may have to be reconsidered
in Step 1. Likewise, if the underlying assumptions of the statistical test are not supported by the
data, then previous steps of the DQA may have to be revisited. The strength of the DQA is that it
is designed to promote an understanding of how well the data satisfy their intended use by
progressing in a logical and efficient manner.
Nevertheless, it should be emphasized that the DQA cannot absolutely prove that one has
or has not achieved the DQOs set forth during the planning phase of a study. This situation
occurs because a decision maker can never know the true value of the item of interest. Data
collection only provides the investigators with an estimate of this, not its true value. Further,
because analytical methods are not perfect, they too can only provide an estimate of the true value
of an environmental sample. Because investigators make a decision based on estimated and not
true values, they run the risk of making a wrong decision (decision error) about the item of
interest.
EPA QA/G-9 Final
QAOO Version 0 - 3 July 2000
-------
0.4 INTENDED AUDIENCE
This guidance is written for a broad audience of potential data users, data analysts, and
data generators. Data users (such as project managers, risk assessors, or principal investigators
who are responsible for making decisions or producing estimates regarding environmental
characteristics based on environmental data) should find this guidance useful for understanding
and directing the technical work of others who produce and analyze data. Data analysts (such as
quality assurance specialists, or any technical professional who is responsible for evaluating the
quality of environmental data) should find this guidance to be a convenient compendium of basic
assessment tools. Data generators (such as analytical chemists, field sampling specialists, or
technical support staff responsible for collecting and analyzing environmental samples and
reporting the resulting data values) should find this guidance useful for understanding how their
work will be used and for providing a foundation for improving the efficiency and effectiveness of
the data generation process.
0.5 ORGANIZATION
This guidance presents background information and statistical tools for performing DQA.
Each chapter corresponds to a step in the DQA and begins with an overview of the activities to be
performed for that step. Following the overviews in Chapters 1, 2, 3, and 4, specific graphical or
statistical tools are described and step-by-step procedures are provided along with examples.
0.6 SUPPLEMENTAL SOURCES
Many of the graphical and statistical tools presented in this guidance are also implemented
in a user-friendly, personal computer software program called Data Quality Evaluation Statistical
Tools (DataQUEST) (G-9D) (EPA, 1996). DataQUEST simplifies the implementation of DQA
by automating many of the recommended statistical tools. DataQUEST runs on most IBM-
compatible personal computers using the DOS operating system; see the DataQUEST User's
Guide for complete information on the minimum computer requirements.
EPA QA/G-9 Final
QAOO Version 0 - 4 July 2000
-------
CHAPTER 1
STEP 1: REVIEW DQOs AND THE SAMPLING DESIGN
THE DATA QUALITY ASSESSMENT PROCESS
Review DQOs and Sampling Design
Conduct Preliminary Data Review
Select the Statistical Test
Verify the Assumptions
Draw Conclusions From the Data
REVIEW DQOs AND SAMPLING DESIGN
Purpose
Review the DQO outputs, the sampling design, and
any data collection documentation for consistency. If
DQOs have not been developed, define the statistical
hypothesis and specify tolerable limits on decision errors.
Activities
• Review Study Objectives
• Translate Objectives into Statistical Hypothesis
• Develop Limits on Decision Errors
• Review Sampling Design
Tools
• Statements of hypotheses
• Sampling design concepts
Step 1: Review DQOs and Sampling Design
Review the objectives of the study.
P If DQOs have not been developed, review Section 1.1.1 and define these objectives.
P If DQOs were developed, review the outputs from the DQO Process.
Translate the data user's objectives into a statement of the primary statistical hypothesis.
P If DQOs have not been developed, review Sections 1.1.2 and 1.2, and Box 1-1,
then develop a statement of the hypothesis based on the data user's objectives.
P If DQOs were developed, translate them into a statement of the primary hypothesis.
Translate the data user's objectives into limits on Type I or Type II decision errors.
P If DQOs have not been developed, review Section 1.1.3 and document the data
user's tolerable limits on decision errors.
P If DQOs were developed, confirm the limits on decision errors.
Review the sampling design and note any special features or potential problems.
P Review the sampling design for any deviations (Sections 1.1.4 and 1.3).
EPA QA/G-9
QAOO Version
1 - 1
Final
July 2000
-------
List of Boxes
Page
Box 1-1: Example Applying the DQO Process Retrospectively 1-8
List of Tables
Page
Table 1-1. Choosing a Parameter of Interest 1-6
Table 1-2. Commonly Used Statements of Statistical Hypotheses 1-12
List of Figures
Page
Figure 1-1. The Data Quality Objectives Process 1-3
EPA QA/G-9 Final
QAOO Version 1 - 2 July 2000
-------
CHAPTER 1
STEP 1: REVIEW DQOs AND THE SAMPLING DESIGN
1.1 OVERVIEW AND ACTIVITIES
DQA begins by reviewing the key outputs from the planning phase of the data life cycle:
the Data Quality Objectives (DQOs), the Quality Assurance (QA) Project Plan, and any
associated documents. The DQOs provide the context for understanding the purpose of the data
collection effort and establish the qualitative and quantitative criteria for assessing the quality of
the data set for the intended use. The sampling design (documented in the QA Project Plan)
provides important information about how to interpret the data. By studying the sampling design,
the analyst can gain an understanding of the assumptions under which the design was developed,
as well as the relationship between these assumptions and the DQOs. By reviewing the methods
by which the samples were collected, measured, and reported, the analyst prepares for the
preliminary data review and subsequent steps of DQA.
Careful planning improves the
representativeness and overall quality of a sampling
design, the effectiveness and efficiency with which
the sampling and analysis plan is implemented, and
the usefulness of subsequent DQA efforts. Given
the benefits of planning, the Agency has developed
the DQO Process which is a logical, systematic
planning procedure based on the scientific method.
The DQO Process emphasizes the planning and
development of a sampling design to collect the
right type, quality, and quantity of data needed to
support the decision. Using both the DQO Process
and the DQA will help to ensure that the decisions
are supported by data of adequate quality; the
DQO Process does so prospectively and the DQA
does so retrospectively.
When DQOs have not been developed
during the planning phase of the study, it is
necessary to develop statements of the data user's
objectives prior to conducting DQA. The primary
purpose of stating the data user's objectives prior
to analyzing the data is to establish appropriate
criteria for evaluating the quality of the data with
respect to their intended use. Analysts who are not
familiar with the DQO Process should refer to the
Guidance for the Data Quality Objectives Process
(QA/G-4) (1994), a book on statistical decision
Step 1. State the Problem
Define the problem; identify the planning team;
examine budget, schedule.
Step 2. Identify the Decision
State decision; identify study question; define
alternative actions.
Step 3. Identify the Inputs to the Decision
Identify information needed for the decision (information
sources, basis for Action Level, sampling/analysis method).
Step 4. Define the Boundaries of the Study
Specify sample characteristics; define
spatial/temporal limits, units of decision making.
Step 5. Develop a Decision Rule
Define statistical parameter (mean, median); specify Action
Level; develop logic for action.
Step 6. Specify Tolerable Limits on Decision Errors
Set acceptable limits for decision errors relative to
consequences (health effects, costs).
Step 7. Optimize the Design for Obtaining Data
Select resource-effective sampling and analysis plan that
meets the performance criteria.
Figure 1-1. The Data Quality Objectives
Process
EPA QA/G-9
QAOO Version
1 -3
Final
July 2000
-------
making using tests of hypothesis, or consult a statistician. The seven steps of the DQO Process
are illustrated in Figure 1.1.
The remainder of this chapter addresses recommended activities for performing this step
of DQA and technical considerations that support these activities. The remainder of this section
describes the recommended activities, the first three of which will differ depending on whether
DQOs have already been developed for the study. Section 1.2 describes how to select the null
and alternative hypothesis and Section 1.3 presents a brief overview of different types of sampling
designs.
1.1.1 Review Study Objectives
In this activity, the objectives of the study are reviewed to provide context for analyzing
the data. If a planning process has been implemented before the data are collected, then this step
reduces to reviewing the documentation on the study objectives. If no planning process was used,
the data user should:
• Develop a concise definition of the problem (DQO Process Step 1) and the decision (DQO
Process Step 2) for which the data were collected. This should provide the fundamental
reason for collecting the environmental data and identify all potential actions that could
result from the data analysis.
Identify if any essential information is missing (DQO Process Step 3). If so, either collect
the missing information before proceeding, or select a different approach to resolving the
decision.
Specify the scale of decision making (any subpopulations of interest) and any boundaries
on the study (DQO Process Step 4) based on the sampling design. The scale of decision
making is the smallest area or time period to which the decision will apply. The sampling
design and implementation may restrict how small or how large this scale of decision
making can be.
1.1.2 Translate Objectives into Statistical Hypotheses
In this activity, the data user's objectives are used to develop a precise statement of the
primary1 hypotheses to be tested using environmental data. A statement of the primary statistical
hypotheses includes a null hypothesis, which is a "baseline condition" that is presumed to be true
in the absence of strong evidence to the contrary, and an alternative hypothesis, which bears the
burden of proof. In other words, the baseline condition will be retained unless the alternative
Throughout this document, the term "primary hypotheses" refers to the statistical hypotheses that correspond to the
data user's decision. Other statistical hypotheses can be formulated to formally test the assumptions that underlie the specific
calculations used to test the primary hypotheses. See Chapter 3 for examples of assumptions underlying primary hypotheses
and Chapter 4 for examples of how to test these underlying assumptions.
EPA QA/G-9 Final
QAOO Version 1 - 4 July 2000
-------
condition (the alternative hypothesis) is thought to be true due to the preponderance of evidence.
In general, such hypotheses consist of the following elements:
a population parameter of interest, which describes the feature of the environment that the
data user is investigating;
• a numerical value to which the parameter will be compared, such as a regulatory or risk-
based threshold or a similar parameter from another place (e.g., comparison to a reference
site) or time (e.g., comparison to a prior time); and
• the relation (such as "is equal to" or "is greater than") that specifies precisely how the
parameter will be compared to the numerical value.
To help the analyst decide what parameter value should be investigated, Table 1-1 compares the
merits of the mean, upper proportion (percentile), and mean. If DQOs were developed, the
statement of hypotheses already should be documented in the outputs of Step 6 of the DQO
Process. If DQOs have not been developed, then the analyst should consult with the data user to
develop hypotheses that address the data user's concerns. Section 1.2 describes in detail how to
develop the statement of hypotheses and includes a list of common encountered hypotheses for
environmental decisions.
1.1.3 Develop Limits on Decision Errors
The goal of this activity is to develop numerical probability limits that express the data
user's tolerance for committing false rejection (Type I) or false acceptance (Type II) decision
errors as a result of uncertainty in the data. A false rejection error occurs when the null
hypothesis is rejected when it is true. A false acceptance decision error occurs when the null
hypothesis is not rejected when it is false. These are the statistical definitions of false rejection
and false acceptance decision errors. Other commonly used phrases include "level of significance"
which is equal to the Type I Error (false rejection) and "complement of power" equal to the Type
II Error (false acceptance). If tolerable decision error rates were not established prior to data
collection, then the data user should:
Specify the gray region where the consequences of a false acceptance decision error are
relatively minor (DQO Process Step 6). The gray region is bounded on one side by the
threshold value and on the other side by that parameter value where the consequences of
making a false acceptance decision error begin to be significant. Establish this boundary
by evaluating the consequences of not rejecting the null hypothesis when it is false and
then place the edge of the gray region where these consequences are severe enough to set
a limit on the magnitude of this false acceptance decision error. The gray region is the
area between this parameter value and the threshold value.
The width of the gray region represents one important aspect of the decision maker's
concern for decision errors. A more narrow gray region implies a desire to detect
EPA QA/G-9 Final
QAOO Version 1 - 5 July 2000
-------
Table 1-1. Choosing a Parameter of Interest
Parameter
Points to Consider
Mean
1. Easy to calculate and estimate a confidence interval.
2. Useful when the standard has been based on consideration of health effects or long-term average exposure.
3. Useful when the data have little variation from sample to sample or season to season.
4. If the data have a large coefficient of variation (greater than about 1.5) testing the mean can require more samples than for testing
an upper percentile in order to provide the same protection to human health and the environment.
5. Can have high false rejection rates with small sample sizes and highly skewed data, i.e., when the contamination levels are
generally low with only occasional short periods of high contamination.
6. Not as powerful for testing attainment when there is a large proportion of less-than-detection-limit values.
7. Is adversely affected by outliers or errors in a few data values.
Upper Proportion
(Percentile)
1. Requiring that an upper percentile be less than a standard can limit the occurrence of samples with high concentrations, depending
on the selected percentile.
2. Unaffected by less-than-detection-limit values, as long as the detection limit is less than the cleanup standard.
3. If the health effects of the contaminant are acute, extreme concentrations are of concern and are best tested by ensuring that a
large portion of the measurements are below a standard.
4. The proportion of the samples that must be below the standard must be chosen.
5. For highly variable or skewed data, can provide similar protection of human health and the environment with a smaller size than
when testing the mean.
6. Is relatively unaffected by a small number of outliers.
Median
1. Has benefits over the mean because it is not as heavily influenced by outliers and highly variable data, and can be used with a
large number of less-than-detection-limit values.
2. Has many of the positive features of the mean, in particular its usefulness of evaluating standards based on health effects and
long-term average exposure.
3. For positively skewed data, the median is lower than the mean and therefore testing the median provides less protection for human
health and the environment than testing the mean.
4. Retains some negative features of the mean in that testing the median will not limit the occurrence of extreme values.
EPA QA/G-9
QAOO Version
1 -6
Final
May 2000
-------
conclusively the condition when the true parameter value is close to the threshold value
("close" relative to the variability in the data).
Specify tolerable limits on the probability of committing false rejection and false
acceptance decision errors (DQO Process Step 6) that reflect the decision maker's
tolerable limits for making an incorrect decision. Select a possible value of the parameter;
then, choose a probability limit based on an evaluation of the seriousness of the potential
consequences of making the decision error if the true parameter value is located at that
point. At a minimum, the decision maker should specify a false rejection decision error
limit at the threshold value (a), and a false acceptance decision error limit at the other
edge of the gray region (P).
An example of the gray region and limits on the probability of committing both false rejection and
false acceptance decision errors are contained in Box 1-1.
If DQOs were developed for the study, the tolerable limits on decision errors will already
have been developed. These values can be transferred directly as outputs for this activity. In this
case, the action level is the threshold value; the false rejection error rate at the action level is the
Type I error rate or a; and the false acceptance error rate at the other bound of the gray region is
the Type II error rate or p.
1.1.4 Review Sampling Design
The goal of this activity is to familiarize the analyst with the main features of the sampling
design that was used to generate the environmental data. The overall type of sampling design and
the manner in which samples were collected or measurements were taken will place conditions
and constraints on how the data must be used and interpreted. Section 1.3 provides additional
information about several different types of sampling designs that are commonly used in
environmental studies.
Review the sampling design documentation with the data user's objectives in mind. Look
for design features that support or contradict those objectives. For example, if the data user is
interested in making a decision about the mean level of contamination in an effluent stream over
time, then composite samples may be an appropriate sampling approach. On the other hand, if the
data user is looking for hot spots of contamination at a hazardous waste site, compositing should
only be used with caution, to avoid "averaging away" hot spots. Also, look for potential
problems in the implementation of the sampling design. For example, verify that each point in
space (or time) had an equal probability of being selected for a simple random sampling design.
Small deviations from a sampling plan may have minimal effect on the conclusions drawn from the
data set. Significant or substantial deviations should be flagged and their potential effect carefully
considered throughout the entire DQA.
EPA QA/G-9 Final
QAOO Version 1 - 7 July 2000
-------
Box 1-1: Example Applying the DQO Process Retrospectively
A waste incineration company was concerned that waste fly ash could contain hazardous levels of
cadmium and should be disposed of in a RCRA landfill. As a result, eight composite samples each
consisting of eight grab samples were taken from each load of waste. The TCLP leachate from these
samples were then analyzed using a method specified in 40 CFR, Pt. 261, App. II. DQOs were not
developed for this problem; therefore, study objectives (Sections 1.1.1 through 1.1.3) should be
developed before the data are analyzed.
1.1.1 Review Study Objectives
P Develop a concise definition of the problem - The problem is defined above.
P Identify if any essential information is missing - It does not appear than any essential information is
missing.
P Specify the scale of decision making - Each waste load is sampled separately and decisions need to
be made for each load. Therefore, the scale of decision making is an individual load.
1.1.2 Translate Objectives into Statistical Hypotheses
Since composite samples were taken, the parameter of interest is the mean cadmium concentration. The
RCRA regulatory standard for cadmium in TCLP leachate is 1.0 mg/L. Therefore, the two hypotheses
are "mean cadmium > 1.0 mg/L" and "mean cadmium < 1.0 mg/L."
There are two possible decision errors 1) to decide the waste is hazardous ("mean > 1.0") when it truly is
not ("mean < 1.0"), and 2) to decide the waste is not hazardous ("mean < 1.0") when it truly is ("mean >
1.0"). The risk of deciding the fly ash is not hazardous when it truly is hazardous is more severe since
potential consequences of this decision error include risk to human health and the environment.
Therefore, this error will be labeled the false rejection error and the other error will be the false
acceptance error. As a result of this decision, the null hypothesis will be that the waste is hazardous
("mean cadmium > 1.0 mg/L") and the alternative hypothesis will be that the waste is not hazardous
("mean cadmium < 1.0 mg/L"). (See Section 1.2 for more information on developing the null and
alternative hypotheses.)
1.1.3 Develop Limits on Decision Errors
1
_ 0.9
J= (D
S -i 0.8
01 £=
"= 0
? < °-7
Q
-------
1.2 DEVELOPING THE STATEMENT OF HYPOTHESES
The full statement of the statistical hypotheses has two major parts: the null hypothesis
(H0) and the alternative hypothesis (HA). In both parts, a population parameter is compared to
either a fixed value (for a one-sample test) or another population parameter (for a two-sample
test). The population parameter is a quantitative characteristic of the population that the data
user wants to estimate using the data. In other words, the parameter describes that feature of the
population that the data user will evaluate when making the decision. Table 1-1 describes several
common statistical parameters.
If the data user is interested in drawing inferences about only one population, then the null
and alternative hypotheses will be stated in terms that relate the true value of the parameter to
some fixed threshold value. A common example of this one-sample problem in environmental
studies is when pollutant levels in an effluent stream are compared to a regulatory limit. If the
data user is interested in comparing two populations, then the null and alternative hypotheses will
be stated in terms that compare the true value of one population parameter to the corresponding
true parameter value of the other population. A common example of this two-sample problem in
environmental studies is when a potentially contaminated waste site is being compared to a
reference area using samples collected from the respective areas. In this situation, the hypotheses
often will be stated in terms of the difference between the two parameters.
The decision on what should constitute the null hypothesis and what should be the
alternative is sometimes difficult to ascertain. In many cases, this problem does not arise because
the null and alternative hypotheses are determined by specific regulation. However, when the null
hypothesis is not specified by regulation, it is necessary to make this determination. The test of
hypothesis procedure prescribes that the null hypothesis is only rejected in favor of the alternative,
provided there is overwhelming evidence from the data that the null hypothesis is false. In other
words, the null hypothesis is considered to be true unless the data show conclusively that this is
not so. Therefore, it is sometimes useful to choose the null and alternative hypotheses in light of
the consequences of possibly making an incorrect decision between the null and alternative
hypotheses. The true condition that occurs with the more severe decision error (not what would
be decided in error based on the data) should be defined as the null hypothesis. For example,
consider the two decision errors: "decide a company does not comply with environmental
regulations when it truly does" and "decide a company does comply with environmental
regulations when it truly does not." If the first decision error is considered the more severe
decision error, then the true condition of this error, "the company does comply with the
regulations" should be defined as the null hypothesis. If the second decision error is considered
the more severe decision error, then the true condition of this error, "the company does not
comply with the regulations" should be defined as the null hypothesis.
An alternative method for defining the null hypothesis is based on historical information.
If a large amount of information exists suggesting that one hypothesis is extremely likely, then this
hypothesis should be defined as the alternative hypothesis. In this case, a large amount of data
may not be necessary to provide overwhelming evidence that the other (null) hypothesis is false.
EPA QA/G-9 Final
QAOO Version 1 - 9 July 2000
-------
For example, if the waste from an incinerator was previously hazardous and the waste process has
not changed, it may be more cost-effective to define the alternative hypothesis as "the waste is
hazardous" and the null hypothesis as "the waste is not hazardous."
Consider a data user who wants to know whether the true mean concentration (|i) of
atrazine in ground water at a hazardous waste site is greater than a fixed threshold value C. If the
data user presumes from prior information that the true mean concentration is at least C due
possibly to some contamination incident, then the data must provide compelling evidence to reject
that presumption, and the hypotheses can be stated as follows:
Narrative Statement of Hypotheses
Statement of Hypotheses Using Standard
Notation
Null Hypothesis (Baseline Condition):
The true mean concentration of atrazine in
ground
water is greater than or equal to the threshold
value C; versus
H
versus
Alternative Hypothesis:
The true mean concentration of atrazine in
ground water is less than the threshold value
C.
H
A
On the other hand, if the data user presumes from prior information that the true mean
concentration is less than C due possibly to the fact that the ground water has not been
contaminated in the past, then the data must provide compelling evidence to reject that
presumption, and the hypotheses can be stated as follows:
Narrative Statement of Hypotheses
Statement of Hypotheses Using Standard
Notation
Null Hypothesis (Baseline Condition):
The true mean concentration of atrazine in
ground
water is less than or equal to the threshold
value C; versus
H0: |i < C;
versus
Alternative Hypothesis:
The true mean concentration of atrazine in
ground water is greater than the threshold
value C.
HA:
EPA QA/G-9
QAOO Version
1 - 10
Final
July 2000
-------
In stating the primary hypotheses, it is convenient to use standard statistical notation, as
shown throughout this document. However, the logic underlying the hypothesis always
corresponds to the decision of interest to the data user.
Table 1-2 summarizes common environmental decisions and the corresponding
hypotheses. In Table 1-2, the parameter is denoted using the symbol "@," and the difference
between two parameters is denoted using "Oj - @2" where @j represents the parameter of the first
population and @2 represents the parameter of the second population. The use of "@" is to avoid
using the terms "population mean" or "population median" repeatedly because the structure of the
hypothesis test remains the same regardless of the population parameter. The fixed threshold
value is denoted "C," and the difference between two parameters is denoted "50" (often the null
hypothesis is defined such that 50= 0).
For the first problem in Table 1-2, only estimates of @ that exceed C can cast doubt on the
null hypothesis. This is called a one-tailed hypothesis test, because only parameter estimates on
one side of the threshold value can lead to rejection of the null hypothesis. The second, fourth,
and fifth rows of Table 1-2 are also examples of one-tailed hypothesis tests. The third and sixth
rows of Table 1-2 are examples of two-tailed tests, because estimates falling both below and
above the null-hypothesis parameter value can lead to rejection of the null hypothesis. Most
hypotheses connected with environmental monitoring are one-tailed because high pollutant levels
can harm humans or ecosystems.
1.3 DESIGNS FOR SAMPLING ENVIRONMENTAL MEDIA
Sampling designs provide the basis for how a set of samples may be analyzed. Different
sampling designs require different analysis techniques and different assessment procedures. There
are two primary types of sampling designs: authoritative (judgment) sampling and probability
sampling. This section describes some of the most common sampling designs.
1.3.1 Authoritative Sampling
With authoritative (judgment) sampling, an expert having knowledge of the site (or
process) designates where and when samples are to be taken. This type of sampling should only
be considered when the objectives of the investigation are not of a statistical nature, for example,
when the objective of a study is to identify specific locations of leaks, or when the study is
focused solely on the sampling locations themselves. Generally, conclusions drawn from
authoritative samples apply only to the individual samples and aggregation may result in severe
bias and lead to highly erroneous conclusions. Judgmental sampling also precludes the use of the
sample for any purpose other than the original one. Thus if the data may be used in further
studies (e.g., for an estimate of variability in a later study), a probabilistic design should be used.
When the study objectives involve estimation or decision making, some form of
probability sampling should be selected. As described below, this does not preclude use of the
expert's knowledge of the site or process in designing a probability-based sampling plan; however,
EPA QA/G-9 Final
QAOO Version 1-11 July 2000
-------
Table 1-2. Commonly Used Statements of Statistical Hypotheses
Type of Decision
Null Hypothesis
Alternative
Hypothesis
Compare environmental conditions to a fixed
threshold value, such as a regulatory standard or
acceptable risk level; presume that the true
condition is less than the threshold value.
H0: 0 < C
HA: 0>C
Compare environmental conditions to a fixed
threshold value; presume that the true condition is
greater than the threshold value.
H0: 0 > C
HA: 0 60
(HA: 0j - 02 > 0)
Compare environmental conditions associated
with two different populations to a fixed threshold
value (60) such as a regulatory standard or
acceptable risk level; presume that the true
condition is greater than the threshold value. If it
is presumed that conditions associated with the
two populations are the same, the threshold value
isO.
H0: 0: - 02 > 60
(Ho: 0: - 02 > 0)
HA: 0j - 02 < 60
(HA: 01-02<0)
Compare environmental conditions associated
with two different populations to a fixed threshold
value (80) such as a regulatory standard or
acceptable risk level; presume that the true
condition is equal to the threshold value. If it is
presumed that conditions associated with the two
populations are the same, the threshold value is 0.
H0: 0: - 02 = 80
(Ho: 0j - 02 = 0)
HA: e^e^do
(HA: 0j - 02 * 0)
EPA QA/G-9
QAOO Version
1 - 12
Final
July 2000
-------
valid statistical inferences require that the plan incorporate some form of randomization in
choosing the sampling locations or sampling times. For example, to determine maximum SO2
emission from a boiler, the sampling plan would reasonably focus, or put most of the weight on,
periods of maximum or near-maximum boiler operation. Similarly, if a residential lot is being
evaluated for contamination, then the sampling plan can take into consideration prior knowledge
of contaminated areas, by weighting such areas more heavily in the sample selection and data
analysis.
1.3.2 Probability Sampling
Probability samples are samples in which every member of the target population (i.e.,
every potential sampling unit) has a known probability of being included in the sample.
Probability samples can be of various types, but in some way, they all make use of randomization,
which allows valid probability statements to be made about the quality of estimates or hypothesis
tests that are derived from the resultant data.
One common misconception of probability sampling procedures is that these procedures
preclude the use of important prior information. Indeed, just the opposite is true. An efficient
sampling design is one that uses all available prior information to stratify the region and set
appropriate probabilities of selection. Another common misconception is that using a probability
sampling design means allowing the possibility that the sample points will not be distributed
appropriately across the region. However, if there is no prior information regarding the areas
most likely to be contaminated, a grid sampling scheme (a type of stratified design) is usually
recommended to ensure that the sampling points are dispersed across the region.
1.3.2.1 Simple Random Sampling
The simplest type of probability sample is the simple random sample where every possible
sampling unit in the target population has an equal chance of being selected. Simple random
samples, like the other samples, can be either samples in time and/or space and are often
appropriate at an early stage of an investigation in which little is known about systematic variation
within the site or process. All of the sampling units should have equal volume or mass, and
ideally be of the same shape if applicable. With a simple random sample, the term "random"
should not be interpreted to mean haphazard; rather, it has the explicit meaning of equiprobable
selection. Simple random samples are generally developed through use of a random number table
or through computer generation of pseudo-random numbers.
1.3.2.2 Sequential Random Sampling
Usually, simple random samples have a fixed sample size, but some alternative approaches
are available, such as sequential random sampling, where the sample sizes are not fixed a priori.
Rather, a statistical test is performed after each specimen's analysis (or after some minimum
number have been analyzed). This strategy could be applicable when sampling and/or analysis is
quite expensive, when information concerning sampling and/or measurement variability is lacking,
EPA QA/G-9 Final
QAOO Version 1-13 July 2000
-------
when the characteristics of interest are stable over the time frame of the sampling effort, or when
the objective of the sampling effort is to test a single specific hypothesis.
1.3.2.3 Systematic Samples
In the case of spatial sampling, systematic sampling involves establishing a two-
dimensional (or in some cases a three-dimensional) spatial grid and selecting a random starting
location within one of the cells. Sampling points in the other cells are located in a deterministic
way relative to that starting point. In addition, the orientation of the grid is sometimes chosen
randomly and various types of systematic samples are possible. For example, points may be
arranged in a pattern of squares (rectangular grid sampling) or a pattern of equilateral triangles
(triangular grid sampling). The result of either approach is a simple pattern of equally spaced
points at which sampling is to be performed.
Systematic sampling designs have several advantages over random sampling and some of
the other types of probability sampling. They are generally easier to implement, for example.
They are also preferred when one of the objectives is to locate "hot spots" within a site or
otherwise map the pattern of concentrations over a site. On the other hand, they should be used
with caution whenever there is a possibility of some type of cyclical pattern in the waste site or
process. Such a situation, combined with the uniform pattern of sampling points, could very
readily lead to biased results.
1.3.2.4 Stratified Samples
Another type of probability sample is the stratified random sample, in which the site or
process is divided into two or more non-overlapping strata, sampling units are defined for each
stratum, and separate simple random samples are employed to select the units in each stratum. (If
a systematic sample were employed within each stratum, then the design would be referred to as a
stratified systematic sample.) Strata should be defined so that physical samples within a stratum
are more similar to each other than to samples from other strata. If so, a stratified random sample
should result in more precise estimates of the overall population parameter than those that would
be obtained from a simple random sample with the same number of sampling units.
Stratification is a way to incorporate prior knowledge and professional judgment into a
probabilistic sampling design. Generally, units that are "alike" or anticipated to be "alike" are
placed together in the same stratum. Units that are contiguous in space (e.g., similar depths) or
time are often grouped together into the same stratum, but characteristics other than spatial or
temporal proximity can also be employed. Media, terrain characteristics, concentration levels,
previous cleanup attempts, and confounding contaminants can be used to create strata.
Advantages of stratified samples over random samples include their ability to ensure more
uniform coverage of the entire target population and, as noted above, their potential for achieving
greater precision in certain estimation problems. Even when imperfect information is used to
form strata, the stratified random sample will generally be more cost-effective than a simple
EPA QA/G-9 Final
QAOO Version 1-14 July 2000
-------
random sample. A stratified design can also be useful when there is interest in estimating or
testing characteristics for subsets of the target population. Because different sampling rates can
be used in different strata, one can oversample in strata containing those subareas of particular
interest to ensure that they are represented in the sample. In general, statistical calculations for
data generated via stratified samples are more complex than for random samples, and certain
types of tests, for example, cannot be performed when stratified samples are employed.
Therefore, a statistician should be consulted when stratified sampling is used.
1.3.2.5 Compositing Physical Samples
When analysis costs are large relative to sampling costs, cost-effective plans can
sometimes be achieved by compositing physical samples or specimens prior to analysis, assuming
that there are no safety hazards or potential biases (for example, the loss of volatile organic
compounds from a matrix) associated with such compositing. For the same total cost,
compositing in this situation would allow a larger number of sampling units to be selected than
would be the case if compositing were not used. Composite samples reflect a physical rather than
a mathematical mechanism for averaging. Therefore, compositing should generally be avoided if
population parameters other than a mean are of interest (e.g., percentiles or standard deviations).
Composite sampling is also useful when the analyses of composited samples are to be used
in a two-staged approach in which the composite-sample analyses are used solely as a screening
mechanism to identify if additional, separate analyses need to be performed. This situation might
occur during an early stage of a study that seeks to locate those areas that deserve increased
attention due to potentially high levels of one or more contaminants.
1.3.2.6 Other Sampling Designs
Adaptive sampling involves taking a sample and using the resulting information to design
the next stage of sampling. The process may continue through several additional rounds of
sampling and analysis. A common application of adaptive sampling to environmental problems
involves subdividing the region of interest into smaller units, taking a probability sample of these
units, then sampling all units that border on any unit with a concentration level greater than some
specified level C. This process is continued until all newly sampled units are below C. The field
of adaptive sampling is currently undergoing active development and can be expected to have a
significant impact on environmental sampling.
Ranked set sampling (RSS) uses the availability of an inexpensive surrogate measurement
when it is correlated with the more expensive measurement of interest. The method exploits this
correlation to obtain a sample which is more representative of the population that would be
obtained by random sampling, thereby leading to more precise estimates of population parameters
than what would be obtained by random sampling. RSS consists of creating n groups, each of
size n (for a total of n2 initial samples), then ranking the surrogate from largest to smallest within
each group. One sample from each group is then selected according to a specified procedure and
these n samples are analyzed for the more expensive measurement of interest.
EPA QA/G-9 Final
QAOO Version 1-15 July 2000
-------
This page is intentionally blank
EPA QA/G-9 Final
QAOO Version 1-16 July 2000
-------
CHAPTER 2
STEP 2: CONDUCT A PRELIMINARY DATA REVIEW
THE DATA QUALITY ASSESSMENT PROCESS
Review DQOs and Sampling Design
Conduct Preliminary Data Review
Select the Statistical Test
Verify the Assumptions
Draw Conclusions From the Data
CONDUCT PRELIMINARY DATA REVIEW
Purpose
Generate statistical quantities and graphical
representations that describe the data. Use this
information to learn about the structure of the data
and identify any patterns or relationships.
Activities
• Review Quality Assurance Reports
• Calculate Basic Statistical Quantities
•Graph the Data
Tools
• Statistical quantities
•Graphical representations
Step 2: Conduct a Preliminary Data Review
Review quality assurance reports.
P Look for problems or anomalies in the implementation of the sample collection and
analysis procedures.
P Examine QC data for information to verify assumptions underlying the Data Quality
Objectives, the Sampling and Analysis Plan, and the QA Project Plans.
Calculate the statistical quantities.
P Consider calculating appropriate percentiles (Section 2.2.1)
P Select measures of central tendency (Section 2.2.2) and dispersion (Section 2.2.3).
P If the data involve two variables, calculate the correlation coefficient (Section 2.2.4).
Display the data using graphical representations.
P Select graphical representations (Section 2.4) that illuminate the structure of the data
set and highlight assumptions underlying the Data Quality Objectives, the Sampling
and Analysis Plan, and the QA Project Plans.
P Use a variety of graphical representations that examine different features of the set.
EPA QA/G-9
QAOO Version
2- 1
Final
July 2000
-------
Boxes
Box 2-1: Directions for Calculating the Measure of Relative Standing (Percentiles) 2-6
Box 2-2: Directions for Calculating the Measures of Central Tendency 2-7
Box 2-3: Example Calculations of the Measures of Central Tendency 2-7
Box 2-4: Directions for Calculating the Measures of Dispersion 2-9
Box 2-5: Example Calculations of the Measures of Dispersion 2-9
Box 2-6: Directions for Calculating Pearson's Correlation Coefficient 2-10
Box 2-7: Directions for Calculating Spearman's Correlation 2-12
Box 2-8: Directions for Estimating the Serial Correlation Coefficient with a Example .... 2-13
Box 2-9: Directions for Generating a Histogram and a Frequency Plot 2-14
Box 2-10: Example of Generating a Histogram and a Frequency Plot 2-15
Box 2-11: Directions for Generating a Stem and Leaf Plot 2-16
Box 2-12: Example of Generating a Stem and Leaf Plot 2-16
Box 2-13: Directions for Generating a Box and Whiskers Plot 2-18
Box 2-14: Example of a Box and Whiskers Plot 2-18
Box 2-15: Directions for Generating a Ranked Data Plot 2-19
Box 2-16: Example of Generating a Ranked Data Plot 2-19
Box 2-17: Directions for Generating a Quantile Plot 2-22
Box 2-18: Example of Generating a Quantile Plot 2-22
Box 2-19: Directions for Constructing a Normal Probability Plot 2-23
Box 2-20: Example of Normal Probability Plot 2-24
Box 2-21: Directions for Generating a Scatter Plot and an Example 2-28
Box 2-22: Directions for Constructing an Empirical Q-Q Plot with an Example 2-31
Figures
Figure 2-1. Example of a Frequency Plot 2-13
Figure 2-2. Example of a Histogram 2-14
Figure 2-3. Example of a Box and Whisker Plot 2-17
Figure 2-4. Example of a Ranked Data Plot 2-20
Figure 2-5. Example of a Quantile Plot of Skewed Data 2-21
Figure 2-6. Normal Probability Paper 2-25
Figure 2-7. Example of Graphical Representations of Multiple Variables 2-26
Figure 2-8. Example of a Scatter Plot 2-27
Figure 2-9. Example of a Coded Scatter Plot 2-28
Figure 2-10. Example of a Parallel Coordinates Plot 2-29
Figure 2-11. Example of a Matrix Scatter Plot 2-29
Figure 2-12. Example of a Time Plot Showing a Slight Downward Trend 2-33
Figure 2-13. Example of a Correlogram 2-34
Figure 2-14. Example of a Posting Plot 2-37
Figure 2-15. Example of a Symbol Plot 2-38
Figure 2-16. The Normal Distribution 2-40
Figure 2-17. The Standard Normal Curve (Z-Curve) 2-40
EPA QA/G-9 Final
QAOO Version 2 - 2 July 2000
-------
CHAPTER 2
STEP 2: CONDUCT A PRELIMINARY DATA REVIEW
2.1 OVERVIEW AND ACTIVITIES
In this step of DQA, the analyst conducts a preliminary evaluation of the data set,
calculates some basic statistical quantities, and examines the data using graphical representations.
A preliminary data review should be performed whenever data are used, regardless of whether
they are used to support a decision, estimate a population parameter, or answer exploratory
research questions. By reviewing the data both numerically and graphically, one can learn the
"structure" of the data and thereby identify appropriate approaches and limitations for using the
data. The DQA software Data Quality Evaluation Statistical Tools (DataQUEST) (G-9D) (EPA,
1996) will perform all of these functions as well as more sophisticated statistical tests.
There are two main elements of preliminary data review: (1) basic statistical quantities
(summary statistics); and (2) graphical representations of the data. Statistical quantities are
functions of the data that numerically describe the data set. Examples include a mean, median,
percentile, range, and standard deviation. They can be used to provide a mental picture of the
data and are useful for making inferences concerning the population from which the data were
drawn. Graphical representations are used to identify patterns and relationships within the data,
confirm or disprove hypotheses, and identify potential problems. For example, a normal
probability plot may allow an analyst to quickly discard an assumption of normality and may
identify potential outliers.
The preliminary data review step is designed to make the analyst familiar with the data.
The review should identify anomalies that could indicate unexpected events that may influence the
analysis of the data. The analyst may know what to look for based on the anticipated use of the
data documented in the DQO Process, the QA Project Plan, and any associated documents. The
results of the review are then used to select a procedure for testing a statistical hypotheses to
support the data user's decision.
2.1.1 Review Quality Assurance Reports
The first activity in conducting a preliminary data review is to review any relevant QA
reports that describe the data collection and reporting process as it actually was implemented.
These QA reports provide valuable information about potential problems or anomalies in the data
set. Specific items that may be helpful include:
• Data validation reports that document the sample collection, handling, analysis,
data reduction, and reporting procedures used;
Quality control reports from laboratories or field stations that document
measurement system performance, including data from check samples, split
samples, spiked samples, or any other internal QC measures; and
EPA QA/G-9 Final
QAOO Version 2 - 3 July 2000
-------
Technical systems reviews, performance evaluation audits, and audits of data
quality, including data from performance evaluation samples.
When reviewing QA reports, particular attention should be paid to information that can be
used to check assumptions made in the DQO Process. Of great importance are apparent
anomalies in recorded data, missing values, deviations from standard operating procedures, and
the use of nonstandard data collection methodologies.
2.1.2 Calculate Basic Statistical Quantities
The goal of this activity is to summarize some basic quantitative characteristics of the data
set using common statistical quantities. Some statistical quantities that are useful to the analyst
include: number of observations; measures of central tendency, such as a mean, median, or mode;
measures of dispersion, such as range, variance, standard deviation, coefficient of variation, or
interquartile range; measures of relative standing, such as percentiles; measures of distribution
symmetry or shape; and measures of association between two or more variables, such as
correlation. These measures can then be used for description, communication, and to test
hypothesis regarding the population from which the data were drawn. Section 2.2 provides
detailed descriptions and examples of these statistical quantities.
The sample design may influence how the statistical quantities are computed. The
formulas given in this chapter are for simple random sampling, simple random sampling with
composite samples, and randomized systematic sampling. If a more complex design is used, such
as a stratified design, then the formulas may need to be adjusted.
2.1.3 Graph the Data
The goal of this step is to identify patterns and trends in the data that might go unnoticed
using purely numerical methods. Graphs can be used to identify these patterns and trends, to
quickly confirm or disprove hypotheses, to discover new phenomena, to identify potential
problems, and to suggest corrective measures. In addition, some graphical representations can be
used to record and store data compactly or to convey information to others. Graphical
representations include displays of individual data points, statistical quantities, temporal data,
spatial data, and two or more variables. Since no single graphical representation will provide a
complete picture of the data set, the analyst should choose different graphical techniques to
illuminate different features of the data. Section 2.3 provides descriptions and examples of
common graphical representations.
At a minimum, the analyst should choose a graphical representation of the individual data
points and a graphical representation of the statistical quantities. If the data set has a spatial or
temporal component, select graphical representations specific to temporal or spatial data in
addition to those that do not. If the data set consists of more than one variable, treat each
variable individually before developing graphical representations for the multiple variables. If the
EPA QA/G-9 Final
QAOO Version 2 - 4 July 2000
-------
sampling plan or suggested analysis methods rely on any critical assumptions, consider whether a
particular type of graph might shed light on the validity of that assumption. For example, if a
small-sample study is strongly dependent on the assumption of normality, then a normal
probability plot would be useful (Section 2.3.6).
The sampling design may influence what data may be included in each representation.
Usually, the graphical representations should be applied to each complete unit of randomization
separately or each unit of randomization should be represented with a different symbol. For
example, the analyst could generate box plots for each stratum instead of generating one box plot
with all the strata.
2.2 STATISTICAL QUANTITIES
2.2.1 Measures of Relative Standing
Sometimes the analyst is interested in knowing the relative position of one of several
observations in relation to all of the observations. Percentiles are one such measure of relative
standing that may also be useful for summarizing data. A percentile is the data value that is
greater than or equal to a given percentage of the data values. Stated in mathematical terms, the
pth percentile is the data value that is greater than or equal to p% of the data values and is less
than or equal to (l-p)% of the data values. Therefore, if V is the pth percentile, then p% of the
values in the data set are less than or equal to x, and (100-p)% of the values are greater than or
equal to x. A sample percentile may fall between a pair of observations. For example, the 75th
percentile of a data set of 10 observations is not uniquely defined. Therefore, there are several
methods for computing sample percentiles, the most common of which is described in Box 2-1.
Important percentiles usually reviewed are the quartiles of the data, the 25th, 50th, and 75th
percentiles. The 50th percentile is also called the sample median (Section 2.2.2), and the 25th and
75th percentile are used to estimate the dispersion of a data set (Section 2.2.3). Also important for
environmental data are the 90th, 95th, and 99th percentile where a decision maker would like to be
sure that 90%, 95%, or 99% of the contamination levels are below a fixed risk level.
A quantile is similar in concept to a percentile; however, a percentile represents a
percentage whereas a quantile represents a fraction. If'x' is the pth percentile, then at least p% of
the values in the data set lie at or below x, and at least (100-p)% of the values lie at or above x,
whereas if x is the p/100 quantile of the data, then the fraction p/100 of the data values lie at or
below x and the fraction (l-p)/100 of the data values lie at or above x. For example, the .95
quantile has the property that .95 of the observations lie at or below x and .05 of the data lie at or
above x. For the example in Box 2-1, 9 ppm would be the .95 quantile and 10 ppm would be the
.99 quantile of the data.
EPA QA/G-9 Final
QAOO Version 2 - 5 July 2000
-------
Box 2-1 : Directions for Calculating the Measure of Relative Standing (Percentiles)
with an Example
Let X.,, X2, ..., Xn represent the n data points. To compute the pth percentile, y(p), first list the data from
smallest to largest and label these points X(1), X(2), . . ., X(n) (so that X(1) is the smallest, X(2) is the
second smallest, and X, n } is the largest). Let t = p/1 00, and multiply the sample size n by t. Divide the
result into the integer part and the fractional part, i.e., let nt = j + g where j is the integer part and g is the
fraction part. Then the pth percentile, y(p), is calculated by:
If 9 = 0, y(p) = (X(j) + X(j + 1))/2
otherwise,
Example: The 90th and 95th percentile will be computed for the following 10 data points (ordered from
smallest to largest) : 4, 4, 4, 5, 5, 6, 7, 7, 8, and 10 ppb.
For the 95th percentile, t = p/1 00 = 95/100= .95 and nt = (10)(.95) = 9.5 = 9 + .5. Therefore,] = 9 and
g = .5. Because g = .5 ? 0, y(95) = X(j + 1) = X(9 + 1) = X(10) = 10 ppm. Therefore, 10 ppm is the 95th
percentile of the above data.
For the 90th percentile, t= p/1 00 = 90/100 = .9 and nt = (10)(.9) = 9. Therefore] = 9 and g = 0. Since g =
0, y(90) = (X(9) + X(10))/2 = (8 + 10) / 2 = 9 ppm.
2.2.2 Measures of Central Tendency
Measures of central tendency characterize the center of a sample of data points. The three
most common estimates are the mean, median, and the mode. Directions for calculating these
quantities are contained in Box 2-2; examples are provided in Box 2-3.
The most commonly used measure of the center of a sample is the sample mean, denoted
by x. This estimate of the center of a sample can be thought of as the "center of gravity" of the
sample. The sample mean is an arithmetic average for simple sampling designs; however, for
complex sampling designs, such as stratification, the sample mean is a weighted arithmetic
average. The sample mean is influenced by extreme values (large or small) and nondetects (see
Section 4.7).
The sample median (x) is the second most popular measure of the center of the data. This
value falls directly in the middle of the data when the measurements are ranked in order from
smallest to largest. This means that l/2 of the data are smaller than the sample median and l/2 of
the data are larger than the sample median. The median is another name for the 50th percentile
(Section 2.2.1). The median is not influenced by extreme values and can easily be used in the case
of censored data (nondetects).
The third method of measuring the center of the data is the mode. The sample mode is the
value of the sample that occurs with the greatest frequency. Since this value may not always
exist, or if it does it may not be unique, this value is the least commonly used. However, the
mode is useful for qualitative data.
EPA QA/G-9 Final
QAOO Version 2 - 6 July 2000
-------
Box 2-2: Directions for Calculating the Measures of Central Tendency
LetX^ X2, ..., Xn represent the n data points.
Sample Mean: The sample mean x is the sum of all the data points divided by the total number of data
points (n):
X = -
n
Sample Median: The sample median (X) is the center of the data when the measurements are ranked in
order from smallest to largest. To compute the sample median, list the data from smallest to largest and
label these points X(1), X(2), . . ., X(n) (so that X(1) is the smallest, X(2) is the second smallest, and X(n) is
the largest).
If the number of data points is odd, then X= -^C-rn+nm
, ..., , , --^*- / i
If the number of data points is even, then X =
Sample Mode: The mode is the value of the sample that occurs with the greatest frequency. The mode
may not exist, or if it does, it may not be unique. To find the mode, count the number of times each
value occurs. The sample mode is the value that occurs most frequently.
Box 2-3: Example Calculations of the Measures of Central Tendency
Using the directions in Box 2-2 and the following 10 data points (in ppm): 4, 5, 6, 7, 4, 10, 4, 5, 7, and 8,
the following is an example of computing the sample mean, median, and mode.
Sample mean:
- 4 + 5 + 6 + 7 + 4+10 + 4 + 5 + 7 + 8 60 ,
X = = — = 6 ppm
10 10
Therefore, the sample mean is 6 ppm.
Sample median: The ordered data are: 4, 4, 4, 5, 5, 6, 7, 7, 8, and 10. Since n=10 is even, the sample
median is
X = X(io/2) + X([io/2]+i) = X(5) + X(6) = 5 + 6 = 5 5
2 2 2
Thus, the sample median is 5.5 ppm.
Sample mode: Computing the number of times each value occurs yields:
4 appears 3 times; 5 appears 2 times; 6 appears 1 time; 7 appears 2 times; 8 appears 1 time; and 10
appears 1 time.
Because the value of 4 ppm appears the most times, it is the mode of this data set.
EPA QA/G-9 Final
QAOO Version 2 - 7 July 2000
-------
2.2.3 Measures of Dispersion
Measures of central tendency are more meaningful if accompanied by information on how
the data spread out from the center. Measures of dispersion in a data set include the range,
variance, sample standard deviation, coefficient of variation, and the interquartile range.
Directions for computing these measures are given in Box 2-4; examples are given in Box 2-5.
The easiest measure of dispersion to compute is the sample range. For small samples, the
range is easy to interpret and may adequately represent the dispersion of the data. For large
samples, the range is not very informative because it only considers (and therefore is greatly
influenced) by extreme values.
The sample variance measures the dispersion from the mean of a data set. A large sample
variance implies that there is a large spread among the data so that the data are not clustered
around the mean. A small sample variance implies that there is little spread among the data so
that most of the data are near the mean. The sample variance is affected by extreme values and by
a large number of nondetects. The sample standard deviation is the square root of the sample
variance and has the same unit of measure as the data.
The coefficient of variation (CV) is a unitless measure that allows the comparison of
dispersion across several sets of data. The CV is often used in environmental applications
because variability (expressed as a standard deviation) is often proportional to the mean.
When extreme values are present, the interquartile range may be more representative of
the dispersion of the data than the standard deviation. This statistical quantity does not depend on
extreme values and is therefore useful when the data include a large number of nondetects.
2.2.4 Measures of Association
Data often include measurements of several characteristics (variables) for each sample
point and there may be interest in knowing the relationship or level of association between two or
more of these variables. One of the most common measures of association is the correlation
coefficient. The correlation coefficient measures the relationship between two variables, such as a
linear relationship between two sets of measurements. However, the correlation coefficient does
not imply cause and effect. The analyst may say that the correlation between two variables is high
and the relationship is strong, but may not say that one variable causes the other variable to
increase or decrease without further evidence and strong statistical controls.
2.2.4.1 Pearson's Correlation Coefficient
The Pearson correlation coefficient measures a linear relationship between two variables.
A linear association implies that as one variable increases so does the other linearly, or as one
EPA QA/G-9 Final
QAOO Version 2 - 8 July 2000
-------
Box 2-4: Directions for Calculating the Measures of Dispersion
LetX.,, X2, ..., Xn represent the n data points.
Sample Range: The sample range (R) is the difference between the largest value and the smallest value
of the sample, i.e., R = maximum - minimum.
Sample Variance: To compute the sample variance (s2), compute:
n ,-=
n-1
Sample Standard Deviation: The sample standard deviation (s) is the square root of the sample
variance, i.e.,
s = \l s2
Coefficient of Variation: The coefficient of variation (CV) is the standard deviation divided by the sample
mean (Section 2.2.2), i.e., CV = s/x. The CV is often expressed as a percentage.
Interquartile Range: Use the directions in Section 2.2.1 to compute the 25th and 75th percentiles of the
data (y(25) and y(75) respectively). The interquartile range (IQR) is the difference between these values,
IQR = y(75) - y(25).
Box 2-5: Example Calculations of the Measures of Dispersion
In this box, the directions in Box 2-4 and the following 10 data points (in ppm): 4, 5, 6, 7, 4, 10, 4, 5, 7,
and 8, are used to calculate the measures of dispersion. From Box 2-2, x = 6 ppm.
Sample Range: R = maximum - minimum =10-4 = 6 ppm
Sample Variance:
[42 + 52+ + ?2+g2] _ (4 +
2 10 10 A 2
s = - = - = 4 ppm
10-1 9
Sample Standard Deviation: s = \Js2 = 4 = 2
ppm
Coefficient of Variation: CV = s IX= 2 ppm 16 ppm = — = 33%
Interguartile Range: Using the directions in Section 2.2.1 to compute the 25th and 75th percentiles of the
data (y(25) and y(75) respectively): y(25) = X(2 + 1) = X(3) = 4 ppm and y(75) = X(7 + 1) = X(8) = 7 ppm.
The interquartile range (IQR) is the difference between these values: IQR = y(75) - y(25) = 7-4 = 3 ppm
EPA QA/G-9 Final
QAOO Version 2 - 9 July 2000
-------
variable decreases the other increases linearly. Values of the correlation coefficient close to +1
(positive correlation) imply that as one variable increases so does the other, the reverse holds for
values close to -1. A value of+1 implies a perfect positive linear correlation, i.e., all the data
pairs lie on a straight line with a positive slope. A value of-1 implies perfect negative linear
correlation. Values close to 0 imply little correlation between the variables. Directions and an
example for calculating Pearson's correlation coefficient are contained in Box 2-6.
Box 2-6: Directions for Calculating Pearson's Correlation Coefficient with an Example
LetX.,, X2, ..., Xn represent one variable of the n data points and let Y.,, Y2, ..., Yn represent a second
variable of the n data points. The Pearson correlation coefficient, r, between X and Y is computed by:
n
i= 1
n n
( \ ^V ^2 / \ ^v \2
/-i ' , ryy2 ,--i ,
J L ZJ i J
« ,= i n
1/2
Example: Consider the following data set (in ppb): Sample 1 — arsenic (X) = 8.0, lead (Y) = 8.0;
Sample 2 - arsenic = 6.0, lead = 7.0; Sample 3 - arsenic = 2.0, lead = 7.0; and Sample 4 - arsenic = 1.0,
lead = 6.0.
n n n n n
\ "V- _ i n \ V - ">9 \ "V"^— 1 fl^ \ V2_ 1 QO \ "V V - ^SvJ?\ _i_ _i_ ^1v/C\ - 1 T^
/_A - 1U, Jl .- Zo, 7-ji ~ ^•"•'5 / /, - lyo, / .A 1 - (oXo) + . . . + \iXO) — 1ZO.
;'= 1 ;'= 1 ;'= 1 ;'= 1 ;'= 1
126_ 02X28)
and r =
[105 - [198 -
4 4
1/2
= 0.865
Since r is close to 1, there is a strong linear relationship between these two contaminants.
The correlation coefficient does not detect nonlinear relationships so it should be used
only in conjunction with a scatter plot (Section 2.3.7.2). A scatter plot can be used to determine
if the correlation coefficient is meaningful or if some measure of nonlinear relationships should be
used. The correlation coefficient can be significantly changed by extreme values so a scatter plot
should be used first to identify such values.
An important property of the correlation coefficient is that it is unaffected by changes in
location of the data (adding or subtracting a constant from all of the X measurements and/or the
Y measurements) and by changes in scale of the data and/or Y values by a positive constant).
Thus linear transformations on the Xs and Ys do not affect the correlation of the measurements.
This is reasonable since the correlation reflects the degree to which linearity between X and Y
measurements occur and the degree of linearity is unaffected by changes in location or scale. For
EPA QA/G-9 Final
QAOO Version 2-10 July 2000
-------
example, if a variable was temperature in Celsius, then the correlation should not change if Celsius
was converted to Fahrenheit.
On the other hand, if nonlinear transformations of the X and/or Y measurements are made,
then the Pearson correlation between the transformed values will differ from the correlation of the
original measurements. For example, if X and Y, respectively, represent PCB and dioxin
concentrations in soil, and x=log (X) and y=log(Y), then the Pearson correlations between X and
Y, X and y x and Y, and x and y, will all be different, in general, since the logarithmic
transformation is a nonlinear transformation.
Pearson's correlation may be sensitive to the presence of one or two extreme values,
especially when sample sizes are small. Such values may result in a high correlation, suggesting a
strong linear trend, when only moderate trend is present. This may happen, for instance, if a
single (X,Y) pair has very high values for both X and Y while the remaining data values are
uncorrelated. Extreme value may also lead to low correlations between X and Y, thus tending to
mask a strong linear trend. This may happen if all the (X, Y) pairs except one (or two) tend to
cluster tightly about a straight line, and the exceptional point has a very large X value paired with
a moderate or small Y value (or vice versa). Because of the influences of extreme values, it is
wise to use a scatter plot (Section 2.3.7.2) in conjunction with a correlation coefficient.
2.2.4.2 Spearman's Rank Correlation Coefficient
An alternative to the Pearson correlation is Spearman's rank correlation coefficient. It is
calculated by first replacing each X value by its rank (i.e., 1 for the smallest X value, 2 for the
second smallest, etc.) and each Y value by its rank. These pairs of ranks are then treated as the
(X,Y) data and Spearman's rank correlation is calculated using the same formulae as for
Pearson's correlation (Box 2-6). Directions and an example for calculating a correlation
coefficient are contained in Box 2-7.
Since meaningful (i.e., monotonic increasing) transformations of the data will not later the
ranks of the respective variables (e.g, the ranks for log (X) will be the same for the ranks for X),
Spearman's correlation will not be altered by nonlinear increasing transformations of the Xs or the
Ys. For instance, the Spearman correlation between PCB and dioxin concentrations (X and Y) in
soil will be the same as the correlations between their logarithms (x and y). This desirable
property and the fact that Spearman's correlation is less sensitive to extreme values than
Pearson's correlation make Spearman's correlation an attractive alternative or complement of
Pearson's correlation coefficient. There are some important theoretical differences between
Pearson's and Spearman's correlation but a full discussion is beyond this guidance. In general,
Pearson's correlation has a higher statistical power than Spearman's, although the latter has some
more varied applications.
2.2.4.3 Serial Correlation Coefficient
When data are truly independent, the correlation between data points is zero. For a
sequence of data points taken serially in time, or one-by-one in a row, the serial correlation
EPA QA/G-9 Final
QAOO Version 2-11 July 2000
-------
Box 2-7: Directions for Calculating Spearman's Correlation Coefficient with an Example
LetX.,, X2, ..., Xn represent a set of ranks of the n data points of one data set and let Y.,, Y2, ..., Yn
represent a set of ranks of a second variable of the n data points. The Spearman correlation coefficient,
r, between X and Y is computed by:
v
~
i= 1 !'= 1
;=!
;=1
1/2
(15)
Example: Consider the following data set (in ppb): Sample 1 — arsenic (X) = 8.0, lead (Y) = 8.0;
Sample 2 - arsenic = 6.0, lead = 7.0; Sample 3 - arsenic = 2.0, lead = 7.0; and Sample 4 - arsenic = 1.0,
lead = 6.0.
Using Arsenic rank the data smallest to largest:
Sample No.
Arsenic
Lead
4
1.0
6.0
3
2.0
7.0
2
6.0
7.0
1
8.0
8.0
Convert the raw data to ranks, any ties being made an average of what ranks should have been
assigned.
Sample No.
Arsenic Rank
Lead Rank
3
2
2.5
2
3
2.5
4
4
4
(X)
(Y)
Note how 7.0 (two lead observations) was converted to the average rank (i.e., ranks 2 and 3, therefore
2.5 each).
;= 1
and r =
,.= 10,
;= 1 ;= 1
...+ (4x4) = 29.5.
;= 1
;= 1
[30 -
[29 5 _
1/2
= 0.948
4 4
Since r is close to 1, there is a strong linear relationship between these two contaminants.
coefficient can be calculated by replacing the sequencing variable by the numbers 1 through n and
calculating Pearon's correlation coefficient with z being the actual data values, and y being the
numbers 1 through n. For example, for a sequence of data collected at a waste site along a
straight transit line, the distances on the transit line of the data points are replaced by the numbers
1 through n, e.g., first 10-foot sample point = 1, the 20-foot sample point = 2, the 30-foot sample
EPA QA/G-9
QAOO Version
2- 12
Final
July 2000
-------
point = 3, etc., for samples taken at 10-foot intervals. Directions for the Serial correlation
coefficient, along with an example, are given in Box 2-8.
Box 2-8: Directions for Estimating the
Serial Correlation Coefficient with a Example
Directions:
LetX.,, X2, . . . , Xn represent the data values collected in sequence over equally spared periods of
time. Label the periods of time 1, 2, ..., n to match the data values. Use the directions in Box 2-6 to calculate
the Pearson's Correlation Coefficient between the data X and the time periods Y.
Example: The following are hourly readings from a discharge monitor. Notice how the actual 24-hour times are
replaced by the numbers 1 through 13.
Time
Reading
Time
Periods
12:00
6.5
1
13:00
6.6
2
14:00
6.7
3
15:00
6.4
4
16:00
6.3
5
17:00
6.4
6
18:00
6.2
7
19:00
6.2
8
20:00
6.3
9
21:00
6.6
10
22:00
6.8
11
23:00
6.9
12
24:00
7.0
13
Using Box 2-6, with the readings being the X values and the Time Periods being the Y values gives a serial
correlation coefficient of 0.432.
2.3 GRAPHICAL REPRESENTATIONS
2.3.1 Histogram/Frequency Plots
10
Two of the oldest methods for summarizing data distributions are the frequency plot
(Figure 2-1) and the histogram (Figure 2-2).
Both the histogram and the frequency plot use
the same basic principles to display the data:
dividing the data range into units, counting the
number of points within the units, and
displaying the data as the height or area within
a bar graph. There are slight differences
between the histogram and the frequency plot.
In the frequency plot, the relative height of the
bars represents the relative density of the data.
In a histogram, the area within the bar
represents the relative density of the data. The
difference between the two plots becomes
more distinct when unequal box sizes are used.
01 6
M
.Q
O
10
15 20 25
Concentration (ppm)
30
35
40
Figure 2-1. Example of a Frequency Plot
EPA QA/G-9
QAOO Version
2- 13
Final
July 2000
-------
The histogram and frequency plot
provide a means of assessing the symmetry and
variability of the data. If the data are symmetric,
then the structure of these plots will be
symmetric around a central point such as a
mean. The histogram and frequency plots will
generally indicate if the data are skewed and the
direction of the skewness.
Directions for generating a histogram
and a frequency plot are contained in Box 2-9
and an example is contained in Box 2-10. When
plotting a histogram for a continuous variable
(e.g., concentration), it is necessary to decide on
an endpoint convention; that is, what to do with cases that fall on the boundary of a box. With
discrete variables, (e.g., family size) the intervals can be centered in between the variables. For
the family size data, the intervals can span between 1.5 and 2.5, 2.5 and 3.5, and so on, so that the
whole numbers that relate to the family size can be centered within the box. The visual
impression conveyed by a histogram or a frequency plot can be quite sensitive to the choice of
interval width. The choice of the number of intervals determines whether the histogram shows
more detail for small sections of the data or whether the data will be displayed more simply as a
smooth overview of the distribution.
1 Percentage of Observations (per ppm)
O M ^ O) 00 |
-
) 5 10 15 20 25 30 35 40
Concentration (ppm)
Figure 2-2. Example of a Histogram
Box 2-9: Directions for Generating a Histogram and a Frequency Plot
LetX.,, X2, ..., Xn represent the n data points. To develop a histogram or a frequency plot:
STEP 1: Select intervals that cover the range of observations. If possible, these intervals should have
equal widths. A rule of thumb is to have between 7 to 11 intervals. If necessary, specify an
endpoint convention, i.e., what to do with cases that fall on interval endpoints.
STEP 2: Compute the number of observations within each interval. For a frequency plot with equal
interval sizes, the number of observations represents the height of the boxes on the frequency
plot.
STEP 3: Determine the horizontal axis based on the range of the data. The vertical axis for a frequency
plot is the number of observations. The vertical axis of the histogram is based on percentages.
STEP 4: For a histogram, compute the percentage of observations within each interval by dividing the
number of observations within each interval (Step 3) by the total number of observations.
STEP 5: For a histogram, select a common unit that corresponds to the x-axis. Compute the number of
common units in each interval and divide the percentage of observations within each interval
(Step 4) by this number. This step is only necessary when the intervals (Step 1) are not of
equal widths.
STEP 6: Using boxes, plot the intervals against the results of Step 5 for a histogram or the intervals
against the number of observations in an interval (Step 2) for a frequency plot.
EPA QA/G-9
QAOO Version
2- 14
Final
July 2000
-------
Consider
17.2 19.1
STEP 1:
STEP 2:
STEP 3:
STEP 4:
STEP 5:
STEP 6:
Box 2-10: Example of Generating a Histogram and a Frequency Plot
the following 22 samples of a contaminant concentration (in ppm): 17.7, 17.4, 22.8, 35.5, 28.6,
<4 72 <4 152 147 149 109 124 124 116 147 102 52 165 and 8 9
This data spans 0-40 ppm. Equally sized intervals of 5 ppm will be used: 0-5 ppm; 5-10
ppm; etc. The endpoint convention will be that values are placed in the highest interval
containing the value. For example, a value of 5 ppm will be placed in the interval 5-10 ppm
instead of 0 - 5 ppm.
The table below shows the number of observations within each interval defined in Step 1 .
The horizontal axis for the data is from 0 to 40 ppm. The vertical axis for the frequency plot is
from 0-10 and the vertical axis for the histogram is from 0% - 10%.
There are 22 observations total, so the number observations shown in the table below will be
divided by 22. The results are shown in column 3 of the table below.
A common unit for this data is 1 ppm. In each interval there are 5 common units so the
percentage of observations (column 3 of the table below) should be divided by 5 (column 4).
The frequency plot is shown in Figure 2-1 and the histogram is shown in Figure 2-2.
# of Obs % of Obs % of Obs
Interval in Interval in Interval per ppm
0 - 5 ppm 2 9.10 1.8
5 -10 ppm 3 13.60 2.7
10 -15 ppm 8 36.36 7.3
15 -20 ppm 6 27.27 5.5
20 -25 ppm 1 4.55 0.9
25 -30 ppm 1 4.55 0.9
30 - 35 ppm 0 0.00 0.0
35 -40 ppm 1 4.55 0.9
2.3.2 Stem-and-LeafPlot
The stem-and-leaf plot is used to show both the numerical values themselves and
information about the distribution of the data. It is a useful method for storing data in a compact
form while, at the same time, sorting the data from smallest to largest. A stem-and-leaf plot can
be more useful in analyzing data than a histogram because it not only allows a visualization of the
data distribution, but enables the data to be reconstructed and lists the observations in the order of
magnitude. However, the stem-and-leaf plot is one of the more subjective visualization
techniques because it requires the analyst to make some arbitrary choices regarding a partitioning
of the data. Therefore, this technique may require some practice or trial and error before a useful
plot can be created. As a result, the stem-and-leaf plot should only be used to develop a picture
of the data and its characteristics. Directions for constructing a stem-and-leaf plot are given in
Box 2-11 and an example is contained in Box 2-12.
Each observation in the stem-and-leaf plot consist of two parts: the stem of the
observation and the leaf. The stem is generally made up of the leading digit of the numerical
values while the leaf is made up of trailing digits in the order that corresponds to the order of
magnitude from left to right. The stem is displayed on the vertical axis and the data points make
EPA QA/G-9 Final
QAOO Version 2-15 July 2000
-------
Box 2-11: Directions for Generating a Stem and Leaf Plot
Let X.,, X2, ..., Xn represent the n data points. To develop a stem-and-leaf plot, complete the following steps:
STEP 1: Arrange the observations in ascending order. The ordered data is usually labeled (from smallest
to largest) X(1),X, 2,, ..., X(n).
STEP 2: Choose either one or more of the leading digits to be the stem values. As an example, for the
value 16, 1 could be used as the stem as it is the leading digit.
STEP 3: List the stem values from smallest to largest at the left (along a vertical axis). Enter the leaf (the
remaining digits) values in order from lowest to highest to the right of the stem. Using the value
16 as an example, if the 1 is the stem then the 6 will be the leaf.
Box 2-12: Example of Generating a Stem and Leaf Plot
Consider the following 22 samples of trifluorine (in ppm): 17.7, 17.4, 22.8, 35.5, 28.6, 17.2 19.1, <4, 7.2,
<4, 15.2, 14.7, 14.9, 10.9, 12.4, 12.4, 11.6, 14.7, 10.2, 5.2, 16.5, and 8.9.
STEP 1: Arrange the observations in ascending order: <4, <4, 5.2, 7.7, 8.9, 10.2, 10.9, 11.6, 12.4, 12.4,
14.7, 14.7, 14.9, 15.2, 16.5, 17.2, 17.4, 17.7, 19.1, 22.8, 28.6, 35.5.
STEP 2: Choose either one or more of the leading digits to be the stem values. For the above data, using
the first digit as the stem does not provide enough detail for analysis. Therefore, the first digit
will be used as a stem; however, each stem will have two rows, one for the leaves 0-4, the other
for the leaves 5-9.
STEP 3: List the stem values at the left (along a vertical axis) from smallest to largest. Enter the leaf (the
remaining digits) values in order from lowest to highest to the right of the stem. The first digit of
the data was used as the stem values; however, each stem value has two leaf rows.
0(0,1,2,3,4) | <4 <4
0(5, 6, 7, 8, 9) | 5.2 7.7 8.9
1 (0, 1, 2, 3, 4) | 0.2 0.9 1.6 2.4 2.4 4.7 4.7 4.9
1 (5, 6, 7, 8, 9) | 5.2 6.5 7.2 7.4 7.7 9.1
2(0,1,2,3,4) | 2.8
2 (5, 6, 7, 8, 9) | 8.6
3(0,1,2,3,4) |
3(5,6,7,8,9) | 5.5
Note: If nondetects are present, place them first in the ordered list, using a symbol such as
-------
up the magnitude from left to right. The stem is displayed on the vertical axis and the data points
make up the leaves. Changing the stem can be accomplished by increasing or decreasing the
digits that are used, dividing the groupings of one stem (i.e., all numbers which start with the
numeral 6 can be divided into smaller groupings), or multiplying the data by a constant factor (i.e.,
multiply the data by 10 or 100). Nondetects can be placed in a single stem.
A stem-and-leaf plot roughly displays the distribution of the data. For example, the stem-
and-leaf plot of normally distributed data is approximately bell shaped. Since the stem-and-leaf
roughly displays the distribution of the data, the plot may be used to evaluate whether the data are
skewed or symmetric. The top half of the stem-and-leaf plot will be a mirror image of the bottom
half of the stem-and-leaf plot for symmetric data. Data that are skewed to the left will have the
bulk of data in the top of the plot and less data spread out over the bottom of the plot.
2.3.3 Box and Whisker Plot
A box and whisker plot or box plot (Figure 2-3) is a schematic diagram useful for
visualizing important statistical quantities of the data. Box plots are useful in situations where it is
not necessary or feasible to portray all the details of a distribution. Directions for generating a
box and whiskers plot are contained in Box 2-13, and an example is contained in Box 2-14.
A box and whiskers plot is composed of a central box divided by a
line and two lines extending out from the box called whiskers. The length of
the central box indicates the spread of the bulk of the data (the central 50%)
while the length of the whiskers show how stretched the tails of the
distribution are. The width of the box has no particular meaning; the plot
can be made quite narrow without affecting its visual impact. The sample
median is displayed as a line through the box and the sample mean is
displayed using a '+' sign. Any unusually small or large data points are
displayed by a '*' on the plot. A box and whiskers plot can be used to
assess the symmetry of the data. If the distribution is symmetrical, then the
box is divided in two equal halves by the median, the whiskers will be the
same length and the number of extreme data points will be distributed
equally on either end of the plot.
2.3.4 Ranked Data Plot
Figure 2-3.
Example of a Box
and Whisker Plot
A ranked data plot is a useful graphical representation that is easy to
construct, easy to interpret, and makes no assumptions about a model for the data. The analyst
does not have to make any arbitrary choices regarding the data to construct a ranked data plot
(such as cell sizes for a histogram). In addition, a ranked data plot displays every data point;
therefore, it is a graphical representation of the data instead of a summary of the data. Directions
for developing a ranked data plot are given in Box 2-15 and an example is given in Box 2-16.
EPA QA/G-9
QAOO Version
2- 17
Final
July 2000
-------
Box 2-13: Directions for Generating a Box and Whiskers Plot
STEP 1 :
STEP 2:
Set the vertical scale of the plot based on the maximum and minimum values of the data set.
Select a width for the box plot keeping in mind that the width is only a visualization tool. Label the
width w; the horizontal scale then ranges from -1/4Wto 1/
Compute the upper quartile (Q(.75), the 75th percentile) and the lower quartile (Q(.25), the 25th
percentile) using Box 2-1. Compute the sample mean and median using Box 2-2. Then, compute
the interquartile range (IQR) where IQR = Q(.75) - Q(.25).
STEP 3: Draw a box through points ( -1/2W, Q(.75) ), ( -1/2W, Q (.25) ), ( 1/2W, Q(.25) ) and ( 1/2W, Q(.75) ).
Draw a line from (1/4W, Q(.5)) to (-1/4W, Q(.5)) and mark point (0, x) with (+).
STEP 4'. Compute the upper end of the top whisker by finding the largest data value X less than
Q(.75) + 1.5( Q(.75) - Q(.25) ). Draw a line from (0, Q(.75)) to (0, X).
Compute the lower end of the bottom whisker by finding the smallest data value Y greater than
Q(.25) - 1.5( Q(.75) - Q(.25) ). Draw a line from (0, Q(.25)) to (0, Y).
STEP 5: For all points X* > X, place an asterisk (*) at the point (0, X*).
For all points Y* < Y, place an asterisk (*) at the point (0, Y*).
Box 2-14: Example of a Box and Whiskers Plot
Consider the following 22 samples of trifluorine (in ppm) listed in order from smallest to largest: 4.0, 6.1, 9.8,
10.7, 10.8, 11.5, 11.6, 12.4, 12.4, 14.6, 14.7, 14.7, 16.5, 17, 17.5, 20.6, 20.8, 25.7, 25.9, 26.5, 32.0, and 35.5.
STEP 1: The data ranges from 4.0 to 35.5 ppm. This is the range of the vertical axis. Arbitrarily, a width of
4 will be used for the horizontal axis.
STEP 2: Using the formulas in Box 2-2, the sample mean = 16.87 and the median = 14.70. Using Box 2-1,
Q(.75) = 20.8 and Q(.25) = 11.5. Therefore, IQR = 20.8 -11.5 =
9.3.
STEP 3: In the figure, a box has been drawn through points (-2, 20.8),
(-2, 11.5), ( 2, 11.5 ), (2, 20.8). A line has been drawn from (-2 ,
14.7 ) to ( 2, 14.7 ), and the point (0, 16.87) has been marked with
a '+' sign.
STEP 4: Q(.75) + 1.5(9.3) = 34.75. The closest data value to this number,
but less than it, is 32.0. Therefore, a line has been drawn in the
figure from ( 0, 20.8) to (0, 32.0).
Q(.25) -1.5( 9.3 ) = -2.45. The closest data value to this number,
but greater than it, is 4.0. Therefore, a line has been drawn in the
figure from ( 0, 4) to ( 0, 11.5).
STEP 5: There is only 1 data value greater than 32.0 which is 35.5.
Therefore, the point ( 0, 35.5) has been marked with an asterisk.
There are no data values less than 4.0.
EPA QA/G-9
QAOO Version
2- 18
Final
July 2000
-------
Box 2-15: Directions for Generating a Ranked Data Plot
LetX.,, X2, ..., Xn represent the n data points. LetX(i), for i=1 to
n, be the data listed in order from smallest to largest so that X,.,}
(i = 1) is the smallest, X(2) (i = 2) is the second smallest, and X(n
) (i = n) is the largest. To generate a ranked data plot, plot the
ordered X values at equally spaced intervals along the horizontal
axis.
Box 2-16: Example of Generating a Ranked Data Plot
Consider the following 22 samples of triflourine (in ppm): 17.7, 17.4, 22.8, 35.5, 28.6, 17.2
191 49 72 40 152 147 149 109 124 124 116 147 102 52 165 and 8 9 The
data listed in order from smallest to largest X(i) along with the ordered number of the
observation (i) are:
1~
2
3
4
5
6
7
8
9
10
11
A ranked data
4.0" 12 14.7"
4.9 13 14.9
5.2 14 15.2
7.7 15 16.5
8.9 16 17.2
10.2 17 17.4
10.9 18 17.7
11.6 19 19.1
12.4 20 22.8
12.4 21 28.6
14.7 22 35.5
plot of this data is a plot of the pairs ( i, X(i)). This plot is shown below:
40
35
30
Q.
5=20
ro
° 15
10
5
0
-
-
"
. • • *
•
. *
I I I I I I I I I I I I I I I I I I I I I I I
Smallest ^ Largest
EPA QA/G-9
QAOO Version
2- 19
Final
July 2000
-------
A ranked data plot is a plot of the data from smallest to largest at evenly spaced intervals
(Figure 2-4). This graphical representation is very similar to the quantile plot described in Section
2.3.5. A ranked data plot is marginally easier to generate than a quantile plot; however, a ranked
data plot does not contain as much information as a quantile plot. Both plots can be used to
determine the density of the data points and the skewness of the data; however, a quantile plot
contains information on the quartiles of the data whereas a ranked data plot does not.
Smallest
Largest
Figure 2-4. Example of a Ranked Data Plot
A ranked data plot can be used to determine the density of the data values, i.e., if all the
data values are close to the center of the data with relatively few values in the tails or if there is a
large amount of values in one tail with the rest evenly distributed. The density of the data is
displayed through the slope of the graph. A large amount of data values has a flat slope, i.e., the
graph rises slowly. A small amount of data values has a large slope, i.e., the graph rises quickly.
Thus the analyst can determine where the data lie, either evenly distributed or in large clusters of
points. In Figure 2-4, the data rises slowly up to a point where the slope increases and the graph
rises relatively quickly. This means that there is a large amount of small data values and relatively
few large data values.
A ranked data plot can be used to determine if the data are skewed or if they are
symmetric. A ranked data plot of data that are skewed to the right extends more sharply at the
top giving the graph a convex shape. A ranked data plot of data that are skewed to the left
increases sharply near the bottom giving the graph a concave shape. If the data are symmetric,
then the top portion of the graph will stretch to upper right corner in the same way the bottom
portion of the graph stretches to lower left, creating a s-shape. Figure 2-4 shows a ranked data
plot of data that are skewed to the right.
EPA QA/G-9
QAOO Version
2-20
Final
July 2000
-------
2.3.5 Quantile Plot
A quantile plot (Figure 2-5) is a graphical representation of the data that is easy to
construct, easy to interpret, and makes no assumptions about a model for the data. The analyst
does not have to make any arbitrary choices regarding the data to construct a quantile plot (such
as cell sizes for a histogram). In addition, a quantile plot displays every data point; therefore, it is
a graphical representation of the data instead of a summary of the data.
0.2 0.4 0.6 0.8
Fraction of Data (f-values)
Figure 2-5. Example of a Quantile Plot of Skewed Data
A quantile plot is a graph of the quantiles (Section 2.2.1) of the data. The basic quantile
plot is visually identical to a ranked data plot except its horizontal axis varies from 0.0 to 1.0, with
each point plotted according to the fraction of the points it exceeds. This allows the addition of
vertical lines indicating the quartiles or, any other quantiles of interest. Directions for developing
a quantile plot are given in Box 2-17 and an example is given in Box 2-18.
A quantile plot can be used to read the quantile information such as the median, quartiles,
and the interquartile range. In addition, the plot can be used to determine the density of the data
points, e.g., are all the data values close to the center with relatively few values in the tails or are
there a large amount of values in one tail with the rest evenly distributed? The density of the data
is displayed through the slope of the graph. A large amount of data values has a flat slope, i.e.,
the graph rises slowly. A small amount of data values has a large slope, i.e., the graph rises
quickly. A quantile plot can be used to determine if the data are skewed or if they are symmetric.
A quantile plot of data that are skewed to the right is steeper at the top right than the bottom left,
as in Figure 2-5. A quantile plot of data that are skewed to the left increases sharply near the
bottom left of the graph. If the data are symmetric then the top portion of the graph will stretch
to the upper right corner in the same way the bottom portion of the graph stretches to the lower
left, creating an s-shape.
EPA QA/G-9
QAOO Version
2-21
Final
July 2000
-------
Box 2-17: Directions for Generating a Quantile Plot
LetX.,, X2, ..., Xn represent the n data points. To obtain a quantile plot, letX(i),
for
i = 1 to n, be the data listed in order from smallest to largest so that X,.,} (i = 1)
is the smallest, X(2) (i = 2) is the second smallest, and X(n) (i = n) is the
largest. For each i, compute the fraction f,= (i - 0.5)/n. The quantile plot is a
plot of the pairs (fh X(i)), with straight lines connecting consecutive points.
Box 2-18: Example of Generating a Quantile Plot
Consider the following 10 data points: 4 ppm, 5 ppm, 6 ppm, 7 ppm, 4 ppm, 10 ppm, 4 ppm, 5 ppm, 7
ppm, and 8 ppm. The data ordered from smallest to largest, X(i), are shown in the first column of the
table below and the ordered number for each observation, i, is shown in the second column. The third
column displays the values f, for each i where f,= (i - 0.5)/n.
4
4
4
5
5
1
2
3
4
5
0.05
0.15
0.25
0.35
0.45
6
7
7
8
10
I
6
7
8
9
10
0.55
0.65
0.75
0.85
0.95
The pairs (fh X(i)) are then plotted to yield the following quantile plot:
10
ro
•%
a 6
0 0.2 0.4 0.6 0.8 1
Fraction of Data (f-values)
Note that the graph curves upward; therefore, the data appear to be skewed to the right.
2.3.6 Normal Probability Plot (Quantile-Quantile Plot)
There are two types of quantile-quantile plots or q-q plots. The first type, an empirical
quantile-quantile plot (Section 2.3.7.4), involves plotting the quantiles of two data variables
against each other. The second type of a quantile-quantile plot, a theoretical quantile-quantile
plot, involves graphing the quantiles of a set of data against the quantiles of a specific distribution.
The following discussion will focus on the most common of these plots for environmental data,
EPA QA/G-9
QAOO Version
2-22
Final
July 2000
-------
the normal probability plot (the normal q-q plot); however, the discussion holds for other q-q
plots. The normal probability plot is used to roughly determine how well the data set is modeled
by a normal distribution. Formal tests are contained in Chapter 4, Section 2. Directions for
developing a normal probability plot are given in Box 2-19 and an example is given in Box 2-20.
A discussion of the normal distribution is contained in Section 2.4.
A normal probability plot is the graph of the quantiles of a data set against the quantiles of
the normal distribution using normal probability graph paper (Figure 2-6). If the graph is linear,
the data may be normally distributed. If the graph is not linear, the departures from linearity give
important information about how the data distribution deviates from a normal distribution.
If the graph of the normal probability plot is not linear, the graph may be used to
determine the degree of symmetry (or asymmetry) displayed by the data. If the data in the upper
tail fall above and the data in the lower tail fall below the quartile line, the data are too slender to
be well modeled by a normal distribution, i.e., there are fewer values in the tails of the data set
than what is expected from a normal distribution. If the data in the upper tail fall below and the
data in the lower tail fall above the quartile line, then the tails of the data are too heavy to be well
modeled using a normal distribution, i.e., there are more values in the tails of the data than what is
expected from a normal distribution. A normal probability plot can be used to identify potential
outliers. A data value (or a few data values) much larger or much smaller than the rest will cause
the other data values to be compressed into the middle of the graph, ruining the resolution.
Box 2-19: Directions for Constructing a Normal Probability Plot
LetX.,, X2, ..., Xn represent the n data points.
STEP 1: For each data value, compute the absolute frequency, AF,. The absolute frequency is the
number of times each value occurs. For distinct values, the absolute frequency is 1 . For
non-distinct observations, count the number of times an observation occurs. For example,
consider the data 1, 2, 3, 3. The absolute frequency of value 1 is 1 and the absolute
frequency of value 2 is 1. The absolute frequency of value 3 is 2 since 3 appears 2 times
in the data set.
STEP 2: Compute the cumulative frequencies, CF,. The cumulative frequency is the number of data
;'
points that are less than or equal to Xh i.e., CFt = /AF .. Using the data given in step
y=i
2, the cumulative frequency for value 1 is 1, the cumulative frequency for value 2 is 2
(1 + 1), and the cumulative frequency for value 3 is 4 (1 + 1+2).
CFi
STEP 3: Compute Y = 100 x - and plot the pairs (Yh X,) using normal probability paper
(Figure 2-6). If the graph of these pairs approximately forms a straight line, then the data
are probably normally distributed. Otherwise, the data may not be normally distributed.
EPA QA/G-9 Final
QAOO Version 2 - 23 July 2000
-------
Box 2-20: Example of Normal Probability Plot
Considerthe following 15 data points: 5, 5, 6, 6, 8, 8, 9, 10, 10, 10, 10, 10, 12, 14, and 15.
STEP 1: Because the value 5 appears 2 times, its absolute frequency is 2. Similarly, the absolute frequency
of 6 is 2, of 8 is 2, of 9 is 1, of 10 is 5, etc. These values are shown in the second column of the
table below.
STEP 2: The cumulative frequency of the data value 8 is 6 because there are 2 values of 5, 2 values of 6, and
2 values of 8. The cumulative frequencies are shown in the 3rd column of the table.
CFt
STEP 3: The values Y. = 100 x ( )for each data point are shown in column 4 of the table below. A
n+ 1
plot of these pairs (Yh X,) using normal probability paper is also shown below.
i
1
2
3
4
5
6
7
8
Individual
Xj
5
6
8
9
10
12
14
15
Absolute
Frequency AFi
2
2
2
1
5
1
1
1
Cumulative
Frequency CFi
2
4
6
7
12
13
14
15
Y:
12.50
25.00
37.50
43.75
75.00
81.25
87.50
93.75
X
18
14
12
10
20 30 40 50 60 70 80
Y
EPA QA/G-9
QAOO Version
2-24
Final
July 2000
-------
0
EPA
QA(
01 0.05
QA/G-9
)0 Version
0.1 0.2 0.5
1
2 5 10
20 30 40 50 60 70 80 90 95 98 99
Figure 2-6. Normal Probability Paper
2-25
99.8 99.9
99.99
Final
May 2000
-------
2.3.7 Plots for Two or More Variables
Data often consist of measurements of several characteristics (variables) for each sample
point in the data set. For example, a data set may consist of measurements of weight, sex, and
age for each animal in a sample or may consist of daily temperature readings for several cities. In
this case, graphs may be used to compare and contrast different variables. For example, the
analyst may wish to compare and contrast the temperature readings for different cities, or
different sample points (each containing several variables) such the height, weight, and sex across
individuals in a study.
To compare and contrast individual data points, some special plots have been developed
to display multiple variables. These plots are discussed in Section 2.3.7.1. To compare and
contrast several variables, collections of the single variable displays described in previous sections
are useful. For example, the analyst may generate box and whisker plots or histograms for each
variable using the same axis for all of the variables. Separate plots for each variable may be
overlaid on one graph, such as overlaying quantile plots for each variable on one graph. Another
useful technique for comparing two variables is to place the stem and leaf plots back to back. In
addition, some special plots have been developed to display two or more variables. These plots
are described in Sections 2.3.7.2 through 2.3.7.4.
2.3.7.1 Plots for Individual Data Points
Since it is difficult to visualize data in more than 2 or 3 dimensions, most of the plots
developed to display multiple variables for individual data points involve representing each
variable as a distinct piece of a two-dimensional figure. Some such plots include Profiles, Glyphs,
and Stars (Figure 2-7). These graphical representations start with a specific symbol to represent
each data point, then modify the various features of the symbol in proportion to the magnitude of
each variable. The proportion of the magnitude is determined by letting the minimum value for
each variable be of length 0, the maximum be of length 1. The remaining values of each variable
are then proportioned based on the magnitude of each value in relation to the maximum and
minimum.
Profile Plot
Glyph Plot
Star Plot
Figure 2-7. Example of Graphical Representations of
Multiple Variables
EPA QA/G-9
QAOO Version
2-26
Final
July 2000
-------
A profile plot starts with a line segment of a fixed length. Then lines spaced an equal
distance apart and extended perpendicular to the line segment represent each variable. A glyph
plot uses a circle of fixed radius. From the perimeter, parallel rays whose sizes are proportional to
the magnitude of the variable extend from the top half of the circle. A star plot starts with a point
where rays spaced evenly around the circle represent each variable and a polygon is then drawn
around the outside edge of the rays.
2.3.7.2
Scatter Plot
For data sets consisting of paired observations where two or more continuous variables
are measured for each sampling point, a scatter plot is one of the most powerful tools for
analyzing the relationship between two or more variables. Scatter plots are easy to construct for
two variables (Figure 2-8) and many computer graphics packages can construct 3-dimensional
scatter plots. Directions for constructing a scatter plot for two variables are given in Box 2-21
along with an example.
40
30
Q.
Q.
^20
"•
10
°c
>K ^
-
*
^ ^ >K
* i*r"% " **
2468
Chromium VI (ppb)
Figure 2-8. Example of a Scatter Plot
A scatter plot clearly shows the relationship between two variables. Both potential
outliers from a single variable and potential outliers from the paired variables may be identified on
this plot. A scatter plot also displays the correlation between the two variables. Scatter plots of
highly linearly correlated variables cluster compactly around a straight line. In addition, nonlinear
patterns may be obvious on a scatter plot. For example, consider two variables where one
variable is approximately equal to the square of the other. A scatter plot of this data would
display a u-shaped (parabolic) curve. Another important feature that can be detected using a
scatter plot is any clustering effect among the data.
2.3.7.3
Extensions of the Scatter Plot
It is easy to construct a 2-dimensional scatter plot by hand and many software packages
can construct a useful 3-dimensional scatter plot. However, with more than 3 variables, it is
difficult to construct and interpret a scatter plot. Therefore, several graphical representations
have been developed that extend the idea of a scatter plot for data consisting of 2 or more
variables.
EPA QA/G-9
QAOO Version
2-27
Final
July 2000
-------
Box 2-21: Directions for Generating a Scatter Plot and an Example
LetX.,, X2, ..., Xn represent one variable of the n data points and let Y.,, Y2, ..., Yn represent a second variable of
the n data points. The paired data can be written as (Xh Y,) for i = 1 , ..., n. To construct a scatter plot, plot the
first variable along the horizontal axis and the second variable along the vertical axis. It does not matter which
variable is placed on which axis.
Example: /
and Chrom
\ scatter plot will be developed for the data below. PCE values are displayed on the vertical axis
um VI values are displayed on the horizontal axis of Figure 2-8.
PCE
(PPb)
14.49
37.21
10.78
18.62
7.44
37.84
13.59
4.31
Chromium
VI (ppb)
3.76
6.92
1.05
6.30
1.43
6.38
5.07
3.56
PCE
(PPb)
2.23
3.51
6.42
2.98
3.04
12.60
3.56
7.72
Chromium
VI (ppb)
0.77
1.24
3.48
1.02
1.15
5.44
2.49
3.01
PCE
(PPb)
4.14
3.26
5.22
4.02
6.30
8.22
1.32
7.73
5.88
Chromium
VI (ppb)
2.36
0.68
0.65
0.68
1.93
3.48
2.73
1.61
1.42
30
20
10
Chromium vs. PCE
Atrazine vs. PCE
Atrazine vs. Chromium IV
The simplest of these graphical representations is a coded scatter plot. In this case, all
possible pairs of data are given a code and plotted on one scatter plot. For example, consider a
data set of 3 variables: variable A, variable B, and variable C. Using the first variable to
designate the horizontal axis, the
analyst may choose to display the
pairs (A, B) using an X, the pairs (A,
C) using a Y, and the pairs (B, C)
using a Z on one scatter plot. All of
the information described above for
a scatter plot is also available on a
coded scatter plot. However, this
method assumes that the ranges of
the three variables are comparable
and does not provide information on
three-way or higher interactions
between the variables. An example
of a coded scatter plot is given in
Figure 2-9.
10
20
Figure 2-9. Example of a Coded Scatter Plot
A parallel coordinate plot also extends the idea of a scatter plot to higher dimensions. The
parallel coordinates method employs a scheme where coordinate axes are drawn in parallel
(instead of perpendicular). Consider a sample point X consisting of values Xt for variable 1, X2
for variable 2, and so on up to Xp for variable p. A parallel coordinate plot is constructed by
placing an axis for each of the p variables parallel to each other and plotting Xt on axis 1, X2 on
axis 2, and so on through Xp on axis p and joining these points with a broken line. This method
EPA QA/G-9
QAOO Version
2-28
Final
July 2000
-------
contains all of the information
available on a scatter plot in
addition to information on 3-way
and higher interactions (e.g.,
clustering among three variables).
However, for p variables one
must construct (p+l)/2 parallel
coordinate plots in order to
display all possible pairs of
variables. An example of a
parallel coordinate plot is given in
Figure 2-10.
A scatter plot matrix is
another useful method of
extending scatter plots to higher
dimensions. In this case, a scatter plot is developed for all possible pairs of the variables which
are then displayed in a matrix format. This method is easy to implement and provides a concise
method of displaying the individual scatter plots. However, this method does not contain
information on 3-way or higher interactions between variables. An example of a scatter plot
matrix is contained in Figure 2-11.
Figure 2-10. Example of a Parallel Coordinates Plot
40
Q.
S
I 20
Chromil
o
0
40
£•30
a.
Q.
5
1 20
Chromil
o o
14
12
s-10
Q.
£ 8
HI
=
5 6
~~r 4
/, , , , \
r
- +
+
++ +
*^ +
"^
10 20 30 40 0 10 20 30 40
Chromium IV (ppb) 14
12
-10
Q.
S 8
HI
=
S 6
\ ^ 4
Chromium IV (ppb)
+
++
A-
:+++
2 4 6 8 10 12 14 0 2 4 6 8 10 12 14
Atrazine (ppb) Atrazine (ppb)
Figure 2-11. Example of a Matrix Scatter Plot
EPA QA/G-9
QAOO Version
2-29
Final
July 2000
-------
2.3.7.4 Empirical Quantile-Quantile Plot
An empirical quantile-quantile (q-q) plot involves plotting the quantiles (Section 2.2.1) of
two data variables against each other. This plot is used to compare distributions of two or more
variables; for example, the analyst may wish to compare the distribution of lead and iron samples
from a drinking water well. This plot is similar in concept to the theoretical quantile-quantile plot
and yields similar information in regard to the distribution of two variables instead of the
distribution of one variable in relation to a fixed distribution. Directions for constructing an
empirical q-q plot with an example are given in Box 2-22.
An empirical q-q plot is the graph of the quantiles of one variable of a data set against the
quantiles of another variable of the data set. This plot is used to determine how well the
distribution of the two variables match. If the distributions are roughly the same, the graph is
linear or close to linear. If the distributions are not the same, than the graph is not linear. Even if
the graph is not linear, the departures from linearity give important information about how the
two data distributions differ. For example, a q-q plot can be used to compare the tails of the two
data distributions in the same manner a normal probability plot was used to compare the tails of
the data to the tails of a normal distribution. In addition, potential outliers (from the paired data)
may be identified on this graph.
2.3.8 Plots for Temporal Data
Data collected over specific time intervals (e.g., monthly, biweekly, or hourly) have a
temporal component. For example, air monitoring measurements of a pollutant may be collected
once a minute or once a day; water quality monitoring measurements of a contaminant level may
be collected weekly or monthly. An analyst examining temporal data may be interested in the
trends over time, correlation among time periods, and cyclical patterns. Some graphical
representations specific to temporal data are the time plot, correlogram, and variogram.
Data collected at regular time intervals are called time series. Time series data may be
analyzed using Box-Jenkins modeling and spectral analysis. Both of these methods require a large
amount of data collected at regular intervals and are beyond the scope of this guidance. It is
recommended that the interested reader consult a statistician.
The graphical representations presented in this section are recommended for all data that
have a temporal component regardless of whether formal statistical time series analysis will be
used to analyze the data. If the analyst uses a time series methodology, the graphical
representations presented below will play an important role in this analysis. If the analyst decides
not to use time series methodologies, the graphical representations described below will help
identify temporal patterns that need to be accounted for in the analysis of the data
EPA QA/G-9 Final
QAOO Version 2-30 July 2000
-------
Box 2-22: Directions for Constructing an Empirical Q-Q Plot with an Example
LetX.,, X2, ..., Xn represent n data points of one variable and let Y.,, Y2, ..., Ym represent a second variable
of m data points. Let X,,}, for i = 1 to n, be the first variable listed in order from smallest to largest so that
X,!) (i = 1) is the smallest, X, 2} (i = 2) is the second smallest, and X, n} (i = n) is the largest. Let Y,,}, for i
= 1 to n, be the second variable listed in order from smallest to largest so that Y,.,} (i = 1) is the smallest,
Y(2) (i = 2) is the second smallest, and Y(m) (i = m) is the largest.
If m = n: If the two variables have the same number of observations, then an empirical q-q plot of the
two variables is simply a plot of the ordered values of the variables. Since n=m, replace m by n. A plot
of the pairs (X(1), Y(1)), (X(2), Y,
x(2). '(2)!
(X(n), Y(n)) is an empirical quantile-quantile plot.
If n > m: If the two variables have a different number of observations, then the empirical quantile-quantile
plot will consist of m (the smaller number) pairs. The empirical q-q plot will then be a plot of the ordered
Y values against the interpolated X values. For i = 1, i = 2, ..., i = m, let v = (n/m)(i - 0.5) + 0.5 and
separate the result into the integer part and the fractional part, i.e., let v = j + g where j is the integer part
and g is the fraction part. If g = 0, plot the pair (Y(i), X(i)). Otherwise, plot the pair (Y(i), (1-g)X(j) + gX(j +
!)). A plot of these pairs is an empirical quantile-quantile plot.
Example: Consider two sets of contaminant readings from two separate drinking water wells at the same
site. The data from well 1 are: 1.32, 3.26, 3.56, 4.02, 4.14, 5.22, 6.30, 7.72, 7.73, and 8.22. The data
from well 2 are: 0.65, 0.68, 0.68, 1.42, 1.61, 1.93, 2.36, 2.49, 2.73, 3.01, 3.48, and 5.44. An empirical
q-q plot will be used to compare the distributions of these two wells. Since there are 10 observations in
well 1, and 12 observations in well, the case for n ? m will be used. Therefore, for i = 1, 2, ..., 10,
compute:
i = 1: v = —(l-.5)+.5 = 1.1 soj=1 andg=.1. Since g*0, plot (1.32,(.9).65+(.1).68)=(1.32,
0.653)
12,
1 = 2: v = —(2-.5)+.5 = 2.3 so j=2 and g=3. Since g^O, plot (3.26,(.7).68+(.3).68)=(3.26, 0.68)
1 = 3: v = —(3-.5)+.5 = 3.5 so j=3 and g=5. Since g^O, plot (3.56,(.5).68+(.5)1.42)=(3.56,1.05)
Continue this process for i =4, 5, 6, 7, 8, 9, and 10 to yield the following 10 data pairs (1.32, 0.653),
(3.26, 0.68), (3.56, 1.05), (4.02, 1.553), (4.14, 1.898), (5.22, 2.373), (6.30, 2.562), (7.72, 2.87), (7.73,
3.339), and (8.22, 5.244). These pairs are plotted below, along with the best fitting regression line.
10
tM
~s
!6
4>
'c ^
ro
O
4 6
QuantilesofWell 1
10
EPA QA/G-9
QAOO Version
2-31
Final
July 2000
-------
The analyst examining temporal environmental data may be interested in seasonal trends,
directional trends, serial correlation, and stationarity. Seasonal trends are patterns in the data that
repeat over time, i.e., the data rise and fall regularly over one or more time periods. Seasonal
trends may be large scale, such as a yearly trend where the data show the same pattern of rising
and falling over each year, or the trends may be small scale, such as a daily trend where the data
show the same pattern for each day. Directional trends are downward or upward trends in the
data which is of importance to environmental applications where contaminant levels may be
increasing or decreasing. Serial correlation is a measure of the extent to which successive
observations are related. If successive observations are related, statistical quantities calculated
without accounting for serial correlation may be biased. Finally, another item of interest for
temporal data is stationarity (cyclical patterns). Stationary data look the same over all time
periods. Directional trends and increasing (or decreasing) variability among the data imply that
the data are not stationary.
Temporal data are sometimes used in environmental applications in conjunction with a
statistical hypothesis test to determine if contaminant levels have changed. If the hypothesis test
does not account for temporal trends or seasonal variations, the data must achieve a "steady state"
before the hypothesis test may be performed. Therefore, the data must be essentially the same for
comparable periods of time both before and after the hypothesized time of change.
Sometimes multiple observations are taken in each time period. For example, the
sampling design may specify selecting 5 samples every Monday for 3 months. If this is the case,
the time plot described in Section 2.3.8.1 may be used to display the data, display the mean
weekly level, display a confidence interval for each mean, or display a confidence interval for each
mean with the individual data values. A time plot of all the data can be used to determine if the
variability for the different time periods changes. A time plot of the means can be used to
determine if the means are possibly changing between time periods. In addition, each time period
may be treated as a distinct variable and the methods of Section 2.3.7 may be applied.
2.3.8.1 Time Plot
One of the simplest plots to generate that provides a large amount of information is a time
plot. A time plot is a plot of the data over time. This plot makes it easy to identify large-scale
and small-scale trends over time. Small-scale trends show up on a time plot as fluctuations in
smaller time periods. For example, ozone levels over the course of one day typically rise until the
afternoon, then decrease, and this process is repeated every day. Larger scale trends, such as
seasonal fluctuations, appear as regular rises and drops in the graph. For example, ozone levels
tend to be higher in the summer than in the winter so ozone data tend to show both a daily trend
and a seasonal trend. A time plot can also show directional trends and increased variability over
time. Possible outliers may also be easily identified using a time plot.
A time plot (Figure 2-12) is constructed by numbering the observations in order by time.
The time ordering is plotted on the horizontal axis and the corresponding observation is plotted
on the vertical axis. The points plotted on a time plot may be joined by lines; however, it is
EPA QA/G-9 Final
QAOO Version 2 - 32 July 2000
-------
recommended that the plotted points not be connected to avoid creating a false sense of
continuity. The scaling of the vertical axis of a time plot is of some importance. A wider scale
tends to emphasize large-scale trends, whereas a smaller scale tends to emphasize small-scale
trends. Using the ozone example above, a wide scale would emphasize the seasonal component
of the data, whereas a smaller scale would tend to emphasize the daily fluctuations. Directions for
constructing a time plot are contained in Box 2-23 along with an example.
20
15
in
> 10
re
re
Q
5
0
• *
* ^ ^
** "" * ** *%** * "" **********
** ** ** * *
) 5 10 15 20 25 30 35 40 45 50
Time
Figure 2-12. Example of a Time Plot Showing a Slight Downward Trend
Box 2-23: Directions for Generating a Time Plot and an Example
LetX.,, X2, ..., Xn represent n data points listed in order by time, i.e., the subscript represents the ordered time
interval. A plot of the pairs (i, X,) is a time plot of this data.
Example: Consider the following 50 daily observations (listed in order by day): 10.05, 11.22, 15.9, 11.15,
10.53, 13.33, 11.81, 14.78, 10.93, 10.31, 7.95, 10.11, 10.27, 14.25, 8.6, 9.18, 12.2, 9.52, 7.59, 10.33, 12.13,
11.31, 10.13, 7.11, 6.72, 8.97, 10.11, 7.72, 9.57, 6.23, 7.25, 8.89, 9.14, 12.34, 9.99, 11.26, 5.57, 9.55, 8.91,
7.11, 6.04, 8.67, 5.62, 5.99, 5.78, 8.66, 5.8, 6.9, 7.7, 8.87. By labeling day 1 as 1, day 2 as 2, and soon, a
time plot is constructed by plotting the pairs (i, X,) where i represents the number of the day and X, represents
the concentration level. A time plot of this data is shown in Figure 2-12.
2.3.8.2
Plot of the Autocorrelation Function (Correlogram)
Serial correlation is a measure of the extent to which successive observations are related.
If successive observations are related, either the data must be transformed or this relationship
must be accounted for in the analysis of the data. The correlogram is a plot that is used to display
serial correlation when the data are collected at equally spaced time intervals. The autocorrelation
function is a summary of the serial correlations of data. The 1st autocorrelation coefficient (r^ is
the correlation between points that are 1 time unit (kj) apart; the 2nd autocorrelation coefficient
(r2) is the correlation between points that are 2 time units (k2) apart; etc. A correlogram (Figure
2-13) is a plot of the sample autocorrelation coefficients in which the values of k versus the values
of rk are displayed. Directions for constructing a correlogram are contained in Box 2-24; example
calculations are contained in Box 2-25. For large sample sizes, a correlogram is tedious to
EPA QA/G-9
QAOO Version
2-33
Final
July 2000
-------
construct by hand; therefore, software like Data Quality Evaluation Statistical Tools
(DataQUEST) (G-9D) should be used.
1
0.75
_ 0.5
tr
0.25
0
-0.25
-0.5
-*
-
*
*. ^
^ ^ ^ ^
* ** ^
** * '* * **
*
0 5 10 15 20 25 30
k
Figure 2-13. Example of a Correlogram
Box 2-24: Directions for Constructing a Correlogram
Let X.,, X2, ..., Xn represent the data points ordered by time for equally spaced time points, i.e., X., was collected
at time 1, X2 was collected at time 2, and so on. To construct a Correlogram, first compute the sample
autocorrelation coefficients. So for k = 0, 1, ..., compute rk where
= - and 8k = T,(Xt-X)(Xt.k-X).
t=k+
Once the rk have been computed, a Correlogram is the graph (k, rk) for k = 0, 1, . . . , and so on. As a
approximation, compute up to approximately k = n/6. Also, note that r0 = 1. Finally, place horizontal lines at
±2//R.
The correlogram is used for modeling time series data and may be used to determine if
serial correlation is large enough to create problems in the analysis of temporal data using other
methodologies besides formal time series methodologies. A quick method for determining if serial
correlation is large is to place horizontal lines at ±2//n on the correlogram (shown as dashed lines
on Figure 2-13). Autocorrelation coefficients that exceed this value require further investigation.
In application, the correlogram is only useful for data at equally spaced intervals. To relax
this restriction, a variogram may be used instead. The variogram displays the same information as
a correlogram except that the data may be based on unequally spaced time intervals. For more
information on the construction and uses of the variogram, consult a statistician.
EPA QA/G-9
QAOO Version
2-34
Final
July 2000
-------
2.3.8.3 Multiple Observations Per Time Period
Sometimes in environmental data collection, multiple observations are taken for each time
period. For example, the data collection design may specify collecting and analyzing 5 samples
from a drinking well every Wednesday for three months. If this is the case, the time plot
described in Section 2.3.8.1. may be used to display the data, display the mean weekly level,
display a confidence interval for each mean, or display a confidence interval for each mean with
the individual data values. A time plot of all the data will allow the analyst to determine if the
variability for the different collection periods varies. A time plot of the means will allow the
analyst to determine if the means may possibly be changing between the collection periods. In
addition, each collection period may be treated as a distinct variable and the methods described in
Section 2.3.7 may be applied.
Box 2-25: Example Calculations for Generating a Correlogram
A correlogram will be constructed using the following four hourly data points: hour 1: 4.5, hour 2: 3.5, hour 3:
2.5, and hour 4: 1.5. Only four data points are used so that all computations may be shown. Therefore, the
idea that no more than n/6 autocorrelation coefficients should be computed will be broken for illustrative
purposes. The first step to constructing a correlogram is to compute the sample mean (Box 2-2) which is 3 for
the 4 points. Then,
_ \ A, _ V\(M - -\!\ = = v ' v ' v ' v ' = 1 9S
g°" gy<-y>(y<-°-y> 4 4
(v3-3)(y2-3) + (v4-3)(y3-3)
= (3.5-3)(4.5-3) +(2.5-3)(3.5-3) +(1.5-3X2.5-3) = 1.25 =
0>4-3)(y2-3)
(2.5-3)(4.5-3) + (1.5-3)(3.5-3) _ -1.5
= ^-^- = (1.5-3)(4.5-3) = -2.25 =
EPA QA/G-9 Final
QAOO Version 2-35 July 2000
-------
Box 2-25: Example Calculations for Generating a Correlogram - Continued
So r, = —- =
0.3125
1.25
= 0.25 , r? = ±1 =
-.375
1.25
= -0.3 , and
-0.5625
1.25
= -0.45.
Remember r0 = 1. Thus, the correlogram of these data is a plot of (0, 1) (1, 0.25), (2, -0.3) and (3, -0.45) with
two horizontal lines at±2//4 (±1). This graph is shown below.
In this case, it appears that the observations are not serially correlated because all of the correlogram points
are within the bounds of ±2//4 (±1.0). In Figure 2-13, if k represents months, then the correlogram shows a
yearly correlation between data points since the points at k=12 and k=24 are out of the bounds of ±2//n. This
correlation will need to be accounted for when the data are analyzed.
1
0.8
0.6
0.4
0.2
o
-0.2
-0.4
-0.6
-0.8
-1
\i/
+ 1.0
-1.0
1 2
k
(Hours)
2.3.9 Plots for Spatial Data
The graphical representations of the preceding sections may be useful for exploring spatial
data. However, an analyst examining spatial data may be interested in the location of extreme
values, overall spatial trends, and the degree of continuity among neighboring locations.
Graphical representations for spatial data include postings, symbol plots, correlograms, h-scatter
plots, and contour plots.
EPA QA/G-9
QAOO Version
2-36
Final
July 2000
-------
The graphical representations presented in this section are recommended for all spatial
data regardless of whether or not geostatistical methods will be used to analyze the data. The
graphical representations described below will help identify spatial patterns that need to be
accounted for in the analysis of the data. If the analyst uses geostatistical methods such as kriging
to analyze the data, the graphical representations presented below will play an important role in
geostatistical analysis.
2.3.9.1
Posting Plots
A posting plot (Figure 2-14) is a map of data locations along with corresponding data
values. Data posting may reveal obvious errors in data location and identify data values that may
be in error. The graph of the sampling locations gives the analyst an idea of how the data were
collected (i.e., the sampling design), areas that may have been inaccessible, and areas of special
interest to the decision maker which may have been heavily sampled. It is often useful to mark
the highest and lowest values of the data to see if there are any obvious trends. If all of the
highest concentrations fall in one region of the plot, the analyst may consider some method such
as post-stratifying the data (stratification after the data are collected and analyzed) to account for
this fact in the analysis. Directions for generating a posting of the data (a posting plot) are
contained in Box 2-26.
4.0 11.6 14.9 17.4 17.7 12.4
28.6 7.7 15.2 35.5 14.7 16.5
14.7 10.9 12.4 22.8 19.1
10.2 5.2 4.9 17.2
2.3.9.2
Figure 2-14. Example of a Posting Plot
Symbol Plots
For large amounts of data, a posting plot may not be feasible and a symbol plot (Figure 2-
15) may be used. A symbol plot is basically the same as a posting plot of the data, except that
instead of posting individual data values, symbols are posted for ranges of the data values. For
example, the symbol '0' could represent all concentration levels less than 100 ppm, the symbol T
could represent all concentration levels between 100 ppm and 200 ppm, etc. Directions for
generating a symbol plot are contained in Box 2-26.
EPA QA/G-9
QAOO Version
2-37
Final
July 2000
-------
022
5 1 3
3 3
7 2
2 4
1 0
Figure 2-15. Example of a Symbol Plot
Box 2-26: Directions for Generating a Posting Plot and a Symbol Plot
with an Example
On a map of the site, plot the location of each sample. At each location, either indicate the value of the
data point (a posting plot) or indicate by an appropriate symbol (a symbol plot) the data range within
which the value of the data point falls for that location, using one unique symbol per data range.
Example: The spatial data displayed in the table below contains both a location (Northing and Easting)
and a concentration level ([c]). The data range from 4.0 to 35.5 so units of 5 were chosen to group the
data:
Range Symbol Range Symbol
0.0- 4.9 0 20.0-24.9 4
5.0- 9.9 1 25.0-29.9 5
10.0-14.9 2 30.0-34.9 6
15.0-19.9 3 35.0-39.9 7
The data values with corresponding symbols then become:
Northing Easting fcl Symbol Northing Easting fcl
Symbol
25.0
25.0
25.0
25.0
20.0
20.0
20.0
20.0
15.0
15.0
15.0
The posting plot of this data
0.0
5.0
10.0
15.0
0.0
5.0
10.0
15.0
0.0
5.0
10.0
4.0
11.6
14.9
17.4
17.7
12.4
28.6
7.7
15.2
35.5
14.7
is displayed in
0
2
2
3
3
2
5
1
3
7
2
Figure 2-14
15.0
15.0
10.0
10.0
10.0
5.0
5.0
5.0
5.0
0.0
0.0
and the symbol
15.0
0.0
5.0
10.0
15.0
0.0
5.0
10.0
15.0
5.0
15.0
plot is
16.5
8.9
14.7
10.9
12.4
22.8
19.1
10.2
5.2
4.9
17.2
displayed
3
1
2
2
2
4
3
2
1
0
3
in Figure 2-15.
EPA QA/G-9
QAOO Version
2-38
Final
July 2000
-------
2.3.9.3 Other Spatial Graphical Representations
The two plots described in Sections 2.3.9.1 and 2.3.9.2 provide information on the
location of extreme values and spatial trends. The graphs below provide another item of interest
to the data analyst, continuity of the spatial data. The graphical representations are not described
in detail because they are used more for preliminary geostatistical analysis. These graphical
representations can be difficult to develop and interpret. For more information on these
representations, consult a statistician.
An h-scatterplot is a plot of all possible pairs of data whose locations are separated by a
fixed distance in a fixed direction (indexed by h). For example, a h-scatter plot could be based on
all the pairs whose locations are 1 meter apart in a southerly direction. A h-scatter plot is similar
in appearance to a scatter plot (Section 2.3.7.2). The shape of the spread of the data in a h-
scatter plot indicates the degree of continuity among data values a certain distance apart in
particular direction. If all the plotted values fall close to a fixed line, then the data values at
locations separated by a fixed distance in a fixed location are very similar. As data values become
less and less similar, the spread of the data around the fixed line increases outward. The data
analyst may construct several h-scatter plots with different distances to evaluate the change in
continuity in a fixed direction.
A correlogram is a plot of the correlations of the h-scatter plots. Because the h-scatter
plot only displays the correlation between the pairs of data whose locations are separated by a
fixed distance in a fixed direction, it is useful to have a graphical representation of how these
correlations change for different separation distances in a fixed direction. The correlogram is such
a plot which allows the analyst to evaluate the change in continuity in a fixed direction as a
function of the distance between two points. A spatial correlogram is similar in appearance to a
temporal correlogram (Section 2.3.8.2). The correlogram spans opposite directions so that the
correlogram with a fixed distance of due north is identical to the correlogram with a fixed distance
of due south.
Contour plots are used to reveal overall spatial trends in the data by interpolating data
values between sample locations. Most contour procedures depend on the density of the grid
covering the sampling area (higher density grids usually provide more information than lower
densities). A contour plot gives one of the best overall pictures of the important spatial features.
However, contouring often requires that the actual fluctuations in the data values are smoothed so
that many spatial features of the data may not be visible. The contour map should be used with
other graphical representations of the data and requires expert judgement to adequately interpret
the findings.
2.4 Probability Distributions
2.4.1 The Normal Distribution
Data, especially measurements, occur in natural patterns that can be considered to be a
distribution of values. In most instances the data values will be grouped around some measure of
EPA QA/G-9 Final
QAOO Version 2-39 July 2000
-------
control tendency such as the mean or median. The spread of the data (as determined by the sum
of the squared distances from data point to the mean) is called the variance (the square root of
this is called the standard deviation). A distribution with a large variance will be more spread out
than one with a small variance (Figure 2-16). When the data values fall in a systematic pattern
around the mean and then taper off rapidly to the tails, it is often a normal distribution or bell-
shaped curve.
Figure 2-16. The Normal
Distribution
Figure 2-17. The Standard
Normal Curve (Z-Curve)
The characteristics of a normal distribution are well known mathematically and when
referred to, usually written as "data are distributed N (//, a2)" where the first characteristic is the
mean (//) and the second, the variance (a2). It may be shown that any normal distribution can be
transformed to a standard normal distribution, N(0,l), and this standard normal referred to as
simply Z (Figure 2-17). It is frequently necessary to refer to the percentiles of a standard normal
and in this guidance document, the subscript to a quoted Z-value will denote the percentile (or
area under the curve, cumulative from the left), see Figure 2-17.
2.4.2 The t-Distribution
The standard normal curve is used when exact information on the mean and variance are
available, but when only estimates from a sample are available, a different distribution applies.
When only information from a random sample on sample mean and sample variance is known for
decision making purposes, a Student's t distribution is appropriate. It resembles a standard
normal but is lower in the center and fatter in the tails. The degree of fatness in the tails is a
function of the degrees of freedom available, which in turn is related to sample size.
2.4.3 The Lognormal Distribution
A commonly met distribution in environmental work is
the lognormal distribution which has a more skewed (lopsided)
shape than a normal, see Figure 2-18. The lognormal is
bounded by zero and has a fatter tail than the normal. It is
related to the normal by the simple relationship: if X is
distributed lognormally, then Y = In (X) is distributed normally.
It is common practice to transform data (and any standard
being tested against) to achieve approximate normality prior to
conducting statistical tests.
Figure 2-18. Three Different
Lognormal Distributions
EPA QA/G-9
QAOO Version
2-40
Final
July 2000
-------
2.4.4 Central Limit Theorem
In nearly all estimation situations in environmental work, the focus of the investigation
centers on the mean of a random sample of observations or measurements. It is rare that true
normality of the observations can be assumed and therefore a question as whether to use
statistical tests based on normality may be considered.
In many cases, the normally-based statistical tests are not overly affected by the lack of
normality as tests are very robust (sturdy) and perform tolerably well unless gross non-normality
is present. In addition, the tests become increasingly tolerant of deviations from normality as the
number of individual samples constituting the sample mean increases. In simple terms, as the size
of the sample increases, the mean of that sample acts increasingly as if it came from a normal
distribution regardless of the true distribution of the individual values. The taking of large
samples "stabilizes" the mean, so it then acts as if normality was present and the statistical test
remain valid.
EPA QA/G-9 Final
QAOO Version 2-41 July 2000
-------
This page is intentionally blank.
EPA QA/G-9 Final
QAOO Version 2 - 42 July 2000
-------
CHAPTER 3
STEP 3: SELECT THE STATISTICAL TEST
THE DATA QUALITY ASSESSMENT PROCESS
Review DQOs and Sampling Design
Conduct Preliminary Data Review
Select the Statistical Test
Verify the Assumptions
Draw Conclusions From the Data
SELECT THE STATISTICAL TEST
Purpose
Select an appropriate procedure for analyzing
data based on the preliminary data review.
Activities
• Select Statistical Hypothesis Test
• Identify Assumptions Underlying Test
Tools
• Hypothesis tests for a single population
• Hypothesis tests for comparing two populations
Step 3: Select the Statistical Test
Select the statistical hypothesis test based on the data user's objectives and the results of the
preliminary data review.
P If the problem involves comparing study results to a fixed threshold, such as a regulatory
standard, consider the hypothesis tests in Section 3.2.
P If the problem involves comparing two populations, such as comparing data from two
different locations or processes, then consider the hypothesis tests in Section 3.3.
Identify the assumptions underlying the statistical test.
P List the key underlying assumptions of the statistical hypothesis test, such as distributional
form, dispersion, independence, or others as applicable.
P Note any sensitive assumptions where relatively small deviations could jeopardize the
validity of the test results.
EPA QA/G-9
QAOO Version
3- 1
Final
July 2000
-------
List of Boxes
Page
Box 3-1: Directions for a One-Sample t-Test 3-7
Box 3-2: An Example of a One-Sample t-Test 3-8
Box 3-3: Directions for a One-Sample t-Test for a Stratified Random Sample 3-9
Box 3-4: An Example of a One-Sample t-Test for a Stratified Random Sample 3-10
Box 3-5: Directions for the Wilcoxon Signed Rank Test 3-12
Box 3-6: An Example of the Wilcoxon Signed Rank Test 3-13
Box 3-7: Directions for the Large Sample Approximation to the Wilcoxon
Signed Rank Test 3-14
Box 3-8: Directions for the Chen Test 3-16
Box 3-9: Example of the Chen Test 3-17
Box 3-10: Directions for the One-Sample Test for Proportions 3-19
Box 3-11: An Example of the One-Sample Test for Proportions 3-20
Box 3-12: Directions for a Confidence Interval for a Mean 3-21
Box 3-13: An Example of a Confidence Interval for a Mean 3-21
Box 3-14: Directions for the Student's Two-Sample t-Test (Equal Variances) 3-24
Box 3-15: An Example of a Student's Two-Sample t-Test (Equal Variances 3-25
Box 3-16: Directions for Satterthwaite's t-Test (Unequal Variances 3-26
Box 3-17: An Example of Satterthwaite's t-Test (Unequal Variances) 3-27
Box 3-18: Directions for a Two-Sample Test for Proportions 3-29
Box 3-19: An Example of a Two-Sample Test for Proportions 3-30
Box 3-20: Directions for the Wilcoxon Rank Sum Test 3-32
Box 3-21: An Example of the Wilcoxon Rank Sum Test 3-33
Box 3-22: Directions for the Large Sample Approximation to the
Wilcoxon Rank Sum Test 3-34
Box 3-23: Directions for a Modified Quantile Test 3-36
Box 3-24: A Example of a Modified Quantile Test 3-37
Box 3-25: Directions for Dunnett's Test for Simple Random and Systematic Samples .... 3-39
Box 3-26: An Example of Dunnett's Test for Simple Random and Systematic Samples ... 3-40
EPA QA/G-9 Final
QAOO Version 3 - 2 July 2000
-------
CHAPTER 3
STEP 3: SELECT THE STATISTICAL TEST
3.1 OVERVIEW AND ACTIVITIES
This chapter provides information that the analyst can use in selecting an appropriate
statistical hypothesis test that will be used to draw conclusions from the data. A brief review of
hypothesis testing is contained in Chapter 1. There are two important outputs from this step: (1)
the test itself, and (2) the assumptions underlying the test that determine the validity of
conclusions drawn from the test results.
This section describes the two primary activities in this step of DQA. The remaining
sections in this chapter contain statistical tests that may be useful for analyzing environmental
data. In the one-sample tests discussed in Section 3.2, data from a population are compared with
an absolute criterion such as a regulatory threshold or action level. In the two-sample tests
discussed in Section 3.3, data from a population are compared with data from another population
(for example, an area expected to be contaminated might be compared with a background area).
For each statistical test, this chapter presents its purpose, assumptions, limitations, robustness,
and the sequence of steps required to apply the test.
The directions for each hypothesis test given in this chapter are for simple random
sampling and randomized systematic sampling designs, except where noted otherwise. If a more
complex design is used (such as a stratified design or a composite random sampling design) then
different formulas are needed, some of which are contained in this chapter.
3.1.1 Select Statistical Hypothesis Test
If a particular test has been specified either in the DQO Process, the QA Project Plan, or
by the particular program or study, the analyst should use the results of the preliminary data
review to determine if this statistical test is legitimate for the data collected. If the test is not
legitimate, the analyst should document why this particular statistical test should not be applied to
the data and then select a different test, possibly after consultation with the decision maker. If a
particular test has not been specified, the analyst should select a statistical test based on the data
user's objectives, preliminary data review, and likely viable assumptions.
3.1.2 Identify Assumptions Underlying the Statistical Test
All statistical tests make assumptions about the data. Parametric tests assume the data
have some distributional form (e.g., the t-test assumes normal distribution), whereas
nonparametric tests do not make this assumption (e.g., the Wilcoxon test only assumes the data
are symmetric but not necessarily normal). However, both parametric and nonparametric tests
may assume that the data are statistically independent or that there are no trends in the data.
While examining the data, the analyst should always list the underlying assumptions of the
statistical hypothesis test, such as distribution, dispersion, or others as applicable.
EPA QA/G-9 Final
QAOO Version 3 - 3 July 2000
-------
Another important feature of statistical tests is their sensitivities (nonrobustness) to
departures from the assumptions. A statistical procedure is called robust if its performance is not
seriously affected by moderate deviations from its underlying assumptions. The analyst should
note any sensitive assumptions where relatively small deviations could jeopardize the validity of
the test results.
3.2 TESTS OF HYPOTHESES ABOUT A SINGLE POPULATION
A one-sample test involves the comparison of a population parameter (e.g., a mean,
percentile, or variance) to a threshold value. Both the threshold value and the population
parameter were specified during Step 1: Review DQOs and Sampling Design. In a one-sample
test, the threshold value is a fixed number that does not vary. If the threshold value was estimated
(and therefore contains variability), a one-sample test is not appropriate. An example of a one-
sample test would be to determine if 95% of all companies emitting sulfur dioxide into the air are
below a fixed discharge level. For this example, the population parameter is a percentage
(proportion) and the threshold value is 95% (.95). Another example is a common Superfund
problem that involves comparing the mean contaminant concentration to a risk-based standard. In
this case, the risk-based standard (which is fixed) is the threshold value and the statistical
parameter is the true mean contaminant concentration level of the site. However, comparing the
mean concentration in an area to the mean concentration of a reference area (background) would
not be a one-sample test because the mean concentration in the reference area would need to be
estimated.
The statistical tests discussed in this section may be used to determine if 6 < 60 or 6 > 60,
where 6 represents either the population mean, median, a percentile, or a proportion and 60
represents the threshold value. Section 3.2.1 discusses tests concerning the population mean,
Section 3.2.2 discusses tests concerning a proportion or percentile, and Section 3.2.2 discusses
tests for a median.
3.2.1 Tests for a Mean
A population mean is a measure of the center of the population distribution. It is one of
the most commonly used population parameters in statistical hypothesis testing because its
distribution is well known for large sample sizes. The hypotheses considered in this section are:
Case 1: H0: ji < C vs. HA: |i> C; and
Case 2: H0: |i > C vs. HA: |i < C
where C represents a given threshold such as a regulatory level, and |i denotes the (true) mean
contaminant level for the population. For example, C may represent the arsenic concentration
level of concern. Then if the mean of the population exceeds C, the data user may wish to take
action.
EPA QA/G-9 Final
QAOO Version 3 - 4 July 2000
-------
The information required for this test (defined in Step 1) includes the null and alternative
hypotheses (either Case 1 or Case 2); the gray region, i.e., a value \il > C for Case 1 or a value \il
< C for Case 2 representing the bound of the gray region; the false rejection error rate a at C; the
false acceptance error rate P at nt; and any additional limits on decision errors. It may be helpful
to label any additional false rejection error limits as a2 at C2, a3 at C3, etc., and to label any
additional false acceptance error limits as P2 at |i2, P3 at |i3, etc. For example, consider the
following decision: determine whether the mean contaminant level at a waste site is greater than
10 ppm. The null hypothesis is H0: |i > 10 ppm and the alternative hypothesis is HA: |i < 10
ppm. A gray region has been set from 10 to 8 ppm, a false rejection error rate of 5% has been set
at 10 ppm, and a false acceptance error rate of 10% has been set at 8 ppm. Thus, C = 10 ppm, \il
= 8 ppm, a = 0.05, and P = 0.1. If an additional false acceptance error rate was set, for example,
an error rate of 1% at 4 ppm, then P2 = .01 and |i2 = 4 ppm.
3.2.1.1 The One-Sample t-Test
PURPOSE
Given a random sample of size n (or a composite sample of size n, each composite
consisting of k aliquots), the one-sample t-test can be used to test hypotheses involving the mean
(|i) of the population from which the sample was selected.
ASSUMPTIONS AND THEIR VERIFICATION
The primary assumptions required for validity of the one-sample t-test are that of a
random sample (independence of the data values) and that the sample mean x is approximately
normally distributed. Because the sample mean and standard deviation are very sensitive to
outliers, the t-test should be preceded by a test for outliers (see Section 4.4).
Approximate normality of the sample mean follows from approximate normality of the
data values. In addition, the Central Limit Theorem states that the sample mean of a random
sample from a population with an unknown distribution will be approximately normally distributed
provided the sample size is large. This means that although the population distribution from
which the data are drawn can be distinctly different from the normal distribution, the distribution
of the sample mean can still be approximately normal when the sample size is relatively large.
Although preliminary tests for normality of the data can and should be done for small sample
sizes, the conclusion that the sample does not follow a normal distribution does not automatically
invalidate the t-test, which is robust to moderate violations of the assumption of normality for
large sample sizes.
LIMITATIONS AND ROBUSTNESS
The t-test is not robust to outliers because the sample mean and standard deviation are
influenced greatly by outliers. The Wilcoxon signed rank test (see Section 3.2.1.2) is more
EPA QA/G-9 Final
QAOO Version 3 - 5 July 2000
-------
robust, but is slightly less powerful. This means that the Wilcoxon signed rank test is slightly less
likely to reject the null hypothesis when it is false than the t-test.
The t-test has difficulty dealing with less-than values, e.g., values below the detection
limit, compared with tests based on ranks or proportions. Tests based on a proportion above a
given threshold (Section 3.2.2) are more valid in such a case, if the threshold is above the
detection limit. It is also possible to substitute values for below detection-level data (e.g., /^ the
detection level) or to adjust the statistical quantities to account for nondetects (e.g., Cohen's
Method for normally or lognormally distributed data). See Chapter 4 for more information on
dealing with data that are below the detection level.
SEQUENCE OF STEPS
Directions for a one-sample t-test for a simple, systematic, and composite random samples
are given in Box 3-1 and an example is given in Box 3-2. Directions for a one-sample t-test for a
stratified random sample are given in Box 3-3 and an example is given in Box 3-4.
EPA QA/G-9 Final
QAOO Version 3 - 6 July 2000
-------
Box 3-1: Directions for a One-Sample t-Test
for Simple and Systematic Random Samples
with or without Compositing
Let X.,, X2, . . . , Xn represent the n data points. These could be either n individual samples or n composite
samples consisting of k aliquots each. These are the steps for a one-sample t-test for Case 1 (H0: u < C);
modifications for Case 2 (H0: u > C) are given in braces {}.
STEP 1: Calculate the sample mean, x (Section 2.2.2), and the standard deviation, s (Section 2.2.3).
STEP 2: Use Table A-1 of Appendix A to find the critical value t.,_,., such that 100(1-a)% of the t distribution with
n -1 degrees of freedom is below !,_„. For example, if a = 0.05 and n = 16, then n-1 =15 and !,_„ =
1.753.
STEP 3: Calculate the sample value t = (X- C) I (s/\fn) .
STEP 4: Compare t with t.,_a.
1) If t > ti-a {t < -t-i-J, the null hypothesis may be rejected. Go to Step 6.
2) If t > t.,_a {t < -t^J, there is not enough evidence to reject the null hypothesis and the false
acceptance error rate should be verified. Go to Step 5.
STEP 5: As the null hypothesis (H0) was not rejected, calculate either the power of the test or the sample size
necessary to achieve the false rejection and false acceptance error rates. To calculate the power,
assume that the true values for the mean and standard deviation are those obtained in the sample
and use a software package like the Decision Error Feasibility Trial (DEFT) software (EPA, 1994) to
generate the power curve of the test.
If only one false acceptance error rate (P) has been specified (at u^, it is possible to calculate the
sample size which achieves the DQOs, assuming the true mean and standard deviation are equal to
the values estimated from the sample, instead of calculating the power of the test. To do this,
s 2(z +z )2
calculates = — + (0.5)zj where zp is the pth percentile of the standard normal
(nrQ2
distribution (Table A-1 of Appendix A). Round m up to the next integer. If m < n, the false
acceptance error rate has been satisfied. If m > n, the false acceptance error rate has not been
satisfied.
STEP 6: The results of the test may be:
1) the null hypothesis was rejected and it seems that the true mean is greater than C {less than C};
2) the null hypothesis was not rejected and the false acceptance error rate was satisfied and it seems
that the true mean is less than C {greater than C}; or
3) the null hypothesis was not rejected and the false acceptance error rate was not satisfied and it
seems that the true mean is less than C {greater than C} but conclusions are uncertain since the
sample size was too small.
Report the results of the test, the sample size, sample mean, standard deviation, t and t1_a.
Note: The calculations for the t-test are the same for both simple random or composite random sampling. The
use of compositing will usually result in a smaller value of "s" than simple random sampling.
EPA QA/G-9 Final
QAOO Version 3 - 7 July 2000
-------
Box 3-2: An Example of a One-Sample t-Test
for a Simple Random or Composite Sample
Consider the following 9 random (or composite samples each of k aliquots) data points: 82.39 ppm,
103.46 ppm, 104.93 ppm, 105.52 ppm, 98.37 ppm, 113.23 ppm, 86.62 ppm, 91.72 ppm, and 108.21
ppm. This data will be used to test the hypothesis: H0: u < 95 ppm vs. HA: u > 95 ppm. The decision
maker has specified a 5% false rejection decision error limit (a) at 95 ppm (C), and a 20% false
acceptance decision error limit (P) at 105 ppm (m).
STEP 1 : Using the directions in Box 2-2 and Box 2-3, it was found that
X = 99.38 ppm and s = 10.41 ppm.
STEP 2: Using Table A-1 of Appendix A, the critical value of the t distribution with 8 degrees of freedom
is t0 95 =1.86.
STEP3: t=L-C= 99-38 - 95 = 1.26
s/fi 10.41/^9
STEP 4: Because 1.26 > 1.86, there is not enough evidence to reject the null hypothesis and the false
acceptance error rate should be verified.
STEP 5: Because there is only one false acceptance error rate, it is possible to use the sample size
formula to determine if the error rate has been satisfied. Therefore,
- - * «"*•-•
0.842)2 +
(95 - 105)2
i.e., 9
Notice that it is customary to round upwards when computing a sample size. Since m=n, the
false acceptance error rate has been satisfied.
STEP 6: The results of the hypothesis test were that the null hypothesis was not rejected but the false
acceptance error rate was satisfied. Therefore, it seems that the true mean is less than 95
ppm.
EPA QA/G-9 Final
QAOO Version 3 - 8 July 2000
-------
Box 3-3: Directions for a One-Sample t-Test
for a Stratified Random Sample
Let h=1 , 2, 3, . . . , L represent the L strata and nh represent the sample size of stratum h. These steps are for a
one-sample t-test for Case 1 (H0: u < C); modifications for Case 2 (H0: u > C) are given in braces { }.
STEP 1: Calculate the stratum weights (Wh) by calculating the proportion of the volume in
V,
stratum h, Wh = - where Vh is the surface area of stratum h multiplied by
h=l
the depth of sampling in stratum h.
STEP 2: For each stratum, calculate the sample stratum mean Xh = and the sample stratum
standard error sh =
,,2
STEPS: Calculate overall mean XST = 2_/V^/i' ar|d variance SST = /]Vh—.
A=i A=I nh
2
STEP 4: Calculate the degrees of freedom (dof): dof =
L Wist
£-
Use Table A-1 of Appendix A to find the critical value t .,_,., so that 100(1-a)% of the t
distribution with the above degrees of freedom (rounded to the next highest integer) is
below t.,_a.
-A. OT-T ~~ \_,
STEP 5: Calculate the sample value: t =
STEP 6: Compare t to t.,_a. If t > t.|_a {t < -t^J, the null hypothesis may be rejected. Go to Step 8. If t > t^,, {t < -
t.,.,.,}, there is not enough evidence to reject the null hypothesis and the false acceptance error rate
should be verified. Go to Step 7.
STEP 7: If the null hypothesis was not rejected, calculate either the power of the test or the sample size
necessary to achieve the false rejection and false acceptance error rates (see Step 5, Box 3-2).
STEP 8: The results of the test may be: 1) the null hypothesis was rejected so it seems that the true mean is
less than C {greater than C}; 2) the null hypothesis was not rejected and the false acceptance error
rate was satisfied and it seems that the true mean is greater than C {less than C}; or 3) the null
hypothesis was not rejected and the false acceptance error rate was not satisfied and it seems that
the true mean is greater than C {less than C} but conclusions are uncertain since the sample size was
too small.
Report the results of the test, as well as the sample size, sample mean, and sample standard
deviation for each stratum, the estimated t, the dof, and !.,_„.
EPA QA/G-9 Final
QAOO Version 3 - 9 July 2000
-------
Box 3-4: An Example of a One-Sample t-Test
for a Stratified Random Sample
Consider a stratified sample consisting of two strata where stratum 1 comprises 10% of the total site
surface area and stratum 2 comprises the other 90%, and 40 samples were collected from stratum 1, and
60 samples were collected from stratum 2. For stratum 1, the sample mean is 23 ppm and the sample
standard deviation is 18.2 ppm. For stratum 2, the sample mean is 35 ppm, and the sample standard
deviation is 20.5 ppm. This information will be used to test the null hypothesis that the overall site mean
is greater than or equal to 40 ppm, i.e., H0: u > 40 ppm (Case 2). The decision maker has specified a
1% false rejection decision limit (a) at 40 ppm and a 20% false acceptance decision error limit (P) at 35
STEP 1: W., = 10/100 = 0.10, W2 = 90/100 = 0.9.
STEP 2: From above, x., = 23 ppm, X2 = 35 ppm, s., = 18.2, and s2 = 20.5. This information was
developed using the equations in step 2 of Box 3-3.
STEP 3: The estimated overall mean concentration is:
_ L _ _ _
XST = YFhXh = W\X\ + W2X2 = (-!X23) + (.9)(35) = 33.8 ppm.
h=l
and the estimated overall variance is:
L 2 T T T T
vv 2 sh (.1) (18.2) (.9) (20.5)
SVT = / "\i — = "~ :— + "~ :— = 5.76
w hTf nh 40 60
STEP 4: The approximate degrees of freedom (dof) is:
(O2 (5 76)2
dof = -±i- = ^-^. = 60.8, i.e.,
.1)4(18.2)4 + (.9)4(20.5)4
(40)239 (60)259
v ' v '
~ 2 2
61
Note how the degrees of freedom has been rounded up to a whole number.
Using Table A-1 of Appendix A, the critical value i^_a of the t distribution with
61 dof is approximately 2.39.
X?T ~ C QQ « _ AH
STEPS: Calculate the sample value t = — - = = -2.58
STEP 6: Because -2.58 < -2.39 the null hypothesis may be rejected.
STEP 7: Because the null hypothesis was rejected, it is concluded that the mean is probably less than
40 ppm. In this example there is no need to calculate the false acceptance rate as the null
hypothesis was rejected and so the chance of making a false acceptance error is zero by
definition.
EPA QA/G-9 Final
QAOO Version 3-10 July 2000
-------
3.2.1.2 The Wilcoxon Signed Rank (One-Sample) Test for the Mean
PURPOSE
Given a random sample of size n (or composite sample size n, each composite consisting
of k aliquots), the Wilcoxon signed rank test can be used to test hypotheses regarding the
population mean or median of the population from which the sample was selected.
ASSUMPTIONS AND THEIR VERIFICATION
The Wilcoxon signed rank test assumes that the data constitute a random sample from a
symmetric continuous population. (Symmetric means that the underlying population frequency
curve is symmetric about its mean/median.) Symmetry is a less stringent assumption than
normality since all normal distributions are symmetric, but some symmetric distributions are not
normal. The mean and median are equal for a symmetric distribution, so the null hypothesis can
be stated in terms of either parameter. Tests for symmetry can be devised which are based on the
chi-squared distribution, or a test for normality may be used. If the data are not symmetric, it may
be possible to transform the data so that this assumption is satisfied. See Chapter 4 for more
information on transformations and tests for symmetry.
LIMITATIONS AND ROBUSTNESS
Although symmetry is a weaker assumption than normality, it is nonetheless a strong
assumption. If the data are not approximately symmetric, this test should not be used. For large
sample sizes (n > 50), the t-test is more robust to violations of its assumptions than the Wilcoxon
signed rank test. For small sample sizes, if the data are not approximately symmetric and are not
normally distributed, this guidance recommends consulting a statistician before selecting a
statistical test or changing the population parameter to the median and applying a different
statistical test (Section 3.2.3).
The Wilcoxon signed rank test may produce misleading results if many data values are the
same. When values are the same, their relative ranks are the same, and this has the effect of
diluting the statistical power of the Wilcoxon test. Box 3-5 demonstrates the method used to
break tied ranks. If possible, results should be recorded with sufficient accuracy so that a large
number of equal values do not occur. Estimated concentrations should be reported for data
below the detection limit, even if these estimates are negative, as their relative magnitude to the
rest of the data is of importance.
SEQUENCE OF STEPS
Directions for the Wilcoxon signed rank test for a simple random sample and a systematic
simple random sample are given in Box 3-5 and an example is given in Box 3-6 for samples sizes
smaller than 20. For sample sizes greater than 20, the large sample approximation to the
Wilcoxon Signed Rank Test should be used. Directions for this test are given in Box 3-7.
EPA QA/G-9 Final
QAOO Version 3-11 July 2000
-------
Box 3-5: Directions for the Wilcoxon Signed Rank Test
for Simple and Systematic Random Samples
Let X.,, X2, . . . , Xn represent the n data points. The following describes the steps for applying the
Wilcoxon signed rank test for a sample size (n) less than 20 for Case 1 (H0: u < C); modifications for
Case 2 (H0: u > C) are given in braces {}. If the sample size is greater than or equal to 20, use Box 3-7.
STEP 1: If possible, assign values to any measurements below the detection limit. If
this is not possible, assign the value "Detection Limit divided by 2" to each
value. Then subtract each observation X, from C to obtain the deviations d, =
C - X,. If any of the deviations are zero delete them and correspondingly
reduce the sample size n.
STEP 2: Assign ranks from 1 to n based on ordering the absolute deviations |dj (i.e.,
magnitude of differences ignoring the sign) from smallest to largest. The rank
1 is assigned to the smallest value, the rank 2 to the second smallest value,
and so forth. If there are ties, assign the average of the ranks which would
otherwise have been assigned to the tied observations.
STEP 3: Assign the sign for each observation to create the signed rank. The sign is
positive if the deviation d, is positive; the sign is negative if the deviation d, is
negative.
STEP 4: Calculate the sum R of the ranks with a positive sign.
STEP 5: Use Table A-6 of Appendix A to find the critical value wa.
If R < wa, {R > n(n+1)/2 - wj, the null hypothesis may be rejected; proceed to Step 7.
Otherwise, there is not enough evidence to reject the null hypothesis, and the false
acceptance error rate will need to be verified; proceed to Step 6.
STEP 6: If the null hypothesis (H0) was not rejected, calculate either the power of the
test or the sample size necessary to achieve the false rejection and false
acceptance error rates using a software package like the DEFT software (EPA
, 1994). For large sample sizes, calculate,
where zp is the pth percentile of the standard normal distribution (Table A-1 of Appendix A). If
1.16m < n, the false acceptance error rate has been satisfied.
STEP 7: The results of the test may be:
1) the null hypothesis was rejected and it seems that the true mean is greater than C {less
than C};
2) the null hypothesis was not rejected and the false acceptance error rate was satisfied and
it seems that the true mean is less than C {greater than C}; or
3) the null hypothesis was not rejected and the false acceptance error rate was not satisfied
and it seems that the true mean is greater than C {less than C} but conclusions are uncertain
since the sample size was too small.
EPA QA/G-9 Final
QAOO Version 3-12 July 2000
-------
Box 3-6: An Example of the Wilcoxon Signed Rank Test
for a Simple Random Sample
Consider the following 10 data points: 974 ppb, 1044 ppb, 1093 ppb, 897 ppb, 879 ppb, 1161 ppb, 839
ppb, 824 ppb, 796 ppb, and one observation below the detection limit of 750 ppb. This data will be used
to test the hypothesis: H0: u > 1000 ppb vs. HA: u < 1000 ppb (Case 2). The decision maker has
specified a 10% false rejection decision error limit (a) at 1000 ppb (C), and a 20% false acceptance
decision error limit (P) at 900 ppb (u^.
STEP 1:
x.
|d,|
rank
s-rank
STEP 2:
STEP 3:
STEP 4:
STEP 5:
STEP 7:
Assign the value 375 ppb (750 divided by 2) to the data point below the detection limit.
Subtract C (1 000) from each of the n observations X, to obtain the deviations d, = 1 000 - X,.
This is shown in row 2 of the table below.
974 1044 1093 897 879 1161 839 824 796 375
26 -44 -93 103 121 -161 161 176 204 625
26 44 93 103 121 161 161 176 204 625
12 2 3 4 5 6.5 6.5 8 9 10
1-2-3 45 -6.5 6.5 8 9 10
Assign ranks from 1 to n based on ordering the absolute deviations |d,| (magnitude ignoring
any negative sign) from smallest to largest. The absolute deviations are listed in row 3 of the
table above. Note that the data have been sorted (rearranged) for clarity so that the absolute
deviations are ordered from smallest to largest.
The rank 1 is assigned to the smallest value, the rank 2 to the second smallest value, and so
forth. Observations 6 and 7 are ties, therefore, the average (6+7)/2 = 6.5 will be assigned to
the two observations. The ranks are shown in row 4.
Assign the sign for each observation to create the signed rank. The sign is positive if the
deviation d, is positive; the sign is negative if the deviation d, is negative. The signed rank is
shown in row 5.
R = 1+4 + 5 + 6.5 + 8 + 9+10 = 43.5.
Table A-6 of Appendix A was used to find the critical value wa where a = 0.10. For this
example, w010 = 15. Since 43.5 > (10x11)72 - 15 = 40, the null hypothesis may be rejected.
The null hypothesis was rejected with a 10% significance level using the Wilcoxon signed rank
test. Therefore, it would seem that the true mean is below 1 000 ppb.
EPA QA/G-9 Final
QAOO Version 3-13 July 2000
-------
Box 3-7: Directions for the Large Sample Approximation to the Wilcoxon Signed Rank Test
for Simple and Systematic Random Samples
Let X.,, X2, . . . , Xn represent the n data points where n is greater than or equal to 20. The following
describes the steps for applying the large sample approximation for the Wilcoxon signed rank test for
Case 1 (H0: u < C); modifications for Case 2 (H0: u > C) are given in braces {}.
STEP 1: If possible, assign values to any measurements below the detection limit. If this is not
possible, assign the value "Detection Limit divided by 2" to each value. Then subtract each
observation X, from C to obtain the deviations d, = C - X,. If any of the deviations are zero
delete them and correspondingly reduce the sample size n.
STEP 2: Assign ranks from 1 to n based on ordering the absolute deviations |d,| (i.e., magnitude of
differences ignoring the sign) from smallest to largest. The rank 1 is assigned to the smallest
value, the rank 2 to the second smallest value, and so forth. If there are ties, assign the
average of the ranks which would otherwise have been assigned to the tied observations.
STEP 3: Assign the sign for each observation to create the signed rank. The sign is positive if the
deviation d, is positive; the sign is negative if the deviation d, is negative.
STEP 4: Calculate the sum R of the ranks with a positive sign.
n(n + 1)
STEPS: Calculate w = — + x \jn(n + l)(2n + 1)124 where p = 1-a {p = a} and zp
4 p
is the pth percentile of the standard normal distribution (Table A-1 of Appendix A).
STEP 6: If R < w {R > w}, the null hypothesis may be rejected. Go to Step 8.
Otherwise, there is not enough evidence to reject the null hypothesis, and the false
acceptance error rate will need to be verified. Go to Step 7.
STEP 7: If the null hypothesis (H0) was not rejected, calculate either the power of the test or the sample
size necessary to achieve the false rejection and false acceptance error rates using a software
package like the DEFT software (EPA , 1994). For large sample sizes, calculate,
m = 11«—i_p^_ + (Q.5)Zl2_a
(MI~ O
where zp is the pth percentile of the standard normal distribution (Table A-1 of Appendix
A). If 1.16m < n, the false acceptance error rate has been satisfied.
STEPS: The results of the test may be:
1) the null hypothesis was rejected and it seems that the true mean is greater {less} than C;
2) the null hypothesis was not rejected and the false acceptance error rate was satisfied and it
seems that the true mean is less than C {greater than C}; or
3) the null hypothesis was not rejected and the false acceptance error rate was not satisfied
and it seems that the true mean is greater than C {less than C} but conclusions are uncertain
since the sample size was too small.
Report the results of the test, the sample size, R, and w.
EPA QA/G-9 Final
QAOO Version 3-14 July 2000
-------
3.2.1.3 The Chen Test
PURPOSE
Environmental data such as concentration measurements are often confined to positive
values and appear to follow a distribution with most of the data values relatively small or near
zero but with a few relatively large data values. Underlying such data is some distribution which
is not symmetrical (like a normal) but which is skewed to the right (like a log normal). Given a
random sample of size 'n' from a right-skewed distribution, the Chen test can be used to compare
the mean (|i) of the distribution with a threshold level or regulatory value. The null hypothesis
has the form H0: ji < C, where C is a given threshold level; the alternative hypothesis is HA: ji > C.
The method is not recommended for testing null hypotheses of the form H0: |i > C against
HA:n
-------
Box 3-8: Directions for the Chen Test
LetX.,, X2, . . . , Xn represent the n data points. Let C denote the threshold level of interest. The null hypothesis
is H0: U < C and the alternative is HA: |J > C, the level of significance is a.
STEP 1: If at most 15% of the data points are below the detection limit (DL) and C is much larger than the
DL, then replace values below the DL with DL/2.
STEP 2: Visually check the assumption of right-skewness by inspecting a histogram or frequency plot for the
data. (The Chen test is appropriate only for data which are skewed to the right.)
STEP 3: Calculate the sample mean, x (Section 2.2.2), and the standard deviation, s (Section 2.2.3).
STEP 4: Calculate the sample skewness b = — — - ' the quantity a = b/6i/n , the t-statistic
(n - V)(n - 2>3 '
t = - — ^ — - , and then compute T = t + a(l + 2t2 ) + 4a2 (t + 2t3 ).
NOTE: The skewness b should be greater than 1 to confirm the data are skewed to the right.
STEP 5: Use the last row of Table A-1 in Appendix A to find the critical value z.|_a such that 100(1-a)% of the
Normal distribution is below !.,_„. For example, if a = 0.05 then z1_a = 1 .645.
STEP 6: Compare t with z.,_a.
1 ) If t > z.,.,, the null hypothesis may be rejected and it seems that the true mean is greater than C;
2) If t > z1_a, there is not enough evidence to reject the null hypothesis so it seems that the true mean
is less than C.
3.2.2 Tests for a Proportion or Percentile
This section considers hypotheses concerning population proportions and percentiles.
A population proportion is the ratio of the number of elements of a population that has some
specific characteristic to the total number of elements. A population percentile represents the
percentage of elements of a population having values less than some threshold C. Thus, if x is the
95th percentile of a population, 95% of the elements of the population have values less than C and
5% of the population have values greater than C.
This section of the guidance covers the following hypothesis: Case 1: H0: P < P0 vs.
HA: P > P0 and Case 2: H0: P > P0 vs. HA: P < P0 where P is a proportion of the population,
and P0 represents a given proportion (0 < P0 < 1). Equivalent hypotheses written in terms of
percentiles are H0: the 100Pth percentile is C or larger for Case 1, and H0: the 100Pth percentile is
C or smaller for Case 2. For example, consider the decision to determine whether the 95th
percentile of a container of waste is less than 1 mg/L cadmium. The null hypothesis in this case is
EPA QA/G-9 Final
QAOO Version 3-16 July 2000
-------
Box 3-9: Example of the Chen Test
Consider the following sample of a contaminant concentration measurements in (in ppm):2.0, 2.0, 5.0, 5.2, 5.9,
6.6, 7.4, 7.4, 9.7, 9.7, 10.2, 11.5, 12.4, 12.7, 14.1, 15.2, 17.7, 18.9, 22.8, 28.6, 30.5, 35.5. We want to test the
null hypothesis that the mean u is less than 10 ppm versus the alternative that it exceeds 10 ppm. A
significance level of 0.05 is to be used.
STEP 1: Since all of the data points exceed the detection limit, there is no need to substitute values for below
the detection limit.
STEP 2: A frequency plot of the 23 data points confirms the right-skewness.
STEP 3: Using boxes 2-2 and 2-4 of Chapter 2, it is found thatX = 13.08 ppm
and s = 8.99 ppm.
0 5 10 15 20 25 30 35 40
23 Concentration (ppm)
, -13.08)3
STEP 4: b = — ^ - r = - — - ;— = 1.14 ,
(fj-l)(fi-2>3 (22)(21)(8.99)3
a = b/6*Jn = 114/6^23 = 0.0396 ,
,
8.997 '
/V23
compute T = t + a(l + 2t2 ) + 4a2 (t + 2t3 ) = 1.965
(The value of 1.14 for skewness confirms that the data are skewed to the right.)
STEP 5: Using the last row of Table A-1 of Appendix A, the critical value z095 of the Normal distribution is
1.645.
STEP 6: Since T > z095 (1.965 > 1.645), the null hypothesis is rejected and we conclude that the true mean is
greater than 10 ppm.
H0: the 95th percentile of cadmium is less than 1 mg/L. Now, instead of considering the
population to consist of differing levels of cadmium, consider the population to consist of a binary
variable that is T if the cadmium level is above 1 mg/L or is '0' if the level is below 1 mg/L. In
this case, the hypothesis may be changed to a test for a proportion so that the null hypothesis
becomes H0: P < .95 where P represents the proportion of 1's (cadmium levels above 1 mg/L) in
the container of waste. Thus, any hypothesis about the proportion of the site below a threshold
can be converted to an equivalent hypothesis about percentiles. Therefore, only hypotheses about
the proportion of the site below a threshold will be discussed in this section. The information
required for this test includes the null and alternative hypotheses, the gray region, the false
rejection error rate a at P0, the false acceptance error rate P at Pl3 and any additional limits on
decision errors. It may be helpful to label any additional false rejection error limits as a2 at Pa2, a3
at Pa3, etc., and any additional false acceptance error limits as P2 at Pp2, P3 at Pp3, etc.
EPA QA/G-9 Final
QAOO Version 3-17 July 2000
-------
3.2.2.1 The One-Sample Proportion Test
PURPOSE
Given a random sample of size n, the one-sample proportion test may be used to test
hypotheses regarding a population proportion or population percentile for a distribution from
which the data were drawn. Note that for P=.5, this test is also called the Sign test.
ASSUMPTIONS AND THEIR VERIFICATION
The only assumption required for the one-sample proportion test is the assumption of a
random sample. To verify this assumption, review the procedures and documentation used to
select the sampling points and ascertain that proper randomization has been used.
LIMITATIONS AND ROBUSTNESS
Since the only assumption is that of a random sample, the procedures are valid for any
underlying distributional shape. The procedures are also robust to outliers, as long as they do not
represent data errors.
SEQUENCE OF STEPS
Directions for the one-sample test for proportions for a simple random sample and a
systematic random sample are given in Box 3-10, an example is given in Box 3-11.
3.2.3 Tests for a Median
A population median (jl) is another measure of the center of the population distribution.
This population parameter is less sensitive to extreme values and nondetects than the sample
mean. Therefore, this parameter is sometimes used instead of the mean when the data contain a
large number of nondetects or extreme values. The hypotheses considered in this section are:
Case 1: H0: jl < C vs. HA: jl > C; and
Case 2: H0: ji > C vs. HA: £< C
where C represents a given threshold such as a regulatory level.
It is worth noting that the median is the 50th percentile, so the methods described in
Section 3.2.2 may be used to test hypotheses concerning the median by letting P0 = 0.50. In this
case, the one-sample test for proportions is also called the Sign Test for a median. The Wilcoxon
signed rank test (Section 3.2.1.2) can also be applied to a median in the same manner as it is
applied to a mean. In
EPA QA/G-9 Final
QAOO Version 3-18 July 2000
-------
Box 3-10: Directions for the One-Sample Test for Proportions
for Simple and Systematic Random Samples
This box describes the steps for applying the one-sample test for proportions for Case 1 (H0: P < P0);
modifications for Case 2 (H0: P > P0) are given in braces { }.
STEP 1: Given a random sample X.,, X2, . . . , Xn of measurements from the population, let
p (small p) denote the proportion of X's that do not exceed C, i.e., p is the number
(k) of sample points that are less than or equal to C, divided by the sample size n.
STEP 2: Compute np, and n(1-p). If both np and n(1-p) are greater than or equal to 5, use
Steps 3 and 4. Otherwise, consult a statistician as analysis may be complex.
p - .5/n - P p + .5/n - PQ
STEP 3: Calculate z = — - for Case 1 or z = — - for
Case 2.
STEP 4: Use Table A-1 of Appendix A to find the critical value z^ such that 100(1-a)% of
the normal distribution is below z^.a. For example, if a = 0.05 then z^.a = 1 .645.
If z > z.,.,, {z < -z-i.J, the null hypothesis may be rejected. Go to Step 6.
If z > z1_cl {z < -z .,_„}, there is not enough evidence to reject the null hypothesis. Therefore, the
false acceptance error rate will need to be verified. Go to Step 5.
STEP 5: To calculate the power of the test, assume that the true values for the mean and
standard deviation are those obtained in the sample and use a statistical software
package like the DEFT software (EPA, 1994) or the DataQUEST software (EPA,
1996) to generate the power curve of the test.
If only one false acceptance error rate (P) has been specified (at P.,), it is possible to calculate
the sample size which achieves the DQOs. To do this, calculate
m =
If m < n, the false acceptance error rate has been satisfied. Otherwise, the false acceptance
error rate has not been satisfied.
STEP 6: The results of the test may be:
1) the null hypothesis was rejected and it seems that the proportion is greater than {less than}
PO;
2) the null hypothesis was not rejected, the false acceptance error rate was satisfied, and it
seems that proportion is less than {greater than} P0; or
3) the null hypothesis was not rejected, the false acceptance error rate was not satisfied, and it
would seem the proportion was less than {greater than} P0, but the conclusions are uncertain
because the sample size was too small.
EPA QA/G-9 Final
QAOO Version 3-19 July 2000
-------
Box 3-1 1 : An Example of the One-Sample Test for Proportions
for a Simple Random Sample
Consider 85 samples of which 1 1 samples have concentrations greater than the clean-up standard. This
data will be used to test the null hypothesis H0: P > .20 vs. HA: P < .20 (Case 2). The decision maker
has specified a 5% false rejection rate (a) for P0 = .2, and a false acceptance rate (P) of 20% for P., =
0.15.
STEP 1: From the data, the observed proportion (p) is p = 11/85 = .1294
STEP 2: np = (85)(.1294) = 1 1 and n(1-p) = (85)(1-.1294) = 74. Since both np and n(1-p) are
greater than or equal to 5, Steps 3 and 4 will be used.
STEP 3: Because H0: P > .20, Case 2 formulas will be used.
P + -5/n ~ Po .1294 + .5/85 - .2 _
JP0(\-P0)/n J. 2(1 -.2)785
STEP 4: Using Table A-1 of Appendix A, it was found that z., 05 = z95 = 1.645. Because z < -z., „
(i.e.,
-1.492 < -1.645), the null hypothesis is not rejected so Step 5 will need to be completed.
STEP 5: To determine whether the test was powerful enough, the sample size necessary to
achieve the DQOs was calculated as follows:
m =
1.64/2(1- .2)+ 1.04/15(1- .15)
= 422.18
.15 - .2
So 423 samples are required, many more than were actually taken.
STEP 6: The null hypothesis was not rejected and the false acceptance error rate was not
satisfied. Therefore, it would seem the proportion is greater than 0.2, but this
conclusion is uncertain because the sample size is too small.
addition, this test is more powerful than the Sign Test for symmetric distributions. Therefore, the
Wilcoxon signed rank test is the preferred test for the median.
3.2.4 Confidence Intervals
In some instances, a test of hypotheses for the estimated parameter (i.e. mean or difference
of two means) is not required, but an idea of the uncertainty of the estimate with respect to the
parameter is needed. The most common type of interval estimate is a confidence interval. A
confidence interval may be regarded as combining an interval around an estimate with a
probabilistic statement about the unknown parameter. When interpreting a confidence interval
statement such as "The 95% confidence interval for the mean is 19.1 to 26.3", the implication is
that the best estimate for the unknown population mean is 22.7 (halfway between 19.1 and 26.3),
and that we are 95% certain that the interval 19.1 to 26.3 captures the unknown population mean.
Box 3-12 gives the directions on how to calculate a confidence interval for the mean, Box 3-13
gives and example of the method.
EPA QA/G-9 Final
QAOO Version 3 - 20 July 2000
-------
The concept of a confidence interval can be shown by a simple example. Suppose a stable
situation producing data without any anomalies was sampled many times. Each time the sample
was taken, the mean and standard deviation was calculated from the sample and a confidence
interval constructed using the method of Box 3-10.
Box 3-12: Directions for a Confidence Interval for a Mean
for Simple and Systematic Random Samples
Let X.,, X2, ..., Xn represent a sample of size n from a population of normally distributed values.
Step 1: Use the directions in Box 2-2 to calculate the sample mean, X. Use the directions in Box 2-3 to
calculate the sample standard deviation, s.
Step 2: Use Table A-1 of Appendix A to find the critical value t^ such that 100(1-a/2)% of the t distribution
with n -1 degrees of freedom is below t^^. For example, if a = 0.10 and n = 16, then n-1 =15 and t
= 1.753.
f v / V
Step 3: The (1-a)100% confidence interval is: X ~"_ to X '
V« \n
Box 3-13: An Example of a Confidence Interval for a Mean
for a Random or Systematic Random Samples
The effluent from a discharge point in a plating manufacturing plant was sampled 7 times over the course of 4
days for the presence of Arsenic with the following results: 8.1, 7.9, 7.9. 8.2, 8.2, 8.0,7.9. The directions in Box
3-12 will be used to develop a 95% confidence interval for the mean.
Stepl: Using Box 2-2, X=8.Q3. Use Box 2-3, s=0.138.
Step 2: Using Table A-1 of Appendix A and 6 degrees of freedom, t^^ = 2.447.
Step 3: The (1-a)100% confidence interval is:
2.447x0.138 2.447x0.138
803 p= to 803 + p= or 7.902 to 8.158.
V7
3.3 TESTS FOR COMPARING TWO POPULATIONS
A two-sample test involves the comparison of two populations or a "before and after"
comparison. In environmental applications, the two populations to be compared may be a
potentially contaminated area with a background area or concentration levels from an upgradient
and a downgradient well. The comparison of the two populations may be based on a statistical
parameter that characterizes the relative location (e.g., a mean or median), or it may be based on a
distribution-free comparison of the two population distributions. Tests that do not assume an
underlying distributions (e.g., normal or lognormal) are called distribution-free or nonparametric
tests. These tests are often more useful for comparing two populations than those that assume a
specific distribution because they make less stringent assumptions. Section 3.3.1 covers tests for
differences in the means of two populations. Section 3.3.2 covers tests for differences in the
EPA QA/G-9 Final
QAOO Version 3-21 July 2000
-------
proportion or percentiles of two populations. Section 3.3.3 describes distribution-free
comparisons of two populations. Section 3.3.4 describes tests for comparing two medians.
Often, a two-sample test involves the comparison of the difference of two population
parameters to a threshold value. For environmental applications, the threshold value is often zero,
representing the case where the data are used to determine which of the two population
parameters is greater than the other. For example, concentration levels from a Superfund site may
be compared to a background site. Then, if the Superfund site levels exceed the background
levels, the site requires further investigation. A two-sample test may also be used to compare
readings from two instruments or two separate populations of people.
If the exact same sampling locations are used for both populations, then the two samples
are not independent. This case should be converted to a one-sample problem by applying the
methods described in Section 3.2 to the differences between the two populations at the same
location. For example, one could compare contaminant levels from several wells after treatment
to contaminant levels from the same wells before treatment. The methods described in Section
3.2 would then be applied to the differences between the before and after treatment contaminant
levels for each well.
3.3.1 Comparing Two Means
Let jij represent the mean of population 1 and |i2 represent the mean of population 2. The
hypotheses considered in this section are:
Case 1: H0: jij - |i2 < 50 vs. HA: jij - |i2 > 60; and
Case 2: H0: ^ - |i2 > 50 vs. HA: ^ - |i2 < 50.
An example of a two-sample test for population means is comparing the mean contaminant level
at a remediated Superfund site to a background site; in this case, 50 would be zero. Another
example is a Record of Decision for a Superfund site which specifies that the remediation
technique must reduce the mean contaminant level by 50 ppm each year. Here, each year would
be considered a separate population and 50 would be 50 ppm.
The information required for these tests includes the null and alternative hypotheses (either
Case 1 or Case 2); the gray region (i.e., a value 6X > 50 for Case 1 or a value 6X < 50 for Case 2
representing the bound of the gray region); the false rejection error rate a at 50; the false
acceptance error rate P at 5t; and any additional limits on decision errors. It may be helpful to
label additional false rejection error limits as a2 at 5a2, a3 at 6a3, etc., and to label additional false
acceptance error limits as P2 at 5p2, P3 at 6p3, etc.
EPA QA/G-9 Final
QAOO Version 3-22 July 2000
-------
3.3.1.1 Student's Two-Sample t-Test (Equal Variances)
PURPOSE
Student's two-sample t-test can be used to compare two population means based on the
independent random samples Xl3 X2, . . . , X,,, from the first population, and Yl3 Y2, . . . , Yn from
the second population. This test assumes the variabilities (as expressed by the variance) of the
two populations are approximately equal. If the two variances are not equal (a test is described in
Section 4.5), use Satterthwaite's t test (Section 3.3.1.2).
ASSUMPTIONS AND THEIR VERIFICATION
The principal assumption required for the two-sample t-test is that a random sample of
size m (Xl3 X2, . . . , XjJ is drawn from population 1, and an independent random sample of size n
(Yl3 Y2, . . . , Yn) is drawn from population 2. Validity of the random sampling and independence
assumptions should be confirmed by reviewing the procedures used to select the sampling points.
The second assumption required for the two-sample t-tests are that the sample means x
(sample 1) and Y (sample 2) are approximately normally distributed. If both m and n are large,
one may make this assumption without further verification. For small sample sizes, approximate
normality of the sample means can be checked by testing the normality of each of the two
samples.
LIMITATIONS AND ROBUSTNESS
The two-sample t-test with equal variances is robust to violations of the assumptions of
normality and equality of variances. However, if the investigator has tested and rejected
normality or equality of variances, then nonparametric procedures may be applied. The t-test is
not robust to outliers because sample means and standard deviations are sensitive to outliers.
SEQUENCE OF STEPS
Directions for the two-sample t-test for a simple random sample and a systematic simple
random sample are given in Box 3-14 and an example in Box 3-15.
3.3.1.2 Satterthwaite's Two-Sample t-Test (Unequal Variances)
Satterthwaite's t-test should be used to compare two population means when the variances
of the two populations are not equal. It requires the same assumptions as the two-sample t-test
(Section 3.3.1.1) except the assumption of equal variances.
Directions for Satterthwaite's t-test for a simple random sample and a systematic simple
random sample are given in Box 3-16 and an example in Box 3-17.
EPA QA/G-9 Final
QAOO Version 3 - 23 July 2000
-------
Box 3-14: Directions for the Student's Two-Sample t-Test (Equal Variances)
for Simple and Systematic Random Samples
This describes the steps for applying the two-sample t-tests for differences between the population
means when the two population variances are equal for Case 1 (H0: U., - u2 < 50). Modifications for Case
2
(H0: U., - u2 > 50) are given in parentheses {}.
STEP 1: Calculate the sample mean X and the sample variance sx2 for sample 1 and compute
the sample mean Y and the sample variance sY2 for sample 2.
STEP 2: Use Section 4.5 to determine if the variances of the two populations are equal. If the
variances of the two populations are not equal, use Satterthwaite's ttest (Section
3.3.1.2). Otherwise, compute the pooled standard deviation
(iw-!)+(«-!)
X- Y- 60
STEP 3: Calculate t = -
II m
Use Table A-1 of Appendix A to find the critical value t.,_a such that 100(1-a)% of the t-
distribution with (m+n-2) degrees of freedom is below t^.
If t > t.,.,, {t < -t^J, the null hypothesis may be rejected. Go to Step 5.
If t > i^ {t < -t^J, there is not enough evidence to reject the null hypothesis. Therefore, the
false acceptance error rate will need to be verified. Go to Step 4.
STEP 4: To calculate the power of the test, assume that the true values for the mean and
standard deviation are those obtained in the sample and use a statistical software
package like the DEFT software (EPA, 1994) or the DataQUEST software (EPA, 1996)
to generate the power curve of the two-sample t-test. If only one false acceptance error
rate (P) has been specified (at 6.,), it is possible to calculate the sample size which
achieves the DQOs, assuming the true mean and standard deviation are equal to the
values estimated from the sample, instead of calculating the power of the test.
Calculate
If m* < m and n* < n, the false acceptance error rate has been satisfied. Otherwise, the false
acceptance error rate has not been satisfied.
STEP 5: The results of the test could be:
1) the null hypothesis was rejected, and it seems U., - u2 > 50 {p., - u2 < 50};
2) the null hypothesis was not rejected, the false acceptance error rate was satisfied, and it
seems U., - u2 < 50 {p., - u2 > 50}; or
3) the null hypothesis was not rejected, the false acceptance error rate was not satisfied, and
it seems U., - u2 < 50 {p., - u2 > 50}, but this conclusion is uncertain because the sample size
EPA QA/G-9 Final
QAOO Version 3-24 July 2000
-------
Box 3-15: An Example of a Student's Two-Sample t-Test (Equal Variances)
for Simple and Systematic Random Samples
At a hazardous waste site, area 1 (cleaned using an in-situ methodology) was compared with a similar
(but relatively uncontaminated) reference area, area 2. If the in-situ methodology worked, then the two
sites should be approximately equal in average contaminant levels. If the methodology did network,
then area 1 should have a higher average than the reference area. Seven random samples were taken
from area 1, and eight were taken from area 2. Because the contaminant concentrations in the two areas
are supposedly equal, the null hypothesis is H0: U., - u2 < 0 (Case 1). The false rejection error rate was
set at 5% and the false acceptance error rate was set at 20% (P) if the difference between the areas is 2.5
ppb.
STEP 1: Sample Mean Sample Variance
Area 1 7.8 ppm 2.1 ppm2
Area 2 6.6 ppm 2.2 ppm2
STEP 2: Methods described in Section 4.5 were used to determine that the variances were
essentially equal. Therefore,
SE=
E
- D2.1 + (8- 1)2.2 =
(7-1) +(8-1)
STEPS: t = 7-8"6-6"0 = 1.5798
1.4676^/1/7+1/8
Table A-1 of Appendix A was used to find that the critical value t095 with (7 + 8 - 2) = 13
degrees of freedom is 1.771.
Because t > i^_a (i.e., 1.5798 > 1.771), there is not enough evidence to reject the null
hypothesis. The false acceptance error rate will need to be verified.
STEP 4: Assuming the true values for the mean and standard deviation are those obtained in the
sample:
2(1.46762)(1.645 + 0.842)2 ^^,^,2 „ noo • c
m = n — - — - — + (0.25)1.645 = 4.938, i.e., 5.
(2.5 - O)2
Because m* < m (7) and n* < n (8), the false acceptance error rate has been satisfied.
STEP 5: The null hypothesis was not rejected and the false acceptance error rate was satisfied.
Therefore, it seems there is no difference between the two areas and that the in-situ
methodology worked as expected.
EPA QA/G-9 Final
QAOO Version 3-25 July 2000
-------
Box 3-16: Directions for Satterthwaite's t-Test (Unequal Variances)
for Simple and Systematic Random Samples
This describes the steps for applying the two-sample t-test for differences between the population means
for Case 1 (H0: U., - u2 < 50). Modifications for Case 2 (H0: U., - u2 > 50) are given in parentheses {}.
STEP 1: Calculate the sample mean X and the sample variance sx2 for sample 1 and compute
the sample mean Y and the sample variance sY2 for sample 2.
STEP 2: Using Section 4.5, test whether the variances of the two populations are equal. If the
variances of the two populations are not equal, compute:
SNE \
o a
y Y
m n
If the variances of the two populations appear approximately equal, use Student's two-
sample t-test (Section 3.3.1.1, Box 3-14).
X- Y- 60
STEP 3: Calculate =
SNE
Use Table A-1 of Appendix A to find the critical value t.,_a such that 100(1-a)% of the t-
distribution with f degrees of freedom is below t.,_a, where
2 2?
/=•
m
sx s¥
m2(m-V) n\n-Y)
(Round f down to the nearest integer.)
If t > t.,.,, {t < -t .,_„}, the null hypothesis may be rejected. Go to Step 5.
If t > t1_cl {t i -t-i.J, there is not enough evidence to reject the null hypothesis and therefore,
the false acceptance error rate will need to be verified. Go to Step 4.
STEP 4: If the null hypothesis (H0) was not rejected, calculate either the power of the test or the
sample size necessary to achieve the false rejection and false acceptance error rates.
To calculate the power of the test, assume that the true values for the mean and
standard deviation are those obtained in the sample and use a statistical software
package to generate the power curve of the two-sample t-test. A simple method to
check on statistical power does not exist.
STEP 5: The results of the test could be:
1) the null hypothesis was rejected, and it seems U., - u2 > 50 {p., - u2 < 50};
2) the null hypothesis was not rejected, the false acceptance error rate was satisfied, and it
seems ^ - u2 < 50 {^ - u2 > 50}; or
3) the null hypothesis was not rejected, the false acceptance error rate was not satisfied,
and it seems U., - u2 < 50 {p., - u2 > 50}, but this conclusion is uncertain because the sample
EPA QA/G-9 Final
QAOO Version 3-26 July 2000
-------
Box 3-17: An Example of Satterthwaite's t-Test (Unequal Variances)
for Simple and Systematic Random Samples
At a hazardous waste site, area 1 (cleaned using an in-situ methodology) was compared with a similar
(but relatively uncontaminated) reference area, area 2. If the in-situ methodology worked, then the two
sites should be approximately equal in average contaminant levels. If the methodology did network,
then area 1 should have a higher average than the reference area. Seven random samples were taken
from area 1, and eight were taken from area 2. Because the contaminant concentrations in the two areas
are supposedly equal, the null hypothesis is H0: U., - u2 < 0 (Case 1). The false rejection error rate was
set at 5% and the false acceptance error rate was set at 20% (P) if the difference between the areas is 2.5
ppb.
STEP 1: Sample Mean Sample Variance
Area 1 9.2 ppm 1.3ppm2
Area 2 6.1 ppm 5.7 ppm2
STEP 2: Using Section 4.5, it was determined that the variances of the two populations were not
equal, and therefore using Satterthwaite's method is appropriate:
SNE = ^1.3/7 + 5.7/8 = 0.9477
STEPS: ,= 9-2-6-1-° =3.271
0.9477
Table A-1 was used with f degrees of freedom, where
/ = —i— '• ! = 10.307 (i.e., 10 degrees of freedom)
1.32 5.72
72(7-l) 82(8-l)
(recall that f is rounded down to the nearest integer), to find t1_c, = 1.812.
Because t > t095 (3.271 > 1.812), the null hypothesis may be rejected.
STEP 5: Because the null hypothesis was rejected, it would appear there is a difference between
the two areas (area 1 being more contaminated than area 2, the reference area) and
that the in-situ methodology has networked as intended.
3.3.2 Comparing Two Proportions or Percentiles
This section considers hypotheses concerning two population proportions (or two
population percentiles); for example, one might use these tests to compare the proportion of
children with elevated blood lead in one urban area compared with the proportion of children with
elevated blood lead in another area. The population proportion is the ratio of the number of
elements in a subset of the total population to the total number of elements, where the subset has
some specific characteristic that the rest of the elements do not. A population percentile
represents the percentage of elements of a population having values less than some threshold
value C.
EPA QA/G-9 Final
QAOO Version 3-27 July 2000
-------
Let Pj represent the true proportion for population 1, and P2 represent the true proportion
of population 2. The hypotheses considered in this section are:
Case 1: H0: P1 - P2 < 50 vs. HA: P1 - P2 > 60; and
Case 2: H0: Pt - P2 > 50 vs. HA: Pt - P2 < 50
where 50 is some numerical value. An equivalent null hypothesis for Case 1, written in terms of
percentiles, is H0: the lOOP/11 percentile minus the 100P2th percentile is C or larger, the reverse
applying to Case 2. Since any hypothesis about the proportion below a threshold can be
converted to an equivalent hypothesis about percentiles (see Section 3.2.2), this guidance will
only consider hypotheses concerning proportions.
The information required for this test includes the null and alternative hypotheses (either
Case 1 or Case 2); the gray region (i.e., a value 5j > 50 for Case 1 or a value 5j < 50 for Case 2,
representing the bound of the gray region); the false rejection error rate a at 60; the false
acceptance error rate P at 5^ and any additional limits on decision errors.
3.3.2.1 Two-Sample Test for Proportions
PURPOSE
The two-sample test for proportions can be used to compare two population percentiles or
proportions and is based on an independent random sample of m (Xl3 X2, . . . , XjJ from the first
population and an independent random sample size n (Yl3 Y2, . . . , Yn) from the second
population.
ASSUMPTIONS AND THEIR VERIFICATION
The principal assumption is that of random sampling from the two populations.
LIMITATIONS AND ROBUSTNESS
The two-sample test for proportions is valid (robust) for any underlying distributional
shape and is robust to outliers, providing they are not pure data errors.
SEQUENCE OF STEPS
Directions for a two-sample test for proportions for a simple random sample and a
systematic simple random sample are given in Box 3-18; an example is provided in Box 3-19.
EPA QA/G-9 Final
QAOO Version 3 - 28 July 2000
-------
Box 3-18: Directions for a Two-Sample Test for Proportions
for Simple and Systematic Random Samples
The following describes the steps for applying the two-sample test for proportions for Case 1 (H0: P., - P2
< 0). Modifications for Case 2 (H0: P., - P2 > 0) are given in braces {}.
STEP 1: Given m random samples X.,, X2, . . . , Xm from the first population, and n samples from the
second population, Y.,, Y2, . . . , Yn, let k., be the number of points from sample 1 which
exceed C, and let k2 be the number of points from sample 2 which exceed C. Calculate the
sample proportions p., = k/m and p2 = k2/n. Then calculate the pooled proportion
p = (kl + k2) / (m + n).
STEP 2: Compute imp.,, m(1-p.|), np2, n(1-p2). If all of these values are greater than or equal to 5,
continue. Otherwise, seek assistance from a statistician as analysis is complicated.
STEP 3: Calculate z = (pl - p2) I Jp(\ - p)(\lm + lln).
Use Table A-1 of Appendix A to find the critical value z^.a such that 100(1-a)% of the normal
distribution is below z^.a. For example, if a = 0.05 then z^.a = 1 .645.
If z > z.,.,, {z < -z .,_„}, the null hypothesis may be rejected. Go to Step 5.
If z > z1_cl {z < -z^J, there is not enough evidence to reject the null hypothesis. Therefore, the
false acceptance error rate will need to be verified. Go to Step 4.
STEP 4: If the null hypothesis (H0) was not rejected, calculate either the power of the test or the
sample size necessary to achieve the false rejection and false acceptance error rates. If only
one false acceptance error rate (P) has been specified at P., - P2, it is possible to calculate
the sample sizes that achieve the DQOs (assuming the proportions are equal to the values
estimated from the sample) instead of calculating the power of the test. To do this, calculate
2(Z, + Z, n
m* = n* = — -^ - li£ — - - - where P =
and zp is the pth percentile of the standard normal distribution (Table A-1 of Appendix A). If
both m and n exceed m*, the false acceptance error rate has been satisfied. If both m and n
are below m*, the false acceptance error rate has not been satisfied.
If m* is between m and n, use a software package like the DEFT software (EPA, 1994) or the
DataQUEST software (EPA, 1996) to calculate the power of the test, assuming that the true
values for the proportions P., and P2 are those obtained in the sample. If the estimated
power is below 1-(3, the false acceptance error rate has not been satisfied.
STEP 5: The results of the test could be:
1) the null hypothesis was rejected, and it seems the difference in proportions is greater
than 0 {less than 0};
2) the null hypothesis was not rejected, the false acceptance error rate was satisfied, and it
seems the difference in proportions is less than or equal to 0 {greater than or equal to 0}; or
3) the null hypothesis was not rejected, the false acceptance error rate was not satisfied,
and it seems the difference in proportions is less than or equal to 0 {greater than or equal to
0}, but this outcome is uncertain because the sample size was probably too small.
EPA QA/G-9 Final
QAOO Version 3-29 July 2000
-------
Box 3-19: An Example of a Two-Sample Test for Proportions
for Simple and Systematic Random Samples
At a hazardous waste site, investigators must determine whether an area suspected to be contaminated
with dioxin needs to be remediated. The possibly contaminated area (area 1) will be compared to a
reference area (area 2) to see if dioxin levels in area 1 are greater than dioxin levels in the reference area.
An inexpensive surrogate probe was used to determine if each individual sample is either "contaminated,"
i.e., over the health standard of 1 ppb, or "clean," i.e., less than the health standard of 1 ppb. The null
hypothesis will be that the proportion or contaminant levels in area 1 is less than or equal to the
proportion in area 2, or H0: P., - P2 < 0 (Case 1). The decision maker is willing to accept a false rejection
decision error rate of 10% (a) and a false-negative decision error rate of 5% (P) when the difference in
proportions between areas exceeds 0.10. A team collected 92 readings from area 1 (of which 12 were
contaminated) and 80 from area 2, the reference area, (of which 10 were contaminated).
STEP 1: The sample proportion for area 1 is p., = 12/92 = 0.130, the sample proportion for area 2 is
p2 = 10/80 = 0.125, and the pooled proportion p = (12 + 10) / (92 + 80 ) = 0.128.
STEP 2: imp., = 12, m(1-p.,) = 80, np2 = 10, n(1-p2) =70. Because these values are greater than or
equal to 5, continue to step 3.
STEPS: z = (0.130 - 0.125) / JO.128(1 - 0.128)(l/92 + 1/80) = 0.098
Table A-1 of Appendix A was used to find the critical value zogo = 1.282.
Because z > zogo (0.098 > 1.282), there is not enough evidence to reject the null hypothesis
and the false acceptance error rate will need to be verified. Go to Step 4.
STEP 4: Because the null hypothesis (H0) was not rejected, calculate the sample size necessary to
achieve the false rejection and false acceptance error rates. Because only one false
acceptance error rate (p = 0.05) has been specified (at a difference of P., - P2 = 0.1), it is
possible to calculate the sample sizes that achieve the DQOs, assuming the proportions are
equal to the values estimated from the sample:
2(1.282+ 1.645)20.1275(1-0.1275) 1on,,. im , ,
m = n = — - = 190.6 (i.e., 191 samples)
n m* P °-115 + °-055
where 0.1275 = P =
2
Because both m and n are less than m*, the false acceptance error rate has not been
satisfied.
STEP 5: The null hypothesis was not rejected, and the false acceptance error rate was not satisfied.
Therefore, it seems that there is no difference in proportions and that the contaminant
concentrations of the investigated area and the reference area are probably the same.
However, this outcome is uncertain because the sample sizes obtained were in all likelihood
too small.
EPA QA/G-9 Final
QAOO Version 3 - 30 July 2000
-------
3.3.3 Nonparametric Comparisons of Two Populations
In many cases, assumptions on distributional characteristics are difficult to verify or
difficult to satisfy for both populations. In this case, several distribution-free test procedures are
available that compare the shape and location of the two distributions instead of a statistical
parameter (such as a mean or median). The statistical tests described below test the null
hypothesis "H0: the distributions of population 1 and population 2 are identical (or, the site is not
more contaminated than background)" versus the alternative hypothesis "HA: part of the
distribution of population 1 is located to the right of the distribution of population 2 (or the site is
more contaminated than background)." Because of the structure of the hypothesis tests, the
labeling of populations 1 and 2 is of importance. For most environmental applications, population
1 is the area of interest (i.e., the potentially contaminated area) and population 2 is the reference
area.
There is no formal statistical parameter of interest in the hypotheses stated above.
However, the concept of false rejection and false acceptance error rates still applies.
3.3.3.1 The Wilcoxon Rank Sum Test
PURPOSE
The Wilcoxon rank sum test can be used to compare two population distributions based
on m independent random samples Xl3 X2, . . . , X,,, from the first population, and n independent
random samples Yl3 Y2, . . . , Yn from the second population. When applied with the Quantile test
(Section 3.3.3.2), the combined tests are most powerful for detecting true differences between
two population distributions.
ASSUMPTIONS AND THEIR VERIFICATION
The validity of the random sampling and independence assumptions should be verified by
review of the procedures used to select the sampling points. The two underlying distributions are
assumed to have the same shape and dispersion, so that one distribution differs by some fixed
amount (or is increased by a constant) when compared to the other distribution. For large
samples, to test whether both site distributions have approximately the same shape, one can create
and compare histograms for the samples.
LIMITATIONS AND ROBUSTNESS
The Wilcoxon rank sum test may produce misleading results if many data values are the
same. When values are the same, their relative ranks are the same, and this has the effect of
diluting the statistical power of the Wilcoxon rank sum test. Estimated concentrations should be
reported for data below the detection limit, even if these estimates are negative, because their
relative magnitude to the rest of the data is of importance. An important advantage of the
Wilcoxon rank sum test is its partial robustness to outliers, because the analysis is conducted in
EPA QA/G-9 Final
QAOO Version 3-31 July 2000
-------
terms of rankings of the observations. This limits the influence of outliers because a given data
point can be no more extreme than the first or last rank.
SEQUENCE OF STEPS
Directions and an example for the Wilcoxon rank sum test are given in Box 3-20 and Box
3-21. However, if a relatively large number of samples have been taken, it is more efficient in
terms of statistical power to use a large sample approximation to the Wilcoxon rank sum test
(Box 3-22) to obtain the critical values of W.
Box 3-20: Directions for the Wilcoxon Rank Sum Test
for Simple and Systematic Random Samples
LetX.,, X2, . . . , Xn represent the n data points from population 1 and Y.,, Y2, . . . , Ym represent the m data points
from population 2 where both n and m are less than or equal to 20. For Case 1, the null hypothesis will be that
population 1 is shifted to the left of population 2 with the alternative that population 1 is either the same as or
shifted to the right of population 2; Case 2 will be that population 1 is shifted to the right of population 2 with the
alternative that population 1 is the same as or shifted to the left of population 2; for Case 3, the null hypothesis
will be that there is no difference between the two populations and the alternative hypothesis will be that
population 1 is shifted either to the right or left of population 2. If either m or n are larger than 20, use Box 3-22.
STEP 1: List and rank the measurements from both populations from smallest to largest,
keeping track of which population contributed each measurement. The rank of 1 is
assigned to the smallest value, the rank of 2 to the second smallest value, and so
forth. If there are ties, assign the average of the ranks that would otherwise have
been assigned to the tied observations.
STEP 2: Calculate R as the sum of the ranks of the data from population 1, then calculate
W=R- H(n+l\
2
STEP 3: Use Table A-7 of Appendix A to find the critical value wa (or wa/2 for Case 3). For
Case 1, reject the null hypothesis if W > nm - wa. For Case 2, reject the null
hypothesis if W < wa. For Case 3, reject the null hypothesis if W > nm - wa/2 or W <
wa/2. If the null hypothesis is rejected, go to Step 5. Otherwise, go to Step 4.
STEP 4: If the null hypothesis (H0) was not rejected, the power of the test or the sample size
necessary to achieve the false rejection and false acceptance error rates should be
calculated. For small samples sizes, these calculations are too complex for this
document.
STEP 5: The results of the test could be:
1) the null hypothesis was rejected and it seems that population 1 is shifted to the right (Case 1), to
the left (Case 2) or to the left or right (Case 3) of population 2.
2) the null hypothesis was not rejected and it seems that population 1 is shifted to the left (Case 1)
or to the right (Case 2) of population 2, or there is no difference between the two populations (Case
3).
EPA QA/G-9 Final
QAOO Version 3-32 July 2000
-------
Box 3-21: An Example of the Wilcoxon Rank Sum Test
for Simple and Systematic Random Samples
At a hazardous waste site, area 1 (cleaned using an in-situ methodology) was compared with a similar
(but relatively uncontaminated) reference area, area 2. If the in-situ methodology worked, then the two
sites should be approximately equal in average contaminant levels. If the methodology did network,
then area 1 should have a higher average than the reference area. The null hypothesis will be that area 1
is shifted to the right of area 2 and the alternative hypothesis will be that there is no difference between
the two areas or that area 1 is shifted to the left of area 2 (Case 2). The false rejection error rate was set
at 10% and the false acceptance error rate was set at 20% (P) if the difference between the areas is 2.5
ppb. Seven random samples were taken from area 1 and eight samples were taken from area 2:
Area 1 Area 2
17,23,26,5 16,20,5,4
13,13,12 8,10,7,3
STEP 1: The data listed and ranked by size are (Area 1 denoted by *):
Data (ppb): 3, 4, 5, 5*, 7, 8, 10, 12*, 13*, 13*, 16, 17*, 20, 23*, 26*
Rank: 1, 2,3.5,3.5*, 5, 6, 7, 8*, 9.5*, 9.5* 11, 12*, 13, 14*, 15*
STEP 2: R = 3.5+ 8 +9.5 +9.5+12+14+15 = 71..5. W= 71.5 - 7(7 + 1)/2 = 43.5
STEPS: Using Table A-7 of Appendix A, a = 0.10 and Wa = 17. Since 43.5 > 17, do
not reject the null hypothesis.
STEP 4: The null hypothesis was not rejected and it would be appropriate to calculate
the probable power of the test. However, because the number of samples is
small, extensive computer simulations are required in order to estimate the
power of this test which is beyond the scope of this guidance.
STEP 5: The null hypothesis was not rejected. Therefore, it is likely that there is no
difference between the investigated area and the reference area, although the
statistical power is low due to the small sample sizes involved.
EPA QA/G-9 Final
QAOO Version 3 - 33 July 2000
-------
Box 3-22: Directions for the Large Sample Approximation
to the Wilcoxon Rank Sum Test for Simple and Systematic Random Samples
LetX.,, X2, . . . , Xn represent the n data points from population 1 and Y.,, Y2, . . . , Ym represent the m data
points from population 2 where both n and m are greater than 20. For Case 1, the null hypothesis will be
that population 1 is shifted to the left of population 2 with the alternative that population 1 is the same as
or shifted to the right of population 2; for Case 2, the null hypothesis will be that population 1 is shifted to
the right of population 2 with the alternative that population 1 is the same as or shifted to the left of
population 2; for Case 3, the null hypothesis will be that there is no difference between the populations
and the alternative hypothesis will be that population 1 is shifted either to the right or left of population 2.
STEP 1: List and rank the measurements from both populations from smallest to
largest, keeping track of which population contributed each measurement.
The rank of 1 is assigned to the smallest value, the rank of 2 to the second
smallest value, and so forth. If there are ties, assign the average of the ranks
that would otherwise have been assigned to the tied observations.
STEP 2: Calculate Was the sum of the ranks of the data from population 1.
mn
STEPS: Calculate w = - +Zp\lmn(n + m + 1)/12 where p = 1 - a for Case
1, p = a for Case 2, and zp is the pth percentile of the standard normal
distribution (Table A-1 of Appendix A). For Case 3, calculate both wa/2 (p =
a/2) and w^^ (p = 1 -a/2).
STEP 4: For Case 1 , reject the null hypothesis if W > \N^a. For Case 2, reject the null
hypothesis if W < wa. For Case 3, reject the null hypothesis if W > w^^ or
W < wa/2. If the null hypothesis is rejected, go to Step 6. Otherwise, go to
Step 5.
STEP 5: If the null hypothesis (H0) was not rejected, calculate either the power of the
test or the sample size necessary to achieve the false rejection and negative
error rates. If only one false acceptance error rate (P) has been specified (at
5.,), it is possible to calculate the sample size that achieves the DQOs,
assuming the true mean and standard deviation are equal to the values
estimated from the sample, instead of calculating the power of the test. If m
and n are large, calculate:
where zp is the pth percentile of the standard normal distribution (Table A-1 of Appendix A). If
1.16m* < m and 1.1 6n* < n, the false acceptance error rate has been satisfied.
STEP 6: The results of the test could be:
1) the null hypothesis was rejected, and it seems that population 1 is shifted to the right
(Case 1 ), to the left (Case 2) or to the left or right (Case 3) of population 2.
2) the null hypothesis was not rejected, the false acceptance error rate was satisfied, and it
seems that population 1 is shifted to the left (Case 1 ) or to the right (Case 2) of population 2,
or there is no difference between the two populations (Case 3).
3) the null hypothesis was not rejected, the false acceptance error rate was not satisfied, and
it seems that population 1 is shifted to the left (Case 1 ) or to the right (Case 2) of population
EPA QA/G-9 Final
QAOO Version 3-34 July 2000
-------
3.3.3.2 The Quantile Test
PURPOSE
The Quantile test can be used to compare two populations based on the independent
random samples Xt, X2, . . ., X,,, from the first population and Yl5 Y2, . . ., Yn from the second
population. When the Quantile test and the Wilcoxon rank sum test (Section 3.3.3.1) are applied
together, the combined tests are the most powerful at detecting true differences between two
populations. The Quantile test is useful in detecting instances where only parts of the data are
different rather than a complete shift in the data. It essentially looks at a certain number of the
largest data values to determine if too many data values from one population are present to be
accounted for by pure chance.
ASSUMPTIONS AND THEIR VERIFICATION
The Quantile test assumes that the data Xl3 X2, . . ., X,,, are a random sample from
population 1, and the data Yl3 Y2, . . ., Yn are a random sample from population 2, and the two
random samples are independent of one another. The validity of the random sampling and
independence assumptions is assured by using proper randomization procedures, either random
number generators or tables of random numbers. The primary verification required is to review
the procedures used to select the sampling points. The two underlying distributions are assumed
to have the same underlying dispersion (variance).
LIMITATIONS AND ROBUSTNESS
The Quantile test is not robust to outliers. In addition, the test assumes either a systematic
(e.g., a triangular grid) or simple random sampling was employed. The Quantile test may not be
used for stratified designs. In addition, exact false rejection error rates are not available, only
approximate rates.
SEQUENCE OF STEPS
The Quantile test is difficult to implement by hand. Therefore, directions are not included
in this guidance but the DataQUEST software (EPA, 1996) can be used to conduct this test.
However, directions for a modified Quantile test that can be implemented by hand are contained
in Box 3-23 and an example is given in Box 3-24.
EPA QA/G-9 Final
QAOO Version 3-35 July 2000
-------
Box 3-23: Directions for a Modified Quantile Test for
Simple and Systematic Random Samples
Let there be 'm' measurements from population 1 (the reference area or group) and 'n' measurement from
population 2 (the test area or group). The Modified Quantile test can be used to detect differences in shape
and location of the two distributions. For this test, the significance level (a) can either be approximately 0.10
or approximately 0.05. The null hypothesis for this test is that the two population are the same (i.e., the test
group is the same as the reference group) and the alternative is that population 2 has larger measurements
than population 1 (i.e., the test group has larger values than the reference group).
STEP 1: Combine the two samples and order them from smallest to largest keeping track
of which sample a value came from.
STEP 2: Using Table A-13 of Appendix A, determine the critical number (C) for a sample
size n from the reference area, sample size m from the test area using the
significance level a. If the Cth largest measurement of the combined population is
the same as others, increase C to include all of these tied values.
STEP 3: If the largest C measurements from the combined samples are all from
population 2 (the test group), then reject the null hypothesis and conclude that
there are differences between the two populations. Otherwise, the null hypothesis
is not rejected and it appears that there is no difference between the two
populations.
3.3.4 Comparing Two Medians
Let jlj represent the median of population 1 and £2 represent the median of population 2.
The hypothesis considered in this section are:
Case 1: H0: j^ - J12 < 50 vs. HA: j^ - J12 > 50; and
Case 2: H0: ^ - £2 > 50 vs. HA: ^ - £2 < 50.
An example of a two-sample test for the difference between two population medians is comparing
the median contaminant level at a Superfund site to the median of a background site. In this case,
50 would be zero.
The median is also the 50th percentile, and, therefore, the methods described in Section
3.3.2 for percentiles and proportions may be used to test hypotheses concerning the difference
between two medians by letting ?! = ?„ = 0.50. The Wilcoxon rank sum test (Section 3.3.3.1) is
also recommended for comparing two medians. This test is more powerful than those for
proportions for symmetric distributions.
EPA QA/G-9 Final
QAOO Version 3-36 July 2000
-------
Box 3-24: A Example of a Modified Quantile Test for
Simple and Systematic Random Samples
At a hazardous waste site a new, cheaper, in-site methodology was compared against an existing
methodology by remediating separate areas of the site using each method. If the new methodology works,
the overall contamination of the area of the site remediated using the new methodology should be the same as
the area of the site remediated using the standard methodology. If the new methodology does network, then
there would be higher contamination values remaining on the area of the site remediated using the new
methodology. The site manager wishes to determine if the new methodology works and has chosen a 5%
significance level. A modified Quantile Test will be used to make this determination based on 7 samples from
the area remediated using the standard methodology (population 1) and 12 samples from the area remediated
using the new methodology (population 2). The sampled values are:
Standard Methodology New Methodology
17, 8, 20, 4, 6, 5, 4 7, 18, 2, 4, 6, 11, 5, 9, 10, 2, 3, 3
STEP 1: Combine the two samples and ordering them from smallest to largest yields:
2* 2* 3* 3* 4 4 4* 5* 5 6 6* 7* 8 9* 10* 11* 17 18* 20
where * denoted samples from the new methodology portion of the site (population 2).
STEP 2: Using Table A-13 of Appendix A with m = 7, n = 12, and a = 0.05, the critical value C = 5. Since
the 5th largest value is 10, there is not need to increase C.
(Note however, if the data were 2* 2* 3* 3* 4 4 4* 5* 5 6 6* 7* 8 9*9* 11* 17 18* 20
then the 5th largest value would have been 9 which is the same tied with 1 other value. In this
case, C would have been raised to 6 to include the tied value.)
STEPS: From Step 1, the 5 largest values are 10* 11* 17 18* 20. Only 3 of these 5
values come from population 2, therefore the null hypothesis can not be rejected
and the site manager concludes that it seems that the new in-situ methodology
works as well as the standard methodology.
3.4 Tests for Comparing Several Populations
3.4.1 Tests for Comparing Several Means
This section describes procedures to test the differences between several sample means
from different populations either against a control population or among themselves. For example,
the test described in this section could be used identify whether or not there are differences
between several drinking water wells or could be used to identify if several downgradient wells
differ from an upgradient well.
In this situation, it would be possible to apply the tests described in Section 3.3.1 multiple
times. However, applying a test multiple times underestimates the true false rejection decision
error rate. Therefore, the test described in this section controls the overall false rejection decision
error rate by making the multiple comparisons simultaneously.
EPA QA/G-9 Final
QAOO Version 3-37 July 2000
-------
3.4.1.1 Dunnett's Test
PURPOSE
Dunnett's test is used to test the difference between sample means from different
populations against a control population. A typical application would involve different cleaned
areas of a hazardous waste site being compared to a reference sample; this reference sample
having been obtained from a uncontaminated part of the hazardous waste site.
ASSUMPTIONS AND THEIR VERIFICATION
Multiple application of any statistical test is inappropriate because the continued use of the
same reference sample violates the assumption that the two samples were obtained independently
for each statistical test. The tests are strongly correlated between themselves with the degree of
correlation depending on the degree of similarity in number of samples used for the control group
and investigated groups. The test is really best suited for approximately equal sample sizes in
both the control group and the groups under investigation.
LIMITATIONS AND ROBUSTNESS
Dunnett's method is the same in operation as the standard two-sample t-test of Section
3.3.1 except for the use of a larger pooled estimate of variance and the need for special t-type
tables (Table A-14 of Appendix A). These tables are for the case of equal number of samples in
the control and each of the investigated groups, but remain valid provided the number of samples
from the investigated group are approximately more than half but less than double the size of the
control group. In this guidance, only the null hypothesis that the mean of the sample populations
is the same as the mean of the control population will be considered.
SEQUENCE OF STEPS
Directions for the use of Dunnett's method for a simple random sample or a systematic
random sample are given in Box 3-25 and an example is contained in Box 3-26.
EPA QA/G-9 Final
QAOO Version 3 - 38 July 2000
-------
Box 3-25: Directions for Dunnett's Test for
Simple Random and Systematic Samples
Let k represent the total number of populations to be compared so there are (k-1) sample populations and a
single control population. Let n.,, n2 ... nk_., represent the sample sizes of each ofthe (k-1) sample populations
and let m represent the sample size ofthe control population. The null hypothesis is H0: ur uc < 0 (i.e., no
difference between the sample means and the control mean) and the alternative hypothesis is HA: ur uc > 0 for
1 =1,2, ..., k-1 where u, represents the mean ofthe ith sample population and uc represents the mean ofthe
control population. Let a represent the chosen significance level for the test.
STEP 1: For each sample population, make sure that approximately 0.5 < m/n, < 2. If not, Dunnett's Test
should not be used.
STEP 2: Calculate the sample mean, x, (Section 2.2.2), and the variance, s2 (Section 2.2.3) for each ofthe k
populations (i.e., i = 1, 2, ... k).
STEP 3: Calculated the pooled standard:
(m—!) + («! —1)+...+(«£_! —1)
STEP 4: For each ofthe k-1 sample populations, compute
Xi - Xc
STEP 5: Use Table A-14 of Appendix A to determine the critical value TD(1_a) where the degrees of freedom is
(m -1) + (n.,-1) + . . . + (nk_.| -1).
STEP 6: Compare t, to TD(1_a) for each of the k-1 sample populations. If t, >TD(1_a) for any of the sample
populations, then reject the null hypothesis and conclude that there are differences between the
means ofthe sample populations and the mean ofthe control populations. Otherwise, conclude that
there is no difference between the sample and control population means.
EPA QA/G-9 Final
QAOO Version 3-39 July 2000
-------
Number of
Samples:
Mean:
Variance:
Ratio: m/n
t;:
Reference
Area
7
10.3
2.5
IAK3
6
11.4
2.6
7/6 = 1.16
1.18
ZBF6
5
12.2
3.3
7/5 = 1.4
1.93
3BG5
6
10.2
3.0
7/6 = 1.16
0.11
4GH2
7
11.4
3.2
7/7 = 1
1.22
5FF3
8
11.9
2.6
7/8 = 0.875
1.84
6GW4
7
12.1
2.8
7/7 = 1
2.00
Box 3-26: An Example of Dunnett's Test for
Simple Random and Systematic Samples
At a hazardous work site, 6 designated areas previously identified as 'hotspots' have been cleaned. In order for
this site to be a potential candidate for the local Brownfields program, it must be demonstrated that these areas
are not longer contaminated. Therefore, the means of these areas will be compared to mean of a reference area
also located on the site using Dunnett's test. The null hypothesis will be that there is no difference between the
means of the 'hotspot' areas and the mean of the reference area. A summary of the data from the areas follows.
STEP 1: Calculate the ratio m/n for each investigated area. These are shown in the 4th row of the table above.
Since all of these ration fall within the range of 0.5 to 2.0, Dunnett's test may be used.
STEP 2: The sample means, x, and the variance, s2 were calculated using Sections 2.2.2 and 2.2.3 of
Chapter 2. These are shown in the 2nd and 3rd row of the table above.
STEP 3: The pooled standard deviation for all 7 areas is:
SD =
(7 -1)2.5 + (6 -1)2.6+.. .+(7 ~ 1)2.8
11.4-10.3
STEP 4: For each 'hotspot' area, t, was computed. For example, t\ = , — = 1.18
1.68^/1/6+1/7
These are shown in the 5th row of the table above.
STEP 5: The degrees of freedom is (7 -1) + (6 -1) + . . . + (7 -1) = 39. So using Table A-14 of Appendix A
with 39 for the degrees of freedom, the critical value TD(095) = 2.37 and TD(090) = 2.03.
STEP 6: Since none of the values in row 5 of the table are greater than either 2.37 or 2.03, it appears that
none of the 'hotspot' areas have contamination levels that are significantly different than the
reference area. Therefore, this site may be a potential candidate to be a Brownsfield site.
NOTE: If an ordinary 2-sample t-test (see Section 3.3.1.1) had been used to compare each 'hotspot' area with
the reference area at the 5% level of significance, areas 2BF6, 5FF3, and 69W4 would have erroneously been
declared different from the reference area, which would probably alter the final conclusion to include the site as a
Brownfields candidate.
EPA QA/G-9
QAOO Version
3-40
Final
July 2000
-------
CHAPTER 4
STEP 4: VERIFY THE ASSUMPTIONS OF THE STATISTICAL TEST
THE DATA QUALITY ASSESSMENT PROCESS
Review DQOs and Sampling Design
Conduct Preliminary Data Review
Select the Statistical Test
Verify the Assumptions
Draw Conclusions From the Data
VERIFY THE ASSUMPTIONS OF THE
STATISTICAL TEST
Purpose
Examine the underlying assumptions of the statistical
hypothesis test in light of the environmental data.
Activities
• Determine Approach for Verifying Assumptions
• Perform Tests of Assumptions
• Determine Corrective Actions
Tools
• Tests of distributional assumptions
• Tests for independence and trends
• Tests for dispersion assumptions
Step 4: Verify the Assumptions of the Statistical Test
! Determine approach for verifying assumptions.
P Identify any strong graphical evidence from the preliminary data review.
P Review (or develop) the statistical model for the data.
P Select the tests for verifying assumptions.
! Perform tests of assumptions.
P Adjust for bias if warranted.
P Perform the calculations required for the tests.
! If necessary, determine corrective actions.
P Determine whether data transformations will correct the problem.
P If data are missing, explore the feasibility of using theoretical justification or
collecting new data.
P Consider robust procedures or nonparametric hypothesis tests.
EPA QA/G-9
QAOO Version
4- 1
Final
July 2000
-------
List of Boxes
Box 4-1: Directions for the Coefficient of Variation Test and an Example 4-9
Box 4-2: Directions for Studentized Range Test and an Example 4-10
Box 4-3: Directions for Geary's Test 4-11
Box 4-4: Example of Geary's Test 4-11
Box 4-5: Directions for the Test for a Correlation Coefficient and an Example 4-15
Box 4-6: "Upper Triangular" Data for Basic Mann-Kendall Trend Test 4-17
Box 4-7: Directions for the Mann-Kendall Trend Test for Small Sample Sizes 4-18
Box 4-8: An Example of Mann-Kendall Trend Test for Small Sample Sizes 4-18
Box 4-9: Directions for the Mann-Kendall Procedure Using Normal Approximation 4-19
Box 4-10: An Example of Mann-Kendall Trend Test by Normal Approximation 4-20
Box 4-11: Data for Multiple Times and Multiple Stations 4-21
Box 4-12: Testing for Comparability of Stations and an Overall Monotonic Trend 4-22
Box 4-13: Directions for the Wald-Wolfowitz Runs Test 4-25
Box 4-14: An Example of the Wald-Wolfowitz Runs Test 4-26
Box 4-15: Directions for the Extreme Value Test (Dixon's Test) 4-28
Box 4-16: An Example of the Extreme Value Test (Dixon's Test) 4-28
Box 4-17: Directions for the Discordance Test 4-29
Box 4-18: An Example of the Discordance Test 4-29
Box 4-19: Directions for Rosner's Test for Outliers 4-30
Box 4-20: An Example of Rosner's Test for Outliers 4-31
Box 4-21: Directions for Walsh's Test for Large Sample Sizes 4-32
Box 4-22: Directions for Constructing Confidence Intervals and Confidence Limits 4-34
Box 4-23: Directions for Calculating an F-Test to Compare Two Variances 4-34
Box 4-24: Directions for Bartletf s Test 4-35
Box 4-25: An Example of Bartletf s Test 4-36
Box 4-26: Directions for Levene's Test 4-37
Box 4-27: An Example of Levene's Test 4-38
Box 4-28: Directions for Transforming Data and an Example 4-40
Box 4-29: Directions for Cohen's Method 4-44
Box 4-30: An Example of Cohen's Method 4-44
Box 4-31: Double Linear Interpolation 4-45
Box 4-32: Directions for Developing a Trimmed Mean 4-46
Box 4-33: An Example of the Trimmed Mean 4-46
Box 4-34: Directions for Developing a Winsorized Mean and Standard Deviation 4-47
Box 4-35: An Example of a Winsorized Mean and Standard Deviation 4-47
Box 4-36: 11 Directions for Aitchison's Method to Adjust Means and Variances 4-48
Box 4-37: An Example of Aitchison's Method 4-48
Box 4-38: Directions for Selecting Between Cohen's Method or Aitchison's Method .... 4-49
Box 4-39: Example of Determining Between Cohen's Method and Aitchison's Method .. 4-49
Box 4-40: Directions for the Rank von Neumann Test 4-52
Box 4-41: An Example of the Rank von Neumann Test 4-53
EPA QA/G-9 Final
QAOO Version 4 - 2 July 2000
-------
CHAPTER 4
STEP 4: VERIFY THE ASSUMPTIONS OF THE STATISTICAL TEST
4.1 OVERVIEW AND ACTIVITIES
In this step, the analyst should assess the validity of the statistical test chosen in step 3 by
examining its underlying assumptions in light of the newly generated environmental data. The
principal thrust of this section is the determination of whether the data support the underlying
assumptions necessary for the selected test, or if modifications to the data are necessary prior to
further statistical analysis.
This determination can be performed quantitatively using statistical analysis of data to
confirm or reject the assumptions that accompany any statistical test. Almost always, however,
the quantitative techniques must be supported by qualitative judgments based on the underlying
science and engineering aspects of the study. Graphical representations of the data, such as those
described in Chapter 2, can provide important qualitative information about the reasonableness of
the assumptions. Documentation of this step is important, especially when subjective judgments
play a pivotal role in accepting the results of the analysis.
If the data support all of the key assumptions of the statistical test, then the DQA
continues to the next step, drawing conclusions from the data (Chapter 5). However, often one
or more of the assumptions will be called into question which may trigger a reevaluation of one of
the previous steps. This iteration in the DQA is an important check on the validity and
practicality of the results.
4.1.1 Determine Approach for Verifying Assumptions
In most cases, assumptions about distributional form, independence, and dispersion can be
verified formally using the statistical tests described in the technical sections in the remainder of
this chapter, although in some situations, information from the preliminary data review may serve
as sufficiently strong evidence to support the assumptions. As part of this activity, the analyst
should identify methods to verify that the type and quantity of data required to perform the
desired test are available. The outputs of this activity should include a list of the specific tests that
will be used to verify the assumptions.
For each statistical test it will be necessary for the investigator to select the "level of
significance." For the specific null hypothesis for the test under consideration, the level of
significance is the chance that this null hypothesis is rejected even though it is true. For example,
if testing for normality of data, the null hypothesis is that the data do indeed exhibit normality.
When a test statistic is computed, choosing a level of significance of 5% is saying that if the null
hypothesis is true then the chance that normally distributed data will produce a statistic more
extreme than that value tabulated is only 1 in 20 (5%).
EPA QA/G-9 Final
QAOO Version 4 - 3 July 2000
-------
The choice of specific level of significance is up to the investigator and is a matter of
experience or personal choice. It does not have to be the same as that chosen in Step 3 (Select
the Statistical Test). If more than a couple of statistical tests are contemplated, it is advisable to
choose a numerically low value for the level of significance to prevent the accumulation of
potential errors. The level of significance for a statistical test is by definition the same as false
rejection error.
The methods and approach chosen for assumption verification depend on the nature of the
study and its documentation. For example, if computer simulation was used to estimate the
theoretical power of the statistical test, then this simulation model should be the basis for
evaluation of the effect of changes to assumptions using estimates calculated from the data to
replace simulation values.
If it is not already part of the design documentation, the analyst may need to formulate a
statistical model that describes the data. In a statistical model, the data are conceptually
decomposed into elements that are assumed to be "fixed" (i.e., the component is either a constant
but unknown feature of the population or is controlled by experimentation) or "random" (i.e., the
component is an uncontrolled source of variation). Which components are considered fixed and
which are random is determined by the assumptions made for the statistical test and by the
inherent structure of the sampling design. The random components that represent the sources of
uncontrolled variation could include several types of measurement errors, as well as other sources
such as temporal and/or spatial components.
In addition to identifying the components that make up an observation and specifying
which are fixed and which are random, the model should also define whether the various
components behave in an additive or multiplicative fashion (or some combination). For example,
if temporal or spatial autocorrelations are believed to be present, then the model needs to identify
the autocorrelation structure (see Section 2.3.8).
4.1.2 Perform Tests of Assumptions
For most statistical tests, investigators will need to assess the reasonableness of
assumptions in relation to the structure of the components making up an observation. For
example, a t-test assumes that the components, or errors, are additive, uncorrelated, and normally
distributed with homogeneous variance. Basic assumptions that should be investigated include:
(1) Is it reasonable to assume that the errors (deviations from the model) are
normally distributed? If adequate data are available, then standard tests for
normality can be conducted (e.g., the Shapiro-Wilk test or the Kolmogorov-
Smirnov test).
(2) Is it reasonable to assume that errors are uncorrelated? While it is natural to
assume that analytical errors imbedded in measurements made on different sample
units are independent, other errors from other sources may not be independent. If
EPA QA/G-9 Final
QAOO Version 4 - 4 July 2000
-------
sample units are "too close together," either in time or space, independence may
not hold. If the statistical test assumes independence and this assumption is not
correct, the proposed false rejection and false acceptance error rates for the
statistical test cannot be verified.
(3) Is it reasonable to assume that errors are additive and have a constant
variability? If sufficient data are available, a plot of the relevant standard
deviations versus mean concentrations may be used to discern if variability tends to
increase with concentration level. If so, transformations of the data may make the
additivity assumption more tenable.
One of the most important assumptions underlying the statistical procedures described
herein is that there is no inherent bias (systematic deviation from the true value) in the data. The
general approach adopted here is that if a long term bias is known to exist, then adjustment for
this bias should be made. If bias is present, then the basic effect is to shift the power curves
associated with a given test to the right or left, depending on the direction of the bias. Thus
substantial distortion of the nominal false rejection and false acceptance decision error rates may
occur and so the level of significance could be very different than that assumed, and the power of
the test be far less than expected. In general, bias cannot be discerned by examination of routine
data; rather, appropriate and adequate QA data are needed, such as performance evaluation data.
If one chooses not to make adjustment for bias on the basis of such data, then one should, at a
minimum, construct the estimated worse-case power curves so as to understand the potential
effects of the bias.
4.1.3 Determine Corrective Actions
Sometimes the assumptions underlying the primary statistical test will not be satisfied and
some type of corrective action will be required before proceeding. In some cases, a
transformation of the data will correct a problem with distributional assumptions. In other cases,
the data for verifying some key assumption may not be available, and existing information may not
support a theoretical justification of the validity of the assumption. In this situation, it may be
necessary to collect additional data to verify the assumptions. If the assumptions underlying a
hypothesis test are not satisfied, and data transformations or other modifications do not appear
feasible, then it may be necessary to consider an alternative statistical test. These include robust
test procedures and nonparametric procedures. Robust test procedures involve modifying the
parametric test by using robust estimators. For instance, as a substitute for a t-test, a trimmed
mean and its associated standard error (Section 4.7.2) might be used to form a t-type statistic.
4.2 TESTS FOR DISTRIBUTIONAL ASSUMPTIONS
Many statistical tests and models are only appropriate for data that follow a particular
distribution. This section will aid in determining if a distributional assumption of a statistical test
is satisfied, in particular, the assumption of normality. Two of the most important distributions
for tests involving environmental data are the normal distribution and the lognormal distribution,
EPA QA/G-9 Final
QAOO Version 4 - 5 July 2000
-------
both of which are discussed in this section. To test if the data follow a distribution other than the
normal distribution or the lognormal distribution, apply the chi-square test discussed in Section
4.2.7 or consult a statistician.
There are many methods available for verifying the assumption of normality ranging from
simple to complex. This section discusses methods based on graphs, sample moments (kurtosis
and skewness), sample ranges, the Shapiro-Wilk test and closely related tests, and goodness-of-fit
tests. Discussions for the simplest tests contain step-by-step directions and examples based on the
data in Table 4-1. These tests are summarized in Table 4-2. This section ends with a comparison
of the tests to help the analyst select a test for normality.
Table 4-1. Data for Examples in Section 4.2
15.63
11.00
11.75
10.45
13.18
10.37
10.54
11.55
11.01
10.23
X=11.57
s= 1.677
Table 4-2. Tests for Normality
Test
Shapiro Wilk W
Test
Filliben's Statistic
Coefficient of
Variation Test
Skewness and
Kurtosis Tests
Geary's Test
Studentized
Range Test
Chi-Square Test
Lilliefors
Kolmogorov-
Smirnoff Test
Section
4.2.2
4.2.3
4.2.4
4.2.5
4.2.6
4.2.6
4.2.7
4.2.7
Sample
Size
< 50
< 100
Any
>50
>50
< 1000
Largea
>50
Recommended Use
Highly recommended.
Highly recommended.
Only use to quickly discard
an assumption of
normality.
Useful for large sample
sizes.
Useful when tables for
other tests are not
available.
Highly recommended
(with some conditions).
Useful for grouped data
and when the comparison
distribution is known.
Useful when tables for
other tests are not
available.
Data-
QUEST
Yes
Yes
Yes
Yes
Yes
Yes
No
No
a The necessary sample size depends on the number of groups formed when implementing this test. Each group
should contain at least 5 observations.
EPA QA/G-9
QAOO Version
4-6
Final
July 2000
-------
The assumption of normality is very important as it is the basis for the majority of
statistical tests. A normal, or Gaussian, distribution is one of the most common probability
distributions in the analysis of environmental data. A normal distribution is a reasonable model of
the behavior of certain random phenomena and can often be used to approximate other probability
distributions. In addition, the Central Limit Theorem and other limit theorems state that as the
sample size gets large, some of the sample summary statistics (e.g., the sample mean) behave as if
they are a normally distributed variable. As a result, a common assumption associated with
parametric tests or statistical models is that the errors associated with data or a model follow a
normal distribution.
The graph of a normally distributed
random variable, a normal curve, is bell-
shaped (see Figure 4-1) with the highest point
located at the mean which is equal to the
median. A normal curve is symmetric about
the mean, hence the part to the left of the
mean is a mirror image of the part to the
right. In environmental data, random errors
occurring during the measurement process
may be normally distributed.
Normal Distribution
Lognormal Distribution
15
20
Figure 4-1. Graph of a Normal and Lognormal
Distribution
Environmental data commonly exhibit
frequency distributions that are non-negative
and skewed with heavy or long right tails. Several standard parametric probability models have
these properties, including the Weibull, gamma, and lognormal distributions. The lognormal
distribution (Figure 4-1) is a commonly used distribution for modeling environmental contaminant
data. The advantage to this distribution is that a simple (logarithmic) transformation will
transform a lognormal distribution into a normal distribution. Therefore, the methods for testing
for normality described in this section can be used to test for lognormality if a logarithmic
transformation has been used.
4.2.1 Graphical Methods
Graphical methods (Section 2.3) present detailed information about data sets that may not
be apparent from a test statistic. Histograms, stem-and-leaf plots, and normal probability plots
are some graphical methods that are useful for determining whether or not data follow a normal
curve. Both the histogram and stem-and-leaf plot of a normal distribution are bell-shaped. The
normal probability plot of a normal distribution follows a straight line. For non-normally
distributed data, there will be large deviations in the tails or middle of a normal probability plot.
Using a plot to decide if the data are normally distributed involves making a subjective
decision. For extremely non-normal data, it is easy to make this determination; however, in many
cases the decision is not straightforward. Therefore, formal test procedures are usually necessary
to test the assumption of normality.
EPA QA/G-9
QAOO Version
4-7
Final
July 2000
-------
4.2.2 Shapiro-Wilk Test for Normality (the W test)
One of the most powerful tests for normality is the W test by Shapiro and Wilk. This test
is similar to computing a correlation between the quantiles of the standard normal distribution and
the ordered values of a data set. If the normal probability plot is approximately linear (i.e., the
data follow a normal curve), the test statistic will be relatively high. If the normal probability plot
contains significant curves, the test statistic will be relatively low.
The W test is recommended in several EPA guidance documents and in many statistical
texts. Tables of critical values for sample sizes up to 50 have been developed for determining the
significance of the test statistic. However, this test is difficult to compute by hand since it requires
two different sets of tabled values and a large number of summations and multiplications.
Therefore, directions for implementing this test are not given in this document, but the test is
contained in the Data Quality Assessment Statistical Toolbox (QA/G-9D) (EPA, 1996).
4.2.3 Extensions of the Shapiro-Wilk Test (Filliben's Statistic)
Because the W test may only be used for sample sizes less than or equal to 50, several
related tests have been proposed. D'Agostino's test for sample sizes between 50 and 1000 and
Royston's test for sample sizes up to 2000 are two such tests that approximate some of the key
quantities or parameters of the W test.
Another test related to the W test is the Filliben statistic, also called the probability plot
correlation coefficient. This test measures the linearity of the points on the normal probability
plot. Similar to the W test, if the normal probability plot is approximately linear (i.e., the data
follow a normal curve), the correlation coefficient will be relatively high. If the normal probability
plot contains significant curves (i.e., the data do not follow a normal curve), the correlation
coefficient will be relatively low. Although easier to compute that the W test, the Filliben statistic
is still difficult to compute by hand. Therefore, directions for implementing this test are not given
in this guidance; however, it is contained in the software, Data Quality Assessment Statistical
Toolbox (QA/G-9D) (EPA, 1996).
4.2.4 Coefficient of Variation
The coefficient of variation (CV) may be used to quickly determine whether or not the
data follow a normal curve by comparing the sample CV to 1. The use of the CV is only valid for
some environmental applications if the data represent a non-negative characteristic such as
contaminant concentrations. If the CV is greater than 1, the data should not be modeled with a
normal curve. However, this method should not be used to conclude the opposite, i.e., do not
conclude that the data can be modeled with a normal curve if the CV is less than 1. This test is to
be used only in conjunction with other statistical tests or when graphical representations of the
data indicate extreme departures from normality. Directions and an example of this method are
contained in Box 4-1.
EPA QA/G-9 Final
QAOO Version 4 - 8 July 2000
-------
Box 4-1: Directions for the Coefficient of Variation Test for
Environmental Data and an Example
Directions
STEP1: Calculate the coefficient of variation (CV): CV = s IX =
«-l;=i
i "
-
«;=!
STEP 2: If CV > 1 .0, conclude that the data are not normally distributed. Otherwise, the test is
inconclusive.
The following example demonstrates using the coefficient of variation to determine that the data in Table
4-1 should not be modeled using a normal curve.
STEP1: Calculate the coefficient of variation (CV): CV = 4 = = 0.145
X 11.571
STEP 2: Since 0.145 > 1.0, the test is inconclusive.
4.2.5 Coefficient of Skewness/Coefficient of Kurtosis Tests
The degree of symmetry (or asymmetry) displayed by a data set is measured by the
coefficient of skewness (g3). The coefficient of kurtosis, g4, measures the degree of flatness of a
probability distribution near its center. Several test methods have been proposed using these
coefficients to test for normality. One method tests for normality by adjusting the coefficients of
skewness and kurtosis to approximate a standard normal distribution for sample sizes greater than
50.
Two other tests based on these coefficients include a combined test based on a chi-squared
(/2) distribution and Fisher's cumulant test. Fisher's cumulant test computes the exact sampling
distribution of g3 and g4; therefore, it is more powerful than previous methods which assume that
the distributions of the two coefficients are normal. Fisher's cumulant test requires a table of
critical values, and these tests require a sample size of greater than 50. Tests based on skewness
and kurtosis are rarely used as they are less powerful than many alternatives.
4.2.6 Range Tests
Almost 100% of the area of a normal curve lies within ±5 standard deviations from the
mean and tests for normality have been developed based on this fact. Two such tests, which are
both simple to apply, are the studentized range test and Geary's test. Both of these tests use a
ratio of an estimate of the sample range to the sample standard deviation. Very large and very
small values of the ratio then imply that the data are not well modeled by a normal curve.
EPA QA/G-9 Final
QAOO Version 4 - 9 July 2000
-------
a. The studentized range test (or w/s test). This test compares the range of the
sample to the sample standard deviation. Tables of critical values for sample sizes up to 1000
(Table A-2 of Appendix A) are available for determining whether the absolute value of this ratio is
significantly large. Directions for implementing this method are given in Box 4-2 along with an
example. The studentized range test does not perform well if the data are asymmetric and if the
tails of the data are heavier than the normal distribution. In addition, this test may be sensitive to
extreme values. Unfortunately, lognormally distributed data, which are common in environmental
applications, have these characteristics. If the data appear to be lognormally distributed, then this
test should not be used. In most cases, the studentized range test performs as well as the Shapiro-
Wilk test and is much easier to apply.
Box 4-2: Directions for Studentized Range Test
and an Example
Directions
STEP 1: Calculate sample range (w) and sample standard deviation (s) using Section 2.2.3.
W (n) ~ (I)
STEP 2: Compare — = —^ = to the critical values given in Table A-2 (labeled a and b).
s s
If w/s falls outside the two critical values then the data do not follow a normal curve.
Example
The following example demonstrates the use of the studentized range test to determine if the data from Table 4-
1 can be modeled using a normal curve.
STEP 1: w = X(n)-X(1)= 15.63-10.23 = 5.40 and s = 1.677.
STEP 2: w/s = 5.4/1.677 = 3.22. The critical values given in Table A-2 are 2.51 and 3.875. Since 3.22 falls
between these values, the assumption of normality is not rejected.
b. Geary's Test. Geary's test uses the ratio of the mean deviation of the sample to
the sample standard deviation. This ratio is then adjusted to approximate a standard normal
distribution. Directions for implementing this method are given in Box 4-3 and an example is
given in Box 4-4. This test does not perform as well as the Shapiro-Wilk test or the studentized
range test. However, since Geary's test statistic is based on the normal distribution, critical values
for all possible sample sizes are available.
EPA QA/G-9 Final
QAOO Version 4-10 July 2000
-------
Box 4-3: Directions for Geary's Test
STEP 1: Calculate the sample mean x, the sample sum of squares (SSS), and the sum of absolute
deviations (SAD):
(b^
_ sss = £Xf - ——- and SAD = T^x,~x\
« ;=1 ' ;=1 n i=l '
SAD
STEP 2: Calculate Geary's test statistic a =
/n(SSS)
a - o7979
STEPS: Test "a" for significance by computing Z= : . Here 0.7979 and 0.2123 are
0.2123/y«
constants used to achieve normality.
STEP 4: Use Table A-1 of Appendix A to find the critical value z.,_a such that 100(1-a)% of the normal
distribution is below z^. For example, if a = 0.05, then z^ = 1.645. Declare "a" to be
sufficiently small or large (i.e., conclude the data are not normally distributed) if Zi > Z.,_a.
Box 4-4: Example of Geary's Test
The following example demonstrates the use of Geary's test to determine if the data from Table 4-1 can
be modeled using a normal curve.
_ i « « _
STEP1: X= -YX.= 11.571, SAD = Y\X.-X\ = 11.694, and
n,,\ ,= i
SSS = - — - = 1364.178 - 1338.88 = 25.298
STEP 2: a = = _ = Q
^10(25.298)
STEPS: z = a735 - a7979 = -Q.934
0.2123/^10
STEP 4: Since Zi > 1.64 (5% significance level), there is not enough information to
conclude that the data do not follow a normal distribution.
EPA QA/G-9 Final
QAOO Version 4-11 July 2000
-------
4.2.7 Goodness-of-Fit Tests
Goodness-of-fit tests are used to test whether data follow a specific distribution, i.e., how
"good" a specified distribution fits the data. In verifying assumptions of normality, one would
compare the data to a normal distribution with a specified mean and variance.
a. Chi-square Test. One classic goodness-of-fit test is the chi-square test which
involves breaking the data into groups and comparing these groups to the expected groups from
the known distribution. There are no fixed methods for selecting these groups and this test also
requires a large sample size since at least 5 observations per group are required to implement this
test. In addition, the chi-square test does not have the power of the Shapiro-Wilk test or some of
the other tests mentioned above.
b. Tests Based on the Empirical Distribution Function. The cumulative
distribution function, denoted by F(x), and the empirical distribution function of the data for a
given sample of size n are defined in Section 2.3.7.4. Since empirical distribution functions
estimate the true F(x) underlying a set of data, and as the cumulative distribution function for a
given type of distribution [e.g., a normal distribution (see Section 2.4) with given mean and
standard deviation] can be computed, a goodness of fit test can be performed using the empirical
distribution function. If the empirical distribution function is "not close to" the given cumulative
distribution function, then there is evidence that the data do not come from the distribution having
that cumulative distribution function.
Various methods have been used to measure the discrepancy between the sample empirical
distribution function and the theoretical cumulative distribution function. These measures are
referred to as empirical distribution function statistics. The best known empirical distribution
function statistic is the Kolmogorov-Smirnov (K-S) statistic. The K-S approach is appropriate if
the sample size exceeds 50 and if F(x) represents a specific distribution with known parameters
(e.g., a normal distribution with mean 100 and variance 30). A modification to the test, called the
Lilliefors K-S test, is appropriate (for n>50) for testing that the data are normally distributed
when the F(x) is based on an estimated mean and variance.
Unlike the K-S type statistics, most empirical distribution function statistics are based on
integrated or average values between the empirical distribution function and cumulative
distribution functions. The two most powerful are the Cramer-von Mises and Anderson-Darling
statistics. Extensive simulations show that the Anderson-Darling empirical distribution function
statistic is just as good as any, including the Shapiro-Wilk statistic, when testing for normality.
However, the Shapiro-Wilk test is applicable only for the case of a normal-distribution cumulative
distribution function, while the Anderson-Darling method is more general.
Most goodness-of-fit tests are difficult to perform manually and are usually included in
standard statistical software. The application of goodness-of-fit tests to non-normal data is
beyond the scope of this guidance and consultation with a statistician recommended.
EPA QA/G-9 Final
QAOO Version 4-12 July 2000
-------
4.2.8 Recommendations
Analysts can perform tests for normality with samples as small as 3. However, the tests
lack statistical power for small sample size. Therefore, for small sample sizes, it is recommended
that a nonparametric statistical test (i.e., one that does not assume a distributional form of the
data) be selected during Step 3 of the DQA in order to avoid incorrectly assuming the data are
normally distributed when there is simply not enough information to test this assumption.
If the sample size is less than 50, then this guidance recommends using the Shapiro-Wilk
W test, wherever practicable. The Shapiro-Wilk W test is one of most powerful tests for
normality and it is recommended in several EPA guidance as the preferred test when the sample
size is less than 50. This test is difficult to implement by hand but can be applied easily using the
Data Quality Assessment Statistical Toolbox (QA/G-9D) (EPA, 1996). If the Shapiro-Wilk W
test is not feasible, then this guidance recommends using either Filliben's statistic or the
studentized range test. Filliben's statistic performs similarly to the Shapiro-Wilk test. The
studentized range is a simple test to perform; however, it is not applicable for non-symmetric data
with large tails. If the data are not highly skewed and the tails are not significantly large
(compared to a normal distribution), the studentized range provides a simple and powerful test
that can be calculated by hand.
If the sample size is greater than 50, this guidance recommends using either the Filliben's
statistic or the studentized range test. However, if critical values for these tests (for the specific
sample size) are not available, then this guidance recommends implementing either Geary's test or
the Lilliefors Kolmogorov-Smirnoff test. Geary's test is easy to apply and uses standard normal
tables similar to Table A-l of Appendix A and widely available in standard textbooks. Lilliefors
Kolmogorov-Smirnoff is more statistically powerful but is also more difficult to apply and uses
specialized tables not readily available.
4.3 TESTS FOR TRENDS
4.3.1 Introduction
This section presents statistical tools for detecting and estimating trends in environmental
data. The detection and estimation of temporal or spatial trends are important for many
environmental studies or monitoring programs. In cases where temporal or spatial patterns are
strong, simple procedures such as time plots or linear regression over time can reveal trends. In
more complex situations, sophisticated statistical models and procedures may be needed. For
example, the detection of trends may be complicated by the overlaying of long- and short-term
trends, cyclical effects (e.g., seasonal or weekly systematic variations), autocorrelations, or
impulses or jumps (e.g., due to interventions or procedural changes).
The graphical representations of Chapter 2 are recommended as the first step to identify
possible trends. A plot of the data versus time is recommended for temporal data, as it may reveal
long-term trends and may also show other major types of trends, such as cycles or impulses. A
EPA QA/G-9 Final
QAOO Version 4-13 July 2000
-------
posting plot is recommended for spatial data to reveal spatial trends such as areas of high
concentration or areas that were inaccessible.
For most of the statistical tools presented below, the focus is on monotonic long-term
trends (i.e., a trend that is exclusively increasing or decreasing, but not both), as well as other
sources of systematic variation, such as seasonality. The investigations of trend in this section are
limited to one-dimensional domains, e.g., trends in a pollutant concentration over time. The
current edition of this document does not address spatial trends (with 2- and 3-dimensional
domains) and trends over space and time (with 3- and 4-dimensional domains), which may involve
sophisticated geostatistical techniques such as kriging and require the assistance of a statistician.
Section 4.3.2 discusses estimating and testing for trends using regression techniques. Section
4.3.3 discusses more robust trend estimation procedures, and Section 4.3.4 discusses hypothesis
tests for detecting trends under several types of situations.
4.3.2 Regression-Based Methods for Estimating and Testing for Trends
4.3.2.1 Estimating a Trend Using the Slope of the Regression Line
The classic procedures for assessing linear trends involve regression. Linear regression is
a commonly used procedure in which calculations are performed on a data set containing pairs of
observations (X;, Y;), so as to obtain the slope and intercept of a line that "best fits" the data. For
temporal trends, the X; values represent time and the Y; values represent the observations, such as
contaminant concentrations. An estimate of the magnitude of trend can be obtained by
performing a regression of the data versus time (or some function of the data versus some
function of time) and using the slope of the regression line as the measure of the strength of the
trend.
Regression procedures are easy to apply; most scientific calculators will accept data
entered as pairs and will calculate the slope and intercept of the best fitting line, as well as the
correlation coefficient r (see Section 2.2.4). However, regression entails several limitations and
assumptions. First of all, simple linear regression (the most commonly used method) is designed
to detect linear relationships between two variables; other types of regression models are
generally needed to detect non-linear relationships such as cyclical or non-monotonic trends.
Regression is very sensitive to extreme values (outliers), and presents difficulties in handling data
below the detection limit, which are commonly encountered in environmental studies. Regression
also relies on two key assumptions: normally distributed errors, and constant variance. It may be
difficult or burdensome to verify these assumptions in practice, so the accuracy of the slope
estimate may be suspect. Moreover, the analyst must ensure that time plots of the data show no
cyclical patterns, outlier tests show no extreme data values, and data validation reports indicate
that nearly all the measurements were above detection limits. Because of these drawbacks,
regression is not recommended as a general tool for estimating and detecting trends, although it
may be useful as an informal, quick, and easy screening tool for identifying strong linear trends.
EPA QA/G-9 Final
QAOO Version 4-14 July 2000
-------
4.3.2.2 Testing for Trends Using Regression Methods
The limitations and assumptions associated with estimating trends based on linear
regression methods apply also to other regression-based statistical tests for detecting trends.
Nonetheless, for situations in which regression methods can be applied appropriately, there is a
solid body of literature on hypothesis testing using the concepts of statistical linear models as a
basis for inferring the existence of temporal trends. The methodology is complex and beyond the
scope of this document.
For simple linear regression, the statistical test of whether the slope is significantly
different from zero is equivalent to testing if the correlation coefficient is significantly different
from zero. Directions for this test are given in Box 4-5 along with an example. This test assumes
a linear relation between Y and X with independent normally distributed errors and constant
variance across all X and Y values. Censored values (e.g., below the detection limit) and outliers
may invalidate the tests.
Box 4-5: Directions for the Test for a Correlation Coefficient
and an Example
Directions
STEP 1: Calculate the correlation coefficient, r (Section 2.2.4).
r
STEP 2: Calculate the t-value t =
\
1 - r2
n - 2
STEP 3: Use Table A-1 of Appendix A to find the critical value t^ such that 100(1-a/2)% of the t
distribution with n - 2 degrees of freedom is below t.,_,.,. For example, if a = 0.10 and n = 17,
then n-2 = 15andt1.a/2= 1.753. Conclude that the correlation is significantly different from
zero if
't' > W-
Example: Consider the following data set (in ppb): for Sample 1, arsenic (X) is 4.0 and lead (Y) is 8.0;
for Sample 2, arsenic is 3.0 and lead is 7.0; for Sample 3, arsenic is 2.0 and lead is 7.0; and for Sample
4, arsenic is 1.0 and lead is 6.0.
STEP 1: In Section 2.2.4, the correlation coefficient r for this data was calculated to be 0.949.
0 949
STEP 2: t = - = 4.26
\
1 - 0.9492
4-2
STEP 3: Using Table A-1 of Appendix A, t^ = 2.920 fora 10% level of significance and 4-2 = 2
degrees of freedom. Therefore, there appears to be a significant correlation between the two
variables lead and arsenic.
EPA QA/G-9 Final
QAOO Version 4-15 July 2000
-------
4.3.3 General Trend Estimation Methods
4.3.3.1 Sen's Slope Estimator
Sen's Slope Estimate is a nonparametric alternative for estimating a slope. This approach
involves computing slopes for all the pairs of ordinal time points and then using the median of
these slopes as an estimate of the overall slope. As such, it is insensitive to outliers and can
handle a moderate number of values below the detection limit and missing values. Assume that
there are n time points (or n periods of time), and let Xj denote the data value for the ith time
point. If there are no missing data, there will be n(n-l)/2 possible pairs of time points (i, j) in
which i > j. The slope for such a pair is called a pairwise slope, by, and is computed as by = (Xj -
Xj) / (i - j). Sen's slope estimator is then the median of the n(n-l)/2 pairwise slopes.
If there is no underlying trend, then a given Xj is as likely to be above another Xj as it is
below. Hence, if there is no underlying trend, there would be an approximately equal number of
positive and negative slopes, and thus the median would be near zero. Due to the number of
calculations required, Sen's estimator is rarely calculated by hand and directions are not given in
this document.
4.3.3.2 Seasonal Kendall Slope Estimator
If the data exhibit cyclic trends, then Sen's slope estimator can be modified to account for
the cycles. For example, if data are available for each month for a number of years, 12 separate
sets of slopes would be determined (one for each month of the year); similarly, if daily
observations exhibit weekly cycles, seven sets of slopes would be determined, one for each day of
the week. In these estimates, the above pairwise slope is calculated for each time period and the
median of all of the slopes is an estimator of the slope for a long-term trend. This is known as the
seasonal Kendall slope estimator. Because of the number of calculations required, this estimator
is rarely calculated by hand.
4.3.4 Hypothesis Tests for Detecting Trends
Most of the trend tests treated in this section involve the Mann-Kendall test or extensions
of it. The Mann-Kendall test does not assume any particular distributional form and
accommodates trace values or values below the detection limit by assigning them a common
value. The test can also be modified to deal with multiple observations per time period and
generalized to deal with multiple sampling locations and seasonality.
4.3.4.1 One Observation per Time Period for One Sampling Location
The Mann-Kendall test involves computing a statistic S, which is the difference between
the number of pairwise slopes (described in 4.3.3.1) that are positive minus the number that are
negative. If S is a large positive value, then there is evidence of an increasing trend in the data. If
S is a large negative value, then there is evidence of a decreasing trend in the data. The null
EPA QA/G-9 Final
QAOO Version 4-16 July 2000
-------
hypothesis or baseline condition for this test is that there is no temporal trend in the data values,
i.e., "H0: no trend". The alternative condition or hypothesis will usually be either "HA: upward
trend" or "H4: downward trend."
The basic Mann-Kendall trend test involves listing the observations in temporal order, and
computing all differences that may be formed between measurements and earlier measurements, as
depicted in Box 4-6. The test statistic is the difference between the number of strictly positive
differences and the number of strictly negative differences. If there is an underlying upward trend,
then these differences will tend to be positive and a sufficiently large value of the test statistic will
suggest the presence of an upward trend. Differences of zero are not included in the test statistic
(and should be avoided, if possible, by recording data to sufficient accuracy). The steps for
conducting the Mann-Kendall test for small sample sizes (i.e., less than 10) are contained in Box
4-7 and an example is contained in Box 4-8.
Box 4-6: "Upper Triangular" Data for Basic Mann-Kendall Trend Test
with a Single Measurement at Each Time Point
Data Table
Original Time
Measurement
X,
X2
xn_2
Xn.!
Original Time
Measurement
X,
X2
xn_2
Xn.!
NOTE: Xy-Yk=
discarded.
t., t2 t3 t4 ... tn_! tn (time from earliest to latest)
X1 X2 X3 X4 ... Xn-1 Xn (actual values recorded)
X3-X2 X4-X2 . . . Xn_!-X2 Xn-X2
Xn-rXn-2 Xn-Xn_2
Xn-Xn_.|
After performing the subtractions this table converts to:
t! t2 t3 t4 ... tn_., tn # Of + # Of -
X., X2 X3 X4 ... Xn_! Xn Differences Differences
(>0) (<0)
'21 '31 '41 • • • '(n-1)1 'n1
'32 '42 • • • '(n-1)2 'n2
Y Y
' (n-1)(n-2) ' n(n-2)
Yn(n-1)
0 do not contribute to either total and are Total # >0 Total # <0
where Yik = sign (XrXk) = + if X; - Xk > 0
= 0 ifXi-Xk = 0
= - ifXi-Xk<0
EPA QA/G-9
QAOO Version
4- 17
Final
July 2000
-------
Box 4-7: Directions for the Mann-Kendall Trend Test for Small Sample Sizes
If the sample size is less than 10 and there is only one datum per time period, the Mann-Kendall Trend Test for
small sample sizes may be used.
STEP 1: List the data in the order collected overtime: X.,, X2, ..., Xn, where X, is the datum at time t,. Assign
a value of DL/2 to values reported as below the detection limit (DL). Construct a "Data Matrix"
similar to the top half of Box 4-6.
STEP 2: Compute the sign of all possible differences as shown in the bottom portion of Box 4-6.
STEP 3: Compute the Mann-Kendall statistic S, which is the number of positive signs minus the number of
negative signs in the triangular table: S = (number of + signs) - (number of - signs).
STEP 4: Use Table A-11 of Appendix A to determine the probability p using the sample size n and the
absolute value of the statistic S. For example, if n=5 and S=8, p=0.042.
STEP 5: For testing the null hypothesis of no trend against H., (upward trend), reject H0 if S > 0 and if p < a.
For testing the null hypothesis of no trend against H2 (downward trend), reject H0 if S < 0 and if p <
a.
Box 4-8: An Example of Mann-Kendall Trend Test for Small Sample Sizes
Consider 5 measurements ordered by the time of their collection: 5, 6, 11, 8, and 10. This data will be used to
test the null hypothesis, H0: no trend, versus the alternative hypothesis H., of an upward trend at an a = 0.05
significance level.
STEP 1: The data listed in order by time are: 5, 6, 11, 8, 10.
STEP 2: A triangular table (see Box 4-6) was used to construct the possible differences. The sum of signs of
the differences across the rows are shown in the columns 7 and 8.
Time 1 2 3 4 5 No. of + No. of
Data 5 6 11 8 10 Signs
Signs
5 + + + + 40
6 + + +30
11 - - 0 2
8 + 1 Q
8 2
STEP 3: Using the table above, 5 = 8-2 = 6.
STEP 4: From Table A-11 of Appendix A for n = 5 and S = 6, p = 0.117.
STEP 5: Since S > 0 but p = 0.117 < 0.05, the null hypothesis is not rejected. Therefore, there is not enough
evidence to conclude that there is an increasing trend in the data.
EPA QA/G-9 Final
QAOO Version 4-18 July 2000
-------
For sample sizes greater than 10, a normal approximation to the Mann-Kendall test is
quite accurate. Directions for this approximation are contained in Box 4-9 and an example is
given in Box 4-10. Tied observations (i.e., when two or more measurements are equal) degrade
the statistical power and should be avoided, if possible, by recording the data to sufficient
accuracy.
4.3.4.2 Multiple Observations per Time Period for One Sampling Location
Often, more than one sample is collected for each time period. There are two ways to
deal with multiple observations per time period. One method is to compute a summary statistic,
such as the median, for each time period and to apply one of the Mann-Kendall trend tests of
Section 4.3.4.1 to the summary statistic. Therefore, instead of using the individual data points in
the triangular table, the summary statistic would be used. Then the steps given in Box 4-7 and
Box 4-9 could be applied to the summary statistics.
An alternative approach is to consider all the multiple observations within a given time
period as being essentially equal (i.e., tied) values within that period. The S statistic is computed
as before with n being the total of all observations. The variance of the S statistic (previously
calculated in step 2) is changed to:
VAR(S) = —
18
w +5) -
p
p=\ q=\ p= 1 q= 1
9n(n-l)(n-2) 2n(n-l)
Box 4-9: Directions for the Mann-Kendall Procedure Using Normal Approximation
If the sample size is 10 or more, a normal approximation to the Mann-Kendall procedure may be used.
STEP 1 : Complete steps 1 , 2, and 3 of Box 4-7.
STEP 2: Calculate the variance of S: V(S) = n(n~V(2n + 5
18
If ties occur, let g represent the number of tied groups and wp represent the number of data points in
the pth group. The variance of S is: V(S) = — [n(n- l)(2« + 5) - 2^ w (w - l)(2w +5)]
18 p=\
STEP 4: Calculate Z = if S > 0, Z = 0 if S = 0, or Z = if S < 0.
STEP 5: Use Table A-1 of Appendix A to find the critical value z^_a such that 100(1-a)% of the normal
distribution is below z^_a. For example, if a=0.05 then 2^=1.645.
STEP 6: For testing the hypothesis, H0 (no trend) against 1) H., (an upward trend) - reject H0 if Z > z^_a, or 2)
H2 (a downward trend) - reject H0 if Z < 0 and the absolute value of Z > z^_a.
EPA QA/G-9 Final
QAOO Version 4-19 July 2000
-------
Box 4-10: An Example of Mann-Kendall Trend Test by Normal Approximation
A test for an upward trend with a=.05 will be based on the 1 1
STEP 1:
Week
Data
10
10
10
5
10
20
18
17
15
24
STEP 2:
STEP 3:
STEP 4:
STEP 5:
STEP 6:
Using Box 4-6, a triangular table was constructed
if the difference is zero, a "+" sign if the difference
negative.
1234567
10 10 10 5 10 20 18
0 0 - 0 + +
0 - 0 + +
0 + +
+ + +
+ +
-
S = (sum of + signs) - (sum of- signs) = 35 - 13 =
There are several observations tied at 10 and 15.
this formula, g=2, t.,=4 for tied values of 1 0, and t2
V(S) = — [11(11- l)(2(ll) + 5) - [4(4-
18
S-l 22-1
^ i nr^r* ^ i** no17* itiwr*" f — —
[V(Sj\* (155.33)
From Table A-1 of Appendix A, z1_05=1 .645.
H! is the alternative of interest. Therefore, since 1
weekly measurements shown below.
of the possible differences. A zero has been used
is positive, and a "-" sign if the difference is
8 9 10 11 No. of
17 15 24 15 + Signs
+ + + + 6
+ + + + 6
+ + + + 6
+ + + + 7
+ + + + 6
+ - 1
+ - 1
+ - 1
+ 0 1
0
35
22
Thus, the formula for tied values will be
=2 for tied values of 15.
-l)(2(4) + 5) + 2(2-l)(2(2) + 5)]]
20
•'•" - 1 (,C\Z.
v> 12.46
.605 is not greater than 1.645, H0 is not
No. of
- Signs
1
1
1
0
0
4
3
2
0
1
13
used. In
= 155.33
rejected.
Therefore, there is not enough evidence to determine that there is an upward trend.
where g represents the number of tied groups, wp represents the number of data points in the p*
group, h is the number of time periods which contain multiple data, and uq is the sample size in the
q* time period.
The preceding variance formula assumes that the data are not correlated. If correlation
within single time periods is suspected, it is preferable to use a summary statistic (e.g., the
median) for each time period and then apply either Box 4-7 or Box 4-9 to the summary statistics.
4.3.4.3 Multiple Sampling Locations with Multiple Observations
The preceding methods involve a single sampling location (station). However,
environmental data often consist of sets of data collected at several sampling locations (see Box
4-11). For example, data are often systematically collected at several fixed sites on a lake or
river, or within a region or basin. The data collection plan (or experimental design) must be
systematic in the sense that approximately the same sampling times should be used at all locations.
EPA QA/G-9 Final
QAOO Version 4 - 20 July 2000
-------
In this situation, it is desirable to express the results by an overall regional summary statement
across all sampling locations. However, there must be consistency in behavioral characteristics
across sites over time in order for a single summary statement to be valid across all sampling
locations. A useful plot to assess the consistency requirement is a single time plot (Section
2.3.8.1) of the measurements from all stations where a different symbol is used to represent each
station.
Box 4-11: Data for Multiple Times and Multiple Stations
Let i = 1, 2, ..., n represent time, k = 1,2, ..., K represent sampling locations, and Xik represent the
measurement at time i for location k. This data can be summarized in matrix form, as shown below.
Stations
1
2
Time
X,
X21
Xn1
81
V(S1)
Z1
X12
X22
Xn, '.
S2
V(S2)
Z2
• • X1K
• • X2K
• • XnK
SK
V(SK)
ZK
where Sk = Mann-Kendall statistic for station k (see STEP 3, Box 4-7),
V(Sk) = variance for S statistic for station k (see STEP 2, Box 4-9), and
If the stations exhibit approximately steady trends in the same direction (upward or
downward), with comparable slopes, then a single summary statement across stations is valid and
this implies two relevant sets of hypotheses should be investigated:
Comparability of stations. H0: Similar dynamics affect all K stations vs. HA: At least
two stations exhibit different dynamics.
Testing for overall monotonic trend. H0*: Contaminant levels do not change over time
vs. HA': There is an increasing (or decreasing) trend consistently exhibited across all
stations.
Therefore, the analyst must first test for homogeneity of stations, and then, if homogeneity is
confirmed, test for an overall monotonic trend.
Ideally, the stations in Box 4-11 should have equal numbers. However, the numbers of
observations at the stations can differ slightly, because of isolated missing values, but the overall
time periods spanned must be similar. This guidance recommends that for less than 3 time
periods, an equal
EPA QA/G-9 Final
QAOO Version 4-21 July 2000
-------
number of observations (a balanced design) is required. For 4 or more time periods, up to 1
missing value per sampling location may be tolerated.
a. One Observation per Time Period. When only one measurement is taken for
each time period for each station, a generalization of the Mann-Kendall statistic can be used to
test the above hypotheses. This procedure is described in Box 4-12.
Box 4-12: Testing for Comparability of Stations and an Overall Monotonic Trend
Let i = 1, 2, ..., n represent time, k = 1, 2, ..., K represent sampling locations, and Xik represent the
measurement at time i for location k. Let a represent the significance level for testing homogeneity and a*
represent the significance level for testing for an overall trend.
STEP 1: Calculate the Mann-Kendall statistic Sk and its variance V(Sk) for each of the K stations using the
methods of Section 4.3.4.1, Box 4-9.
STEP 2: For each of the K stations, calculate Zk = Sk/JV(Sk).
_ K
STEPS: Calculate the average Z = ^Zk/K.
k=\
K _
STEP 4: Calculate the homogeneity chi-square statistic 5^ = ^ Zk - K Z .
k=\
STEP 5: Using a chi-squared table (Table A-8 of Appendix A), find the critical value for x2 with (K-1 ) degrees
of freedom at an a significance level. For example, for a significance level of 5% and 5 degrees of
freedom, x2(5) = 11.07, i.e., 11.07 is the cut point which puts 5% of the probability in the upper tail of
a chi-square variable with 5 degrees of freedom.
STEP 6: If x2h X2(K-i), tne stations are not homogeneous (i.e., different dynamics at different stations) at the
significance level a. Therefore, individual a*-level Mann-Kendall tests should be conducted at each
station using the methods presented in Section 4.3.4.1.
STEP 7: Using a chi-squared table (Table A-8 of Appendix A), find the critical value for x2 with 1 degree of
freedom at an a significance level. If
then reject H0* and conclude that there is a significant (upward or downward) monotonic trend
across all stations at significance level a*. The signs of the Sk indicate whether increasing or
decreasing trends are present. If
K Z2 < X^1} ,
there is not significant evidence at the a' level of a monotonic trend across all stations. That is, the
stations appear approximately stable over time.
EPA QA/G-9 Final
QAOO Version 4 - 22 July 2000
-------
b. Multiple Observations per Time Period. If multiple measurements are taken at
some times and station, then the previous approaches are still applicable. However, the variance
of the statistic Sk must be calculated using the equation for calculating V(S) given in Section
4.3.4.2. Note that Sk is computed for each station, so n, wp, g, h, and uq are all station-specific.
4.3.4.4 One Observation for One Station with Multiple Seasons
Temporal data are often collected over extended periods of time. Within the time
variable, data may exhibit periodic cycles, which are patterns in the data that repeat over time
(e.g., the data may rise and fall regularly over the months in a year or the hours in a day). For
example, temperature and humidity may change with the season or month, and may affect
environmental measurements. (For more information on seasonal cycles, see Section 2.3.8). In
the following discussion, the term season represents one time point in the periodic cycle, such as a
month within a year or an hour within a day.
If seasonal cycles are anticipated, then two approaches for testing for trends are the
seasonal Kendall test and Sen's test for trends. The seasonal Kendall test may be used for large
sample sizes, and Sen's test for trends may be used for small sample sizes. If different seasons
manifest similar slopes (rates of change) but possibly different intercepts, then the Mann-Kendall
technique of Section 4.3.4.3 is applicable, replacing time by year and replacing station by season.
The seasonal Kendall test, which is an extension of the Mann-Kendall test, involves
calculating the Mann-Kendall test statistic, S, and its variance separately for each "season" (e.g.,
month of the year, day of the week). The sum of the S's and the sum of their variances are then
used to form an overall test statistic that is assumed to be approximately normally distributed for
larger size samples.
For data at a single site, collected at multiple seasons within multiple years, the techniques
of Section 4.3.4.3 can be applied to test for homogeneity of time trends across seasons. The
methodology follows Boxes 4-11 and 4-12 exactly except that "station" is replaced by "season"
and the inferences refer to seasons.
4.3.5 A Discussion on Tests for Trends
This section discusses some further considerations for choosing among the many tests for
trends. All of the nonparametric trend tests and estimates use ordinal time (ranks) rather than
cardinal time (actual time values, such as month, day or hour) and this restricts the interpretation
of measured trends. All of the Mann-Kendall Trend Tests presented are based on certain pairwise
differences in measurements at different time points. The only information about these differences
that is used in the Mann-Kendall calculations is their signs (i.e., whether they are positive or
negative) and therefore are generalizations of the sign test. Mann-Kendall calculations are
relatively easy and simply involve counting the number of cases in which X; + j exceeds X; and the
number of cases in which X; exceeds Xj+j. Information about magnitudes of these differences is
not used by the Mann-Kendall methods and this can adversely affect the statistical power when
only limited amounts of data are available.
EPA QA/G-9 Final
QAOO Version 4 - 23 July 2000
-------
There are, however, nonparametric methods based on ranks that takes such magnitudes
into account and still retains the benefit of robustness to outliers. These procedures can be
thought of as replacing the data by their ranks and then conducting parametric analyses. These
include the Wilcoxon rank sum test and its many generalizations. These methods are more
resistant to outliers than parametric methods; a point can be no more extreme than the smallest or
largest value.
Rank-based methods, which make fuller use of the information in the data than the Mann-
Kendall methods, are not as robust with respect to outliers as the sign and the Mann-Kendall
tests. They are, however, more statistically powerful than the sign test and the Mann-Kendall
methods; the Wilcoxon test being a case in point. If the data are random samples from normal
distributions with equal variances, then the sign test requires approximately 1.225 times as many
observations as the Wilcoxon rank sum test to achieve a given power at a given significance level.
This kind of tradeoff between power and robustness exemplifies the analyst's evaluation process
leading to the selection of the best statistical procedure for the current situation. Further
statistical tests will be developed in future editions of this guidance.
4.3.6 Testing for Trends in Sequences of Data
There are cases where it is desirable to see if a long sequence (for example, readings from
a monitoring station) could be considered random variation or correlated in some way, that is, if
consecutive results are attributable to random chance. An everyday example would be to
determine if a basketball player exhibited "hot streaks" during the season when shooting a basket
from the free-throw line. One test to make this determination is the Wald-Wolfowitz test. This
test can only be used if the data are binary, i.e., there are only two potential values. For example,
the data could either be 'Yes/No', '0/1', or 'black/white'. Directions for the Wald-Wofowitz test
are given in Box 4-13 and an example in Box 4-14.
4.4 OUTLIERS
4.4.1 Background
Outliers are measurements that are extremely large or small relative to the rest of the data
and, therefore, are suspected of misrepresenting the population from which they were collected.
Outliers may result from transcription errors, data-coding errors, or measurement system
problems such as instrument breakdown. However, outliers may also represent true extreme
values of a distribution (for instance, hot spots) and indicate more variability in the population
than was expected. Not removing true outliers and removing false outliers both lead to a
distortion of estimates of population parameters.
Statistical outlier tests give the analyst probabilistic evidence that an extreme value
(potential outlier) does not "fit" with the distribution of the remainder of the data and is therefore
a statistical outlier. These tests should only be used to identify data points that require further
investigation. The
EPA QA/G-9 Final
QAOO Version 4 - 24 July 2000
-------
Box 4-13: Directions for the Wald-Wolfowitz Runs Test
Consider a sequence of two values and let n denote the number of observations of one value and m denote the
number of observations of the other value. Note that if is customary for n < m (i.e., n denotes the value that
occurs the least amount of times. This test is used to test the null hypothesis that the sequence is random
against the alternative hypothesis that the data in the sequence are correlated or may come from different
populations.
STEP 1: List the data in the order collected and identify which will be the 'n' values, and which will be the 'm'
values.
STEP 2: Bracket the sequences within the series. A sequence is a group of consecutive values. For
example, consider the data AAABAABBBBBBBABB. The following are sequences in the data
{AAA} {B} {AA} {BBBBBB} {A} {BB}
In the example above, the smallest sequence is has one data value and the largest sequence has 6.
STEP 3: Count the number of sequences for the 'n' values and call it T. For the example sequence, the 'n'
values are 'A' since there are 6 A's and 9 B's, and T = 3: {AAA}, {AA}, and {A}.
STEP 4: If T is less than the critical value from Table A-12 of Appendix A for the specified significance level a,
then reject the null hypothesis that the sequence is random in favor of the alternative that the data
are correlated amongst themselves or possibly came from different distributions. Otherwise,
conclude the sequence is random. In the example above, 3 < 6 (where 6 is the critical value from
Table A-12 using n=6, m=9, and a = 0.01) so the null hypothesis that the sequence is random is
rejected.
tests alone cannot determine whether a statistical outlier should be discarded or corrected within a
data set; this decision should be based on judgmental or scientific grounds.
There are 5 steps involved in treating extreme values or outliers:
1. Identify extreme values that may be potential outliers;
2. Apply statistical test;
3. Scientifically review statistical outliers and decide on their disposition;
4. Conduct data analyses with and without statistical outliers; and
5. Document the entire process.
Potential outliers may be identified through the graphical representations of Chapter 2 (step 1
above). Graphs such as the box and whisker plot, ranked data plot, normal probability plot, and
time plot can all be used to identify observations that are much larger or smaller than the rest of
the data. If potential outliers are identified, the next step is to apply one of the statistical tests
described in the following sections. Section 4.4.2 provides recommendations on selecting a
statistical test for outliers.
EPA QA/G-9 Final
QAOO Version 4 - 25 July 2000
-------
Box 4-14: An Example of the Wald-Wolfowitz Runs Test
This is a set of monitoring data from the main discharge station at a chemical manufacturing plant. The permit
states that the discharge should have a pH of 7.0 and should never be less than 5.0. So the plant manager has
decided to use a pH of 6.0 to an indicate potential problems. In a four-week period the following values were
recorded:
6.5 6.6 6.4 6.2 5.9 5.8 5.9 6.2 6.2 6.3 6.6 6.6 6.7 6.4
6.2 6.3 6.2 5.8 5.9 5.8 6.1 5.9 6.0 6.2 6.3 6.2
STEP 1: Since the plant manager has decided that a pH of 6.0 will indicate trouble the data have been
replaced with a binary indicator. If the value is greater than 6.0, the value will be replaced by a 1;
otherwise the value will be replaced by a 0. So the data are now:
11110001111111111000100111
As there are 8 values of'O'and 19 values of '1', n = 8 and m = 19.
STEP 2: The bracketed sequence is: {1 1 1 1} {0 0 0} {1 1 1 1 1 1 1 1 1 1} {0 0 0} {1} {0 0 } {1 1 1}
STEP 3: T = 3: {000}, {000}, and {00}
STEP 4: Since 3 < 9 (where 9 is the critical value from Table A-12 using a = 0.05) so the null hypothesis that
the sequence is random is rejected.
If a data point is found to be an outlier, the analyst may either: 1) correct the data point;
2) discard the data point from analysis; or 3) use the data point in all analyses. This decision
should be based on scientific reasoning in addition to the results of the statistical test. For
instance, data points containing transcription errors should be corrected, whereas data points
collected while an instrument was malfunctioning may be discarded. One should never discard an
outlier based solely on a statistical test. Instead, the decision to discard an outlier should be based
on some scientific or quality assurance basis. Discarding an outlier from a data set should be done
with extreme caution, particularly for environmental data sets, which often contain legitimate
extreme values. If an outlier is discarded from the data set, all statistical analysis of the data
should be applied to both the full and truncated data set so that the effect of discarding
observations may be assessed. If scientific reasoning does not explain the outlier, it should not be
discarded from the data set.
If any data points are found to be statistical outliers through the use of a statistical test,
this information will need to be documented along with the analysis of the data set, regardless of
whether any data points are discarded. If no data points are discarded, document the
identification of any "statistical" outliers by documenting the statistical test performed and the
possible scientific reasons investigated. If any data points are discarded, document each data
point, the statistical test performed, the scientific reason for discarding each data point, and the
effect on the analysis of deleting the data points. This information is critical for effective peer
review.
EPA QA/G-9 Final
QAOO Version 4 - 26 July 2000
-------
4.4.2 Selection of a Statistical Test
There are several statistical tests for determining whether or not one or more observations
are statistical outliers. Step by step directions for implementing some of these tests are described
in Sections 4.4.3 through 4.4.6. Section 4.4.7 describes statistical tests for multivariate outliers.
If the data are normally distributed, this guidance recommends Rosner's test when the
sample size is greater than 25 and the Extreme Value test when the sample size is less than 25. If
only one outlier is suspected, then the Discordance test may be substituted for either of these
tests. If the data are not normally distributed, or if the data cannot be transformed so that the
transformed data are normally distributed, then the analyst should either apply a nonparametric
test (such as Walsh's test) or consult a statistician. A summary of this information is contained in
Table 4-3.
Table 4-3. Recommendations for Selecting a Statistical Test for Outliers
Sample
Size
n < 25
n < 50
n> 25
n > 50
Test
Extreme Value Test
Discordance Test
Rosner's Test
Walsh's Test
Section
4.4.3
4.4.4
4.4.5
4.4.6
Assumes
Normality
Yes
Yes
Yes
No
Multiple
Outliers
No/Yes
No
Yes
Yes
4.4.3 Extreme Value Test (Dixon's Test)
Dixon's Extreme Value test can be used to test for statistical outliers when the sample size
is less than or equal to 25. This test considers both extreme values that are much smaller than the
rest of the data (case 1) and extreme values that are much larger than the rest of the data (case 2).
This test assumes that the data without the suspected outlier are normally distributed; therefore, it
is necessary to perform a test for normality on the data without the suspected outlier before
applying this test. If the data are not normally distributed, either transform the data, apply a
different test, or consult a statistician. Directions for the Extreme Value test are contained in Box
4-15; an example of this test is contained in Box 4-16.
This guidance recommends using this test when only one outlier is suspected in the data.
If more than one outlier is suspected, the Extreme Value test may lead to masking where two or
more outliers close in value "hide" one another. Therefore, if the analyst decides to use the
Extreme Value test for multiple outliers, apply the test to the least extreme value first.
EPA QA/G-9
QAOO Version
4-27
Final
July 2000
-------
Box 4-15: Directions for the Extreme Value Test
(Dixon's Test)
STEP 1 : Let X( 1(, X( 2 ),..., X( n ) represent the data ordered from smallest to largest. Check that the data
without the suspect outlier are normally distributed, using one of the methods of Section 4.2. If
normality fails, either transform the data or apply a different outlier test.
STEP 2: X< 1 ; is a Potential Outlier (case 1): Compute the test statistic C, where
c = Xm ~ XW for3,n,7, C = X(3) ~ X(l) for 11, n, 13,
X(n) ~ X(l) X(n-l) ~ X(l)
X - X X - X
C = — £) - ^L- for 8 < n < 10, C = — £> - ~ X(n~2) for 11. n. 13,
X(n) ~ X(\) X(n) ~ X(2)
C = X(n) ~ X("~l) for8. n. 10, C = *<"> ~ X(n~2) for 14 , n < 25
X(n) ~ X(2) X(n) (3)
STEP 5: If C exceeds the critical value from Table A-3 of Appendix A for the specified significance level a, X(
n) is an outlier and should be further investigated.
Box 4-16: An Example of the Extreme Value Test
(Dixon's Test)
The data in order of magnitude from smallest to largest are: 82.39, 86.62, 91.72, 98.37, 103.46, 104.93,
105.52, 108.21, 113.23, and 150.55 ppm. Because the largest value (150.55) is much larger than the other
values, it is suspected that this data point might be an outlier which is Case 2 in Box 4-15.
STEP 1: A normal probability plot of the data shows that there is no reason to suspect that the data (without
the extreme value) are not normally distributed. The studentized range test (Section 4.2.6) also
shows that there is no reason to suspect that the data are not normally distributed. Therefore, the
Extreme Value test may be used to determine if the largest data value is an outlier.
STEp4. c = x(n) ~ Vi) = 150.55 - 113.23 = 37.32 = Q 5g4
X(n} - X(2) 150.55 - 86.62 63.93
STEP 5: Since C = 0.584 > 0.477 (from Table A-3 of Appendix A with n=10), there is evidence that X(n) is an
outlier at a 5% significance level and should be further investigated.
EPA QA/G-9 Final
QAOO Version 4 - 28 July 2000
-------
4.4.4 Discordance Test
The Discordance test can be used to test if one extreme value is an outlier. This test
considers two cases: 1) where the extreme value (potential outlier) is the smallest value of the
data set, and 2) where the extreme value (potential outlier) is the largest value of the data set.
The Discordance test assumes that the data are normally distributed; therefore, it is necessary to
perform a test for normality before applying this test. If the data are not normally distributed
either transform the data, apply a different test, or consult a statistician. Note that the test
assumes that the data without the outlier are normally distributed; therefore, the test for normality
should be performed without the suspected outlier. Directions and an example of the Discordance
test are contained in Box 4-17 and Box 4-18.
Box 4-17: Directions for the Discordance Test
STEP 1: Let X( 1}, X( 2),..., X( n} represent the data ordered from smallest to largest. Check that the data
without the suspect outlier are normally distributed, using one of the methods of Section 4.2. If
normality fails, either transform the data or apply a different outlier test.
STEP 2: Compute the sample mean, x (Section 2.2.2), and the sample standard deviation, s (Section 2.2.3).
If the minimum value X(1) is a suspected outlier, perform Steps 3 and 4. If the maximum value X(n) is
a suspected outlier, perform Steps 5 and 6.
~~ (i1
STEP 3: IfX,!; is a Potential Outlier (case 1): Compute the test statistic D = =
s
STEP 4: If D exceeds the critical value from Table A-4, X(., (is an outlier and should be further investigated.
X - X
STEPS: If X;n) is a Potential Outlier (case 2): Compute the test statistic D = —^
s
STEP 6: If D exceeds the critical value from Table A-4, X(n) is an outlier and should be further investigated.
Box 4-18: An Example of the Discordance Test
The ordered data are 82.39, 86.62, 91.72, 98.37, 103.46, 104.93, 105.52, 108.21, 113.23, and 150.55 ppm.
Because the largest value of this data set (150.55) is much larger than the rest, it may be an outlier.
STEP 1: A normal probability plot of the data shows that there is no reason to suspect that the data (without
the extreme value) are not normally distributed. The studentized range test (Section 4.2.6) also
shows that there is no reason to suspect that the data are not normally distributed. Therefore, the
Discordance test may be used to determine if the largest data value is an outlier.
STEP 2: x = 104.5 ppm and s = 18.922 ppm.
STEP 5: D - *<"> " * - 15°-55 " 104-5° = 2.43
s 18.92
STEP 6: Since D = 2.43 > 2.176 (from Table A-4 of Appendix A with n = 10), there is evidence that X(n) is an
outlier at a 5% significance level and should be further investigated.
EPA QA/G-9 Final
QAOO Version 4 - 29 July 2000
-------
4.4.5 Rosner's Test
A parametric test developed by Rosner can be used to detect up to 10 outliers for sample
sizes of 25 or more. This test assumes that the data are normally distributed; therefore, it is
necessary to perform a test for normality before applying this test. If the data are not normally
distributed either transform the data, apply a different test, or consult a statistician. Note that the
test assumes that the data without the outlier are normally distributed; therefore, the test for
normality may be performed without the suspected outlier. Directions for Rosner's test are
contained in Box 4-19 and an example is contained in Box 4-20.
Rosner's test is not as easy to apply as the preceding tests. To apply Rosner's test, first
determine an upper limit r0 on the number of outliers (r0 < 10), then order the r0 extreme values
from most extreme to least extreme. Rosner's test statistic is then based on the sample mean and
sample
Box 4-19: Directions for Rosner's Test for Outliers
STEP 1: LetX.,, X2, . . . , Xn represent the ordered data points. By inspection, identify the maximum number
of possible outliers, r0. Check that the data are normally distributed, using one of the methods of
Section 4.2.
STEP 2: Compute the sample mean x, and the sample standard deviation, s, for a]i the data. Label these
values x(0) and s(0), respectively. Determine the observation farthest from x(0) and label this
observation y( ° '. Delete y( ° ' from the data and compute the sample mean, labeled x( 1 ', and the
sample standard deviation, labeled s( 1 '. Then determine the observation farthest from x( 1 ' and
label this observation y(1 '. Delete y(1 ' and compute x(2) and s(2). Continue this process until r0
extreme values have been eliminated.
In summary, after the above process the analyst should have
[X(°\ s<0>,/°>]; [XW, s]; ..., [X(r°~l\ s(r°~ l\ y(r«~ l)} where
(l) l
X(l) = — — jc., s(l) = [— — £fr.-x(0)2]1/2, and y'1' is the farthest value fromx'1'.
n - ij, i J n- ij, i J
(Note, the above formulas for x( ' ' and s( ' ' assume that the data have been renumbered after each
observation is deleted.)
(r-l) _ ^(r-1)
STEPS: To test if there are V outliers in the data, compute: Rr = — - and compare Rr
to Ar in Table A-5 of Appendix A. If Rr > \, conclude that there are r outliers.
First, test if there are r0 outliers (compare R to A ). If not, test if there are r0-1 outliers
'o-1 'o-1
(compare .ft to A ). If not, test if there are r0-2 outliers, and continue, until either it is
'o- 1 'o-1
determined that there are a certain number of outliers or that there are no outliers at all.
EPA QA/G-9 Final
QAOO Version 4-30 July 2000
-------
Box 4-20: An Example of Rosner's Test for Outliers
STEP 1: Consider the following 32 data points (in ppm) listed in order from smallest to largest: 2.07, 40.55,
84.15, 88.41, 98.84, 100.54, 115.37, 121.19, 122.08, 125.84, 129.47, 131.90, 149.06, 163.89,
166.77, 171.91, 178.23, 181.64, 185.47, 187.64, 193.73, 199.74, 209.43, 213.29, 223.14, 225.12,
232.72, 233.21, 239.97, 251.12, 275.36, and 395.67.
A normal probability plot of the data shows that there is no reason to suspect that the data (without
the suspect outliers) are not normally distributed. In addition, this graph identified four potential
outliers: 2.07, 40.55, 275.36, and 395.67. Therefore, Rosner's test will be applied to see if there are
4 or fewer (r0 = 4) outliers.
STEP 2: First the sample mean and sample standard deviation were computed for the entire data set (x(0) and
s(0)). Using subtraction, it was found that 395.67 was the farthest data p_pint from x(0), so y(0) =
395.67. Then 395.67 was deleted from the data and the sample mean, x(1), and the sample standard
deviation, s(1), were computed. Using subtraction, it was found that 2.07 was the farthest value from
x(1). This value was then dropped from the data and the process was repeated again on 40.55 to
yield x(2), s(2), and y(2) and x(3), s(3), and y(3). These values are summarized below.
i x('' s(''
~0 169.923 75.133
1 162.640 63.872
2 167.993 57.460
3 172.387 53.099
STEP 3: To apply Rosner's test, it is first necessary to test if there are 4 outliers by computing
|275.36 - 172.387
5(3) 53.099
and comparing R4 to A4 in Table A-5 of Appendix A with n = 32. Since R4 = 1.939 2 A4 = 2.89, there
are not 4 outliers in the data set. Therefore, it will next be tested if there are 3 outliers by computing
I (2) _ x(2) |4055 _ 167.9931
- L.
s (2) 57.460
218
and comparing R3 to A3 in Table A-5 with n = 32. Since R3 = 2.218 2 A3 = 2.91, there are not 3
outliers in the data set. Therefore, it will next be tested if there are 2 outliers by computing
R, =
|2.07 - 162.640
2 sm 63.872
and comparing R2 to A2 in Table A-5 with n = 32. Since R2 = 2.514 2 A3 = 2.92, there are not 2
outliers in the data set. Therefore, it will next be tested if there is 1 outlier by computing
|395.67 - 169.923
s
(0) 75.133
and comparing R1 to A., in Table A-5 with n = 32. Since R1 = 3.005 > A.,= 2.94, there is evidence at a
5% significance level that there is 1 outlier in the data set. Therefore, observation 355.67 is a
statistical outlier and should be further investigated.
EPA QA/G-9 Final
QAOO Version 4-31 July 2000
-------
standard deviation computed without the r = r0 extreme values. If this test statistic is greater than
the critical value given in Table A-5 of Appendix A, there are r0 outliers. Otherwise, the test is
performed again without the r = r0 - 1 extreme values. This process is repeated until either
Rosner's test statistic is greater than the critical value or r = 0.
4.4.6 Walsh's Test
A nonparametric test was developed by Walsh to detect multiple outliers in a data set.
This test requires a large sample size: n > 220 for a significance level of a = 0.05, and n > 60 for a
significance level of a = 0.10. However, since the test is a nonparametric test, it may be used
whenever the data are not normally distributed. Directions for the test by Walsh for large sample
sizes are given in Box 4-21.
Box 4-21: Directions for Walsh's Test for Large Sample Sizes
Let X(.,), X( 2), . . . , X( n) represent the data ordered from smallest to largest. If n < 60, do not apply this
test. If 60 < n < 220, then a = 0.10. If n > 220, then a = 0.05.
STEP 1: Identify the number of possible outliers, r. Note that r can equal 1.
STEP 2: Compute c = [fin], k = r + c, b2 = 1 /a, and a = 1 + bv(c~b ^c ~ l\
c-b2- 1
where [ ] indicates rounding the value to the largest possible integer (i.e., 3.24 becomes 4).
STEP 3: The r smallest points are outliers (with a a% level of significance) if
STEP 4: The r largest points are outliers (with a a% level of significance) if
X(n+l-r) ~ (^+a)X(n-r) + aX(n+l-k)>®
STEP 5: If both of the inequalities are true, then both small and large outliers are indicated.
4.4.7 Multivariate Outliers
Multivariate analysis, such as factor analysis and principal components analysis, involves
the analysis of several variables simultaneously. Outliers in multivariate analysis are then values
that are extreme in relationship to either one or more variables. As the number of variables
increases, identifying potential outliers using graphical representations becomes more difficult. In
addition, special procedures are required to test for multivariate outliers. Details of these
procedures are beyond the scope of this guidance. However, procedures for testing for
multivariate outliers are contained in statistical textbooks on multivariate analysis.
EPA QA/G-9 Final
QAOO Version 4-32 July 2000
-------
4.5 TESTS FOR DISPERSIONS
Many statistical tests make assumptions on the dispersion (as measured by variance) of
data; this section considers some of the most commonly used statistical tests for variance
assumptions. Section 4.5.1 contains the methodology for constructing a confidence interval for a
single variance estimate from a sample. Section 4.5.2 deals with the equality of two variances, a
key assumption for the validity of a two-sample t-test. Section 4.5.3 describes Bartlett's test and
Section 4.5.4 describes Levene's test. These two tests verify the assumption that two or more
variances are equal, a requirement for a standard two-sample t-test, for example. The analyst
should be aware that many statistical tests only require the assumption of approximate equality
and that many of these tests remain valid unless gross inequality in variances is determined.
4.5.1 Confidence Intervals for a Single Variance
This section discusses confidence intervals for a single variance or standard deviation for
analysts interested in the precision of variance estimates. This information may be necessary for
performing a sensitivity analysis of the statistical test or analysis method. The method described
in Box 4-22 can be used to find a two-sided 100(l-a)% confidence interval. The upper end point
of a two-sided 100(l-a)% confidence interval is a 100(l-a/2)% upper confidence limit, and the
lower end point of a two-sided 100(l-a)% confidence interval is a 100(l-a/2)% lower confidence
limit. For example, the upper end point of a 90% confidence interval is a 95% upper confidence
limit and the lower end point is a 95% lower confidence limit. Since the standard deviation is the
square root of the variance, a confidence interval for the variance can be converted to a
confidence interval for the standard deviation by taking the square roots of the endpoints of the
interval. This confidence interval assumes that the data constitute a random sample from a
normally distributed population and can be highly sensitive to outliers and to departures from
normality.
4.5.2 The F-Test for the Equality of Two Variances
An F-test may be used to test whether the true underlying variances of two populations
are equal. Usually the F-test is employed as a preliminary test, before conducting the two-sample
t-test for the equality of two means. The assumptions underlying the F-test are that the two
samples are independent random samples from two underlying normal populations. The F-test for
equality of variances is highly sensitive to departures from normality. Directions for implementing
an F-test with an example are given in Box 4-23.
4.5.3 Bartlett's Test for the Equality of Two or More Variances
Bartlett's test is a means of testing whether two or more population variances of normal
distributions are equal. In the case of only two variances, Bartlett's test is equivalent to the F-test.
Often in practice unequal variances and non-normality occur together and Bartlett's test is itself
sensitive
EPA QA/G-9 Final
QAOO Version 4-33 July 2000
-------
Box 4-22: Directions for Constructing Confidence Intervals and
Confidence Limits for the Sample Variance and Standard Deviation with an Example
Directions: Let X.,, X2, . . . , Xn represent the n data points.
STEP 1: Calculate the sample variance s2 (Section 2.2.3).
STEP 2: For a 100(1-a)% two-sided confidence interval, use Table A-8 of Appendix A to find the cutoffs L and
: ( 1 - a/2 ) '
U such that L = x2a/2 and U = x2< 1.a/2) with (n-1) degrees of freedom (dof).
(n-l}s2 (n-l}s2
STEPS: A 100(1-a)% confidence interval forthe true underlying variance is: - — to —.
_/_j LJ
i 9 I
A 100(1-a)% confidence interval forthe true standard deviation is: —to '
L M u
Example: Ten samples were analyzed for lead: 46.4, 46.1, 45.8, 47, 46.1, 45.9, 45.8, 46.9, 45.2, 46 ppb.
STEP 1: Using Section 2.2.3, s2 = 0.286.
STEP 2: Using Table A-8 of Appendix A and 9 dof, L = x205/2 = X2025 = 19-02 and U = x2(i-.05/2) = X2.975 = 2.70.
STEP 3: A 95% confidence interval forthe variance is: ^ 1)0-286 fo (10-1)0.286 ^ Q^5 to Q g54
19.02 2.70
A 95% confidence interval forthe standard deviation is: A/0.135 = .368 to t/0.954 = .977.
Box 4-23: Directions for Calculating an F-Test to Compare
Two Variances with an Example
Directions: Let X.,, X2, . . . , Xm represent the m data points from population 1 and Y.,, Y2, . . . , Yn
represent the n data points from population 2. To perform an F-test, proceed as follows.
STEP 1: Calculate the sample variances sx2 (for the X's) and sY2 (for the Y's ) (Section 2.2.3).
STEP 2: Calculate the variance ratios Fx = sx2/sY2 and FY = sY2/sx2. Let F equal the larger of these two
values. If F = Fx, then let k = m -1 and q = n -1. If F = Fy, then let k = n -1 and q = m -1.
STEP 3: Using Table A-9 of Appendix A of the F distribution, find the cutoff U = f^k, q). If F > U,
conclude that the variances of the two populations are not the same.
Example: Manganese concentrations were collected from 2 wells. The data are Well X: 50, 73, 244,
202 ppm; and Well Y: 272, 171, 32, 250, 53 ppm. An F-test will be used to test if the variances are
equal.
STEP1: For Well X, sx2 = 9076. For Well Y, sY2 = 12125.
STEP 2: Fx = sx2/sY2 = 9076 /12125 = 0.749. FY = sY2/sx2 = 12125 / 9076 = 1.336. Since, FY > Fx, F =
FY = 1.336, k = 5 -1 = 4 and q = 4 - 1 = 3.
STEP 3: Using Table A-9 of Appendix A of the F distribution with a = 0.05, L = f,..^ 4, 3) = 15.1.
Since 1.336 < 15.1, there is no evidence that the variability of the two wells is different.
EPA QA/G-9 Final
QAOO Version 4-34 July 2000
-------
to departures from normality. With long-tailed distributions, the test too often rejects equality
(homogeneity) of the variances.
Bartletfs test requires the calculation of the variance for each sample, then calculation of a
statistic associated with the logarithm of these variances. This statistic is compared to tables and
if it exceeds the tabulated value, the conclusion is that the variances differ as a complete set. It
does not mean that one is significantly different from the others, nor that one or more are larger
(smaller) than the rest. It simply implies the variances are unequal as a group. Directions for
Bartlett's test are given in Box 4-24 and an example is given in Box 4-25.
4.5.4 Levene's Test for the Equality of Two or More Variances
Levene's test provides an alternative to Bartlett's test for homogeneity of variance (testing
for differences among the dispersions of several groups). Levene's test is less sensitive to
departures from normality than Bartlett's test and has greater power than Bartlett's for non-normal
data. In addition, Levene's test has power nearly as great as Bartlett's test for normally distributed
data. However, Levene's test is more difficult to apply than Bartlett's test since it involves
applying an analysis of variance (ANOVA) to the absolute deviations from the group means.
Directions and an example of Levene's test are contained in Box 4-26 and Box 4-27, respectively.
Box 4-24: Directions for Bartlett's Test
Consider k groups with a sample size of n, for each group. Let N represent the total number of samples,
i.e., let N = n., + n2 + . . . + nk. For example, consider two wells where 4 samples have been taken from
well 1 and 3 samples have been taken from well 2. In this case, k = 2, n., = 4, n2 = 3, and N = 4 + 3 = 7.
STEP 1: For each of the k groups, calculate the sample variances, s2 (Section 2.2.3).
1 k
STEP 2: Compute the pooled variance across groups: s = /iX- ~ I)-?,
(N-k) ,= i
k
STEPS: Compute the test statistic: TS = (N - k) ln(s^2) - Y^n,~ *) ln(5f)
;= 1
where "In" stands for natural logarithms.
STEP 4: Using a chi-squared table (Table A-8 of Appendix A), find the critical value for x2 with (k-1)
degrees of freedom at a predetermined significance level. For example, for a significance
level of 5% and 5 degrees of freedom, x2 = 11.1. If the calculated value (TS) is greater than
the tabulated value, conclude that the variances are not equal at that significance level.
EPA QA/G-9 Final
QAOO Version 4-35 July 2000
-------
Box 4-25: An Example of Bartlett
Manganese concentrations were collected from 6 wells over a 4 month
following table. Before analyzing the data, it is important to determine
Bartlett's test will be used to make this determination.
STEP
STEP
STEP
STEP
1 : For each of the 6 wells, the sample means and variances
bottom rows of the table below.
Sampling Date Well 1 Well 2
January 1 50
February 1 73
March 1 244 46
April 1 202 77
n;(N=17) 4 2
X; 142.25 61.50
s2 9076.37 480.49
o- o 2 1 Yin 1 IT 2
p (N-k) fff ' ' (17-6)
3:
TS = (17-6) ln(751837.27) - , (4-
Well 3
272
171
32
53
4
132
12455
[(4-1)9076 -
l)ln(9076) H
's Test
period. The data are shown in the
if the variances of the six wells are equal.
were calculated. These
Well 4
34
3940
2
1987
7628243
,-,(3
Well 5
48
54
2
51.00
17.98
are shown
Well 6
68
991
54
3
371.00
288348
in the
751837.27
- ... + (3- l)ln(288348) , =
4: The critical x2 value with 6-1=5 degrees of freedom at the 5% significance level
Table A-8 of Appendix A). Since 43.16 is larger than 11.1, it is concluded that the
(s2, . . . , s|) are not homogeneous at the 5% significance level.
43.16
is 11.1 (from
six variances
EPA QA/G-9
QAOO Version
4-36
Final
July 2000
-------
Box 4-26: Directions for Levene's Test
Consider k groups with a sample size of n, for the ith group. Let N represent the total number of samples, i.e.,
let N = n., + n2 + . . . + nk. For example, consider two wells where 4 samples have been taken from well 1 and 3
samples have been taken from well 2. In this case, k = 2, n., = 4, n2 = 3, and N = 4 + 3 = 7.
STEP 1: For each of the k groups, calculate the group mean, x , (Section 2.2.2), i.e., calculate:
_ 1 "l _ 1 "2 1 "k
x\ = — T$ir X2= — Y$2j> •••' xk= — Yfkj-
STEP 2: Compute the absolute residuals z.. = JC.. - X where Xy represents the jth value of the i
group. For each of the k groups, calculate the means, zh of these residuals, i.e., calculate:
" "
1 k "' 1 k
= — / / z .. = — / n.z..
\r *• — ' *• — ^ (/ AT -^ — ^ I I
Also calculate the overall mean residual as z =
-\T '-^1 '-^1 II -\T
i= 1
STEP 3: Compute the following sums of squares for the absolute residuals:
k "i - £ -2 _
oo _ V^V~^2 _ z oo _ V^• / _ Z _
TOTAL / > / »n ~5 ^^GROUPS 2-J ~Tf ^^ERROR ~ ^^TOTAL ' ^^GROUPS-
1)
STEP 4: Compute / =
STEP 5: Using Table A-9 of Appendix A, find the critical value of the F-distribution with (k-1) numerator
degrees of freedom, (N-k) denominator degrees of freedom, and a desired level of significance (a).
For example, if a = 0.05, the numerator degrees of freedom is 5, and the denominator degrees of
freedom is 18, then using Table A-9, F = 2.77. Iff is greater than F, reject the assumptions of equal
variances.
EPA QA/G-9 Final
QAOO Version 4-37 July 2000
-------
Box 4-27: An Example of Levene's Test
Four months of data on arsenic concentration were collected from six wells at a Superfund site. This data set is
shown in the table below. Before analyzing this data, it is important to determine if the variances of the six wells
are equal. Levene's test will be used to make this determination.
STEP 1 : The group mean for each well (x ,) is shown in the last row of the table below.
STEP 2:
STEP 3:
STEP 4:
STEP 5:
Arsenic Concentration (ppm)
Month Well 1 Well 2 Well 3 Well 4 Well 5 Well 6
1 22.90 2.00 2.0 7.84 24.90 0.34
2 3.09 1.25 109.4 9.30 1.30 4.78
3 35.70 7.80 4.5 25.90 0.75 2.85
4 4.18 52.00 2.5 2.00 27.00 1.20
Group Means x .,=16.47 x X3=29.6 x X5=13.49 x
2=15.76 4=11.26 6=2.29
To compute the absolute residuals zy in each well, the value 16.47 will be subtracted from Well 1
data, 15.76 from Well 2 data, 29.6 from Well 3 data, 11.26 from Well 4 data, 13.49 from Well 5
data, and 2.29 from Well 6 data. The resulting values are shown in the following table with the new
well means (z~) and the total mean z.
Residual Arsenic Concentration (ppm)
Month Well 1 Well 2 Well 3 Well 4 Well 5 Well 6
1 6.43 13.76 27.6 3.42 11.41 1.95
2 13.38 14.51 79.8 1.96 12.19 2.49
3 19.23 7.96 25.1 14.64 12.74 0.56
4 12.29 36.24 27.1 9.26 13.51 1.09
Residual Means z.,=12.83 z2=18.12 z3=39.9 z4=7.32 z5=12.46 z6=1.52
Total Residual Mean z = (1/6)(12.83 + 18.12 + 39.9 + 7.32 + 12.46+ 1.52) = 15.36
The sum of squares are: SSTOML = 6300.89, SS^^s = 3522.90, and SSERROR = 2777.99.
, ^WELLS' (k~V 3522.97(6-1) ^ ^
SSERRORl(N-k) 2777.997(24-6)
Using Table A-9 of Appendix A, the F statistic for 5 and 1 8 degrees of freedom with a = 0.05 is 2.77.
Since f=4.56 exceeds F05=2.77, the assumption of equal variances should be rejected.
EPA QA/G-9
QAOO Version
4-38
Final
July 2000
-------
4.6 TRANSFORMATIONS
Most statistical tests and procedures contain assumptions about the data to which they will
be applied. For example, some common assumptions are that the data are normally distributed;
variance components of a statistical model are additive; two independent data sets have equal
variance; and a data set has no trends over time or space. If the data do not satisfy such
assumptions, then the results of a statistical procedure or test may be biased or incorrect.
Fortunately, data that do not satisfy statistical assumptions may often be converted or transformed
mathematically into a form that allows standard statistical tests to perform adequately.
4.6.1 Types of Data Transformations
Any mathematical function that is applied to every point in a data set is called a
transformation. Some commonly used transformations include:
Logarithmic (Log X or Ln X): This transformation may be used when the original
measurement data follow a lognormal distribution or when the variance at each level of the
data is proportional to the square of the mean of the data points at that level. For
example, if the variance of data collected around 50 ppm is approximately 250, but the
variance of data collected around 100 ppm is approximately 1000, then a logarithmic
transformation may be useful. This situation is often characterized by having a constant
coefficient of variation (ratio of standard deviation to mean) over all possible data values.
The logarithmic base (for example, either natural or base 10) needs to be consistent
throughout the analysis. If some of the original values are zero, it is customary to add a
small quantity to make the data value non-zero as the logarithm of zero does not exist.
The size of the small quantity depends on the magnitude of the non-zero data and the
consequences of potentially erroneous inference from the resulting transformed data. As a
working point, a value of one tenth the smallest non-zero value could be selected. It does
not matter whether a natural (In) or base 10 (log) transformation is used because the two
transformations are related by the expression ln(X) = 2.303 log(X). Directions for
applying a logarithmic transformation with an example are given in Box 4-28.
Square Root (x): This transformation may be used when dealing with small whole
numbers, such as bacteriological counts, or the occurrence of rare events, such as
violations of a standard over the course of a year. The underlying assumption is that the
original data follow a Poisson-like distribution in which case the mean and variance of the
data are equal. It should be noted that the square root transformation overcorrects when
very small values and zeros appear in the original data. In these cases, \/X+ 1 is often
used as a transformation.
EPA QA/G-9 Final
QAOO Version 4-39 July 2000
-------
Box 4-28: Directions for Transforming Data and an Example
Let X.,, X2, . . . , Xn represent the n data points. To apply a transformation, simply apply the transforming
function to each data point. When a transformation is implemented to make the data satisfy some
statistical assumption, it will need to be verified that the transformed data satisfy this assumption.
Example: Transforming Loqnormal Data
A logarithmic transformation is particularly useful for pollution data. Pollution data are often skewed,
thus the log-transformed data will tend to be symmetric. Consider the data set shown below with 15 data
points. The frequency plot of this data (below) shows that the data are possibly lognormally distributed.
If any analysis performed with this data assumes normality, then the data may be logarithmically
transformed to achieve normality. The transformed data are shown in column 2. A frequency plot of the
transformed data (below) shows that the transformed data appear to be normally distributed.
Observed
X
0.22
3.48
6.67
2.53
1.11
0.33
1.64
1.37
« 6
? 5
O
•5 3
Transformed
ln(X)
-1.51
1.25
1.90
0.93
0.10
-1.11
0.50
- 0.31
Observed
X
0.47
0.67
0.75
0.60
0.99
0.90
0.26
Transformed
- ln(X)
-0.76
-0.40
-0.29
-0.51
-0.01
-0.11
-1.35
2 3
Observed Values
«6
O
"S *i
o
•5 3
2
-101
Transformed Values
EPA QA/G-9
QAOO Version
4-40
Final
July 2000
-------
Inverse Sine (Arcsine X): This transformation may be used for binomial proportions
based on count data to achieve stability in variance. The resulting transformed data are
expressed in radians (angular degrees). Special tables must be used to transform the
proportions into degrees.
Box-Cox Transformations: This transformation is a complex power transformation that
takes the original data and raises each data observation to the power lambda (A). A
logarithmic transformation is a special case of the Box-Cox transformation. The rationale
is to find A such that the transformed data have the best possible additive model for the
variance structure, the errors are normally distributed, and the variance is as constant as
possible over all possible concentration values. The Maximum Likelihood technique is
used to find A such that the residual error from fitting the theorized model is minimized.
In practice, the exact value of A is often rounded to a convenient value for ease in
interpretation (for example, A = -1.1 would be rounded to -1 as it would then have the
interpretation of a reciprocal transform). One of the drawbacks of the Box-Cox
transformation is the difficulty in physically interpreting the transformed data.
4.6.2 Reasons for Transforming Data
By transforming the data, assumptions that are not satisfied in the original data can be
satisfied by the transformed data. For instance, a right-skewed distribution can be transformed to
be approximately Gaussian (normal) by using a logarithmic or square-root transformation. Then
the normal-theory procedures can be applied to the transformed data. If data are lognormally
distributed, then apply procedures to logarithms of the data. However, selecting the correct
transformation may be difficult. If standard transformations do not apply, it is suggested that the
data user consult a statistician.
Another important use of transformations is in the interpretation of data collected under
conditions leading to an Analysis of Variance (ANOVA). Some of the key assumptions needed
for analysis (for example, additivity of variance components) may only be satisfied if the data are
transformed suitably. The selection of a suitable transformation depends on the structure of the
data collection design; however, the interpretation of the transformed data remains an issue.
While transformations are useful for dealing with data that do not satisfy statistical
assumptions, they can also be used for various other purposes. For example, transformations are
useful for consolidating data that may be spread out or that have several extreme values. In
addition, transformations can be used to derive a linear relationship between two variables, so that
linear regression analysis can be applied. They can also be used to efficiently estimate quantities
such as the mean and variance of a lognormal distribution. Transformations may also make the
analysis of data easier by changing the scale into one that is more familiar or easier to work with.
Once the data have been transformed, all statistical analysis must be performed on the
transformed data. No attempt should be made to transform the data back to the original form
because this can lead to biased estimates. For example, estimating quantities such as means,
EPA QA/G-9 Final
QAOO Version 4-41 July 2000
-------
variances, confidence limits, and regression coefficients in the transformed scale typically leads to
biased estimates when transformed back into original scale. However, it may be difficult to
understand or apply results of statistical analysis expressed in the transformed scale. Therefore, if
the transformed data do not give noticeable benefits to the analysis, it is better to use the original
data. There is no point in working with transformed data unless it adds value to the analysis.
4.7 VALUES BELOW DETECTION LIMITS
Data generated from chemical analysis may fall below the detection limit (DL) of the
analytical procedure. These measurement data are generally described as not detected, or
nondetects, (rather than as zero or not present) and the appropriate limit of detection is usually
reported. In cases where measurement data are described as not detected, the concentration of
the chemical is unknown although it lies somewhere between zero and the detection limit. Data
that includes both detected and non-detected results are called censored data in the statistical
literature.
There are a variety of ways to evaluate data that include values below the detection limit.
However, there are no general procedures that are applicable in all cases. Some general
guidelines are presented in Table 4-4. Although these guidelines are usually adequate, they should
be implemented cautiously.
Table 4-4. Guidelines for Analyzing Data with Nondetects
Percentage of
Nondetects
< 15%
15% -50%
> 50% - 90%
Section
4.7.1
4.7.2
4.7.3
Statistical Analysis
Method
Replace nondetects with
DL/2, DL, or a very small
number.
Trimmed mean, Cohen's
adjustment, Winsorized
mean and standard deviation.
Use tests for proportions
(Section 3. 2.2)
All of the suggested procedures for analyzing data with nondetects depend on the amount
of data below the detection limit. For relatively small amounts below detection limit values,
replacing the nondetects with a small number and proceeding with the usual analysis may be
satisfactory. For moderate amounts of data below the detection limit, a more detailed adjustment
is appropriate. In situations where relatively large amounts of data below the detection limit exist,
EPA QA/G-9 Final
QAOO Version 4 - 42 July 2000
-------
one may need only to consider whether the chemical was detected as above some level or not.
The interpretation of small, moderate, and large amounts of data below the DL is subjective.
Table 4-4 provides percentages to assist the user in evaluating their particular situation.
However, it should be recognized that these percentages are not hard and fast rules, but should be
based on judgement.
In addition to the percentage of samples below the detection limit, sample size influences
which procedures should be used to evaluate the data. For example, the case where 1 sample out
of 4 is not detected should be treated differently from the case where 25 samples out of 100 are
not detected. Therefore, this guidance suggests that the data analyst consult a statistician for the
most appropriate way to evaluate data containing values below the detection level.
4.7.1 Less than 15% Nondetects - Substitution Methods
If a small proportion of the observations are not detected, these may be replaced with a
small number, usually the detection limit divided by 2 (DL/2), and the usual analysis performed.
As a guideline, if 15% or fewer of the values are not detected, replace them with the method
detection limit divided by two and proceed with the appropriate analysis using these modified
values. If simple substitution of values below the detection limit is proposed when more than
15% of the values are reported as not detected, consider using nonparametric methods or a test of
proportions to analyze the data. If a more accurate method is to be considered, see Cohen's
Method (Section 4.7.2.1).
4.7.2 Between 15-50% Nondetects
4.7.2.1 Cohen's Method
Cohen's method provides adjusted estimates of the sample mean and standard deviation
that accounts for data below the detection level. The adjusted estimates are based on the
statistical technique of maximum likelihood estimation of the mean and variance so that the fact
that the nondetects are below the limit of detection but may not be zero is accounted for. The
adjusted mean and standard deviation can then be used in the parametric tests described in
Chapter 3 (e.g., the one sample t-test of Section 3.2.1). However, if more than 50% of the
observations are not detected, Cohen's method should not be used. In addition, this method
requires that the data without the nondetects be normally distributed and the detection limit is
always the same. Directions for Cohen's method are contained in Box 4-29; an example is given
in Box 4-30.
EPA QA/G-9 Final
QAOO Version 4 - 43 July 2000
-------
Box 4-29: Directions for Cohen's Method
Let X.,, X2, . . . , Xn represent the n data points with the first m values representing the data points above the
detection limit (DL). Thus, there are (n-m) data points are below the DL.
STEP 1: Compute the sample mean xd from the data above the detection limit:
m ,=
STEP 2: Compute the sample variance s2dfrom the data above the detection limit:
m- 1
s
STEP 3: Compute h = ^—^ and y = ——
" (Xd-DL)2
STEP 4: Use h and y in Table A-10 of Appendix A to determine A. For example, if h = 0.4 and y = 0.30, then
A = 0.6713. If the exact value of h and y do not appear in the table, use double linear interpolation
(Box 4-31) to estimate A.
STEP 5: Estimate the corrected sample mean, x, and sample variance, s2, to account for the data below the
detection limit, as follows: X = Xd - l(Xd - DL) and s2 = s] + l(Xd - DL)2.
Box 4-30: An Example of Cohen's Method
Sulfate concentrations were measured for 24 data points. The detection limit was 1,450 mg/L and 3 of the 24
values were below the detection level. The 24 values are 1850, 1760, < 1450 (ND), 1710, 1575, 1475, 1780,
1790, 1780, < 1450 (ND), 1790, 1800, < 1450 (ND), 1800, 1840, 1820, 1860, 1780, 1760, 1800, 1900, 1770,
1790, 1780 mg/L. Cohen's Method will be used to adjust the sample mean for use in a t-test to determine if the
mean is greater than 1600 mg/L.
STEP 1: The sample mean of the m = 21 values above the detection level is Xd = 1771.9
STEP 2: The sample variance of the 21 quantified values is s2d= 8593.69.
STEP 3: h = (24 - 21)/24 = 0.125 and Y = 8593.697(1771.9 - 1450)2 = 0.083
STEP 4: Table A-10 of Appendix A was used for h = 0.125 and y = 0.083 to find the value of A. Since the
table does not contain these entries exactly, double linear interpolation was used to estimate A =
0.149839 (see Box 4-31).
STEP 5: The adjusted sample mean and variance are then estimated as follows:
X= 1771.9 - 0.149839(1771.9-1450) = 1723.67 and
s2 = 8593.69 + 0.149839(1771.9-1450)2 = 24119.95
EPA QA/G-9 Final
QAOO Version 4 - 44 July 2000
-------
Box 4-31: Double Linear Interpolation
The details of the double linear interpolation are provided to assist in the use of Table A-10 of Appendix
A. The desired value for A corresponds to y = 0.083 and, h = 0.125 from Box 4-30, Step 3. The values
from Table A-10 for interpolatation are:
Y h = 0.10 h = 0.15
0.05 0.11431 0.17925
0.10 0.11804 0.18479
There are 0.05 units between 0.10 and 0.15 on the h-scale and 0.025 units between 0.10 and 0.125.
Therefore, the value of interest lies (0.025/0.05)100% = 50% of the distance along the interval between
0.10 and 0.15. To linearly interpolate between tabulated values on the h axis for y = 0.05, the range
between the values must be calculated, 0.17925 -0.11431 = 0.06494; the value that is 50% of the
distance along the range must be computed, 0.06494 x 0.50 = 0.03247; and then that value must be
added to the lower point on the tabulated values, 0.11431 + 0.03247 = 0.14678. Similarly for y = 0.10,
0.18479 - 0.11804 = 0.06675, 0.06675 x 0.50 = 0.033375, and 0.11804 + 0.033375 = 0.151415.
On the Y-axis there are 0.033 units between 0.05 and 0.083 and there are 0.05 units between 0.05 and
0.10. The value of interest (0.083) lies (0.033/0.05 x 100) = 66% of the distance along the interval
between 0.05 and 0.10, so 0.151415 - 0.14678 = 0.004635, 0.004635 * 0.66 = 0.003059. Therefore,
A = 0.14678 + 0.003059 = 0.149839.
4.7.2.2 Trimmed Mean
Trimming discards the data in the tails of a data set in order to develop an unbiased
estimate of the population mean. For environmental data, nondetects usually occur in the left tail
of the data so trimming the data can be used to adjust the data set to account for nondetects when
estimating a mean. Developing a 100p% trimmed mean involves trimming p% of the data in both
the lower and the upper tail. Note that p must be between 0 and .5 since p represents the portion
deleted in both the upper and the lower tail. After np of the largest values and np of the smallest
values are trimmed, there are n(l-2p) data values remaining. Therefore, the proportion trimmed
is dependent on the total sample size (n) since a reasonable amount of samples must remain for
analysis. For approximately symmetric distributions, a 25% trimmed mean (the midmean) is a
good estimator of the population mean. However, environmental data are often skewed (non-
symmetric) and in these cases a 15% trimmed mean performance may be a good estimator of the
population mean. It is also possible to trim the data only to replace the nondetects. For example,
if 3% of the data are below the detection limit, a 3% trimmed mean could be used to estimate the
population mean. Directions for developing a trimmed mean are contained in Box 4-32 and an
example is given in Box 4-33. A trimmed variance is rarely calculated and is of limited use.
4.7.2.3 Winsorized Mean and Standard Deviation
Winsorizing replaces data in the tails of a data set with the next most extreme data value.
For environmental data, nondetects usually occur in the left tail of the data. Therefore,
winsorizing can be used to adjust the data set to account for nondetects. The mean and standard
deviation can then be computed on the new data set. Directions for winsorizing data (and
revising the sample size) are contained in Box 4-34 and an example is given in Box 4-35
EPA QA/G-9 Final
QAOO Version 4 - 45 July 2000
-------
4.7.2.4 Atchison's Method
Previous adjustments to the mean and variance assumed that the data values really were
present but could not be recorded or "seen" as they were below the detection limit. In other
words, if the detection limit had been substantially lower, the data values would have been
recorded. There are however, cases where the data values are below the detection limit because
they are actually zero, the contaminant or chemical of concern being entirely absent. Such data
sets are actually a mixture - partly the assumed distribution (for example, a normal distribution)
and partly a number of zero values. Aitchison's Method is used in this situation to adjust the
mean and variance for the zero values.
Box 4-32: Directions for Developing a Trimmed Mean
LetX.,, X2, . . . , Xn represent the n data points. To develop a 100p% trimmed mean (0 < p < 0.5):
STEP 1 : Let t represent the integer part of the product np. For example, if p = .25 and n = 1 7,
np = (.25)(17) = 4.25, so t = 4.
STEP 2: Delete the t smallest values of the data set and the t largest values of the data set.
_ i n-2t
STEPS: Compute the arithmetic mean of the remaining n - 2t values: X =
n - 2t ,,
This value is the estimate of the population mean.
Box 4-33: An Example of the Trimmed Mean
Sulfate concentrations were measured for 24 data points. The detection limit was 1,450 mg/L and 3 of
the 24 values were below this limit. The 24 values listed in order from smallest to largest are: < 1450
(ND), < 1450 (ND), < 1450 (ND), 1475, 1575, 1710, 1760, 1760, 1770, 1780, 1780, 1780, 1780, 1790,
1790, 1790, 1800, 1800, 1800, 1820, 1840, 1850, 1860, 1900 mg/L. A 15% trimmed mean will be used
to develop an estimate of the population mean that accounts for the 3 nondetects.
STEP 1: Since np = (24)(.15) = 3.6, t = 3.
STEP 2: The 3 smallest values of the data set and the 3 largest values of the data set were deleted.
The new data set is: 1475, 1575, 1710, 1760, 1760, 1770, 1780, 1780, 1780, 1780, 1790,
1790, 1790, 1800, 1800, 1800, 1820, 1840 mg/L.
STEP 3: Compute the arithmetic mean of the remaining n-2t values:
X= i (1475 + ... + 1840) = 1755.56
24 - (2)(3)
Therefore, the 15% trimmed mean is 1755.56 mg/L.
EPA QA/G-9 Final
QAOO Version 4 - 46 July 2000
-------
Box 4-34: Directions for Developing a Winsorized
Mean and Standard Deviation
Let X.,, X2, . . . , Xn represent the n data points and m represent the number of data points above the
detection limit (DL), and hence n-m below the DL.
STEP 1: List the data in order from smallest to largest, including nondetects. Label these points X,.,}, X,
2)l. . ., X(n) (so that X(1) is the smallest, X(2) is the second smallest, and X(n) is the largest).
STEP 2: Replace the n-m nondetects with X, m +.,} and replace the n-m largest values with X, n _ m}.
STEP 3: Using the revised data set, compute the sample mean, x, and the sample standard deviation,
s:
X = - yXt and s
n
i= 1
sf) - nx2
n-1
STEP 4: The Winsorized mean x w is equal to x. The Winsorized standard deviation is
s(n- 1)
s =
(2m-n- 1)
Box 4-35: An Example of a Winsorized
Mean and Standard Deviation
Sulfate concentrations were measured for 24 data points. The detection limit was 1,450 mg/L and 3 of
the 24 values were below the detection level. The 24 values listed in order from smallest to largest are: <
1450 (ND), < 1450 (ND), < 1450 (ND), 1475, 1575, 1710, 1760, 1760, 1770, 1780, 1780, 1780, 1780,
1790, 1790, 1790, 1800, 1800, 1800, 1820, 1840, 1850, 1860, 1900 mg/L.
STEP 1: The data above are already listed from smallest to largest. There are n=24 samples, 21 above
DL, and n-m=3 nondetects.
STEP 2: The 3 nondetects were replaced with X, 4 ( and the 3 largest values were replaced with X, 21 >.
The resulting data set is: 1475, 1475, 1475, 1475, 1575, 1710, 1760, 1760, 1770, 1780, 1780,
1780, 1780, 1790, 1790, 1790, 1800, 1800, 1800, 1820, 1840, 1840, 1840, 1840 mg/L
STEP 3: For the new data set, x = 1731 mg/L and s = 128.52 mg/L.
STEP 4: The Winsorized mean xw = 1731 mg/L. The Winsorized sample standard deviation is:
s __ 128.52(24-1) = n388
2(21) -24-1
Aitchison's method for adjusting the mean and variance of the above the detection level
values works quite well provided the percentage of non-detects is between 15-50% of the total
number of values. Care must be taken when using Aitchison's adjustment to the mean and
standard deviation as the mean is reduced and variance increased. With such an effect it may
become very difficult to use the adjusted data for tests of hypotheses or for predicative purposes.
As a diagnostic tool, the relevance of Aitchison' adjustment can lead to an evaluation of the data
to determine if two populations are being sampled simultaneously: one population being
represented by a normal distribution, the other being simply blanks. In some circumstances, for
EPA QA/G-9 Final
QAOO Version 4 - 47 July 2000
-------
example investigating a hazardous site, it may be possible to relate the position of the sample
through a Posting Plot (Section 2.3.9) and determine if the target population has not been
adequately stratified. Directions for Aitchison's method are contained in Box 4-36; an example
(with a comparison to Cohen's method) is contained in Box 4-37.
Box 4-36: 11 Directions for Aitchison's Method to Adjust Means and Variances
Let X.,, X2, . .,Xm, . . . , Xn represent the data points where the first m values are above the detection limit (DL)
and the remaining (n-m) data points are below the DL
STEP 1: Using the data above the detection level, compute the sample mean,
1 \ m
Xd = — y Xj and the sample variance, sd = ——
m-l
STEP 2: Estimate the corrected sample mean, X = —Xd
n
2 m-l ^ , m(n-m)—2
j — ^ d d
and the sample variance n — i n(n — i)
Box 4-37: An Example of Aitchison's Method
The following data consist of 10 Methylene Chloride samples: 1.9,1.3, <1, 2.0, 1.9, <1, <1, <1, 1.6, and 1.7.
There are 7 values above the detection limit and 3 below, so m = 7 and n - m=3. Aitchison's method will be
used to estimate the mean and sample variance of this data.
STEP1: Xd=~2^ Xt =-(1.9 + 1.3 + 2.0 + 1.9 + 1.6 + 1.6 + 1.7) = 1.714
and s2d = -^ - ^ - = 0.05809
d 7-1
STEP 2: The corrected sample mean is then X = —(1.714) = 1.2
7 _ i
and the sample variance s2 = - (0 05809) H -- — (1 714) = 0 7242
10-1 3(3-1)
EPA QA/G-9 Final
QAOO Version 4 - 48 July 2000
-------
4.7.2.5
Selecting Between Atchison's Method or Cohen's Method
To determine if a data set is better adjusted by Cohen's method or Aitchison's method, a
simple graphical procedure using Normal Probability Paper (Section 2.3) can be used. Examples
for this procedures are given in Box 4-38 and an example is contained in Box 4-39.
Box 4-38: Directions for Selecting Between
Cohen's Method or Aitchison's Method
LetX.,, X2, . .,Xm, . . . , Xn represent the data points with the first m values are above the detection limit (DL) and
the remaining (n-m) data points are below the DL.
STEP 1: Use Box 2-19 to construct a Normal Probability Plot of all the data but only plot the values belonging
to those above the detection level. This is called the Censored Plot.
STEP 2: Use Box 2-19 to construct a Normal Probability Plot of only those values above the detection level.
This called the Detects only Plot.
STEP 3: If the Censored Plot is more linear than the Detects Only Plot, use Cohen's Method to estimate the
sample mean and variance. If the Detects Only Plot is more linear than the Censored Plot, then use
Aitchison's Method to estimate the sample mean and variance.
Box 4-39: Example of Determining Between
Cohen's Method and Aitchison's Method
In this example, 10 readings of Chlorobenzene were obtained from a monitoring well and submitted for
consideration for a permit: < 1, 1.9, 1.4, 1.5, < 1, 1.2, <1, 1.3, 1.9, 2.1 ppm. The data can be thought to be
independent readings.
Step 1: Using the directions in Box 2-19 the following is the
Censored Plot:
STEP 2: Using the directions in Box 2-19 the following is the
Detects only Plot:
STEP 3: Since the Censored Plots is more linear than the Detects Only Plot, Cohen's Method should be used
to estimate the sample mean and variance.
EPA QA/G-9
QAOO Version
4-49
Final
July 2000
-------
4.7.3 Greater than 50% Nondetects - Test of Proportions
If more than 50% of the data are below the detection limit but at least 10% of the
observations are quantified, tests of proportions may be used to test hypotheses using the data.
Thus, if the parameter of interest is a mean, consider switching the parameter of interest to some
percentile greater than the percent of data below the detection limit. For example, if 67% of the
data are below the DL, consider switching the parameter of interest to the 75th percentile. Then
the method described in 3.2.2 can be applied to test the hypothesis concerning the 75th percentile.
It is important to note that the tests of proportions may not be applicable for composite samples.
In this case, the data analyst should consult a statistician before proceeding with analysis.
If very few quantified values are found, a method based on the Poisson distribution may be
used as an alternative approach. However, with a large proportion of nondetects in the data, the
data analyst should consult with a statistician before proceeding with analysis.
4.7.4 Recommendations
If the number of sample observations is small (n<20), maximum likelihood methods can
produce biased results since it is difficult to assure that the underlying distribution appropriate and
the solutions to the likelihood equation for the parameters of interest are statistically consistent
only if the number of samples is large. Additionally, most methods will yield estimated parameters
with large estimation variance, which reduces the power to detect import differences from
standards or between populations. While these methods can be applied to small data sets, the user
should be cautioned that they will only be effective in detecting large departures from the null
hypothesis.
If the degree of censoring (the percentage of data below the detection limit) is relatively
low, reasonably good estimates of means, variances and upper percentiles can be obtained.
However, if the rate of censoring is very high (greater than 50%) then little can be done
statistically except to focus on some upper quantile of the contaminant distribution, or on some
proportion of measurements above a certain critical level that is at or above the censoring limit.
When the numerical standard is at or below one of the censoring levels and a one-sample
test is used, the most useful statistical method is to test whether the proportion of a population is
above (below) the standard is too large, or to test whether and upper quantile of the population
distribution is above the numerical standard. Table 4-5 gives some recommendation on which
statistical parameter to use when censoring is present in data sets for different sizes of the
coefficient of variation (Section 2.2.3).
When comparing two data sets with different censoring levels (i.e., different detection
limits), it is recommended that all data be censored at the highest censoring value present and a
non parametric test such as the Wilcoxon Rank Sum Test (Section 3.3.3.1) used to compare the
two data sets. There is a corresponding loss of statistical power but this can be minimized
through the use of large samples.
EPA QA/G-9 Final
QAOO Version 4-50 July 2000
-------
Table 4-5. Guidelines for Recommended Parameters for Different
Coefficient of Variations and Censoring
Assumed Coefficient
of Variation (CV)
Large: CV> 1.5
Medium: 0.530%)
Upper Percentile
Upper Percentile
Median
4.8 INDEPENDENCE
The assumption of independence of data is key to the validity of the false rejection and
false acceptance error rates associated with the selected statistical test. When data are truly
independent between themselves, the correlation between data points is by definition zero and the
selected statistical tests work with the desired chosen decision error rates (given appropriate the
assumptions have been satisfied). When correlation (usually positive) exists, the effectiveness of
statistical tests is diminished. Environmental data are particularly susceptible to correlation
problems due to the fact that such environmental data are collected under a spatial pattern (for
example a grid) or sequentially over time (for example, daily readings from a monitoring station).
The reason non-independence is an issue for statistical testing situations is that if
observations are positively correlated over time or space, then the effective sample size for a test
tends to be smaller than the actual sample size - i.e., each additional observation does not provide
as much "new" information because its value is partially determined by (or a function of) the value
of adjacent observations. This smaller effective sample size means that the degrees of freedom for
the test statistic is less, or equivalently, the test is not as powerful as originally thought. In
addition to affecting the false acceptance error rate, applying the usual tests to correlated data
tends to result in a test whose actual significance level (false rejection error rate) is larger than the
nominal error rate.
When observations are correlated, estimates of the variance that are used in test statistic
formulas are often understated. For example, consider the mean of a series of n temporally-
ordered observations. If these observations are independent, then the variance of the mean is
o2/n, where a2 is the variance of individual observations (see Section 2.2.3). However, if the
observations are not independent and the correlation between successive observations is p (for
example, the correlation between the first and second observation is p, between first and third
observations is p2, between first and fourth observations is p3, etc.), then the variance of the mean
increases to
EPA QA/G-9
QAOO Version
4-51
Final
July 2000
-------
VAR(X) = o\l + q\where q=~
n k=\
which will tend to be larger than o2/n if the correlations (on average) are positive. If one conducts
a t-test at the significance level, using the usual formula for the estimated variance (Box 2-3), then
the actual significance level can be approximately double what was expected even for relatively
low values of p.
One of the most effective ways to determine statistical independence is through use of the
Rank von Neumann Test. Directions for this test are given in Box 4-40 and an example in
contained in Box 4-41. Compared to other tests of statistical independence, the rank von
Neumann test has been shown to be more powerful over a wide variety of cases. It is also a
reasonable test when the data really follow a Normal distribution. In that case, the efficiency of
the test is always close to 90 percent when compared to the von Neumann ratio computed on the
original data instead of the ranks. This means that very little effectiveness is lost by always using
the ranks in place of the original concentrations; the rank von Neumann ration should still
correctly detect non-independent data.
Box 4-40: Directions for the Rank von Neumann Test
LetX.,, X2, . . . , Xn represent the data values collected in sequence over equally spaced periods of time.
Step 1. Order the data measurements from smallest to largest and assign a unique rank (H) to each
measurement (See Box 3-20 Then list the observations and their corresponding ranks in the order that
sampling occurred(i.e., by sampling event or time order.)
Step 2. Using the list of ranks, rh for the sampling periods i=1, 2, ..., n, compute the rank von
Neumann ratio:
n(n2 -1)/12
Step 3: Use Table A-15 of Appendix A to determine the lower critical point of the rank von Neumann ration
using the sample size, n, and the desired significance level, a. If the computed ratio, v, is smaller than
this critical point, conclude that the data series is strongly auto correlated. If not, the data may be
mildly correlated, but there is no statistically significant evidence to reject the hypothesis of
independence. Therefore, the data should be regarded as independent in subsequent statistical testing.
Note: if the rank von Neumann ratio test indicates significant evidence of dependence in the data, a
statistician should be consulted before further analysis is performed.
EPA QA/G-9 Final
QAOO Version 4 - 52 July 2000
-------
Th
St
St
St
Box 4-41: An Example of the Rank von Neumann Test
le following are hourly readings from a discharge monitor: hourly readings from a discharge monitor.
fime:
Reading
Rank
ep 1: Tht
ep2: v
ep 3: Us
the
12:00
6.5
7
3 ranks
n
I
/' =
13:00
6.6
8.5
14:00
6.7
10
are displayed
(r- ~ r- ,
2 ' l~l
15:00
6.4
5.5
16:00
6.3
3.5
17:00
6.4
5.5
18:00
6.2
1.5
in the table above and the time
)2
(8.5- 7)2 + (10-8
19:00
6.2
1.5
periods
5)2 + .
20:00
6.3
3.5
21:00
6.6
8.5
22:00 23:00
6.8 6.9
11 12
24:00
7.0
13
were labeled 1 through 13.
..+ (13- 12)2 _
«O2-1)/12 13(132-1)/12
ng Table A-15 of Appendix A with a = 0.05 , the lower critical point is 1.17.
hypothesis that the data are independent must be rejected.
- U.t/J
Since v = 0.473 < 1.17,
EPA QA/G-9
QAOO Version
4-53
Final
July 2000
-------
This page is intentionally blank.
EPA QA/G-9 Final
QAOO Version 4 - 54 July 2000
-------
CHAPTER 5
STEP 5: DRAW CONCLUSIONS FROM THE DATA
THE DATA QUALITY ASSESSMENT PROCESS
Review DQOs and Sampling Design
Conduct Preliminary Data Review
Select the Statistical Test
Verify the Assumptions
Draw Conclusions From the Data
DRAW CONCLUSIONS FROM THE DATA
Purpose
Conduct the hypothesis test and interpret the results
in the context of the data user's objectives.
Activities
• Perform the Statistical Hypothesis Test
• Draw Study Conclusions.
• Evaluate Performance of the Sampling Design
Tools
• Issues in hypothesis testing related to understanding
and communicating the test results
Step 5: Draw Conclusions from the Data
! Perform the calculations for the statistical hypothesis test.
P Perform the calculations and document them clearly.
P If anomalies or outliers are present in the data set, perform the calculations
with and without the questionable data.
! Evaluate the statistical test results and draw conclusions.
P If the null hypothesis is rejected, then draw the conclusions and document the
analysis.
P If the null hypothesis is not rejected, verify whether the tolerable limits on
false acceptance decision errors have been satisfied. If so, draw conclusions
and document the analysis; if not, determine corrective actions, if any.
P Interpret the results of the test.
! Evaluate the performance of the sampling design if the design is to be used
again.
P Evaluate the statistical power of the design over the full range of parameter
values; consult a statistician as necessary.
EPA QA/G-9
QAOO Version
5- 1
Final
July 2000
-------
List of Boxes
Box 5-1: Checking Adequacy of Sample Size for a One-
Sample t-Test for Simple Random Sampling 5-5
Box 5-2: Example of Power Calculations for the One-Sample Test of a Single Proportion . . 5 - 6
Box 5-3: Example of a Comparison of Two Variances
which is Statistically but not Practically Significant 5-9
Box 5-4: Example of a Comparison of Two Biases 5-10
List of Figures
Page
Figure 5-1. Illustration of Unbiased versus Biased Power Curves 5-11
EPA QA/G-9 Final
QAOO Version 5 - 2 July 2000
-------
CHAPTER 5
STEP 5: DRAW CONCLUSIONS FROM THE DATA
5.1 OVERVIEW AND ACTIVITIES
In this final step of the DQA, the analyst performs the statistical hypothesis test and draws
conclusions that address the data user's objectives. This step represents the culmination of the
planning, implementation, and assessment phases of the data operations. The data user's planning
objectives will have been reviewed (or developed retrospectively) and the sampling design
examined in Step 1. Reports on the implementation of the sampling scheme will have been
reviewed and a preliminary picture of the sampling results developed in Step 2. In light of the
information gained in Step 2, the statistical test will have been selected in Step 3. To ensure that
the chosen statistical methods are valid, the key underlying assumptions of the statistical test will
have been verified in Step 4. Consequently, all of the activities conducted up to this point should
ensure that the calculations performed on the data set and the conclusions drawn here in Step 5
address the data user's needs in a scientifically defensible manner. This chapter describes the main
activities that should be conducted during this step. The actual procedures for implementing
some commonly used statistical tests are described in Step 3, Select the Statistical Test.
5.1.1 Perform the Statistical Hypothesis Test
The goal of this activity is to conduct the statistical hypothesis test. Step-by-step
directions for several commonly used statistical tests are described in Chapter 3. The calculations
for the test should be clearly documented and easily verifiable. In addition, the documentation of
the results of the test should be understandable so that the results can be communicated
effectively to those who may hold a stake in the resulting decision. If computer software is used
to perform the calculations, ensure that the procedures are adequately documented, particularly if
algorithms have been developed and coded specifically for the project.
The analyst should always exercise best professional judgment when performing the
calculations. For instance, if outliers or anomalies are present in the data set, the calculations
should be performed both with and without the questionable data to see what effect they may
have on the results.
5.1.2 Draw Study Conclusions
The goal of this activity is to translate the results of the statistical hypothesis test so that
the data user may draw a conclusion from the data. The results of the statistical hypothesis test
will be either:
(a) reject the null hypothesis, in which case the analyst is concerned about a possible
false rejection decision error; or
EPA QA/G-9 Final
QAOO Version 5 - 3 July 2000
-------
(b) fail to reject the null hypothesis, in which case the analyst is concerned about a
possible false acceptance decision error.
In case (a), the data have provided the evidence needed to reject the null hypothesis, so
the decision can be made with sufficient confidence and without further analysis. This is because
the statistical test based on the classical hypothesis testing philosophy, which is the approach
described in prior chapters, inherently controls the false rejection decision error rate within the
data user's tolerable limits, provided that the underlying assumptions of the test have been verified
correctly.
In case (b), the data do not provide sufficient evidence to reject the null hypothesis, and
the data must be analyzed further to determine whether the data user's tolerable limits on false
acceptance decision errors have been satisfied. One of two possible conditions may prevail:
(1) The data do not support rejecting the null hypothesis and the false acceptance
decision error limits were satisfied. In this case, the conclusion is drawn in favor
of the null hypothesis, since the probability of committing a false acceptance
decision error is believed to be sufficiently small in the context of the current study
(see Section 5.2).
(2) The data do not support rejecting the null hypothesis, and the false acceptance
decision error limits were not satisfied. In this case, the statistical test was not
powerful enough to satisfy the data user's performance criteria. The data user may
choose to tolerate a higher false acceptance decision error rate than previously
specified and draw the conclusion in favor of the null hypothesis, or instead take
some form of corrective action, such as obtaining additional data before drawing a
conclusion and making a decision.
When the test fails to reject the null hypothesis, the most thorough procedure for verifying
whether the false acceptance decision error limits have been satisfied is to compute the estimated
power of the statistical test, using the variability observed in the data. Computing the power of
the statistical test across the full range of possible parameter values can be complicated and
usually requires specialized software. Power calculations are also necessary for evaluating the
performance of a sampling design. Thus, power calculations will be discussed further in Section
5.1.3.
A simpler method can be used for checking the performance of the statistical test. Using
an estimate of variance obtained from the actual data or upper 95% confidence limit on variance,
the sample size required to satisfy the data user's objectives can be calculated retrospectively. If
this theoretical sample size is less than or equal to the number of samples actually taken, then the
test is sufficiently powerful. If the required number of samples is greater than the number actually
collected, then additional samples would be required to satisfy the data user's performance criteria
for the statistical test. An example of this method is contained in Box 5-1. The equations
EPA QA/G-9 Final
QAOO Version 5 - 4 July 2000
-------
Box 5-1: Checking Adequacy of Sample Size for a One-
Sample t-Test for Simple Random Sampling
In Box 3-1, the one-sample t-test was used to test the hypothesis H0: u < 95 ppm vs. HA: u > 95 ppm. DQOs
specified that the test should limit the false rejection error rate to 5% and the false acceptance error rate to 20%
if the true mean were 105 ppm. A random sample of size n = 9 had sample mean x = 99.38 ppm and standard
deviation s = 10.41 ppm. The null hypothesis was not rejected. Assuming that the true value of the standard
deviation was equal to its sample estimate 10.41 ppm, it was found that a sample size of 9 would be required,
which validated the sample size of 9 which had actually been used.
The distribution of the sample standard deviation is skewed with a long right tail. It follows that the chances
are greater than 50% that the sample standard deviation will underestimate the true standard deviation. In
such a case it makes sense to build in some conservatism, for example, by using an upper 90% confidence
limit for o in Step 5 of Box 3-12. Using Box 4-22 and n -1 = 8 degrees of freedom, it is found that L = 3.49, so
that an upper 90% confidence limit for the true standard deviation is
sj[(n - 1)/Z] = 10.41^/8/3.49 = 15.76
Using this value for s in Step 5 of Box 3-1 leads to the sample size estimate of 17. Hence, a sample size of at
least 17 should be used to be 90% sure of achieving the DQOs. Since it is generally desirable to avoid the
need for additional sampling, it is advisable to conservatively estimate sample size in the first place. In cases
where DQOs depend on a variance estimate, this conservatism is achieved by intentionally overestimating the
variance.
required to perform these calculations have been provided in the detailed step-by-step instructions
for each hypothesis test procedure in Chapter 3.
5.1.3 Evaluate Performance of the Sampling Design
If the sampling design is to be used again, either in a later phase of the current study or in
a similar study, the analyst will be interested in evaluating the overall performance of the design.
To evaluate the sampling design, the analyst performs a statistical power analysis that describes
the estimated power of the statistical test over the range of possible parameter values. The power
of a statistical test is the probability of rejecting the null hypothesis when the null hypothesis is
false. The estimated power is computed for all parameter values under the alternative hypothesis
to create a power curve. A power analysis helps the analyst evaluate the adequacy of the
sampling design when the true parameter value lies in the vicinity of the action level (which may
not have been the outcome of the current study). In this manner, the analyst may determine how
well a statistical test performed and compare this performance with that of other tests.
The calculations required to perform a power analysis can be relatively complicated,
depending on the complexity of the sampling design and statistical test selected. Box 5-2
illustrates power calculations for a test of a single proportion, which is one of the simpler cases.
A further discussion of power curves (performance curves) is contained in the Guidance for Data
Quality Objectives (QA/G-4) (EPA 1994).
EPA QA/G-9 Final
QAOO Version 5 - 5 July 2000
-------
Box 5-2: Example of Power Calculations for the One-Sample Test of a Single Proportion
This box illustrates power calculations for the test of H0: P > .20 vs. HA: P < .20, with a false rejection error rate
of 5% when P=.20 presented in Boxes 3-10 and 3-11 The power of the test will be calculated assuming P., =
.15 and before any data are available. Since nP., and n(1-P.,) both exceed 4, the sample size is large enough
for the normal approximation, and the test can be carried out as in steps 3 and 4 of Box 3-10
STEP 1: Determine the general conditions for rejection of the null hypothesis. In this case, the null
hypothesis is rejected if the sample proportion is sufficiently smaller than P0. (Clearly, a sample
proportion above P0 cannot cast doubt on H0.) By steps 3 and 4 of Box 3-10 and 3-3 H0 is rejected
if
P + -5/n ~ Po ^
Here p is the sample proportion, Q0 = 1 - P0, n is the sample size, and z^_a is the critical value such
that 100(1-a)% of the standard normal distribution is below z^_a. This inequality is true if
p + .5/n < P0 - z^JP^JTt.
STEP 2: Determine the specific conditions for rejection of the null hypothesis if P., (=1-Q.,) is the true value of
the proportion P. The same operations as are used in step 3 of Box 3-10 are performed on both
sides of the above inequality. However, P0 is replaced by P., since it is assumed that P., is the true
proportion. These operations make the normal approximation applicable. Hence, rejection occurs if
p + .5/n - P1 P0 - P1 - Zl_tt^P0Q0/n _ 20 - .15 - ,..-,.vv.^v.uy, „. = _Q
STEP 3: Find the probability of rejection if P., is the true proportion. By the same reasoning that led to the
test in steps 3 and 4 of Boxes 3-10 and 3-11 the quantity on the left-hand side of the above
inequality is a standard normal variable. Hence the power at P., = .15 (i.e., the probability of
rejection of H0 when .15 is the true proportion) is the probability that a standard normal variable is
less than -0.55. In this case, the probability is approximately 0.3 (using the last line from Table A-1
of Appendix A) which is fairly small.
5.2 INTERPRETING AND COMMUNICATING THE TEST RESULTS
Sometimes difficulties may arise in interpreting or explaining the results of a statistical test.
One reason for such difficulties may stem from inconsistencies in terminology; another may be due
to a lack of understanding of some of the basic notions underlying hypothesis tests. As an
example, in explaining the results to a data user, an analyst may use different terminology than
that appearing in this guidance. For instance, rather than saying that the null hypothesis was or
was not rejected, analysts may report the result of a test by saying that their computer output
shows a p-value of 0.12. What does this mean? Similar problems of interpretation may occur
when the data user attempts to understand the practical significance of the test results or to
explain the test results to others. The following paragraphs touch on some of the philosophical
issues related to hypothesis testing which may help in understanding and communicating the test
results.
EPA QA/G-9 Final
QAOO Version 5 - 6 July 2000
-------
5.2.1 Interpretation of p-Values
The classical approach for hypothesis tests is to prespecify the significance level of the
test, i.e., the Type I decision error rate a. This rate is used to define the decision rule associated
with the hypothesis test. For instance, in testing whether the population mean |i exceeds a
threshold level (e.g., 100 ppm), the test statistic may depend on x, an estimate of |i. Obtaining an
estimate X that is greater than 100 ppm may occur simply by chance even if the true mean |i is less
than or equal to 100; however, if X is "much larger" than 100 ppm, then there is only a small
chance that the null hypothesis H0 (|i < 100 ppm) is true. Hence the decision rule might take the
form "reject H0 if x exceeds 100 + C", where C is a positive quantity that depends on a (and on
the variability of x). If this condition is met, then the result of the statistical test is reported as
"reject H0"; otherwise, the result is reported as "do not reject H0."
An alternative way of reporting the result of a statistical test is to report its p-value, which
is defined as the probability, assuming the null hypothesis to be true, of observing a test result at
least as extreme as that found in the sample. Many statistical software packages report p-values,
rather than adopting the classical approach of using a prespecified false rejection error rate. In the
above example, for instance, the p-value would be the probability of observing a sample mean as
large as X (or larger) if in fact the true mean was equal to 100 ppm. Obviously, in making a
decision based on the p-value, one should reject H0 when p is small and not reject it if p is large.
Thus the relationship between p-values and the classical hypothesis testing approach is that one
rejects H0 if the p-value associated with the test result is less than a. If the data user had chosen
the false rejection error rate as 0.05 a priori and the analyst reported a p-value of 0.12, then the
data user would report the result as "do not reject the null hypothesis;" if the p-value had been
reported as 0.03, then that person would report the result as "reject the null hypothesis." An
advantage of reporting p-values is that they provide a measure of the strength of evidence for or
against the null hypothesis, which allows data users to establish their own false rejection error
rates. The significance level can be interpreted as that p-value (a) that divides "do not reject H0"
from "reject H0."
5.2.2 "Accepting" vs. "Failing to Reject" the Null Hypothesis
As noted in the paragraphs above, the classical approach to hypothesis testing results in
one of two conclusions: "reject H0" (called a significant result) or "do not reject H0" (a
nonsignificant result). In the latter case one might be tempted to equate "do not reject H0" with
"accept H0." This terminology is not recommended, however, because of the philosophy
underlying the classical testing procedure. This philosophy places the burden of proof on the
alternative hypothesis, that is, the null hypothesis is rejected only if the evidence furnished by the
data convinces us that the alternative hypothesis is the more likely state of nature. If a
nonsignificant result is obtained, it provides evidence that the null hypothesis could sufficiently
account for the observed data, but it does not imply that the hypothesis is the only hypothesis that
could be supported by the data. In other words, a highly nonsignificant result (e.g., a p-value of
0.80) may indicate that the null hypothesis provides a reasonable model for explaining the data,
but it does not necessarily imply that the null hypothesis is true. It may, for example, simply
EPA QA/G-9 Final
QAOO Version 5 - 7 July 2000
-------
indicate that the sample size was not large enough to establish convincingly that the alternative
hypothesis was more likely. When the phrase "accept H0" is encountered, it must be considered
as "accepted with the preceding caveats."
5.2.3 Statistical Significance vs. Practical Significance
There is an important distinction between these two concepts. Statistical significance
simply refers to the result of the hypothesis test: Was the null hypothesis rejected? The likelihood
of achieving a statistically significant result depends on the true value of the population parameter
being tested (for example, |i), how much that value deviates from the value hypothesized under
the null hypothesis (for example, |i0), and on the sample size. This dependence on (|i - |i0) is
depicted by the power curve associated with the test (Section 5.1.3). A steep power curve can be
achieved by using a large sample size; this means that there will be a high likelihood of detecting
even a small difference. On the other hand, if small sample sizes are used, the power curve will be
less steep, meaning that only a very large difference between ji and |i0 will be detectable with high
probability. Hence, suppose one obtains a statistically significant result but has no knowledge of
the power of the test. Then it is possible, in the case of the steep power curve, that one may be
declaring significance (claiming ji > |i0, for example) when the actual difference, from a practical
standpoint, may be inconsequential. Or, in the case of the slowly increasing power curve, one
may not find a significant result even though a "large" difference between ji and |i0 exists. Neither
of these situations is desirable: in the former case, there has been an excess of resources
expended, whereas in the latter case, a false acceptance error is likely and has occurred.
But how large a difference between the parameter and the null value is of real importance?
This relates to the concept of practical significance. Ideally, this question is asked and answered
as part of the DQO process during the planning phase of the study. Knowing the magnitude of
the difference that is regarded as being of practical significance is important during the design
stage because this allows one, to the extent that prior information permits, to determine a
sampling plan of type and size that will make the magnitude of that difference commensurate with
a difference that can be detected with high probability. From a purely statistical design
perspective, this can be considered to be main purpose of the DQO process. With such planning,
the likelihood of encountering either of the undesirable situations mentioned in the prior
paragraph can be reduced. Box 5-3 contains an example of a statistically significant but fairly
inconsequential difference.
5.2.4 Impact of Bias on Test Results
Bias is defined as the difference between the expected value of a statistic and a population
parameter. It is relevant when the statistic of interest (e.g., a sample average x) is to be used as
an estimate of the parameter (e.g., the population mean ji). For example, the population
parameter of interest may be the average concentration of dioxin within the given bounds of a
hazardous waste site, and the statistic might be the sample average as obtained from a random
sample of points within those bounds. The expected value of a statistic can be interpreted as
supposing one repeatedly implemented the particular sampling design a very large number of
EPA QA/G-9 Final
QAOO Version 5 - 8 July 2000
-------
times and calculated the statistic of interest in each case. The average of the statistic's values
would then be regarded as its expected value. Let E denote the expected value of x and denote
the relationship between the expected value and the parameter, |i, as E = |i + b where b is the
bias. For instance, if the bias occurred due to incomplete recovery of an analyte (and no
adjustment is made), then b = (R-100)|o/100, where R denotes the percent recovery. Bias may
also occur for other reasons, such as lack of coverage of the entire target population (e.g., if only
the drums within a storage site that are easily accessible are eligible for inclusion in the sample,
then inferences to the entire group of drums may be biased). Moreover, in cases of incomplete
coverage, the magnitude and direction of the bias may be unknown. An example involving
comparison of the biases of two measurement methods is contained in Box 5-4.
Box 5-3: Example of a Comparison of Two Variances
which is Statistically but not Practically Significant
The quality control (QC) program associated with a measurement system provides important information on
performance and also yields data which should be taken into account in some statistical analyses. The QC
program should include QC check samples, i.e., samples of known composition and concentration which are
run at regular frequencies. The term precision refers to the consistency of a measurement method in repeated
applications under fixed conditions and is usually equated with a standard deviation. The appropriate standard
deviation is one which results from applying the system to the same sample over a long period of time.
This example concerns two methods for measuring ozone in ambient air, an approved method and a new
candidate method. Both methods are used once per week on a weekly basis for three months. Based on 13
analyses with each method of the mid-range QC check sample at 100 ppb, the null hypothesis of the equality
of the two variances will be tested with a false rejection error rate of 5% or less. (If the variances are equal,
then the standard deviations are equal.) Method 1 had a sample mean of 80 ppb and a standard deviation of 4
ppb. Method 2 had a mean of 90 ppb and a standard deviation of 8 ppb. The Shapiro-Wilks test did not reject
the assumption of normality for either method. Applying the F-test of Box 4-23, the F ratio is 82/42 = 2. Using
12 degrees of freedom for both the numerator and denominator, the F ratio must exceed 3.28 in order to reject
the hypothesis of equal variances (Table A-9 of Appendix A). Since 4 > 3.28, the hypothesis of equal variances
is rejected, and it is concluded that method 1 is significantly more precise than method 2.
In an industrialized urban environment, the true ozone levels at a fixed location and time of day are known to
vary over a period of months with a coefficient of variation of at least 100%. This means that the ratio of the
standard deviation (SD) to the mean at a given location is at least 1. For a mean of 100 ppb, the standard
deviation over time for true ozone values at the location would be at least 100 ppb. Relative to this degree of
variability, a difference between measurement error standard deviations of 4 or 8 ppb is negligible. The overall
variance, incorporating the true process variability and measurement error, is obtained by adding the individual
variances. For instance, if measurement error standard deviation is 8 ppb, then the total variance is (100
ppb)(100 ppb) + (8 ppb)(8 ppb). Taking the square root of the variance gives a corresponding total standard
deviation of 100.32 ppb. For a measurement error standard deviation of 4 ppb, the total standard deviation
would be 100.08 ppb. From a practical standpoint, the difference in precision between the two methods is
insignificant for the given application, despite the finding that there is a statistically significant difference
between the variances of the two methods.
EPA QA/G-9 Final
QAOO Version 5 - 9 July 2000
-------
Box 5-4: Example of a Comparison of Two Biases
This example is a continuation of the ozone measurement comparison described in Box 5-3. Let x and
sx denote the sajriple mean and standard deviation of measurement method 1 applied to the QC check
sample, andjet Y and SY denote the sample mean and standard deviation of method 2. Then x = 80 ppb,
sx = 4 ppb, Y = 90 ppb and SY = 8 ppb. The estimated biases are x - T = 80 -100 = -20 ppb for method
1, and Y - T = 90-100 = 10 ppb for method 2, since 100 ppb is the true value T. That is, method 1 seems
to underestimate by 20 ppb, and method 2 seems to underestimate by 10 ppb. Let U., and u2 be the
underlying mean concentrations for measurement methods 1 and 2 applied to the QC check sample.
These means correspond to the average results which would obtain by applying each method a large
number of times to the QC check sample, over a long period of time.
A two-sample t-test (Boxes 3-14 and 3-16) can be used to test for a significant difference between these
two biases. In this case, a two-tailed test of the null hypothesis H0: U., - u2 = 0 against the alternative HA:
U., - u2t 0 is appropriate, because there is no a priori reason (in advance of data collection) to suspect
that one measurement method is superior to the other. (In general, hypotheses should not be tailored to
data.) Note that the difference between the two biases is the same as the difference (p., - u2) between the
two underlying means of the measurement methods. The test will be done to limit the false rejection
error rate to 5% if the two means are equal.
STEP 1: x = 80 ppb, sx = 4 ppb, Y = 90 ppb, SY = 8 ppb.
STEP 2: From Box 5-3, it is known that the methods have significantly different variances, so that
Sattherthwaite's t-test should be used. Therefore,
JNE
2
SX ^
m
2
. SY
n
42 82
— + — = 2.48
13 13
STEP 3: / =
sx
m
4
SX
m2(m - 1)
SY
n
2
4
SY
n2 (n - 1)
13
44
132 12
13
2
84
132 12
= 17.65.
STEP 4:
Rounding down to the nearest integer gives f = 17. For a two-tailed test, the critical value is t
1-0/2 = t.975 = 2.110, from Table A-1 of Appendix A.
t =
X - Y
80 - 90
2.48
= -4.032
STEP 5: For a two-tailed test, compare *t* with t^ = 2.11. Since 4.032 > 2.11, reject the null
hypothesis and conclude that there is a significant difference between the two method biases,
in favor of method 2.
This box illustrates a situation involving two measurement methods where one method is more precise,
but also more biased, than the other. If no adjustment for bias is made, then for many purposes, the
less biased, more variable method is preferable. However, proper bias adjustment can make both
methods unbiased, so that the more precise method becomes the preferred method. Such adjustments
can be based on QC check sample results, if the QC check samples are regarded as representative of
environmental samples involving sufficiently similar analytes and matrices.
EPA QA/G-9
QAOO Version
5- 10
Final
July 2000
-------
1
OJ
| 0.8
ro ID
p c 06
11
If °'4
£• 8
J5 m 0.2
o
Q_
0
8
-
Unbiased
-
— -T^/7
/
1
I I
/ /
/ /
^^S
/^^
/
/
"""
/ Biased Upward
0 100 120 140 160
True Value of the Parameter
In the context of hypothesis testing, the impact of bias can be quite severe in some
circumstances. This can be illustrated by comparing the power curve of a test when bias is not
present with a power curve for the same test when bias is present. The basic influence of bias is
to shift the former "no bias" curve to
the right or left, depending on the
direction of the bias. If the bias is
constant, then the second curve will be
an exact translation of the former
curve; if not, there will be a change in
the shape of the second curve in
addition to the translation. If the
existence of the bias is unknown, then
the former power curve will be
regarded as the curve that determines
the properties of the test when in fact Figure 5_L illustration of Unbiased versus Biased
the second curve will be the one that Power Curves
actually represents the test's power.
For example, in Figure 5-1 when the true value of the parameter is 120, the "no bias" power is
0.72 but the true power (the biased power) is only 0.4, a substantial difference. Since bias is
not impacted by changing the sample size, while the precision of estimates and the power of tests
increases with sample size, the relative importance of bias becomes more pronounced when the
sample size increases (i.e., when one makes the power curve steeper). Similarly, if the same
magnitude of bias exists for two different sites, then the impact on testing errors will be more
severe for the site having the smaller inherent variability in the characteristic of interest (i.e., when
bias represents a larger portion of total variability).
To minimize the effects of bias: identify and document sources of potential bias; adopt
measurement procedures (including specimen collection, handling, and analysis procedures) that
minimize the potential for bias; make a concerted effort to quantify bias whenever possible; and
make appropriate compensation for bias when possible.
5.2.5 Quantity vs. Quality of Data
The above conclusions imply that, if compensation for bias cannot be made and if
statistically-based decisions are to be made, then there will be situations in which serious
consideration should be given to using an imprecise (and perhaps relatively inexpensive) chemical
method having negligible bias as compared to using a very precise method that has even a
moderate degree of bias. The tradeoff favoring the imprecise method is especially relevant when
the inherent variability in the population is very large relative to the random measurement error.
For example, suppose a mean concentration for a given spatial area (site) is of interest and
that the coefficient of variation (CV) characterizing the site's variability is 100%. Let method A
denote an imprecise method, with measurement-error CV of 40%, and let method B denote a
highly precise method, with measurement-error CV of 5%. The overall variability, or total
EPA QA/G-9
QAOO Version
5- 11
Final
July 2000
-------
variability, can essentially be regarded as the sum of the spatial variability and the measurement
variability. These are obtained from the individual CVs in the form of variances. As CV equals
standard deviation divided by mean, it follows that the site standard deviation is then the CV times
the mean. Thus, for the site, the variance is l.OO2 x mean2; for method A, the variance is 0.402 x
mean2; and for method B, the variance is 0.052 x mean2. The overall variability when using
method A is then (l.OO2 x mean2) + (0.402 x mean2) = 1.16 x mean2, and when using method B,
the variance is (l.OO2 x mean2) + (0.052 x mean2) = 1.0025 mean2. It follows that the overall CV
when using each method is then (1.077 x mean) / mean = 107.7% for method A, and (1.001 x
mean) / mean = 100.1% for method B.
Now consider a sample of 25 specimens from the site. The precision of the sample mean
can then be characterized by the relative standard error (RSE) of the mean (which for the simple
random sample situation is simply the overall CV divided by the square root of the sample size).
For Method A, RSE = 21.54%; for method B, RSE = 20.02%. Now suppose that the imprecise
method (Method A) is unbiased, while the precise method (Method B) has a 10% bias (e.g., an
analyte percent recovery of 90%). An overall measure of error that reflects how well the sample
mean estimates the site mean is the relative root mean squared error (RRMSE):
RRMSE = \I(RB)2 + (RSE)2
where RB denotes the relative bias (RB = 0 for Method A since it is unbiased and RB = ±10% for
Method B since it is biased) and RSE is as defined above. The overall error in the estimation of
the population mean (the RRMSE) would then be 21.54% for Method A and 22.38% for Method
B. If the relative bias for Method B was 15% rather than 10%, then the RRMSE for Method A
would be 21.54% and the RRMSE for Method B would be 25.02%, so the method difference is
even more pronounced. While the above illustration is portrayed in terms of estimation of a mean
based on a simple random sample, the basic concepts apply more generally.
This example serves to illustrate that a method that may be considered preferable from a
chemical point of view [e.g., 85 or 90% recovery, 5% relative standard deviation (RSD)] may not
perform as well in a statistical application as a method with less bias and greater imprecision (e.g.,
zero bias, 40% RSD), especially when the inherent site variability is large relative to the
measurement-error RSD.
5.2.6 "Proof of Safety" vs. "Proof of Hazard"
Because of the basic hypothesis testing philosophy, the null hypothesis is generally
specified in terms of the status quo (e.g., no change or action will take place if null hypothesis is
not rejected). Also, since the classical approach exercises direct control over the false rejection
error rate, this rate is generally associated with the error of most concern (for further discussion
of this point, see Section 1.2). One difficulty, therefore, may be obtaining a consensus on which
error should be of most concern. It is not unlikely that the Agency's viewpoint in this regard will
differ from the viewpoint of the regulated party. In using this philosophy, the Agency's ideal
approach is not only to set up the direction of the hypothesis in such a way that controlling the
EPA QA/G-9 Final
QAOO Version 5-12 July 2000
-------
false rejection error protects the health and environment but also to set it up in a way that
encourages quality (high precision and accuracy) and minimizes expenditure of resources in
situations where decisions are relatively "easy" (e.g., all observations are far from the threshold
level of interest).
In some cases, how one formulates the hypothesis testing problem can lead to very
different sampling requirements. For instance, following remediation activities at a hazardous
waste site, one may seek to answer "Is the site clean?" Suppose one attempts to address this
question by comparing a mean level from samples taken after the remediation with a threshold
level (chosen to reflect "safety"). If the threshold level is near background levels that might have
existed in the absence of the contamination, then it may be very difficult (i.e., require enormous
sample sizes) to "prove" that the site is "safe." This is because the concentrations resulting from
even a highly efficient remediation under such circumstances would not be expected to deviate
greatly from such a threshold. A better approach for dealing with this problem may be to
compare the remediated site with a reference ("uncontaminated") site, assuming that such a site
can be determined.
To avoid excessive expense in collecting and analyzing samples for a contaminant,
compromises will sometimes be necessary. For instance, suppose that a significance level of 0.05
is to be used; however, the affordable sample size may be expected to yield a test with power of
only 0.40 at some specified parameter value chosen to have practical significance (see Section
5.2.3). One possible way that compromise may be made in such a situation is to relax the
significance level, for instance, using a = 0.10, 0.15, or 0.20. By relaxing this false rejection rate,
a higher power (i.e., a lower false acceptance rate P) can be achieved. An argument can be made,
for example, that one should develop sampling plans and determine sample sizes in such a way
that both the false rejection and false acceptance errors are treated simultaneously and in a
balanced manner (for example, designing to achieve a = P = 0.15) instead of using the traditional
approach of fixing the false rejection error rate at 0.05 or 0.01 and letting P be determined by the
sample size. This approach of treating the false rejection and false acceptance errors
simultaneously is taken in the DQO Process and it is recommended that several different scenarios
of a and P be investigated before a decision on specific values for a and P are selected.
EPA QA/G-9 Final
QAOO Version 5-13 July 2000
-------
This page is intentionally blank.
EPA QA/G-9 Final
QAOO Version 5-14 July 2000
-------
APPENDIX A
STATISTICAL TABLES
EPA QA/G-9 Final
QAOO Version A - 1 July 2000
-------
LIST OF TABLES
Table No. Page
A-l: CRITICAL VALUES OF STUDENT'S t DISTRIBUTION A - 3
A-2: CRITICAL VALUES FOR THE STUDENTIZED RANGE TEST A - 4
A-3: CRITICAL VALUES FOR THE EXTREME VALUE TEST (DIXONS TEST) A - 5
A-4: CRITICAL VALUES FOR DISCORDANCE TEST A - 6
A-5: APPROXIMATE CRITICAL VALUES Ar FOR ROSNERS TEST A - 7
A-6: QUANTILES OF THE WILCOXON SIGNED RANKS TEST A - 9
A-7: CRITICAL VALUES FOR THE RANK-SUM TEST A - 10
A-8: PERCENTILES OF THE CHI-SQUARE DISTRIBUTION A - 12
A-9: PERCENTILES OF THE F DISTRIBUTION A - 13
A-10: VALUES OF THE PARAMETER FOR COHENS ESTIMATES
ADJUSTING FOR NONDETECTED VALUES A - 18
A-l 1: PROBABILITIES FOR THE SMALL-SAMPLE
MANN-KENDALL TEST FOR TREND A - 19
A-12. QUANTILES FOR THE WALD-WOLFOWITZ TEST FOR RUNS A - 20
A-13. MODIFIED QUANTILE TEST CRITICAL NUMBERS A - 23
A-14. DUNNETT'S TEST (ONE TAILED) A - 27
A-15. APPROXIMATE a-LEVEL CRITICAL POINTS FOR
RANK VON NEUMANN RATIO TEST A - 29
EPA QA/G-9 Final
QAOO Version A - 2 July 2000
-------
TABLE A-l: CRITICAL VALUES OF STUDENT'S t DISTRIBUTION
Degrees of
Freedom
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
40
60
120
00
1-
.70
0.727
0.617
0.584
0.569
0.559
0.553
0.549
0.546
0.543
0.542
0.540
0.539
0.538
0.537
0.536
0.535
0.534
0.534
0.533
0.533
0.532
0.532
0.532
0.531
0.531
0.531
0.531
0.530
0.530
0.530
0.529
0.527
0.526
0.524
.75
1.000
0.816
0.765
0.741
0.727
0.718
0.711
0.706
0.703
0.700
0.697
0.695
0.694
0.692
0.691
0.690
0.689
0.688
0.6880
.687
0.686
0.686
0.685
0.685
0.684
0.684
0.684
0.683
0.683
0.683
0.681
0.679
0.677
0.674
.80
1.376
1.061
0.978
0.941
0.920
0.906
0.896
0.889
0.883
0.879
0.876
0.873
0.870
0.868
0.866
0.865
0.863
0.862
0.861
0.860
0.859
0.858
0.858
0.857
0.856
0.856
0.855
0.855
0.854
0.854
0.851
0.848
0.845
0.842
.85
1.963
1.386
1.250
1.190
1.156
1.134
1.119
1.108
1.100
1.093
1.088
1.083
1.079
1.076
1.074
1.071
1.069
1.067
1.066
1.064
1.063
1.061
1.060
1.059
1.058
1.058
1.057
1.056
1.055
1.055
1.050
1.046
1.041
1.036
.90
3.078
1.886
1.638
1.533
1.476
1.440
1.415
1.397
1.383
1.372
1.363
1.356
1.350
1.345
1.34
1.337
1.333
1.330
1.328
1.325
1.323
1.321
1.319
1.318
1.316
1.315
1.314
1.313
1.311
1.310
1.303
1.296
1.289
1.282
.95
6.314
2.920
2.353
2.132
2.015
1.943
1.895
1.860
1.833
1.812
1.796
1.782
1.771
1.761
1.753
1.746
1.740
1.734
1.729
1.725
1.721
1.717
1.714
1.711
1.708
1.706
1.703
1.701
1.699
1.697
1.684
1.671
1.658
1.645
.975
12.706
4.303
3.182
2.776
2.571
2.447
2.365
2.306
2.262
2.228
2.201
2.179
2.160
2.145
2.131
2.120
2.110
2.101
2.093
2.086
2.080
2.074
2.069
2.064
2.060
2.056
2.052
2.048
2.045
2.042
2.021
2.000
1.980
1.960
.99
31.821
6.965
4.541
3.747
3.365
3.143
2.998
2.896
2.821
2.764
2.718
2.681
2.650
2.624
2.602
2.583
2.567
2.552
2.539
2.528
2.518
2.508
2.500
2.492
2.485
2.479
2.473
2.467
2.462
2.457
2.423
2.390
2.358
2.326
.995
63.657
9.925
5.841
4.604
4.032
3.707
3.499
3.355
3.250
3.169
3.106
3.055
3.012
2.977
2.947
2.921
2.898
2.878
2.861
2.845
2.831
2.819
2.807
2.797
2.787
2.779
2.771
2.763
2.756
2.750
2.704
2.660
2.617
2.576
Note: The last row of the table (°
= zng, = 1.645.
degrees of freedom) gives the critical values fora standard normal distribution (z), e.g., t.
EPA QA/G-9
QAOO Version
A-3
Final
July 2000
-------
TABLE A-2: CRITICAL VALUES FOR THE STUDENTIZED RANGE TEST
n
o
J
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
25
30
35
40
45
50
55
60
65
70
75
80
85
90
95
100
150
200
500
1000
a
1.737
1.87
2.02
2.15
2.26
2.35
2.44
2.51
2.58
2.64
2.70
2.75
2.80
2.84
2.88
2.92
2.96
2.99
3.15
3.27
3.38
3.47
3.55
3.62
3.69
3.75
3.80
3.85
3.90
3.94
3.99
4.02
4.06
4.10
4.38
4.59
5.13
5.57
0.01
b
2.000
2.445
2.803
3.095
3.338
3.543
3.720
3.875
4.012
4.134
4.244
4.34
4.44
4.52
4.60
4.67
4.74
4.80
5.06
5.26
5.42
5.56
5.67
5.77
5.86
5.94
6.01
6.07
6.13
6.18
6.23
6.27
6.32
6.36
6.64
6.84
7.42
7.80
Level of Significance «
0.05
a
1.758
1.98
2.15
2.28
2.40
2.50
2.59
2.67
2.74
2.80
2.86
2.92
2.97
3.01
3.06
3.10
3.14
3.18
3.34
3.47
3.58
3.67
3.75
3.83
3.90
3.96
4.01
4.06
4.11
4.16
4.20
4.24
4.27
4.31
4.59
4.78
5.47
5.79
b
1.999
2.429
2.753
3.012
3.222
3.399
3.552
3.685
3.80
3.91
4.00
4.09
4.17
4.24
4.31
4.37
4.43
4.49
4.71
4.89
5.04
5.16
5.26
5.35
5.43
5.51
5.57
5.63
5.68
5.73
5.78
5.82
5.86
5.90
6.18
6.39
6.94
7.33
a
1.782
2.04
2.22
2.37
2.49
2.59
2.68
2.76
2.84
2.90
2.96
3.02
3.07
3.12
3.17
3.21
3.25
3.29
3.45
3.59
3.70
3.79
3.88
3.95
4.02
4.08
4.14
4.19
4.24
4.28
4.33
4.36
4.40
4.44
4.72
4.90
5.49
5.92
0.10
b
1.997
2.409
2.712
2.949
3.143
3.308
3.449
3.57
3.68
3.78
3.87
3.95
4.02
4.09
4.15
4.21
4.27
4.32
4.53
4.70
4.84
4.96
5.06
5.14
5.22
5.29
5.35
5.41
5.46
5.51
5.56
5.60
5.64
5.68
5.96
6.15
6.72
7.11
EPA QA/G-9
QAOO Version
A-4
Final
July 2000
-------
TABLE A-3: CRITICAL VALUES FOR THE EXTREME VALUE TEST
(DIXON'S TEST)
n
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
Level of Significance «
0.10
0.886
0.679
0.557
0.482
0.434
0.479
0.441
0.409
0.517
0.490
0.467
0.492
0.472
0.454
0.438
0.424
0.412
0.401
0.391
0.382
0.374
0.367
0.360
0.05
0.941
0.765
0.642
0.560
0.507
0.554
0.512
0.477
0.576
0.546
0.521
0.546
0.525
0.507
0.490
0.475
0.462
0.450
0.440
0.430
0.421
0.413
0.406
0.01
0.988
0.889
0.780
0.698
0.637
0.683
0.635
0.597
0.679
0.642
0.615
0.641
0.616
0.595
0.577
0.561
0.547
0.535
0.524
0.514
0.505
0.497
0.489
EPA QA/G-9
QAOO Version
A-5
Final
July 2000
-------
TABLE A-4: CRITICAL VALUES FOR DISCORDANCE TEST
n
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
Level of Significance «
0.01
1.155
1.492
1.749
1.944
2.097
2.221
2.323
2.410
2.485
2.550
2.607
2.659
2.705
2.747
2.785
2.821
2.854
2.884
2.912
2.939
2.963
2.987
3.009
3.029
3.049
3.068
3.085
3.103
3.119
3.135
0.05
1.153
1.463
1.672
1.822
1.938
2.032
2.110
2.176
2.234
2.285
2.331
2.371
2.409
2.443
2.475
2.504
2.532
2.557
2.580
2.603
2.624
2.644
2.663
2.681
2.698
2.714
2.730
2.745
2.759
2.773
n
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
Level of Significance «
0.01
3.150
3.164
3.178
3.191
3.204
3.216
3.228
3.240
3.251
3.261
3.271
3.282
3.292
3.302
3.310
3.319
3.329
3.336
0.05
2.786
2.799
2.811
2.823
2.835
2.846
2.857
2.866
2.877
2.887
2.896
2.905
2.914
2.923
2.931
2.940
2.948
2.956
EPA QA/G-9
QAOO Version
A-6
Final
July 2000
-------
TABLE A-5: APPROXIMATE CRITICAL VALUES Ar FOR ROSNER'S TEST
n
25
26
27
28
29
30
31
r
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
a
0.05
2.82
2.80
2.78
2.76
2.73
2.59
2.84
2.82
2.80
2.78
2.76
2.62
2.86
2.84
2.82
2.80
2.78
2.65
2.88
2.86
2.84
2.82
2.80
2.68
2.89
2.88
2.86
2.84
2.82
2.71
2.91
2.89
2.88
2.86
2.84
2.73
2.92
2.91
2.89
2.88
2.86
2.76
0.01
3.14
3.11
3.09
3.06
3.03
2.85
3.16
3.14
3.11
3.09
3.06
2.89
3.18
3.16
3.14
3.11
3.09
2.93
3.20
3.18
3.16
3.14
3.11
2.97
3.22
3.20
3.18
3.16
3.14
3.00
3.24
3.22
3.20
3.18
3.16
3.03
3.25
3.24
3.22
3.20
3.18
3.06
n
32
33
34
35
36
37
38
r
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
a
0.05
2.94
2.92
2.91
2.89
2.88
2.78
2.95
2.94
2.92
2.91
2.89
2.80
2.97
2.95
2.94
2.92
2.91
2.82
2.98
2.97
2.95
2.94
2.92
2.84
2.99
2.98
2.97
2.95
2.94
2.86
3.00
2.99
2.98
2.97
2.95
2.88
3.01
3.00
2.99
2.98
2.97
2.91
0.01
3.27
3.25
3.24
3.22
3.20
3.09
3.29
3.27
3.25
3.24
3.22
3.11
3.30
3.29
3.27
3.25
3.24
3.14
3.32
3.30
3.29
3.27
3.25
3.16
3.33
3.32
3.30
3.29
3.27
3.18
3.34
3.33
3.32
3.30
3.29
3.20
3.36
3.34
3.33
3.32
3.30
3.22
n
39
40
41
42
43
44
45
r
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
a
0.05
3.03
3.01
3.00
2.99
2.98
2.91
3.04
3.03
3.01
3.00
2.99
2.92
3.05
3.04
3.03
3.01
3.00
2.94
3.06
3.05
3.04
3.03
3.01
2.95
3.07
3.06
3.05
3.04
3.03
2.97
3.08
3.07
3.06
3.05
3.04
2.98
3.09
3.08
3.07
3.06
3.05
2.99
0.01
3.37
3.36
3.34
3.33
3.32
3.24
3.38
3.37
3.36
3.34
3.33
3.25
3.39
3.38
3.37
3.36
3.34
3.27
3.40
3.39
3.38
3.37
3.36
3.29
3.41
3.40
3.39
3.38
3.37
3.30
3.43
3.41
3.40
3.39
3.38
3.32
3.44
3.43
3.41
3.40
3.39
3.33
EPA QA/G-9
QAOO Version
A-7
Final
July 2000
-------
TABLE A-5: APPROXIMATE CRITICAL VALUES Ar FOR ROSNER'S TEST
n
46
47
48
49
50
60
r
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
a
0.05
3.09
3.09
3.08
3.07
3.06
3.00
3.10
3.09
3.09
3.08
3.07
3.01
3.11
3.10
3.09
3.09
3.08
3.03
3.12
3.11
3.10
3.09
3.09
3.04
3.13
3.12
3.11
3.10
3.09
3.05
3.20
3.19
3.19
3.18
3.17
3.14
0.01
3.45
3.44
3.43
3.41
3.40
3.34
3.46
3.45
3.44
3.43
3.41
3.36
3.46
3.46
3.45
3.44
3.43
3.37
3.47
3.46
3.46
3.45
3.44
3.38
3.48
3.47
3.46
3.46
3.45
3.39
3.56
3.55
3.55
3.54
3.53
3.49
n
70
80
90
100
150
200
r
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
1
2
3
4
5
10
a
0.05
3.26
3.25
3.25
3.24
3.24
3.21
3.31
3.30
3.30
3.29
3.29
3.26
3.35
3.34
3.34
3.34
3.33
3.31
3.38
3.38
3.38
3.37
3.37
3.35
3.52
3.51
3.51
3.51
3.51
3.50
3.61
3.60
3.60
3.60
3.60
3.59
0.01
3.62
3.62
3.61
3.60
3.60
3.57
3.67
3.67
3.66
3.66
3.65
3.63
3.72
3.71
3.71
3.70
3.70
3.68
3.75
3.75
3.75
3.74
3.74
3.72
3.89
3.89
3.89
3.88
3.88
3.87
3.98
3.98
3.97
3.97
3.97
3.96
n
250
300
350
400
450
500
r
1
5
10
1
5
10
1
5
10
1
5
10
1
5
10
1
5
10
a
0.05
3.67
3.67
3.66
3.72
3.72
3.71
3.77
3.76
3.76
3.80
3.80
3.80
3.84
3.83
3.83
3.86
3.86
3.86
0.01
4.04
4.04
4.03
4.09
4.09
4.09
4.14
4.13
4.13
4.17
4.17
4.16
4.20
4.20
4.20
4.23
4.23
4.22
EPA QA/G-9
QAOO Version
A-8
Final
July 2000
-------
TABLE A-6: QUANTILES OF THE WILCOXON SIGNED RANKS TEST
n
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
w.01
0
0
0
1
2
4
6
8
10
13
16
20
24
28
33
38
44
W.05
0
1
3
4
6
9
11
14
18
22
26
31
36
42
48
54
61
WJO
1
3
4
6
9
11
15
18
22
27
32
37
43
49
56
63
70
w.20
3
4
6
9
12
15
19
23
28
33
39
45
51
58
66
74
82
EPA QA/G-9
QAOO Version
A-9
Final
July 2000
-------
TABLE A-7: CRITICAL VALUES FOR THE RANK-SUM TEST
n
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
a
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
0.05
0.10
m
2
0
0
0
1
0
1
1
2
1
2
1
2
2
3
2
3
2
4
2
4
3
5
3
5
4
5
4
6
4
6
4
7
3
0
1
1
2
1
2
2
3
3
4
3
5
4
6
5
6
5
7
6
8
6
9
7
10
8
11
8
11
9
12
10
13
4
0
1
1
2
2
4
3
5
4
6
5
7
6
8
7
10
8
11
9
12
10
13
11
14
12
16
13
17
15
18
16
19
5
1
2
2
3
3
5
5
6
6
8
7
9
9
11
10
13
12
14
13
16
14
18
16
19
17
21
19
23
20
24
21
26
6
1
2
3
4
4
6
6
8
8
10
9
12
111
4
13
16
15
18
17
20
18
22
20
24
22
26
24
28
26
30
27
32
7
1
2
3
5
5
7
7
9
9
12
12
14
14
17
16
19
18
22
20
24
22
27
25
29
27
32
29
34
31
37
34
39
8
2
3
4
6
6
8
9
11
11
14
14
17
16
20
19
23
21
25
24
28
27
31
29
34
32
37
34
40
37
43
40
46
9
2
3
5
6
7
10
10
13
13
16
16
19
19
23
22
26
25
29
28
32
31
36
34
39
37
42
40
46
43
49
46
53
10
2
4
5
7
8
11
12
14
15
18
18
22
21
25
25
29
28
33
32
37
35
40
38
44
42
48
45
52
49
55
52
59
11
2
4
6
8
9
12
13
16
17
20
20
24
24
28
28
32
32
37
35
41
39
45
43
49
47
53
51
58
55
62
58
66
12
3
5
6
9
10
13
14
18
18
22
22
27
27
31
31
36
35
40
39
45
43
50
48
54
52
59
56
64
61
68
65
73
13
3
5
7
10
11
14
16
19
20
24
25
29
29
34
34
39
38
44
43
49
48
54
52
59
57
64
62
69
66
75
71
80
14
4
5
8
11
12
16
17
21
22
26
27
32
32
37
37
42
42
48
47
53
52
59
57
64
62
70
67
75
72
81
78
86
15
4
6
8
11
13
17
19
23
24
28
29
34
34
40
40
46
45
52
51
58
56
64
62
69
67
75
73
81
78
87
84
93
16
4
6
9
12
15
18
20
24
26
30
31
37
37
43
43
49
49
55
55
62
61
68
66
75
72
81
78
87
84
94
90
100
17
4
7
10
13
16
19
21
26
27
32
34
39
40
46
46
53
52
59
58
66
65
73
71
80
78
86
84
93
90
100
97
107
18
5
7
10
14
17
21
23
28
29
35
36
42
42
49
49
56
56
63
62
70
69
78
76
85
83
92
89
99
96
107
103
114
19
5
8
11
15
18
22
24
29
31
37
38
44
45
52
52
59
59
67
66
74
73
82
81
90
88
98
95
105
102
113
110
121
20
5
8
12
16
19
23
26
31
33
39
40
47
48
55
55
63
63
71
70
79
78
87
85
95
93
103
101
111
108
120
116
128
EPA QA/G-9
QAOO Version
A- 10
Final
July 2000
-------
TABLE A-7: CRITICAL VALUES FOR THE RANK-SUM TEST
n
18
19
20
a
0.05
0.10
0.05
0.10
0.05
n 11)
m
2
5
1
5
8
5
8
3
10
14
11
15
12
ifi
4
17
21
18
22
19
9^
5
23
28
24
29
26
^1
6
29
35
31
37
33
^Q
7
36
42
38
44
40
47
8
42
49
45
52
48
ss
9
49
56
52
59
55
«
10
56
63
59
67
63
71
11
62
70
66
74
70
7Q
12
69
78
73
82
78
87
13
76
85
81
90
85
QS
14
83
92
88
98
93
im
15
89
99
95
105
101
1 1 1
16
96
107
102
113
108
190
17
103
114
110
121
116
198
18
110
121
117
129
124
^^,£,
19
117
129
124
136
131
144
20
124
136
131
144
139
1S9
EPA QA/G-9
QAOO Version
A- 11
Final
July 2000
-------
TABLE A-8: PERCENTILES OF THE CHI-SQUARE DISTRIBUTION
V
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
40
50
60
70
80
90
100
1-a
.005
0.04393
0.0100
0.072
0.207
0.412
0.676
0.989
1.34
1.73
2.16
2.60
3.07
3.57
4.07
4.60
5.14
5.70
6.26
6.84
7.43
8.03
8.64
9.26
9.89
10.52
11.16
11.81
12.46
13.12
13.79
20.71
27.99
35.53
43.28
51.17
59.20
67.33
.010
0.03157
0.0201
0.115
0.297
0.554
0.872
1.24
1.65
2.09
2.56
3.05
3.57
4.11
4.66
5.23
5.81
6.41
7.01
7.63
8.26
8.90
9.54
10.20
10.86
11.52
12.20
12.88
13.56
14.26
14.95
22.16
29.71
37.48
45.44
53.54
61.75
70.06
.025
0.03982
0.0506
0.216
0.484
0.831
1.24
1.69
2.18
2.70
3.25
3.82
4.40
5.01
5.63
6.26
6.91
7.56
8.23
8.91
9.59
10.28
10.98
11.69
12.40
13.12
13.84
14.57
15.31
16.05
16.79
24.43
32.36
40.48
48.76
57.15
65.65
74.22
.050
0.02393
0.103
0.352
0.711
1.145
1.64
2.17
2.73
3.33
3.94
3.57
5.23
5.89
6.57
7.26
7.96
8.67
9.39
10.12
10.85
11.59
12.34
13.09
13.85
14.61
15.38
16.15
16.93
17.71
18.49
26.51
34.76
43.19
51.74
60.39
69.13
77.93
.100
0.0158
0.211
0.584
1.064
1.61
2.20
2.83
3.49
4.17
4.87
5.58
6.30
7.04
7.79
8.55
9.31
10.09
10.86
11.65
12.44
13.24
14.04
14.85
15.66
16.47
17.29
18.11
18.94
19.77
20.60
29.05
37.69
46.46
53.33
64.28
73.29
82.36
.900
2.71
4.61
6.25
7.78
9.24
10.64
12.02
13.36
14.68
15.99
17.28
18.55
19.81
21.06
22.31
23.54
24.77
25.99
27.20
28.41
29.62
30.81
32.01
33.20
34.38
35.56
36.74
37.92
39.09
40.26
51.81
63.17
74.40
85.53
96.58
107.6
118.5
.950
3.84
5.99
7.81
9.49
11.07
12.59
14.07
15.51
16.92
18.31
19.68
21.03
22.36
23.68
25.00
26.30
27.59
28.87
30.14
31.41
32.67
33.92
35.17
36.42
37.65
38.89
40.11
41.34
42.56
43.77
55.76
67.50
79.08
90.53
101.9
113.1
124.3
.975
5.02
7.38
9.35
11.14
12.83
14.45
16.01
17.53
19.02
20.48
21.92
23.34
24.74
26.12
27.49
28.85
30.19
31.53
32.85
34.17
35.48
36.78
38.08
39.36
40.65
41.92
43.19
44.46
45.72
46.98
59.34
71.42
83.30
95.02
106.6
118.1
129.6
.990
6.63
9.21
11.34
13.28
15.09
16.81
18.48
20.09
21.67
23.21
24.73
26.22
27.69
29.14
30.58
32.00
33.41
34.81
36.19
37.57
38.93
40.29
41.64
42.98
44.31
45.64
46.96
48.28
49.59
50.89
63.69
76.15
88.38
100.4
112.3
124.1
135.8
.995
7.88
10.60
12.84
14.86
16.75
18.55
20.28
21.96
23.59
25.19
26.76
28.30
29.82
31.32
32.80
34.27
35.72
37.16
38.58
40.00
41.40
42.80
44.18
45.56
46.93
48.29
49.64
50.99
52.34
53.67
66.77
79.49
91.95
104.2
116.3
128.3
140.2
EPA QA/G-9
QAOO Version
A- 12
Final
July 2000
-------
TABLE A-9: PERCENTILES OF THE F DISTRIBUTION
Degrees
Freedom
for
Denom-
inator
1 .50
.90
.95
.975
.99
2 .50
.90
.95
.975
.99
3 .50
.90
.95
.975
.99
4 .50
.90
.95
.975
.99
.999
Degrees of Freedom for Numerator
1
1.00
39.9
161
648
4052
0.667
8.53
18.5
38.5
98.5
0.585
5.54
10.1
17.4
34.1
0.549
4.54
7.71
12.2
21.2
74.1
2
1.50
49.5
200
800
5000
1.00
9.00
19.0
39.0
99.0
0.881
5.46
9.55
16.0
30.8
0.828
4.32
6.94
10.6
18.0
61.2
3
1.71
53.6
216
864
5403
1.13
9.16
19.2
39.2
99.2
1.00
5.39
9.28
15.4
29.5
0.941
4.19
6.59
9.98
16.7
56.2
4
1.82
55.8
225
900
5625
1.21
9.24
19.2
39.2
99.2
1.06
5.34
9.12
15.1
28.7
1.00
4.11
6.39
9.60
16.0
53.4
5
1.89
57.2
230
922
5764
1.25
9.29
19.3
39.3
99.3
1.10
5.31
9.01
14.9
28.2
1.04
4.05
6.26
9.36
15.5
51.7
6
1.94
58.2
234
937
5859
1.28
9.33
19.3
39.3
99.3
1.13
5.28
8.94
14.7
27.9
1.06
4.01
6.16
9.20
15.2
50.5
7
1.98
58.9
237
948
5928
1.30
9.35
19.4
39.4
99.4
1.15
5.27
8.89
14.6
27.7
1.08
3.98
6.09
9.07
15.0
49.7
8
2.00
59.4
239
957
5981
1.32
9.37
19.4
39.4
99.4
1.16
5.25
8.85
14.5
27.5
1.09
3.95
6.04
8.98
14.8
49.0
9
2.03
59.9
241
963
6022
1.33
9.38
19.4
39.4
99.4
1.17
5.24
8.81
14.5
27.3
1.10
3.94
6.00
8.90
14.7
48.5
10
2.04
60.2
242
969
6056
1.34
9.39
19.4
39.4
99.4
1.18
5.23
8.79
14.4
27.2
1.11
3.92
5.96
8.84
14.5
48.1
12
2.07
60.7
244
977
6106
1.36
9.41
19.4
39.4
99.4
1.20
5.22
8.74
14.3
27.1
1.13
3.90
5.91
8.75
14.4
47.4
15
2.09
61.2
246
985
6157
1.38
9.42
19.4
39.4
99.4
1.21
5.20
8.70
14.3
26.9
1.14
3.87
5.86
8.66
14.2
46.8
20
2.12
61.7
248
993
6209
1.39
9.44
19.4
39.4
99.4
1.23
5.18
8.66
14.2
26.7
1.15
3.84
5.80
8.56
14.0
46.1
24
2.13
62.0
249
997
6235
1.40
9.45
19.5
39.5
99.5
1.23
5.18
8.64
14.1
26.6
1.16
3.83
5.77
8.51
13.9
45.8
30
2.15
62.3
250
1001
6261
1.41
9.46
19.5
39.5
99.5
1.24
5.17
8.62
14.1
26.5
1.16
3.82
5.75
8.46
13.8
45.4
60
2.17
62.8
252
1010
6313
1.43
9.47
19.5
39.5
99.5
1.25
5.15
8.57
14.0
26.3
1.18
3.79
5.69
8.36
13.7
44.7
120
2.18
63.1
253
1014
6339
1.43
9.48
19.5
39.5
99.5
1.26
5.14
8.55
13.9
26.2
1.18
3.78
5.66
8.31
13.6
44.4
CO
2.20
63.3
254
1018
6366
1.44
9.49
19.5
39.5
99.5
1.27
5.13
8.53
13.9
26.1
1.19
3.76
5.63
8.26
13.5
44.1
EPA QA/G-9
QAOO Version
A- 13
Final
July 2000
-------
TABLE A-9: PERCENTILES OF THE F DISTRIBUTION
Degrees
Freedom
for
Denom-
inator
5 .50
.90
.95
.975
.99
.999
6 .50
.90
.95
.975
.99
.999
7 .50
.90
.95
.975
.99
.999
8 .50
.90
.95
.975
.99
.999
Degrees of Freedom for Numerator
1
0.528
4.06
6.61
10.0
16.3
47.2
0.515
3.78
5.99
8.81
22.8
35.5
0.506
3.59
5.59
8.07
12.2
29.2
0.499
3.46
5.32
7.57
11.3
25.4
2
0.799
3.78
5.79
8.43
13.3
37.1
0.780
3.46
5.14
7.26
10.9
27.0
.0767
3.26
4.74
6.54
9.55
21.7
0.757
3.11
4.46
6.06
8.65
18.5
3
0.907
3.62
5.41
7.76
12.1
33.2
0.886
3.29
4.76
6.60
9.78
23.7
0.871
3.07
4.35
5.89
8.45
18.8
0.860
2.92
4.07
5.42
7.59
15.8
4
0.965
3.52
5.19
7.39
11.4
31.1
0.942
3.18
4.53
6.23
9.15
21.9
0.926
2.96
4.12
5.52
7.85
17.2
0.915
2.81
3.84
5.05
7.01
14.4
5
1.00
3.45
5.05
7.15
11.0
29.8
0.977
3.11
4.39
5.99
8.75
20.8
0.960
2.88
3.97
5.29
7.46
16.2
0.948
2.73
3.69
4.82
6.63
13.5
6
1.02
3.40
4.95
6.98
10.7
28.8
1.00
3.05
4.28
5.82
8.47
20.0
0.983
2.83
3.87
5.12
7.19
15.5
0.971
2.67
3.58
4.65
6.37
12.9
7
1.04
3.37
4.88
6.85
10.5
28.2
1.02
3.01
4.21
5.70
8.26
19.5
1.00
2.78
3.79
4.99
6.99
15.0
0.988
2.62
3.50
4.53
6.18
12.4
8
1.05
3.34
4.82
6.76
10.3
27.6
1.03
2.98
4.15
5.60
8.10
19.0
1.01
2.75
3.73
4.90
6.84
14.6
1.00
2.59
3.44
4.43
6.03
12.0
9
1.06
3.32
4.77
6.68
10.2
27.2
1.04
2.96
4.10
5.52
7.98
18.7
1.02
2.72
3.68
4.82
6.72
14.5
1.01
2.56
3.39
4.36
5.91
11.8
10
1.07
3.39
4.74
6.62
10.1
26.9
1.05
2.94
4.06
5.46
7.87
18.4
1.03
2.70
3.64
4.76
6.62
14.1
1.02
2.54
3.35
4.30
5.81
11.5
12
1.09
3.27
4.68
6.52
9.89
26.4
1.06
2.90
4.00
5.37
7.72
18.0
1.04
2.67
3.57
4.67
6.47
13.7
1.03
2.50
3.28
4.20
5.67
11.2
15
1.10
3.24
4.62
6.43
9.72
25.9
1.07
2.87
3.94
5.27
7.56
17.6
1.05
2.63
3.51
4.57
6.31
13.3
1.04
2.46
3.22
4.10
5.52
10.8
20
1.11
3.21
4.56
6.33
9.55
25.4
1.08
2.84
3.87
5.17
7.40
17.1
1.07
2.59
3.44
4.47
6.16
12.9
1.05
2.42
3.15
4.00
5.36
10.5
24
1.12
3.19
4.53
6.28
9.47
25.1
1.09
2.82
3.84
5.12
7.31
16.9
1.07
2.58
3.41
4.42
6.07
12.7
1.06
2.40
3.12
3.95
5.28
10.3
30
1.12
3.17
4.50
6.23
9.38
24.9
1.10
2.80
3.81
5.07
7.23
16.7
1.08
2.56
3.38
4.36
5.99
12.5
1.07
2.38
3.08
3.89
5.20
10.1
60
1.14
3.14
4.43
6.12
9.20
24.3
1.11
2.76
3.74
4.96
7.06
16.2
1.09
2.51
3.30
4.25
5.82
12.1
1.08
2.34
3.01
3.78
5.03
9.73
120
1.14
3.12
4.40
6.07
9.11
24.1
1.12
2.74
3.70
4.90
6.97
16.0
1.10
2.49
3.27
4.20
5.74
11.9
1.08
2.32
2.97
3.73
4.95
9.53
CO
1.15
3.11
4.37
6.02
9.02
23.8
1.12
2.72
3.67
4.85
6.88
15.7
1.10
2.47
3.23
4.14
5.65
11.7
1.09
2.29
2.93
3.67
4.86
9.33
EPA QA/G-9
QAOO Version
A- 14
Final
July 2000
-------
TABLE A-9: PERCENTILES OF THE F DISTRIBUTION
Degrees
Freedom
for
Denom-
inator
9 .50
.90
.95
.975
.99
.999
10 .50
.90
.95
.975
.99
.999
12 .50
.90
.95
.975
.99
.999
15 .50
.90
.95
.975
.99
.999
Degrees of Freedom for Numerator
1
0.494
3.36
5.12
7.21
10.6
22.9
0.490
3.29
4.96
6.94
10.0
21.0
0.484
3.18
4.75
6.55
9.33
18.6
0.478
3.07
4.54
6.20
8.68
16.6
2
0.749
3.01
4.26
5.71
8.02
16.4
0.743
2.92
4.10
5.46
7.56
14.9
0.735
2.81
3.89
5.10
6.93
13.0
0.726
2.70
3.68
4.77
6.36
11.3
3
0.852
2.81
3.86
5.08
6.99
13.9
0.845
2.73
3.71
4.83
6.55
12.6
0.835
2.61
3.49
4.47
5.95
10.8
0.826
2.49
3.29
4.15
5.42
9.34
4
0.906
2.69
3.63
4.72
6.42
12.6
0.899
2.61
3.48
4.47
5.99
11.3
0.888
2.48
3.26
4.12
5.41
9.63
0.878
2.36
3.06
3.80
4.89
8.25
5
0.939
2.61
3.48
4.48
6.06
11.7
0.932
2.52
3.33
4.24
5.64
10.5
0.921
2.39
3.11
3.89
5.06
8.89
0.911
2.27
2.90
3.58
4.56
7.57
6
0.962
2.55
3.37
4.32
5.80
11.1
0.954
2.46
3.22
4.07
5.39
9.93
0.943
2.33
3.00
3.73
4.82
8.38
0.933
2.21
2.79
3.41
4.32
7.09
7
0.978
2.51
3.29
4.20
5.61
10.7
0.971
2.41
3.14
3.95
5.20
9.52
0.959
2.28
2.91
3.61
4.64
8.00
0.949
2.16
2.71
3.29
4.14
6.74
8
0.990
2.47
3.23
4.10
5.47
10.4
0.983
2.38
3.07
3.85
5.06
9.20
0.972
2.24
2.85
3.51
4.50
7.71
0.960
2.12
2.64
3.20
4.00
6.47
9
1.00
2.44
3.18
4.03
5.35
10.1
0.992
2.35
3.02
3.78
4.94
8.96
0.981
2.21
2.80
3.44
4.39
7.48
0.970
2.09
2.59
3.12
3.89
6.26
10
1.01
2.42
3.14
3.96
5.26
9.89
1.00
2.32
2.98
3.72
4.85
8.75
0.989
2.19
2.75
3.37
4.30
7.29
0.977
2.06
2.54
3.06
3.80
6.08
12
1.01
2.38
3.07
3.87
5.11
9.57
1.01
2.28
2.91
3.62
4.71
8.45
1.00
2.15
2.69
3.28
4.16
7.00
0.989
2.02
2.48
2.96
3.67
5.81
15
1.03
2.34
3.01
3.77
4.96
9.24
1.02
2.24
2.84
3.52
4.56
8.13
1.01
2.10
2.62
3.18
4.01
6.71
1.00
1.97
2.40
2.86
3.52
5.54
20
1.04
2.30
2.94
3.67
4.81
8.90
1.03
2.20
2.77
3.42
4.41
7.80
1.02
2.06
2.54
3.07
3.86
6.40
1.01
1.92
2.33
2.76
3.37
5.25
24
1.05
2.28
2.90
3.61
4.73
8.72
1.04
2.18
2.74
3.37
4.33
7.64
1.03
2.04
2.51
3.02
3.78
6.25
1.02
1.90
2.29
2.70
3.29
5.10
30
1.05
2.25
2.86
3.56
4.65
8.55
1.05
2.16
2.70
3.31
4.25
7.47
1.03
2.01
2.47
2.96
3.70
6.09
1.02
1.87
2.25
2.64
3.21
4.95
60
1.07
2.21
2.79
3.45
4.48
8.19
1.06
2.11
2.62
3.20
4.08
7.12
1.05
1.96
2.38
2.85
3.54
5.76
1.03
1.82
2.16
2.52
3.05
4.64
120
1.07
2.18
2.75
3.39
4.40
8.00
1.06
2.08
2.58
3.14
4.00
6.94
1.05
1.93
2.34
2.79
3.45
5.59
1.04
1.79
2.11
2.46
2.96
4.48
CO
1.08
2.16
2.71
3.33
4.31
7.81
1.07
2.06
2.54
3.08
3.91
6.76
1.06
1.90
2.30
2.72
3.36
5.42
1.05
1.76
2.07
2.40
2.87
4.31
EPA QA/G-9
QAOO Version
A- 15
Final
July 2000
-------
TABLE A-9: PERCENTILES OF THE F DISTRIBUTION
Degrees
Freedom
for
Denom-
inator
20 .50
.90
.95
.975
.99
.999
24 .50
.90
.95
.975
.99
.999
30 .50
.90
.95
.975
.99
.999
60 .50
.90
.95
.975
.99
.999
Degrees of Freedom for Numerator
1
0.472
2.97
4.35
5.87
8.10
14.8
0.469
2.93
4.26
5.72
7.82
14.0
0.466
2.88
4.17
5.57
7.56
13.3
0.461
2.79
4.00
5.29
7.08
12.0
2
0.718
2.59
3.49
4.46
5.85
9.95
0.714
2.54
3.40
4.32
6.66
9.34
0.709
2.49
3.32
4.18
5.39
8.77
0.701
2.39
3.15
3.93
4.98
7.77
3
0.816
2.38
3.10
3.86
4.94
8.10
0.812
2.33
3.01
3.72
4.72
7.55
0.807
2.28
2.92
3.59
4.51
7.05
0.798
2.18
2.76
3.34
4.13
6.17
4
0.868
2.25
2.87
3.51
4.43
7.10
0.863
2.19
2.78
3.38
4.22
6.59
0.858
2.14
2.69
3.25
4.02
6.12
0.849
2.04
2.53
3.01
3.65
5.31
5
0.900
2.16
2.71
3.29
4.10
6.46
0.895
2.10
2.62
3.15
3.90
5.98
0.890
2.05
2.53
3.03
3.70
5.53
0.880
1.95
2.37
2.79
3.34
4.76
6
0.922
2.09
2.60
3.13
3.87
6.02
0.917
2.04
2.51
2.99
3.67
5.55
0.912
1.98
2.42
2.87
3.47
5.12
0.901
1.87
2.25
2.63
3.12
4.37
7
0.938
2.04
2.51
3.01
3.70
5.69
0.932
1.98
2.42
2.87
3.50
5.23
0.927
1.93
2.33
2.75
3.30
4.82
0.917
1.82
2.17
2.51
2.95
4.09
8
0.950
2.00
2.45
2.91
3.56
5.44
0.944
1.94
2.36
2.78
3.36
4.99
0.939
1.88
2.27
2.65
3.17
4.58
0.928
1.77
2.10
2.41
2.82
3.86
9
0.959
1.96
2.39
2.84
3.46
5.24
0.953
1.91
2.30
2.70
3.26
4.80
0.948
1.85
2.21
2.57
3.07
4.39
0.937
1.74
2.04
2.33
2.72
3.69
10
0.966
1.94
2.35
2.77
3.37
5.08
0.961
1.88
2.25
2.64
3.17
4.64
0.955
1.82
2.16
2.51
2.98
4.24
0.945
1.71
1.99
2.27
2.63
3.54
12
0.977
1.89
2.28
2.68
3.23
4.82
0.972
1.83
2.18
2.54
3.03
4.39
0.966
1.77
2.09
2.41
2.84
4.00
0.956
1.66
1.92
2.17
2.50
3.32
15
0.989
1.84
2.20
2.57
3.09
4.56
0.983
1.78
2.11
2.44
2.89
4.14
0.978
1.72
2.01
2.31
2.70
3.75
0.967
1.60
1.84
2.06
2.35
3.08
20
1.00
1.79
2.12
2.46
2.94
4.29
0.994
1.73
2.03
2.33
2.74
3.87
0.989
1.62
1.93
2.20
2.55
3.49
0.978
1.54
1.75
1.94
2.20
2.83
24
1.01
1.77
2.08
2.41
2.86
4.15
1.00
1.70
1.98
2.27
2.66
3.74
0.994
1.64
1.89
2.14
2.47
3.36
0.983
1.51
1.70
1.88
2.12
2.69
30
1.01
1.74
2.04
2.35
2.78
4.00
1.01
1.67
1.94
2.21
2.58
3.59
1.00
1.61
1.84
2.07
2.39
3.22
0.989
1.48
1.65
1.82
2.03
2.55
60
1.02
1.68
1.95
2.22
2.61
3.70
1.02
1.61
1.84
2.08
2.40
3.29
1.01
1.54
1.74
1.94
2.21
2.92
1.00
1.40
1.53
1.67
1.84
2.25
120
1.03
1.64
1.90
2.16
2.52
3.54
1.02
1.57
1.79
2.01
2.31
3.14
1.02
1.50
1.68
1.87
2.11
2.76
1.01
1.35
1.47
1.58
1.73
2.08
CO
1.03
1.61
1.84
2.09
2.42
3.38
1.03
1.53
1.73
1.94
2.21
2.97
1.02
1.46
1.62
1.79
2.01
2.59
1.01
1.29
1.39
1.48
1.60
1.89
EPA QA/G-9
QAOO Version
A- 16
Final
July 2000
-------
TABLE A-9: PERCENTILES OF THE F DISTRIBUTION
Degrees
Freedom
for
Denom-
inator
20.90
.95
.975
.99
.999
•*> .90
.95
.975
.99
.999
Degrees of Freedom for Numerator
1
2.75
3.92
5.15
6.85
11.4
2.71
3.84
5.02
6.63
10.8
2
2.35
3.07
3.80
4.79
7.32
2.30
3.00
3.69
4.61
6.91
3
2.13
2.68
3.23
3.95
5.78
2.08
2.60
3.12
3.78
5.42
4
1.99
2.45
2.89
3.48
4.95
1.94
2.37
2.79
3.32
4.62
5
1.90
2.29
2.67
3.17
4.42
1.85
2.21
2.57
3.02
4.10
6
1.82
2.18
2.52
2.96
4.04
1.77
2.10
22.41
2.80
3.74
7
1.77
2.09
2.39
2.79
3.77
1.72
2.01
2.29
2.64
3.47
8
1.72
2.02
2.30
2.66
3.55
1.67
1.94
2.19
2.51
3.27
9
1.68
1.96
2.22
2.56
3.38
1.63
1.88
2.11
2.41
3.10
10
1.65
1.91
2.16
2.47
3.24
1.60
1.83
2.05
2.32
2.96
12
1.60
1.83
2.05
2.34
3.02
1.55
1.75
1.94
2.18
2.74
15
1.55
1.75
1.95
2.19
2.78
1.49
1.67
1.83
2.04
2.51
20
1.48
1.66
1.82
2.03
2.53
1.42
1.57
1.71
1.88
2.27
24
1.45
1.61
1.76
1.95
2.40
1.38
1.52
1.64
1.79
2.13
30
1.41
1.55
1.69
1.86
2.26
1.34
1.46
1.57
1.70
1.99
60
1.32
1.43
1.53
1.66
1.95
1.24
1.32
1.39
1.47
1.66
120
1.26
1.35
1.43
1.53
1.77
1.17
1.22
1.27
1.32
1.45
00
1.19
1.25
1.31
1.38
1.54
1.00
1.00
1.00
1.00
1.00
EPA QA/G-9
QAOO Version
A- 17
Final
July 2000
-------
TABLE A-10: VALUES OF THE PARAMETER A FOR COHEN'S ESTIMATES
ADJUSTING FOR NONDETECTED VALUES
Y
.00
.05
.10
.15
.20
.25
.30
.35
.40
.45
.50
.55
.60
.65
.70
.75
.80
.85
.90
.95
1.00
.01
.010100
.010551
.010950
.011310
.011642
.011952
.012243
.012520
.012784
.013036
.013279
.013513
.013739
.013958
.014171
.014378
.014579
.014773
.014967
.015154
.015338
.02
.020400
.021294
.022082
.022798
.023459
.024076
.024658
.025211
.025738
.026243
.026728
.027196
.027849
.028087
.028513
.029927
.029330
.029723
.030107
.030483
.030850
.03
.030902
.032225
.033398
.034466
.035453
.036377
.037249
.038077
.038866
.039624
.040352
.041054
.041733
.042391
.043030
.043652
.044258
.044848
.045425
.045989
.046540
.04
.041583
.043350
.044902
.046318
.047829
.048858
.050018
.051120
.052173
.053182
.054153
.055089
.055995
.056874
.057726
.058556
.059364
.060153
.060923
.061676
.062413
.05
.052507
.054670
.056596
.058356
.059990
.061522
.062969
.064345
.065660
.066921
.068135
.069306
.070439
.071538
.072505
.073643
.074655
.075642
.075606
.077549
.078471
.06
.063625
.066159
.068483
.070586
.072539
.074372
.076106
.077736
.079332
.080845
.082301
.083708
.085068
.086388
.087670
.088917
.090133
.091319
.092477
.093611
.094720
h
.07
.074953
.077909
.080563
.083009
.085280
.087413
.089433
.091355
.093193
.094958
.096657
.098298
.099887
.10143
.10292
.10438
.10580
.10719
.10854
.10987
.11116
.08
.08649
.08983
.09285
.09563
.09822
.10065
.10295
.10515
.10725
.10926
.11121
.11208
.11490
.11666
.11837
.12004
.12167
.12225
.12480
.12632
.12780
.09
.09824
.10197
.10534
.10845
.11135
.11408
.11667
.11914
.12150
.12377
.12595
.12806
.13011
.13209
.13402
.13590
.13775
.13952
.14126
.14297
.14465
.10
.11020
.11431
.11804
.12148
.12469
.12772
.13059
.13333
.13595
.13847
.14090
.14325
.14552
.14773
.14987
.15196
.15400
.15599
.15793
.15983
.16170
.15
.17342
.17925
.18479
.18985
.19460
.19910
.20338
.20747
.21129
.21517
.21882
.22225
.22578
.22910
.23234
.23550
.23858
.24158
.24452
.24740
.25022
.20
.24268
.25033
.25741
.26405
.27031
.27626
.28193
.28737
.29250
.29765
.30253
.30725
.31184
.31630
.32065
.32489
.32903
.33307
.33703
.34091
.34471
Y
.00
.05
.10
.15
.20
.25
.30
.35
.40
.45
.50
.55
.60
.65
.70
.75
.80
.85
.90
.95
1.00
.25
.31862
.32793
.33662
.34480
.35255
.35993
.36700
.37379
.38033
.38665
.39276
.39679
.40447
.41008
.41555
.42090
.42612
.43122
.43622
.44112
.44592
.30
.4021
.4130
.4233
.4330
.4422
.4510
.4595
.4676
.4735
.4831
.4904
.4976
.5045
.5114
.5180
.5245
.5308
.5370
.5430
.5490
.5548
.35
.4941
.5066
.5184
.5296
.5403
.5506
.5604
.5699
.5791
.5880
.5967
.6061
.6133
.6213
.6291
.6367
.6441
.6515
.6586
.6656
.6724
.40
.5961
.6101
.6234
.6361
.6483
.6600
.6713
.6821
.6927
.7029
.7129
.7225
.7320
.7412
.7502
.7590
.7676
.7781
.7844
.7925
.8005
.45
.7096
.7252
.7400
.7542
.7673
.7810
.7937
.8060
.8179
.8295
.8408
.8517
.8625
.8729
.8832
.8932
.9031
.9127
.9222
.9314
.9406
.50
.8388
.8540
.8703
.8860
.9012
.9158
.9300
.9437
.9570
.9700
.9826
.9950
1.007
1.019
1.030
1.042
1.053
1.064
1.074
1.085
1.095
h
.55
.9808
.9994
1.017
1.035
1.051
1.067
1.083
1.098
1.113
1.127
1.141
1.155
1.169
1.182
1.195
1.207
1.220
1.232
1.244
1.255
1.287
.60
1.145
1.166
1.185
1.204
1.222
1.240
1.257
1.274
1.290
1.306
1.321
1.337
1.351
1.368
1.380
1.394
1.408
1.422
1.435
1.448
1.461
.65
1.336
1.358
1.379
1.400
1.419
1.439
1.457
1.475
1.494
1.511
1.528
1.545
1.561
1.577
1.593
1.608
1.624
1.639
1.653
1.668
1.882
.70
1.561
1.585
1.608
1.630
1.651
1.672
1.693
1.713
1.732
1.751
1.770
1.788
1.806
1.824
1.841
1.851
1.875
1.892
1.908
1.924
1.940
.80
2.176
2.203
2.229
2.255
2.280
2.305
2.329
2.353
2.376
2.399
2.421
2.443
2.465
2.486
2.507
2.528
2.548
2.568
2.588
2.607
2.626
.90
3.283
3.314
3.345
3.376
3.405
3.435
3.464
3.492
3.520
3.547
3.575
3.601
3.628
3.654
3.679
3.705
3.730
3.754
3.779
3.803
3.827
EPA QA/G-9
QAOO Version
A- 18
Final
July 2000
-------
TABLE A-ll: PROBABILITIES FOR THE SMALL-SAMPLE
MANN-KENDALL TEST FOR TREND
s
0
2
4
6
8
10
12
14
16
18
20
22
24
26
28
30
32
34
36
n
458 9
0.625 0.592 0.548 0.540
0.375 0.408 0.452 0.460
0.167 0.242 0.360 0.381
0.042 0.117 0.274 0.306
0.042 0.199 0.238
0.0083 0.138 0.179
0.089 0.130
0.054 0.090
0.031 0.060
0.016 0.038
0.0071 0.022
0.0028 0.012
0.00087 0.0063
0.00019 0.0029
0.000025 0.0012
0.00043
0.00012
0.000025
0.0000028
S
1
3
5
7
9
11
13
15
17
19
21
23
25
27
29
31
33
35
37
39
41
43
45
n
67 10
0.500 0.500 0.500
0.360 0.386 0.431
0.235 0.281 0.364
0.136 0.191 0.300
0.068 0.119 0.242
0.028 0.068 0.190
0.0083 0.035 0.146
0.0014 0.015 0.108
0.0054 0.078
0.0014 0.054
0.00020 0.036
0.023
0.014
0.0083
0.0046
0.0023
0.0011
0.00047
0.00018
0.000058
0.000015
0.0000028
0.00000028
EPA QA/G-9
QAOO Version
A- 19
Final
July 2000
-------
TABLE A-12. QUANTILES FOR THE WALD-WOLFOWITZ TEST FOR RUNS
n
4
in
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
w,.01
-
-
-
-
-
-
-
3
3
3
3
3
4
4
4
4
4
w
" 0.05
-
-
3
3
3
3
4
4
4
4
5
5
5
5
5
5
5
w
"0.10
3
3
4
4
4
4
5
5
5
5
6
6
6
6
6
6
6
n
5
in
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
w,.01
3
3
3
3
3
4
4
4
4
5
5
5
5
5
5
5
w
" 0.05
4
4
4
4
4
4
5
5
5
6
6
6
6
6
6
6
w
"0.10
4
4
5
5
5
5
6
6
6
6
7
7
7
7
7
7
n
6
m
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
w,.01
3
4
4
4
4
5
5
5
5
5
6
6
6
6
6
w
" 0.05
4
5
5
5
6
6
6
7
7
7
7
7
7
7
7
w
"0.10
5
6
6
6
7
7
7
8
8
8
8
8
8
8
8
n
7
m
8
9
10
11
12
13
14
15
16
17
18
19
20
w,.01
4
4
4
5
5
5
5
6
6
6
7
7
7
w
" 0.05
5
5
6
6
6
7
7
7
7
8
8
8
8
Wo.!
0
6
6
7
7
7
7
8
8
8
8
9
9
9
EPA QA/G-9
QAOO Version
A-20
Final
July 2000
-------
TABLE A-12. QUANTILES FOR THE WALD-WOLFOWITZ TEST FOR RUNS
n
8
in
8
9
10
11
12
13
14
15
16
17
18
19
20
w,.01
5
5
5
6
6
6
6
6
7
7
7
7
7
w
" 0.05
6
6
7
7
7
7
8
8
8
8
9
9
9
w
"0.10
6
7
7
8
8
8
8
9
9
9
10
10
10
n
9
in
9
10
11
12
13
14
15
16
17
18
19
20
w,.01
5
6
6
6
7
7
7
7
8
8
8
8
w
" 0.05
6
7
7
7
8
8
9
9
9
9
10
10
w
"0.10
7
8
8
8
9
9
9
10
10
10
11
11
n
10
m
10
11
12
13
14
15
16
17
18
19
20
w,.01
7
7
7
7
8
8
8
8
9
9
9
w
" 0.05
8
8
8
9
9
9
9
10
10
11
11
w
"0.10
9
9
9
10
10
10
11
11
11
12
12
n
11
m
11
12
13
14
15
16
17
18
19
20
w,.01
7
7
7
8
8
8
9
9
9
9
w
" 0.05
8
8
9
9
9
10
10
10
11
11
Wo.!
0
9
9
10
10
10
11
11
11
12
12
12
12
13
14
15
16
17
7
8
8
8
9
9
9
10
10
10
11
11
10
10
10
11
12
12
13
13
14
15
16
17
18
9
9
10
10
10
10
10
11
11
11
11
11
11
12
12
12
12
13
14
14
15
16
17
18
19
9
9
10
10
10
11
11
11
11
12
12
13
12
12
13
13
13
14
15
15
16
17
18
19
20
8
9
10
10
11
12
10
11
12
12
13
14
11
12
13
14
15
16
EPA QA/G-9
QAOO Version
A-21
Final
July 2000
-------
TABLE A-12. QUANTILES FOR THE WALD-WOLFOWITZ TEST FOR RUNS
n
16
in
18
19
20
16
17
18
19
20
w,.01
9
10
10
10
11
11
12
13
w
" 0.05
11
12
12
12
13
13
14
14
w
"0.10
12
13
13
13
14
14
15
16
n
in
19
20
w,.01
10
11
w
" 0.05
12
12
w
"0.10
13
13
17
17
18
19
20
11
11
12
12
13
13
14
14
14
15
16
16
n
m
20
w,.01
11
w
" 0.05
13
w
"0.10
14
18
18
19
20
13
13
14
14
14
15
16
16
16
n
m
w,.01
w
" 0.05
Wo.!
0
19
19
20
14
14
16
16
17
17
20
20
143
16
17
When n or m is greater than 20 the Wp quantile is given by:
I , 2mn \ 7rl 2mn(2mn -m- n)
= \-\ -- + LP - -^— - -
m + n
where Zp is the appropriate quantile from the standard normal (see last row of Table A-1).
EPA QA/G-9
QAOO Version
A-22
Final
July 2000
-------
TABLE A-13. MODIFIED QUANTILE TEST CRITICAL NUMBERS
LEVEL OF SIGNIFICANCE ( ) APPROXIMATELY 0.10
m = number of measurements population 2
n = number of measurements population 1
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
5
3
3
2
2
2
2
2
2
2
2
2
2
2
2
2
2
6
3
3
2
2
2
2
2
2
2
2
2
2
2
2
2
2
7
4
3
3
3
3
2
2
2
2
2
2
2
2
2
2
2
8
4
4
3
3
3
2
2
2
2
2
2
2
2
2
2
2
90
5
4
4
3
3
3
2
2
2
2
2
2
2
2
2
2
10
5
4
4
4
3
3
3
2
2
2
2
2
2
2
2
2
11
5
4
4
4
4
3
3
3
2
2
2
2
2
2
2
2
12
6
5
4
4
4
3
3
3
3
2
2
2
2
2
2
2
13
6
5
5
4
4
4
3
3
3
3
2
2
2
2
2
2
14
7
6
5
5
4
4
4
3
3
3
3
2
2
2
2
2
15
7
6
6
5
5
4
4
4
3
3
3
3
2
2
2
2
16
7
7
6
5
5
4
4
4
4
3
3
3
3
2
2
2
17
8
7
6
5
5
4
4
4
4
4
3
3
3
3
2
2
18
8
8
7
6
5
5
5
4
4
4
4
3
3
3
3
2
19
8
8
7
6
5
5
5
5
4
4
4
4
3
3
3
3
20
8
8
7
6
6
5
5
5
4
4
4
4
3
3
3
3
EPA QA/G-9
QAOO Version
A-23
Final
July 2000
-------
TABLE A-13. MODIFIED QUANTILE TEST CRITICAL NUMBERS (CONTINUED)
LEVEL OF SIGNIFICANCE ( ) APPROXIMATELY 0.10
m = number of measurements population 2
n = number of measurements population 1
25
30
35
40
45
50
55
60
65
70
75
80
85
90
95
100
3
3
3
3
3
2
2
2
2
2
2
2
2
2
2
2
30
4
3
3
3
3
3
3
2
2
2
2
2
2
2
2
2
35
4
4
3
3
3
3
3
3
3
2
2
2
2
2
2
2
40
5
4
4
4
3
3
3
3
3
3
2
2
2
2
2
2
45
5
5
4
4
4
3
3
3
3
3
3
2
2
2
2
2
50
5
5
4
4
4
4
3
3
3
3
3
3
3
3
2
2
55
6
5
5
4
4
4
4
3
3
3
3
3
3
3
3
3
60
6
6
5
5
4
4
4
4
3
3
3
3
3
3
3
3
65
7
6
5
5
4
4
4
4
4
3
3
3
3
3
3
3
70
7
6
6
5
5
4
4
4
4
4
3
3
3
3
3
3
75
8
7
6
5
5
5
4
4
4
4
4
3
3
3
3
3
80
8
7
6
6
5
5
4
4
4
4
4
4
3
3
3
3
85
8
7
6
6
5
5
5
4
4
4
4
4
4
3
3
3
90
9
8
7
6
6
5
5
5
4
4
4
4
4
4
3
3
95
9
8
7
6
6
5
5
5
5
4
4
4
4
4
3
3
100
10
8
7
7
6
6
5
5
5
4
4
4
4
4
4
4
EPA QA/G-9
QAOO Version
A-24
Final
July 2000
-------
TABLE A-13. MODIFIED QUANTILE TEST CRITICAL NUMBERS (CONTINUED)
LEVEL OF SIGNIFICANCE (<*) APPROXIMATELY 0.05
m = number of measurements population 2
n = number of measurements population 1
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
5
4
4
3
3
3
3
3
3
3
2
2
2
2
2
2
2
6
4
4
4
3
3
3
3
3
3
3
2
2
2
2
2
2
7
5
4
4
4
3
3
3
3
3
3
3
2
2
2
2
2
8
5
5
4
4
4
3
3
3
3
3
3
3
2
2
2
2
90
6
5
5
4
4
4
3
3
3
3
3
3
3
2
2
2
10
6
5
5
5
4
4
4
3
3
3
3
3
3
3
2
2
11
6
5
5
5
5
4
4
4
3
3
3
3
3
3
3
2
12
7
6
5
5
5
4
4
4
4
3
3
3
3
3
3
3
13
7
6
6
5
5
5
4
4
4
4
3
3
3
3
3
3
14
8
7
6
6
5
5
5
4
4
4
4
3
3
3
3
3
15
8
7
7
6
6
5
5
5
4
4
4
4
3
3
3
3
16
8
8
7
6
6
5
5
5
5
4
4
4
4
3
3
3
17
9
8
7
6
6
6
5
5
5
5
4
4
4
4
3
3
18
9
9
8
7
6
6
6
5
5
5
5
4
4
4
4
3
19
10
9
8
7
6
6
6
6
5
5
5
5
4
4
4
4
20
10
9
8
7
6
6
6
6
5
5
5
5
4
4
4
4
EPA QA/G-9
QAOO Version
A-25
Final
July 2000
-------
TABLE A-13. MODIFIED QUANTILE TEST CRITICAL NUMBERS (CONTINUED)
LEVEL OF SIGNIFICANCE (<*) APPROXIMATELY 0.05
m = number of measurements population 2
n = number of measurements population 1
25
30
35
40
45
50
55
60
65
70
75
80
85
90
95
100
4
4
3
3
3
3
3
3
2
2
2
2
2
2
2
2
30
4
4
4
4
4
3
3
3
3
3
3
3
2
2
2
2
35
5
5
4
4
4
4
3
3
3
3
3
3
3
2
2
2
40
6
5
5
4
4
4
4
3
3
3
3
3
3
3
2
2
45
7
6
5
5
4
4
4
4
3
3
3
3
3
3
3
2
50
7
6
6
5
5
4
4
4
4
3
3
3
3
3
3
3
55
8
7
6
5
5
5
4
4
4
4
3
3
3
3
3
3
60
8
7
6
6
5
5
5
4
4
4
4
3
3
3
3
3
65
8
8
7
6
6
5
5
5
4
4
4
4
3
3
3
3
70
9
8
7
7
6
5
5
5
5
4
4
4
4
3
3
3
75
9
9
8
7
6
6
5
5
5
5
5
4
4
4
3
3
80
10
9
8
7
7
6
6
5
5
5
5
5
4
4
4
3
85
10
9
8
8
7
6
6
6
5
5
5
5
5
4
4
4
90
11
10
9
8
7
7
6
6
5
5
5
5
5
5
4
4
95
11
10
9
8
8
7
6
6
6
5
5
5
5
5
5
4
100
12
11
10
9
8
7
7
6
6
6
5
5
5
5
5
5
EPA QA/G-9
QAOO Version
A-26
Final
July 2000
-------
TABLE A-14. DUNNETT'S TEST (ONE TAILED)
TOTAL NUMBER OF INVESTIGATED GROUPS (K - 1)
Degrees of Freedom
2
3
4
5
6
7
8
9
10
12
16
a
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
2
3.80
2.54
2.94
2.13
2.61
1.96
2.44
1.87
2.34
1.82
2.27
1.78
2.22
1.75
2.18
1.73
2.15
1.71
2.11
1.69
2.06
1.66
3
4.34
2.92
3.28
2.41
2.88
2.20
2.68
2.09
2.56
2.02
2.48
1.98
2.42
1.94
2.37
1.92
2.34
1.90
2.29
1.87
2.23
1.83
4
4.71
3.20
3.52
2.61
3.08
2.37
2.85
2.24
2.71
2.17
2.82
2.11
2.55
2.08
2.50
2.05
2.47
2.02
2.41
1.99
2.34
1.95
5
5.08
3.40
3.70
2.76
3.22
2.50
2.98
2.36
2.83
2.27
2.73
2.22
2.66
2.17
2.60
2.14
2.56
2.12
2.50
2.08
2.43
2.04
6
5.24
3.57
3.85
2.87
3.34
2.60
3.08
2.45
2.92
2.36
2.81
2.30
2.74
2.25
2.68
2.22
2.64
2.19
2.58
2.16
2.50
2.11
7
5.43
3.71
3.97
2.97
3.44
2.68
3.16
2.53
3.00
2.43
2.89
2.37
2.81
2.32
2.75
2.28
2.70
2.26
2.64
2.22
2.56
2.17
8
5.60
3.83
4.08
3.06
3.52
2.75
3.24
2.59
3.06
2.49
2.95
2.42
2.87
2.38
2.81
2.34
2.76
2.31
2.69
2.27
2.61
2.22
9
5.75
3.94
4.17
3.13
3.59
2.82
3.30
2.65
3.12
2.54
3.00
2.47
2.92
2.42
2.86
2.39
2.81
2.35
2.74
2.31
2.65
2.26
10
5.88
4.03
4.25
3.20
3.66
2.87
3.36
2.70
3.17
2.59
3.05
2.52
2.96
2.47
2.90
2.43
2.85
2.40
2.78
2.35
2.69
2.30
12
6.11
4.19
4.39
3.31
3.77
2.97
3.45
2.78
3.26
2.67
3.13
2.59
3.04
2.54
2.97
2.50
2.92
2.46
2.84
2.42
2.75
2.36
14
6.29
4.32
4.51
3.41
3.86
3.05
3.53
2.86
3.33
2.74
3.20
2.66
3.11
2.60
3.04
2.56
2.98
2.52
2.90
2.47
2.81
2.41
16
6.45
4.44
4.61
3.49
3.94
3.11
3.60
2.92
3.48
2.79
3.26
2.71
3.16
2.65
3.09
2.61
3.03
2.57
2.95
2.52
2.85
2.46
EPA QA/G-9
QAOO Version
A-27
Final
July 2000
-------
TABLE A-14. DUNNETT'S TEST (ONE TAILED) (CONTINUED)
TOTAL NUMBER OF INVESTIGATED GROUPS (K - 1)
Degrees of Freedom
20
24
30
40
50
60
70
80
90
100
120
00
a
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
.05
.10
2
2.03
1.64
2.01
1.63
1.99
1.62
1.97
1.61
1.96
1.61
1.95
1.60
1.95
1.60
1.94
1.60
1.94
1.60
1.93
1.59
1.93
1.59
1.92
1.58
3
2.19
1.81
2.17
2.80
2.15
1.79
2.13
1.77
2.11
1.77
2.10
1.76
2.10
1.76
2.10
1.76
2.09
1.76
2.08
1.75
2.08
1.75
2.06
1.73
4
2.30
1.93
2.28
1.91
2.25
1.90
2.23
1.88
2.22
1.88
2.21
1.87
2.21
1.87
2.20
1.87
2.20
1.86
2.18
1.85
2.18
1.85
2.16
1.84
5
2.39
2.01
2.36
2.00
2.34
1.98
2.31
1.96
2.29
1.96
2.28
1.95
2.28
1.95
2.28
1.95
2.27
1.94
2.27
1.93
2.26
1.93
2.23
1.92
6
2.46
2.08
2.43
2.06
2.40
2.05
2.37
2.03
2.32
2.02
2.34
2.01
2.34
2.01
2.34
2.01
2.33
2.00
2.33
1.99
2.32
1.99
2.29
1.98
7
2.51
2.14
2.48
2.12
2.45
2.10
2.42
2.08
2.41
2.07
2.40
2.06
2.40
2.06
2.39
2.06
2.39
2.06
2.38
2.05
2.37
2.05
2.34
2.03
8
2.56
2.19
2.53
2.17
2.50
2.15
2.47
2.13
2.45
2.12
2.44
2.11
2.44
2.11
2.43
2.10
2.43
2.10
2.42
2.09
2.41
2.09
2.38
2.07
9
2.60
2.23
2.57
2.21
2.54
2.19
2.51
2.17
2.49
2.16
2.48
2.15
2.48
2.15
2.47
2.15
2.47
2.14
2.46
2.14
2.45
2.13
2.42
2.11
10
2.64
2.26
2.60
2.24
2.57
2.22
2.54
2.20
2.52
2.19
2.51
2.18
2.51
2.18
2.50
2.18
2.50
2.17
2.49
2.17
2.48
2.16
2.45
2.14
12
2.70
2.33
2.66
2.30
2.63
2.28
2.60
2.26
2.58
2.25
2.57
2.24
2.56
2.24
2.55
2.23
2.55
2.23
2.54
2.22
2.53
2.22
2.50
2.20
14
2.75
2.38
2.72
2.35
2.68
2.33
2.65
2.31
2.63
2.30
2.61
2.29
2.61
2.29
2.60
2.28
2.60
2.28
2.59
2.27
2.58
2.27
2.55
2.24
16
2.80
2.42
2.76
2.40
2.72
2.37
2.69
2.35
2.67
2.34
2.65
2.33
2.65
2.33
2.64
2.32
2.63
2.31
2.63
2.31
2.62
2.31
2.58
2.28
EPA QA/G-9
QQAOO Version
A-28
Final
May 2000
-------
TABLE A-15. APPROXIMATE a-LEVEL CRITICAL POINTS FOR
RANK VON NEUMANN RATIO TEST
n/a
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
32
34
36
38
40
42
44
46
48
50
55
60
65
70
75
80
85
90
95
100
.050
0.70
0.80
0.86
0.93
0.98
1.04
1.08
1.11
1.14
1.17
1.19
1.21
1.24
1.26
1.27
1.29
1.31
1.32
1.33
1.35
1.36
1.37
1.38
1.39
1.40
1.41
1.43
1.45
1.46
1.48
1.49
1.50
1.51
1.52
1.53
1.54
1.56
1.58
1.60
1.61
1.62
1.64
1.65
1.66
1.66
1.67
.100
0.60
0.97
1.11
1.14
1.18
1.23
1.26
1.29
1.32
1.34
1.36
1.38
1.40
1.41
1.43
1.44
1.45
1.46
1.48
1.49
1.50
1.51
1.51
1.52
1.53
1.54
1.55
1.57
1.58
1.59
1.60
1.61
1.62
1.63
1.63
1.64
1.66
1.67
1.68
1.70
1.71
1.71
1.72
1.73
1.74
1.74
EPA QA/G-9
QAOO Version
A-29
Final
July 2000
-------
This page is intentionally blank.
EPA QA/G-9 Final
QAOO Version A - 30 July 2000
-------
APPENDIX B
REFERENCES
EPA QA/G-9 Final
QAOO Version B - 1 July 2000
-------
APPENDIX B: REFERENCES
This appendix provides references for the topics and procedures described in this
document. The references are broken into three groups: Primary, Basic Statistics Textbooks, and
Secondary. This classification does not refer in any way to the subject matter content but to the
relevance to the intended audience for this document, ease in understanding statistical concepts
and methodologies, and accessability to the non-statistical community. Primary references are
those thought to be of particular benefit as hands-on material, where the degree of sophistication
demanded by the writer seldom requires extensive training in statistics; most of these references
should be on an environmental statistician's bookshelf. Users of this document are encouraged to
send recommendations on additional references to the address listed in the Foreword.
Some sections within the chapters reference materials found in most introductory statistics
books. This document uses Walpole and Myers (1985), Freedman, Pisani, Purves, and Adhakari
(1991), Mendenhall (1987), and Dixon and Massey (1983). Table B-l (at the end of this
appendix) lists specific chapters in these books where topics contained in this guidance may be
found. This list could be extended much further by use of other basic textbooks; this is
acknowledged by the simple statement that further information is available from introductory text
books.
Some important books specific to the analysis of environmental data include: Gilbert
(1987), an excellent all-round handbook having strength in sampling, estimation, and hot-spot
detection; Gibbons (1994), a book specifically concentrating on the application of statistics to
groundwater problems with emphasis on method detection limits, censored data, and the detection
of outliers; and Madansky (1988), a slightly more theoretical volume with important chapters on
the testing for Normality, transformations, and testing for independence. In addition, Ott (1995)
describes modeling, probabilistic processes, and the Lognormal distribution of contaminants, and
Berthouex and Brown (1994) provide an engineering approach to problems including estimation,
experimental design and the fitting of models.
B.I CHAPTER 1
Chapter 1 establishes the framework of qualitative and quantitative criteria against which
the data that has been collected will be assessed. The most important feature of this chapter is the
concept of the test of hypotheses framework which is described in any introductory textbook. A
non-technical exposition of hypothesis testing is also to be found in U.S. EPA (1994a, 1994b)
which provides guidance on planning for environmental data collection. An application of the
DQO Process to geostatistical error management may be found in Myers (1997).
A full discussion of sampling methods with the attendant theory are to be found in Gilbert
(1987) and a shorter discussion may be found in U.S. EPA (1989). Cochran (1966) and Kish
(1965) also provide more advanced theoretical concepts but may require the assistance of a
statistician for full comprehension. More sophisticated sampling designs such as composite
EPA QA/G-9 Final
QAOO Version B - 2 July 2000
-------
sampling, adaptive sampling, and ranked set sampling, will be discussed in future Agency
guidance.
B.2 CHAPTER 2
Standard statistical quantities and graphical representations are discussed in most
introductory statistics books. In addition, Berthouex & Brown (1994) and Madansky (1988) both
contain thorough discussions on the subject. There are also several textbooks devoted exclusively
to graphical representations, including Cleveland (1993), which may contain the most applicable
methods for environmental data, Tufte (1983), and Chambers, Cleveland, Kleiner and Tukey
(1983).
Two EPA sources for temporal data that keep theoretical discussions to a minimum are
U.S. EPA (1992a) and U.S. EPA (1992b). For a more complete discussion on temporal data,
specifically time series analysis, see Box and Jenkins (1970), Wei (1990), or Ostrum (1978).
These more complete references provide both theory and practice; however, the assistance of a
statistician may be needed to adapt the methodologies for immediate use. Theoretical discussions
of spatial data may be found in Journel and Huijbregts (1978), Cressie (1993), and Ripley (1981).
B.3 CHAPTER 3
The hypothesis tests covered in this edition of the guidance are well known and straight-
forward; basic statistics texts cover these subjects. Besides basic statistical text books, Berthouex
& Brown (1994), Hardin and Gilbert (1993), and U.S. EPA (1989, 1994c) may be useful to the
reader. In addition, there are some statistics books devoted specifically to hypothesis testing, for
example, see Lehmann (1991). These books may be too theoretical for most practitioners, and
their application to environmental situations may not be obvious.
The statement in this document that the sign test requires approximately 1.225 times as
many observations as the Wilcoxon rank sum test to achieve a given power at a given significance
level is attributable to Lehmann (1975).
B.4 CHAPTER 4
This chapter is essentially a compendium of statistical tests drawn mostly from the primary
references and basic statistics textbooks. Gilbert (1987) and Madansky (1988) have an excellent
collection of techniques and U.S. EPA (1992a) contains techniques specific to water problems.
For Normality (Section 4.2), Madansky (1988) has an excellent discussion on tests as does
Shapiro (1986). For trend testing (Section 4.3), Gilbert (1987) has an excellent discussion on
statistical tests and U.S. EPA (1992b) provides adjustments for trends and seasonality in the
calculation of descriptive statistics.
EPA QA/G-9 Final
QAOO Version B - 3 July 2000
-------
There are several very good textbooks devoted to the treatment of outliers (Section 4.4).
Two authoritative texts are Barnett and Lewis (1978) and Hawkins (1980). Additional
information is also to be found in Beckman and Cook (1983) and Tietjen and Moore (1972).
Several useful software programs are available on the statistical market including U.S. EPA's
GEO-EASE and Scout, both developed by the Environmental Monitoring Systems Laboratory,
Las Vegas, Nevada and described in U.S. EPA (1991) and U.S. EPA (1993b), respectively.
Tests for dispersion (Section 4.5) are described in the basic textbooks and examples are to
be found in U.S. EPA (1992a). Transformation of data (Section 4.6) is a sensitive topic and
thorough discussions may be found in Gilbert (1987), and Dixon and Massey (1983). Equally
sensitive is the analysis of data where some values are recorded as non-detected (Section 4.7);
Gibbons (1994) and U.S. EPA (1992a) have relevant discussions and examples.
B.5 CHAPTER 5
Chapter 5 discusses some of the philosophical issues related to hypothesis testing which
may help in understanding and communicating the test results. Although there are no specific
references for this chapter, many topics (e.g., the use of p-values) are discussed in introductory
textbooks. Future editions of this guidance will be expanded by incorporating practical
experiences from the environmental community into this chapter.
B.6 LIST OF REFERENCES
B.6.1 Primary References
Berthouex, P.M., and L.C. Brown, 1994. Statistics for Environmental Engineers. Lewis, Boca
Raton, FL.
Gilbert, R.O., 1987. Statistical Methods for Environmental Pollution Monitor ing. John Wiley,
New York, NY.
Gibbons, R. D., 1994. Statistical Methods for Groundwater Monitoring. John Wiley, New
York, NY.
Madansky, A., 1988. Prescriptions for Working Statisticians. Springer-Verlag, New York, NY.
Ott, W.R., 1995. Environmental Statistics and Data Analysis. Lewis, Boca Raton, FL.
U.S. Environmental Protection Agency, 1996. The Data Quality Evaluation Statistical Toolbox
(DataQUEST) Software, EPA QA/G-9D. Office of Research and Development.
U.S. Environmental Protection Agency, 1994a. Guidance for the Data Quality Objectives
Process (EPA QA/G4). EPA/600/R-96/055. Office of Research and Development.
EPA QA/G-9 Final
QAOO Version B - 4 July 2000
-------
U.S. Environmental Protection Agency, 1994b. The Data Quality Objectives Decision Error
Feasibility Trials (DEFT) Software (EPA QA/G-4D). EPA/600/R-96/056. Office of
Research and Development.
U.S. Environmental Protection Agency , 1992a. Guidance Document on the Statistical Analysis
of Ground-Water Monitoring Data at RCRA Facilities. EPA/530/R-93/003. Office of
Solid Waste. (NTIS: PB89-151026)
B.6.2 Basic Statistics Textbooks
Dixon, W.J., and FJ. Massey, Jr., 1983. Introduction to Statistical Analysis (Fourth Edition).
McGraw-Hill, New York, NY.
Freedman, D., R. Pisani, R. Purves, and A. Adhikari, 1991. Statistics. W.W. Norton & Co.,
New York, NY.
Mendenhall, W., 1987. Introduction to Probability and Statistics (Seventh Edition). PWS-Kent,
Boston, MA.
Walpole, R., and R. Myers, 1985. Probability and Statistics for Engineers and Scientists (Third
Ed.). MacMillan, New York, NY.
B.6.3 Secondary References
Aitchison, J., 1955. On the distribution of a positive random variable having a discrete probability
mass at the origin. Journal of American Statistical Association 50(272):901-8
Barnett, V., and T. Lewis, 1978. Outliers in Statistical Data. John Wiley, New York, NY.
Beckman, R.J., and R.D. Cook, 1983. Outlier s, Technometrics 25:119-149.
Box[, G.E.P., and G.M. Jenkins, 1970. Time Series Analysis, Forecasting, and Control.
Holden-Day, San Francisco, CA.
Chambers, J.M, W.S. Cleveland, B. Kleiner, and P.A. Tukey, 1983. Graphical Methods for Data
Analysis. Wadsworth & Brooks/Cole Publishing Co., Pacific Grove, CA.
Chen, L., 1995. Testing the mean of skewed distributions, Journal of the American Statistical
Association 90:767-772.
Cleveland, W.S., 1993. Visualizing Data. Hobart Press, Summit, NJ.
Cochran, W. G., 1966. Sampling Techniques (Third Edition). John Wiley, New York, NY.
EPA QA/G-9 Final
QAOO Version B - 5 July 2000
-------
Cohen, A.C., Jr. 1959. Simplified estimators for the normal distribution when samples are singly
censored or truncated, Technometrics 1:217-237.
Conover, W.J., 1980. PracticalNonparametric Statistics (Second Edition). John Wiley, New
York, NY.
Cressie, N., 1993. Statistics for Spatial Data. John Wiley, New York, NY.
D'Agostino, R.B., 1971. An omnibus test of normality for moderate and large size samples,
Biometrika 58:341-348.
David, H.A., H.O. Hartley, and E.S. Pearson, 1954. The distribution of the ratio, in a single
normal sample, of range to standard deviation, Biometrika 48:41-55.
Dixon, W.J., 1953. Processing data for outliers, Biometrika 9:74-79.
Filliben, J.J., 1975. The probability plot correlation coefficient test for normality, Technometrics
17:111-117.
Geary, R.C., 1947. Testing for normality, Biometrika 34:209-242.
Geary, R.C., 1935. The ratio of the mean deviation to the standard deviation as a test of
normality, Biometrika 27:310-32.
Grubbs, F.E., 1969. Procedures for detecting outlying observations in samples, Technometrics
11:1-21.
Hardin, J.W., and R.O. Gilbert, 1993. Comparing Statistical Tests for Detecting Soil
Contamination Greater than Background, Report to U.S. Department of Energy, PNL-
8989, UC-630, Pacific Northwest Laboratory, Richland, WA.
Hawkins, D.M., 1980. Identification of Outliers. Chapman and Hall, New York, NY.
Hochberg, Y., and A. Tamhane, 1987. Multiple Comparison Procedures. John Wiley, New
York, NY.
Journel, A.G., and C.J. Huijbregts, 1978. Mining Geostatistics. Academic Press, London.
Kish, L., 1965. Survey Sampling. John Wiley, New York, NY.
Kleiner, B., and J.A. Hartigan, 1981. Representing points in many dimensions by trees and
castles, Journal of the American Statistical Association 76:260.
EPA QA/G-9 Final
QAOO Version B - 6 July 2000
-------
Lehmann, E.L., 1991. Testing Statistical Hypotheses. Wadsworth & Brooks/Cole Publishing
Co., Pacific Grove, CA.
Lehmann, E.L., 1975. Nonparametrics: Statistical Methods Based on Ranks. Holden-Day, Inc.,
San Francisco, CA.
Lilliefors, H.W., 1969. Correction to the paper "On the Kolmogorov-Smirnov test for normality
with mean and variance unknown," Journal of the American Statistical Association
64:1702.
Lilliefors, H.W., 1967. On the Kolmogorov-Smirnov test for normality with mean and variance
unknown, Journal of the American Statistical Association 64:399-402.
Myers, J.C., 1997. Geostatistical Error Management., John Wiley, New York, NY.
Ostrum, C.W., 1978. Time Series Analysis (Second Edition). Sage University Papers Series, Vol
9. Beverly Hills and London.
Ripley, B.D., 1981. Spatial Statistics. John Wiley and Sons, Somerset, NJ.
Rosner, B., 1975. On the detection of many outliers, Technometrics 17:221-227'.
Royston, J.P., 1982. An extension of Shapiro and Wilk's W test for normality to large samples,
Applied Statistics 31:161-165.
Sen, P.K., 1968a. Estimates of the regression coefficient based on Kendall's tau, Journal of the
American Statistical Association 63:1379-1389.
Sen, P.K., 1968b. On a class of aligned rank order tests in two-way layouts, Annals of
Mathematical Statistics 39:1115-1124.
Shapiro, S., 1986. Volume 3: How to Test Normality and Other Distributional Assumptions.
American Society for Quality Control, Milwaukee, WI.
Shapiro, S., and M.B. Wilk, 1965. An analysis of variance test for normality (complete samples),
Biometrika 52:591-611.
Siegel, J.H., R.W. Goldwyn, and H.P. Friedman, 1971. Pattern and process of the evolution of
human septic shock, Surgery 70:232.
Stefansky, W., 1972. Rejecting outliers in factorial designs, Technometrics 14:469-478.
Stephens, M.A., 1974. EDF statistics for goodness-of-fit and some comparisons, Journal of the
American Statistical Association 90:730-737.
EPA QA/G-9 Final
QAOO Version B - 7 July 2000
-------
Tietjen, G.L., and R.M. Moore, 1972. Some Grubbs-type statistics for the detection of several
outliers, Technometrics 14:583-597.
Tufte, E.R., 1983. The Visual Display of Quantitative Information. Graphics Press, Chesire,
CN.
Tufte, E.R., 1990. Envisioning Information. Graphics Press, Chesire, CN.
U. S. Environmental Protection Agency, 1994c. Methods for Evaluating the Attainments of
Cleanup Standards: Volume 3: Reference-Based Standards. EPA/230/R-94-004. Office
of Policy, Planning, and Evaluation. (NTIS: PB94-176831)
U.S. Environmental Protection Agency, 1993 a. The Data Quality Objectives Process for
Superfund: Interim Final Guidance. EPA/540/R-93/071. Office of Emergency and
Remedial Response.
U.S. Environmental Protection Agency, 1993b. Scout: A Data Analysis Program.
Environmental Monitoring Systems Laboratory, Office of Research and Development.
(NTIS: PB93-505303)
U. S. Environmental Protection Agency, 1992b. Methods for Evaluating the Attainments of
Cleanup Standards: Volume 2: Ground Water. EPA/230/R-92/014. Office of Policy,
Planning, and Evaluation. (NTIS: PB94-138815)
U.S. Environmental Protection Agency, 1991. GEO-EAS 1.2.1. User's Guide. EPA 600/8-
91/008. Environmental Monitoring Systems Laboratory, Office of Research and
Development. (NTIS: PB93-504967)
U. S. Environmental Protection Agency, 1989. Methods for Evaluating the Attainments of
Cleanup Standards: Volume 1: Soils and Solid Media. EPA/230/02-89-042. Office of
Policy, Planning, and Evaluation. (NTIS: PB89-234959;
Walsh, I.E., 1958. Large sample nonparametric rejection of outlying observations, Annals of the
Institute of Statistical Mathematics 10:223-232.
Walsh, I.E., 1953. Correction to "Some nonparametric tests of whether the largest observations
of a set are too large or too small," Annals of Mathematical Statistics 24:134-135.
Walsh, I.E., 1950. Some nonparametric tests of whether the largest observations of a set are too
large or too small, Annals of Mathematical Statistics 21:583-592.
Wang, Peter C.C., 1978. Graphical Representation of Multivariate Data. Academic Press, New
York, NY.
EPA QA/G-9 Final
QAOO Version B - 8 July 2000
-------
Wegman, Edward J., 1990. Hyperdimensional data analysis using parallel coordinates, Journal of
the American Statistical Association 85: 664.
Wei, W.S., 1990. Time Series Analysis (SecondEdition). Addison Wesley, Menlo Park, CA.
EPA QA/G-9 Final
QAOO Version B - 9 July 2000
-------
United States Office of Research and EPA/600/R-98/018
Environmental Protection Development February 1998
Agency Washington, B.C. 20460
EPA GUIDANCE FOR
QUALITY ASSURANCE
PROJECT PLANS
EPA QA/G-5
-------
FOREWORD
The U.S. Environmental Protection Agency (EPA) has developed the Quality Assurance Project
Plan (QAPP) as an important tool for project managers and planners to document the type and quality of
data needed for environmental decisions and to use as the blueprint for collecting and assessing those
data from environmental programs. The development, review, approval, and implementation of the
QAPP is part of the mandatory Agency-wide Quality System that requires all organizations performing
work for EPA to develop and operate management processes and structures for ensuring that data or
information collected are of the needed and expected quality for their desired use. The QAPP is an
integral part of the fundamental principles of quality management that form the foundation of the
Agency's Quality System and the requirements for a QAPP are contained in EPA QA/R-5, EPA
Requirements for Quality Assurance Project Plans for Environmental Data Operations.
This document is one of the U.S. Environmental Protection Agency Quality System Series
requirements and guidance documents. These documents describe the EPA policies and procedures for
planning, implementing, and assessing the effectiveness of the Quality System. Requirements
documents (identified as EPA/R-x) establish criteria and mandatory specifications for quality assurance
(QA) and quality control (QC) activities. Guidance documents (identified as EPA QA/G-x) provide
suggestions and recommendations of a nonmandatory nature for using the various components of the
Quality System. This guidance document contains advice and recommendations on how to meet the
requirements of EPA QA/R-5. In addition to this guidance document on writing a QAPP, other EPA
documents are available to assist the QAPP writer; these are discussed in Appendix A. Effective use of
this document assumes that appropriate management systems for QA and QC have been established by
the implementing organization and are operational. For requirements and guidance on the structure of
this management system, refer to Appendix A.
Questions regarding this document or other documents from the Quality System Series may be
directed to:
U.S. EPA
Quality Staff (2811R)
Office of Environmental Information
401 M Street, SW
Washington, DC 20460
Phone: (202)564-6830
Fax: (202) 565-2441
All requirements and guidance documents are available on the EPA's Quality Staff website:
http://www.epa.gov/quality
EPA QA/G-5 1 QA98 (1/00)
-------
EPA QA/G-5
QA98
-------
TABLE OF CONTENTS
CHAPTER I. INTRODUCTION 1
OVERVIEW 1
PURPOSE OF QA PLANNING 2
CHAPTER II. QAPP REQUIREMENTS 3
EPA POLICY ON QAPPS 3
QAPP GROUPS AND ELEMENTS 3
QAPP RESPONSIBILITIES 5
CHAPTER III. QAPP ELEMENTS 7
A. PROJECT MANAGEMENT 7
Al Title and Approval Sheet 7
A2 Table of Contents and Document Control Format 7
A3 Distribution List 8
A4 Project/Task Organization 8
A5 Problem Definition/Background 10
A6 Project/Task Description and Schedule 11
A7 Quality Objectives and Criteria for Measurement Data 12
A8 Special Training Requirements/Certification 13
A9 Documentation and Records 14
B. MEASUREMENT/DATA ACQUISITION 17
Bl Sampling Process Design (Experimental Design) 17
B2 Sampling Methods Requirements 19
B3 Sample Handling and Custody Requirements 23
B4 Analytical Methods Requirements 28
B5 Quality Control Requirements 30
B6 Instrument/Equipment Testing, Inspection, and Maintenance Requirements . . 32
B7 Instrument Calibration and Frequency 33
B8 Inspection/Acceptance Requirements for Supplies and Consumables 35
B9 Data Acquisition Requirements (Non-Direct Measurements) 37
BIO Data Management 38
C. ASSESSMENT/OVERSIGHT 41
Cl Assessments and Response Actions 41
C2 Reports to Management 44
D. DATA VALIDATION AND USABILITY 45
Dl Data Review, Validation, and Verification Requirements 45
D2 Validation and Verification Methods 47
D3 Reconciliation with Data Quality Objectives 48
EPA QA/G-5 iii QA98
-------
CHAPTER IV. QAPP REVISIONS AND RELATED GUIDANCE 49
QAPP REVISIONS 49
COMPARISON WITH PREVIOUS GUIDANCE (QAMS 005/80) 49
APPENDIX A. CROSSWALKS BETWEEN QUALITY ASSURANCE DOCUMENTS A-l
AA1. Relationship Between E4 and EPA Quality System A-l
AA2. Crosswalk Between QA/R-5 and QAMS-005/80 A-3
AA3. Crosswalk Between EPA QA/R-5 and ISO 9000 A-4
AA4. Crosswalk Between the DQO Process and the QAPP A-5
AA5. EPA Quality Assurance Documents A-7
APPENDIX B. GLOSSARY OF QUALITY ASSURANCE AND RELATED TERMS B-l
APPENDIX C. CHECKLISTS USEFUL IN QUALITY ASSURANCE REVIEW C-l
AC1. Sample Handling, Preparation, and Analysis Checklist C-l
AC2. QAPP Review Checklist C-5
ACS. Chain-of-Custody Checklist C-8
APPENDIX D. DATA QUALITY INDICATORS D-l
ADI. Principal DQIs: PARCC D-l
AD2. Other Data Quality Indicators D-5
APPENDIX E. QUALITY CONTROL TERMS E-l
AE1. Quality Control Operations E-l
AE2. Quality Control Requirements in Existing Programs E-3
APPENDIX F. SOFTWARE FOR THE DEVELOPMENT AND PREPARATION
OF A QUALITY ASSURANCE PROJECT PLAN F-l
AF1. Overview of Potential Need for Software in QAPP Preparation F-l
AF2. Existing Software F-3
AF3. Software Availability and Sources F-5
APPENDIX G. ISSUES IN DATA MANAGEMENT G-l
AG1. Introduction G-l
AG2. Regulatory and Policy Framework G-l
AG3. QA Planning for Information Systems G-3
AG4. References G-14
EPA QA/G-5 IV QA98
-------
LIST OF FIGURES
Figure 1. QA Planning and the Data Life Cycle 2
Figure 2. An Example of a Table of Contents and a Distribution List 9
Figure 3. An Example of a Project Organization Chart 10
Figure 4. The DQO Process 12
Figure 5. An Example of a Sample Log Sheet 25
Figure 6. An Example of a Sample Label 26
Figure 7. An Example of a Custody Seal 26
Figure 8. An Example of a Chain-Of-Custody Record 27
Figure 9. Example of a Record for Consumables 36
Figure 10. Example of Inspection/Acceptance Testing Requirements 36
Figure 11. Example of a Log for Tracking Supplies and Consumables 36
Figure AA1. Relationships Among EPA Quality System Documents at the Program Level A-9
Figure AA2. Relationships Among EPA Quality System Documents at the Project Level A-10
Figure ADI. Measurement Bias and Random Measurement Uncertainties. Shots at a Target D-3
LIST OF TABLES
Table 1. Project Quality Control Checks 31
Table AA1. Numbering System for EPA's Quality System Documents A-7
Table AA2. Quality System Documents A-8
Table AD 1. Principal Types of Error D-6
Table AE1. Comparison of QC Terms E-6
Table AE2. QC Requirements for Programs E-13
Table AE3. QC Requirements for Methods E-16
Table API. Software Available to Meet QAPP Development Needs F-4
Table AG1. Project Scope and Risks G-5
Table AG2. Software Development Life Cycle G-7
EPA QA/G-5 V QA98
-------
LIST OF ACRONYMS
ACS American Chemical Society
ADQ Audit of Data Quality
CFR Code of Federal Regulations
DQA Data Quality Assessment
DQI Data Quality Indicator
DQO Data Quality Objective
EPA Environmental Protection Agency
ISO International Organization for Standardization
MSR Management Systems Review
NIST National Institute of Standards and Technology
OSHA Occupational Safety and Health Administration
PARCC Precision, Accuracy, Representativeness, Comparability, and Completeness
PE Performance Evaluation
QA Quality Assurance
QAD Quality Assurance Division
QAMS Quality Assurance Management Staff (now QAD)
QAPP Quality Assurance Project Plan
QC Quality Control
RCRA Resource Conservation and Recovery Act
SOP Standard Operating Procedure
SRM Standard Reference Material
TSA Technical Systems Audit
EPA QA/G-5
VI
QA98
-------
CHAPTER I
INTRODUCTION
OVERVIEW
This document presents detailed guidance on how to develop a Quality Assurance Project Plan
(QAPP) for environmental data operations performed by or for the U.S. Environmental Protection
Agency (EPA). This guidance discusses how to address and implement the specifications in
Requirements for QA Project Plans for Environmental Data Operations (EPA QA/R-5).
The QAPP is the critical planning document for any environmental data collection operation
because it documents how quality assurance (QA) and quality control (QC) activities will be
implemented during the life cycle of a program, project, or task. The QAPP is the blueprint for
identifying how the quality system of the organization performing the work is reflected in a particular
project and in associated technical goals. QA is a system of management activities designed to ensure
that the data produced by the operation will be of the type and quality needed and expected by the data
user. QA is acknowledged to be a management function emphasizing systems and policies, and it aids
the collection of data of needed and expected quality appropriate to support management decisions in a
resource-efficient manner.
In order to obtain environmental data for decision making, a project should be conducted in three
phases: planning, implementation, and assessment. The first phase involves the development of Data
Quality Objectives (DQOs) using the DQO Process or a similar structural systematic planning process.
The DQOs provide statements about the expectations and requirements of the data user (such as the
decision maker). In the second phase, the QAPP translates these requirements into measurement
performance specifications and QA/QC procedures for the data suppliers to provide the information
needed to satisfy the data user's needs. This guidance links the results of the DQO Process with the
QAPP to complete documentation of the planning process. Once the data have been collected and
validated in accordance with the elements of the QAPP, the data should be evaluated to determine
whether the DQOs have been satisfied. In the assessment phase, the Data Quality Assessment (DQA)
Process applies statistical tools to determine whether the data meet the assumptions made during
planning and whether the total error in the data is small enough to support a decision within tolerable
decision error rates expressed by the decision maker. Plans for data validation and DQA are discussed in
the final sections of the QAPP. Thus, the activities addressed and documented in the QAPP cover the
entire project life cycle, integrating elements of the planning, implementation, and assessment phases.
A QAPP is composed of four sections of project-related information called "groups," which are
subdivided into specific detailed "elements." The degree to which each QAPP element should be
addressed will be dependent on the specific project and can range from "not applicable" to extensive
documentation. This document provides a discussion and background of the elements of a QAPP that
will typically be necessary. There is no Agency-wide template for QAPP format; however, QAD
encourages organizational consistency in the presentation and content of the elements contained within
the QAPP. The final decision on the specific need for these elements for project-specific QAPPs will be
made by the overseeing or sponsoring EPA organization(s). The Agency encourages the specific
tailoring of implementation documents within the EPA's general QA framework on a project-specific
basis.
EPA QA/G-5 1 QA98
-------
PLANNING
Data Quality Objectives Process
Quality Assurance Project Plan Development
IMPLEMENTATION
Field Data Collection and Associated
Quality Assurance/Quality Control Activities
ASSESSMENT
Data Validation
Data Quality Assessment
QA PLANNING FOR
DATA COLLECTION
Data Quality Objectives Process
i
r OUTPUTS ^
/Data ,
Quality /
Objectives //
i
r
// Data /
/ Collection /
Design /
, INPUTS 1
r
Quality Assurance Project Plan
Development
^
r
Quality Assurance
Project Plan
^
Figure 1. QA Planning and the Data Life Cycle.
PURPOSE OF QA PLANNING
The EPA Quality System is a structured and documented management system describing the
policies, objectives, principles, organization, responsibilities, accountability, and implementation plan of
an organization for ensuring quality in its work processes, products, and services. The Agency's Quality
System is described in EPA QA/G-0, The EPA Quality System.
EPA policy requires that all projects involving the generation, acquisition, and use of
environmental data be planned and documented and have an Agency-approved QAPP prior to the start of
data collection. The primary purpose of the QAPP is to provide an overview of the project, describe the
need for the measurements, and define QA/QC activities to be applied to the project, all within a single
document. The QAPP should be detailed enough to provide a clear description of every aspect of the
project and include information for every member of the project staff, including samplers, lab staff, and
data reviewers. The QAPP facilitates communication among clients, data users, project staff,
management, and external reviewers. Effective implementation of the QAPP assists project managers in
keeping projects on schedule and within the resource budget. Agency QA policy is described in the
Quality Manual and EPA QA/R-1, EPA Quality System Requirements for Environmental Programs.
EPA QA/G-5
QA98
-------
CHAPTER II
QAPP REQUIREMENTS
EPA POLICY ON QAPPS
It is EPA's internal policy requirement1 that the collection of environmental data by or for the
Agency be supported by a QA program, or quality system. The authority for this requirement for work
done for EPA through extramural agreements may be found in 48 CFR, Chapter 15, Part 1546 for
contractors, and 40 CFR, Parts 30, 31, and 35 for financial assistance recipients, and may be included in
negotiated interagency agreements and consent agreements in enforcement actions.
A key component of this mandatory quality system is the development, review, approval, and
implementation of the QAPP. A QAPP must address all of the elements contained in QA/R-5 unless
otherwise specified by the EPA QA Manager responsible for the data collection. The format of the
QAPP is decided by the QA approving authority prior to preparation of the QAPP.
The QAPP is the logical product of the planning process for any data collection, as it documents
how QA and QC activities will be planned and implemented. To be complete, the QAPP must meet
certain specifications for detail and coverage, but the extent of detail is dependent on the type of project,
the data to be collected, and the decisions to be made. Overall, the QAPP must provide sufficient detail
to demonstrate that:
the project's technical and quality objectives are identified and agreed upon,
• the intended measurements or data acquisition methods are consistent with project
objectives,
• the assessment procedures are sufficient for determining if data of the type and quality
needed and expected are obtained, and
• any potential limitations on the use of the data can be identified and documented.
Documents prepared prior to the QAPP (e.g., standard operating procedures [SOPs], test plans, and
sampling plans) can be appended or, in some cases, incorporated by reference.
QAPP GROUPS AND ELEMENTS
The elements of a QAPP are categorized into "groups" according to their function.
Specifications for each element are found in EPA Requirements for Quality Assurance Project Plans
(EPA QA/R-5). Summaries of each requirement of the elements from that document are contained in a
box at the beginning of each specific element description. The elements of a QAPP are:
Group A: Project Management
This group of QAPP elements covers the general areas of project management, project history
and objectives, and roles and responsibilities of the participants. The following 9 elements ensure that
'EPA Order 5360.1, Policy and Program Requirements to Implement the Mandatory Quality Assurance
Program, was issued originally in April 1984 and will be revised in 1998.
EPA QA/G-5 3 QA98
-------
the project's goals are clearly stated, that all participants understand the goals and the approach to be
used, and that project planning is documented:
Al Title and Approval Sheet
A2 Table of Contents and Document Control Format
A3 Distribution List
A4 Project/Task Organization and Schedule
A5 Problem Definition/Background
A6 Project/Task Description
A7 Quality Objectives and Criteria for Measurement Data
A8 Special Training Requirements/Certification
A9 Documentation and Records
Group B: Measurement/Data Acquisition
This group of QAPP elements covers all of the aspects of measurement system design and
implementation, ensuring that appropriate methods for sampling, analysis, data handling, and QC are
employed and will be thoroughly documented:
Bl Sampling Process Design (Experimental Design)
B2 Sampling Methods Requirements
B3 Sample Handling and Custody Requirements
B4 Analytical Methods Requirements
B5 Quality Control Requirements
B6 Instrument/Equipment Testing, Inspection, and Maintenance Requirements
B7 Instrument Calibration and Frequency
B8 Inspection/Acceptance Requirements for Supplies and Consumables
B9 Data Acquisition Requirements (Non-Direct Measurements)
BIO Data Management
Group C: Assessment/Oversight
The purpose of assessment is to ensure that the QAPP is implemented as prescribed. This group
of QAPP elements addresses the activities for assessing the effectiveness of the implementation of the
project and the associated QA/QC activities:
C1 Assessments and Response Actions
C2 Reports to Management
Group D: Data Validation and Usability
Implementation of Group D elements ensures that the individual data elements conform to the
specified criteria, thus enabling reconciliation with the project's objectives. This group of elements
covers the QA activities that occur after the data collection phase of the project has been completed:
Dl Data Review, Validation, and Verification Requirements
D2 Validation and Verification Methods
D3 Reconciliation with Data Quality Objectives
EPA QA/G-5 4 QA98
-------
QAPP RESPONSIBILITIES
QAPPs may be prepared by EPA organizations and by groups outside EPA including contractors,
assistance agreement holders, or other Federal agencies under interagency agreements. Generally, all
QAPPs prepared by non-EPA organizations must be approved by EPA for implementation. Writing a
QAPP is often a collaborative effort within an organization, or among organizations, and depends on the
technical expertise, writing skills, knowledge of the project, and availability of the staff. Organizations
are encouraged to involve technical project staff and the QA Manager or the QA Officer in this effort to
ensure that the QAPP has adequate detail and coverage.
None of the environmental data collection work addressed by the QAPP may be started until the
initial QAPP has been approved by the EPA Project Officer and the EPA QA Manager and then
distributed to project personnel except under circumstances requiring immediate action to protect human
health and the environment or to operations conducted under police power. In some cases, EPA may
grant conditional or partial approval to a QAPP to permit some work to begin while noncritical
deficiencies in it are being resolved. However, the QA Manager should be consulted to determine the
length of time and nature of the work that may continue and the type of work that may be performed
under a conditionally approved QAPP. Some organizations have defined and outlined these terms as:
• Approval: No remaining identified deficiencies exist in the QAPP and the project may
commence.
• Partial Approval: Some activities identified in the QAPP still contain critical
deficiencies while other activities are acceptable. If the acceptable activities are not
contingent upon the completion of the activities with the deficiencies, a partial approval
may be granted to allow those activities to proceed. Work will continue to resolve the
portions of the QAPP that contain deficiencies.
• Conditional Approval: Approval of the QAPP or portions thereof will be granted upon
agreement to implement specific conditions, specific language, etc. by entities required
to approve the QAPP in order to expedite the initiation of field work. In most situations,
the conditional approval is upgraded to final approval upon receipt, review, and sign off
by all entities of the revised/additional QAPP pages.
The organizational group performing the work is responsible for implementing the approved
QAPP. This responsibility includes ensuring that all personnel involved in the work have copies of or
access to the approved QAPP along with all other necessary planning documents. In addition, the group
must ensure that these personnel understand their requirements prior to the start of data generation
activities.
Moreover, organizations are responsible for keeping the QAPP current when changes to
technical aspects of the project change. QAPPs must be revised to incorporate such changes and the
QAPP must be re-examined to determine the impact of the changes. Any revisions to the QAPP must be
re-approved and distributed to all participants in the project.
EPA QA/G-5 5 QA98
-------
EPA QA/G-5
QA98
-------
CHAPTER III
QAPP ELEMENTS
A PROJECT MANAGEMENT
The following project management elements address the procedural aspects of project
development and what to include in the QAPP project background, task description, and quality
objectives elements. Summaries from R-5 are contained in the text box following the title of each
element.
Al TITLE AND APPROVAL SHEET
Include title of plan; name of the organization(s); and names, titles, signatures of appropriate
approving officials, and their approval dates.
The title and approval sheet includes the title of the QAPP; the name(s) of the organization(s)
implementing the project; and the names, titles, and signatures, and the signature dates of the appropriate
approving officials. The approving officials typically include: the organization's Technical Project
Manager, the organization's Quality Assurance Officer or Manager, the EPA (or other funding agency)
Technical Project Manager/Project Officer, Laboratory Directors, Laboratory QA Officers, the EPA (or
other funding agency) Quality Assurance Officer or Manager, and other key staff, such as the QA Officer
of the prime contractor when a QAPP is prepared by a subcontractor organization.
The purpose of the approval sheet is to enable officials to document their approval of the QAPP.
The title page (along with the organization chart) also identifies the key project officials for the work.
The title and approval sheet should also indicate the date of the revision and a document number, if
appropriate.
A2 TABLE OF CONTENTS AND DOCUMENT CONTROL FORMAT
List sections, figures, tables, references, and appendices.
The table of contents lists all the elements, references, and appendices contained in a QAPP,
including a list of tables and a list of figures that are used in the text. The major headings for most
QAPPs should closely follow the list of required elements; an example is shown in Figure 2. While the
exact format of the QAPP does not have to follow the sequence given here, it is generally more
convenient to do so, and it provides a standard format to the QAPP reviewer. Moreover, consistency in
the format makes the document more familiar to users, who can expect to find a specific item in the same
place in every QAPP.
The table of contents of the QAPP may include a document control component. This information
should appear in the upper right-hand corner of each page of the QAPP when document control format is
desired. For example:
EPA QA/G-5 7 QA98
-------
Project No. or Name
Element or Section No.
Revision No.
Revision Date
Section/Element Page of.
This component, together with the distribution list (see element A3), facilitates control of the
document to help ensure that the most current QAPP is in use by all project participants. Each revision
of the QAPP should have a different revision number and date.
A3 DISTRIBUTION LIST
List all the individuals and their organizations who will receive copies of the approved QAPP
and any subsequent revisions. Include all persons who are responsible for implementation
(including managers), the QA managers, and representatives of all groups involved.
All the persons and document files designated to receive copies of the QAPP, and any planned
future revisions, need to be listed in the QAPP. This list, together with the document control
information, will help the project manager ensure that all key personnel in the implementation of the
QAPP have up-to-date copies of the plan. A typical distribution list appears in Figure 2.
A4 PROJECT/TASK ORGANIZATION
Identify the individuals or organizations participating in the project and discuss their specific
roles and responsibilities. Include principal data users, the decision makers, the project QA
manager, and all persons responsible for implementation.
Ensure that the project QA manager is independent of the unit generating the data.
Provide a concise organization chart showing the relationships and the lines of communication
among all project participants; other data users who are outside of the organization generating
the data; and any subcontractor relationships relevant to environmental data operations.
A4.1 Purpose/Background
The purpose of the project organization is to provide EPA and other involved parties with a clear
understanding of the role that each party plays in the investigation or study and to provide the lines of
authority and reporting for the project.
A4.2 Roles and Responsibilities
The specific roles, activities, and responsibilities of participants, as well as the internal lines of
authority and communication within and between organizations, should be detailed. The position of the
QA Manager or QA Officer should be described. Include the principal data users, the decision maker,
project manager, QA manager, and all persons responsible for implementation of the QAPP. Also
included should be the person responsible for maintaining the QAPP and any individual approving
EPA QA/G-5 8 QA98
-------
CONTENTS
Section
List of Tables iv
List of Figures v
A Project Management 1
1 Project/Task Organization 1
2 Problem Definition/Background 3
3 Project/Task Description 4
4 Data Quality Objectives 7
4.1 Project Quality Objectives 7
4.2 Measurement Performance Criteria 8
5 Documentation and Records 10
B Measurement Data Acquisition 11
6 Sampling Process Design 11
7 Analytical Methods Requirements 13
7.1 Organics 13
7.2 Inorganics 14
7.3 Process Control Monitoring 15
8 Quality Control Requirements 16
8.1 Field QC Requirements 16
8.2 Laboratory QC Requirements 17
9 Instrument Calibration and Frequency 19
10 Data Acquisition Requirements 20
11 Data Management 22
C Assessment/Oversight 23
12 Assessment and Response Actions 23
12.1 Technical Systems Audits 23
12.2 Performance Evaluation Audits 23
13 Reports to Management 24
D Data Validation and Usability 24
14 Data Review, Validation, and Verification Requirements 24
15 Reconciliation with Data Quality Objectives 26
15.1 Assessment of Measurement Performance 26
15.2 Data Quality Assessment 27
Distribution List
N. Wentworth, EPA/ORD (Work Assignment Manager)*
B. Waldron, EPA/ORD (QA Manager)
J. Warren, State University (Principal Investigator)
T. Dixon, State University (QA Officer)
G. Johnson, State University (Field Activities)
F. Haeberer, State University (Laboratory Activities)
B. Odom, State University (Data Management)
E. Renard, ABC Laboratories (Subcontractor Laboratory)
P. Lafornara, ABC Laboratories (QA Manager Subcontractor Laboratory)
indicates approving authority
Figure 2. An Example of a Table of Contents and a Distribution List
EPA QA/G-5 9 QA98
-------
deliverables other than the project manager. A concise chart showing the project organization, the lines
of responsibility, and the lines of communication should be presented; an example is given in Figure 3.
For complex projects, it may be useful to include more than one chart—one for the overall project (with
at least the primary contact) and others for each organization. Where direct contact between project
managers and data users does not occur, such as between a project consultant for a potentially
responsible party and the EPA risk assessment staff, the organization chart should show the route by
which information is exchanged.
EPA Work Assic
*N. We
2O2-56
Office of Researc
nment Manager
ntwo rth
n & Development
Principal Investigator
J. Warren
202-564-6876
State University
Engineering Department
EPA QA Manager
B. Waldron
2O2-2564-683O
ce of Research & Development
communication only
Project QA Officer
T. Dixon, post doctoral fellow
State University
Chemistry Department
Fie,d Activities Laboratory Activities ^^OdorT^'
G. Johnson, graduate student h" Haebe_rel". graduate assistant professor
91 9-541 -761 2 2O2 564 6872 202-564-6881
State University State Unlvefs^ M ^^ Un^ersi^
Engineering Department Chemistry Department Mathemat.cs Departmen
*approving authority
Subcontractor
ABC Laboratories
(GC/MS Analyses Only)
Laboratory Manager
E. Renard
908-321-4355
t QA Manager
P. Lafornara
9O8-9O6-6988
Figure 3. An Example of a Project Organization Chart
AS PROBLEM DEFINITION/BACKGROUND
State the specific problem to be solved or decision to be made and include sufficient
background information to provide a historical and scientific perspective for this particular
project.
A5.1 Purpose/Background
The background information provided in this element will place the problem in historical
perspective, giving readers and users of the QAPP a sense of the project's purpose and position relative
to other project and program phases and initiatives.
EPA QA/G-5
10
QA98
-------
A5.2 Problem Statement and Background
This discussion must include enough information about the problem, the past history, any
previous work or data, and any other regulatory or legal context to allow a technically trained reader to
make sense of the project objectives and activities. This discussion should include:
a description of the problem as currently understood, indicating its importance and
programmatic, regulatory, or research context;
• a summary of existing information on the problem, including any conflicts or
uncertainties that are to be resolved by the project;
a discussion of initial ideas or approaches for resolving the problem there were
considered before selecting the approach described in element A6, "Project/Task
Description"; and
• the identification of the principal data user or decision maker (if know).
Note that the problem statement is the first step of the DQO Process and the decision specification is the
second step of the DQO Process.
A6 PROJECT/TASK DESCRIPTION AND SCHEDULE
Provide a description of the work to be performed and the schedule for implementation.
Include measurements that will be made during the course of the project; applicable technical,
regulatory, or program-specific quality standards, criteria, or objectives; any special personnel
and equipment requirements; assessment tools needed; a schedule for work to be performed;
and project and quality records required, including types of reports needed.
A6.1 Purpose/Background
The purpose of the project/task description element is to provide the participants with a
background understanding of the project and the types of activities to be conducted, including the
measurements that will be taken and the associated QA/QC goals, procedures, and timetables for
collecting the measurements.
A6.2 Description of the Work to be Performed
(1) Measurements that are expected during the course of the project. Describe the
characteristic or property to be studied and the measurement processes and techniques
that will be used to collect data.
(2) Applicable technical quality standards or criteria. Cite any relevant regulatory
standards or criteria pertinent to the project. For example, if environmental data are
collected to test for compliance with a permit limit standard, the standard should be cited
and the numerical limits should be given in the QAPP. The DQO Process refers to these
limits as "action levels," because the type of action taken by the decision maker will
depend on whether the measured levels exceed the limit (Step 5 of the DQO Process).
(3) Any special personnel and equipment requirements that may indicate the
complexity of the project. Describe any special personnel or equipment required for
the specific type of work being planned or measurements being taken.
EPA QA/G-5 11 QA98
-------
(4) The assessment techniques needed for the project. The degree of quality assessment
activity for a project will depend on the project's complexity, duration, and objectives. A
discussion of the timing of each planned assessment and a brief outline of the roles of the
different parties to be involved should be included.
(5) A schedule for the work performed. The anticipated start and completion dates for the
project should be given. In addition, this discussion should include an approximate
schedule of important project milestones, such as the start of environmental
measurement activities.
(6) Project and quality records required, including the types of reports needed. An
indication of the most important records should be given.
A7 QUALITY OBJECTIVES AND CRITERIA FOR MEASUREMENT DATA
Describe the project quality objectives and measurement performance criteria.
A7.1 Purpose/Background
The purpose of this element is to document the DQOs of the project and to establish performance
criteria for the mandatory systematic planning process and measurement system that will be employed in
generating the data.
A7.2 Specifying Quality Objectives
This element of the QAPP should discuss the
desired quality of the final results of the study to ensure that
the data user's needs are met. The Agency strongly
recommends using the DQO Process (see Figure 4), a
systematic procedure for planning data collection activities,
to ensure that the right type, quality, and quantity of data are
collected to satisfy the data user's needs. DQOs are
qualitative and quantitative statements that:
clarify the intended use of the data,
• define the type of data needed to support the
decision,
• identify the conditions under which the data
should be collected, and
• specify tolerable limits on the probability of
making a decision error due to uncertainty
in the data.
Data Quality Indicators (DQIs) can be evolved from DQOs
for a sampling activity through the use of the DQO Process _,. A _,, _.„„ _
/A A• T\\ c A u +u + f+u nr\r\ Figure 4. The DQO Process
(Appendix D). figure 4 shows the seven steps or the DQO
Process, which is explained in detail in EPA QA/G-4, Guidance for the Data Quality Objectives Process.
1. State the Problem
J,
2. Identify the Decision
^,
3. Identify Inputs to the Decision
J,
4. Define the Study Boundaries
^,
5. Develop a Decision Rule
J,
6. Specify Limits on Decision Errors
1 ¥
7. Optimize the Design for Obtaining Data
EPA QA/G-5
12
QA98
-------
Appendix A.4 provides a crosswalk between the requirements of the QAPP and the DQO outputs. The
QAPP should include a reference for a full discussion of the proposed DQOs.
For exploratory research, sometimes the goal is to develop questions that may be answered by
subsequent work. Therefore, researchers may modify activities advocated in QA/G-4 to define decision
errors (see EPA QA/G-4R, Data Quality Objectives for Researchers).
A7.3 Specifying Measurement Performance Criteria
While the quality objectives state what the data user's needs are, they do not provide sufficient
information about how these needs can be satisfied. The specialists who will participate in generating
the data need to know the measurement performance criteria that must be satisfied to achieve the overall
quality objectives. One of the most important features of the QAPP is that it links the data user's quality
objectives to verifiable measurement performance criteria. Although the level of rigor with which this is
done and documented will vary widely, this linkage represents an important advancement in the
implementation of QA. Once the measurement performance criteria have been established, sampling and
analytical methods criteria can be specified under the elements contained in Group B.
A8 SPECIAL TRAINING REQUIREMENTS/CERTIFICATION
Identify and describe any specialized training or certification requirements and discuss how
such training will be provided and how the necessary skills will be assured and documented.
A8.1 Purpose/Background
The purpose of this element is to ensure that any specialized training requirements necessary to
complete the projects are known and furnished and the procedures are described in sufficient detail to
ensure that specific training skills can be verified, documented, and updated as necessary.
A8.2 Training
Requirements for specialized training for nonroutine field sampling techniques, field analyses,
laboratory analyses, or data validation should be specified. Depending on the nature of the
environmental data operation, the QAPP may need to address compliance with specifically mandated
training requirements. For example, contractors or employees working at a Superfund site need
specialized training as mandated by the Occupational Safety and Health (OSF£A) regulations. If
hazardous materials are moved offsite, compliance with the training requirements for shipping hazardous
materials as mandated by the Department of Transportation (DOT) in association with the International
Air Transportation Association may be necessary. This element of the QAPP should show that the
management and project teams are aware of specific health and safety needs as well as any other
organizational safety plans.
A8.3 Certification
Usually, the organizations participating in the project that are responsible for conducting training
and health and safety programs are also responsible for ensuring certification. Training and certification
should be planned well in advance for necessary personnel prior to the implementation of the project.
EPAQA/G-5 13 QA98
-------
All certificates or documentation representing completion of specialized training should be maintained in
personnel files.
A9 DOCUMENTATION AND RECORDS
Itemize the information and records that must be included in the data report package and
specify the desired reporting format for hard copy and electronic forms, when used.
Identify any other records and documents applicable to the project, such as audit reports,
interim progress reports, and final reports, that will be produced.
Specify or reference all applicable requirements for the final disposition of records and
documents, including location and length of retention period.
A9.1 Purpose/Background
This element defines which records are critical to the project and what information needs to be
included in reports, as well as the data reporting format and the document control procedures to be used.
Specification of the proper reporting format, compatible with data validation, will facilitate clear, direct
communication of the investigation.
A9.2 Information Included in the Reporting Packages
The selection of which records to include in a data reporting package must be determined based
on how the data will be used. Different "levels of effort" require different supporting QA/QC
documentation. For example, organizations conducting basic research have different reporting
requirements from organizations collecting data in support of litigation or in compliance with permits.
When possible, field and laboratory records should be integrated to provide a continuous reporting track.
The following are examples of different records that may be included in the data reporting package.
A9.2.1 Field Operation Records
The information contained in these records documents overall field operations and generally
consists of the following:
• Sample collection records. These records show that the proper sampling protocol was
performed in the field. At a minimum, this documentation should include the names of
the persons conducting the activity, sample number, sample collection points, maps and
diagrams, equipment/method used, climatic conditions, and unusual observations.
Bound field notebooks are generally used to record raw data and make references to
prescribed procedures and changes in planned activities. They should be formatted to
include pre-numbered pages with date and signature lines.
• Chain-of-custody records. Chain-of-custody records document the progression of
samples as they travel from the original sampling location to the laboratory and finally to
their disposal area. (See Appendix C for an example of a chain-of-custody checklist.)
EPA QA/G-5 14 QA98
-------
• QC sample records. These records document the generation of QC samples, such as
field, trip, and equipment rinsate blanks and duplicate samples. They also include
documentation on sample integrity and preservation and include calibration and
standards' traceability documentation capable of providing a reproducible reference
point. Quality control sample records should contain information on the frequency,
conditions, level of standards, and instrument calibration history.
• General field procedures. General field procedures record the procedures used in the
field to collect data and outline potential areas of difficulty in gathering specimens.
• Corrective action reports. Corrective action reports show what methods were used in
cases where general field practices or other standard procedures were violated and
include the methods used to resolve noncompliance.
If applicable, to show regulatory compliance in disposing of waste generated during the data operation,
procedures manifest and testing contracts should be included in the field procedures section.
A9.2.2 Laboratory Records
The following list describes some of the laboratory-specific records that should be compiled if
available and appropriate:
Sample Data. These records contain the times that samples were analyzed to verify that
they met the holding times prescribed in the analytical methods. Included should be the
overall number of samples, sample location information, any deviations from the SOPs,
time of day, and date. Corrective action procedures to replace samples violating the
protocol also should be noted.
• Sample Management Records. Sample management records document sample receipt,
handling and storage, and scheduling of analyses. The records verify that the chain-of-
custody and proper preservation were maintained, reflect any anomalies in the samples
(such as receipt of damaged samples), note proper log-in of samples into the laboratory,
and address procedures used to ensure that holding time requirements were met.
Test Methods. Unless analyses are performed exactly as prescribed by SOPs, this
documentation will describe how the analyses were carried out in the laboratory. This
includes sample preparation and analysis, instrument standardization, detection and
reporting limits, and test-specific QC criteria. Documentation demonstrating laboratory
proficiency with each method used could be included.
* QA/QC Reports. These reports will include the general QC records, such as initial
demonstration of capability, instrument calibration, routine monitoring of analytical
performance, calibration verification, etc. Project-specific information from the QA/QC
checks such as blanks (field, reagent, rinsate, and method), spikes (matrix, matrix spike
replicate, analysis matrix spike, and surrogate spike), calibration check samples (zero
check, span check, and mid-range check), replicates, splits, and so on should be included
in these reports to facilitate data quality analysis.
EPAQA/G-5 15 QA98
-------
A9.2.3 Data Handling Records
These records document protocols used in data reduction, verification, and validation. Data
reduction addresses data transformation operations such as converting raw data into reportable quantities
and units, use of significant figures, recording of extreme values, blank corrections, etc. Data
verification ensures the accuracy of data transcription and calculations, if necessary, by checking a set of
computer calculations manually. Data validation ensures that QC criteria have been met.
A9.3 Data Reporting Package Format and Documentation Control
The format of all data reporting packages must be consistent with the requirements and
procedures used for data validation and data assessment described in Sections B, C, and D of the QAPP.
All individual records that represent actions taken to achieve the objective of the data operation and the
performance of specific QA functions are potential components of the final data reporting package. This
element should discuss how these various components will be assembled to represent a concise and
accurate record of all activities impacting data quality. The discussion should detail the recording
medium for the project, guidelines for hand-recorded data (e.g., using indelible ink), procedures for
correcting data (e.g., single line drawn through errors and initialed by the responsible person), and
documentation control. Procedures for making revisions to technical documents should be clearly
specified and the lines of authority indicated.
A9.4 Data Reporting Package Archiving and Retrieval
The length of storage for the data reporting package may be governed by regulatory
requirements, organizational policy, or contractual project requirements. This element of the QAPP
should note the governing authority for storage of, access to, and final disposal of all records.
A9.5 References
Kanare, Howard M. 1985. Writing the Laboratory Notebook. Washington, DC: American Chemical Society.
U.S. Environmental Protection Agency. 1993. Guidance on Evaluation, Resolution, and Documentation of Analytical Problems
Associated-with Compliance Monitoring. EPA/821/B-93/001.
EPAQA/G-5 16 QA98
-------
B MEASUREMENT/DATA ACQUISITION
Bl SAMPLING PROCESS DESIGN (EXPERIMENTAL DESIGN)
Describe the experimental design or data collection design for the project.
Classify all measurements as critical or non-critical.
Bl.l Purpose/Background
The purpose of this element is to describe all the relevant components of the experimental
design; define the key parameters to be estimated; indicate the number and type of samples expected;
and describe where, when, and how samples are to be taken. The level of detail should be sufficient that
a person knowledgeable in this area could understand how and why the samples will be collected. This
element provides the main opportunity for QAPP reviewers to ensure that the "right" samples will be
taken. Strategies such as stratification, compositing, and clustering should be discussed, and diagrams or
maps showing sampling points should be included. Most of this information should be available as
outputs from the final steps of the planning (DQO) process.
In addition to describing the design, this element of the QAPP should discuss the following:
a schedule for project sampling activities,
• a rationale for the design (in terms of meeting DQOs),
the sampling design assumptions,
• the procedures for locating and selecting environmental samples,
a classification of measurements as critical or noncritical, and
• the validation of any nonstandard sampling/measurement methods.
Elements B1.2 through B1.8 address these subjects.
B1.2 Scheduled Project Activities, Including Measurement Activities
This element should give anticipated start and completion dates for the project as well as
anticipated dates of major milestones, such as the following:
schedule of sampling events;
• schedule for analytical services by offsite laboratories;
schedule for phases of sequential sampling (or testing), if applicable;
• schedule of test or trial runs; and
schedule for peer review activities.
The use of bar charts showing time frames of various QAPP activities to identify both potential
bottlenecks and the need for concurrent activities is recommended.
B1.3 Rationale for the Design
The objectives for an environmental study should be formulated in the planning stage of any
investigation. The requirements and the rationale of the design for the collection of data are derived
EPAQA/G-5 17 QA98
-------
from the quantitative outputs of the DQO Process. The type of design used to collect data depends
heavily on the key characteristic being investigated. For example, if the purpose of the study is to
estimate overall average contamination at a site or location, the characteristic (or parameter) of interest
would be the mean level of contamination. This information is identified in Step 5 of the DQO Process.
The relationship of this parameter to any decision that has to be made from the data collected is obtained
from Steps 2 and 3 of the DQO Process (see Figure 4).
The potential range of values for the parameter of interest should be considered during
development of the data collection methodology and can be greatly influenced by knowledge of potential
ranges in expected concentrations. For example, the number of composite samples needed per unit area
is directly related to the variability in potential contaminant levels expected in that area.
The choice between a probability-based (statistical) data collection design or a nonrandom
(judgmental) data collection methodology depends on the ultimate use of the data being collected. This
information is specified in Steps 5 and 6 of the DQO Process. Adherence to the data collection design
chosen in Step 7 of the DQO Process directly affects the magnitude of potential decision error rates
(false positive rate and false negative rate) established in Step 6 of the DQO Process. Any procedures for
coping with unanticipated data collection design changes also should be briefly discussed.
B1.4 Design Assumptions
The planning process usually recommends a specific data collection method (Step 7 of the DQO
Process), but the effectiveness of this methodology rests firmly on assumptions made to establish the
data collection design. Typical assumptions include the homogeneity of the medium to be sampled (for
example, sludge, fine silt, or wastewater effluent), the independence in the collection of individual
samples (for example, four separate samples rather than four aliquots derived from a single sample), and
the stability of the conditions during sample collection (for example, the effects of a rainstorm during
collection of wastewater from an industrial plant). The assumptions should have been considered during
the DQO Process and should be summarized together with a contingency plan to account for exceptions
to the proposed sampling plan. An important part of the contingency plan is documenting the procedures
to be adopted in reporting deviations or anomalies observed after the data collection has been completed.
Examples include an extreme lack of homogeneity within a physical sample or the presence of analytes
that were not mentioned in the original sampling plan. Chapter 1 of EPA QA/G-9 provides an overview
of sampling plans and the assumptions needed for their implementation. EPA QA/G-5S provides
guidance on the construction of sampling plans to meet the requirements generated by the DQO Process.
B1.5 Procedures for Locating and Selecting Environmental Samples
The most appropriate plan for a particular sampling application will depend on: the practicality
and feasibility (e.g., determining specific sampling locations) of the plan, the key characteristic (the
parameter established in Step 5 of the DQO Process) to be estimated, and the implementation resource
requirements (e.g., the costs of sample collection, transportation, and analysis).
This element of the QAPP should also describe the frequency of sampling and specific sample
locations (e.g., sample port locations and traverses for emissions source testing, well installation designs
for groundwater investigations) and sampling materials. When decisions on the number and location of
samples will be made in the field, the QAPP should describe how these decisions will be driven whether
by actual observations or by field screening data. When locational data are to be collected, stored, and
transmitted, the methodology used must be described (or referenced) and include the following:
EPAQA/G-5 18 QA98
-------
procedures for rinding prescribed sample locations,
• contingencies for cases where prescribed locations are inaccessible,
location bias and its assessment, and
• procedures for reporting deviations from the sampling plan.
When appropriate, a map of the sample locations should be provided and locational map
coordinates supplied. EPA QA/G-5S provides nonmandatory guidance on the practicality of
constructing sampling plans and references to alternative sampling procedures.
B1.6 Classification of Measurements as Critical or Noncritical
All measurements should be classified as critical (i.e., required to achieve project objectives or
limits on decision errors, Step 6 of the DQO Process) or noncritical (for informational purposes only or
needed to provide background information). Critical measurements will undergo closer scrutiny during
the data gathering and review processes and will have first claim on limited budget resources. It is also
possible to include the expected number of samples to be tested by each procedure and the acceptance
criteria for QC checks (as described in element B5, "Quality Control Requirements").
B1.7 Validation of Any Nonstandard Methods
For nonstandard sampling methods, sample matrices, or other unusual situations, appropriate
method validation study information may be needed to confirm the performance of the method for the
particular matrix. The purpose of this validation information is to assess the potential impact on the
representativeness of the data generated. For example, if qualitative data are needed from a modified
method, rigorous validation may not be necessary. Such validation studies may include round-robin
studies performed by EPA or by other organizations. If previous validation studies are not available,
some level of single-user validation study or ruggedness study should be performed during the project
and included as part of the project's final report. This element of the QAPP should clearly reference any
available validation study information.
B2 SAMPLING METHODS REQUIREMENTS
Describe the procedures for collecting samples and identify the sampling methods and
equipment. Include any implementation requirements, support facilities, sample preservation
requirements, and materials needed. Describe the process for preparing and decontaminating
sampling equipment, including disposing decontamination by-products; selecting and
preparing sample containers, sample volumes, preservation methods, and maximum holding
times for sampling and/or analysis.
Describe specific performance requirements for the method. Address what to do when a
failure in the sampling occurs, who is responsible for corrective action, and how the
effectiveness of the corrective action shall be determined and documented.
B2.1 Purpose/Background
Environmental samples should reflect the target population and parameters of interest. As with
all other considerations involving environmental measurements, sampling methods should be chosen
with respect to the intended application of the data. Just as methods of analysis vary in accordance with
EPAQA/G-5 19 QA98
-------
project needs, sampling methods can also vary according to these requirements. Different sampling
methods have different operational characteristics, such as cost, difficulty, and necessary equipment. In
addition, the sampling method can materially affect the representativeness, comparability, bias, and
precision of the final analytical result.
In the area of environmental sampling, there exists a great variety of sample types. It is beyond
the scope of this document to provide detailed advice for each sampling situation and sample type.
Nevertheless, it is possible to define certain common elements that are pertinent to many sampling
situations with discrete samples (see EPA QA/G-5S).
If a separate sampling and analysis plan is required or created for the project, it should be
included as an appendix to the QAPP. The QAPP should simply refer to the appropriate portions of the
sampling and analysis plan for the pertinent information and not reiterate information.
B2.2 Describe the Sample Collection, Preparation, and Decontamination Procedures
(1) Select and describe appropriate sampling methods from the appropriate compendia of methods.
For each parameter within each sampling situation, identify appropriate sampling methods from
applicable EPA regulations, compendia of methods, or other sources of methods that have been
approved by EPA. When EPA-sanctioned procedures are available, they will usually be
selected. When EPA-sanctioned procedures are not available, standard procedures from other
organizations and disciplines may be used. A complete description of non-EPA methods should
be provided in (or attached to) the QAPP. Procedures for sample homogenization of nonaqueous
matrices may be described in part (2) as a technique for assuring sample representativeness. In
addition, the QAPP should specify the type of sample to be collected (e.g., grab, composite,
depth-integrated, flow- weighted) together with the method of sample preservation.
(2) Discuss sampling methods' requirements. Each medium or contaminant matrix has its own
characteristics that define the method performance and the type of material to be sampled.
Investigators should address the following:
• actual sampling locations,
choice of sampling method/collection,
• delineation of a properly shaped sample,
inclusion of all particles within the volume sampled, and
• subsampling to reduce the representative field sample into a representative laboratory
aliquot.
Having identified appropriate and applicable methods, it is necessary to include the
requirements for each method in the QAPP. If there is more than one acceptable sampling
method applicable to a particular situation, it may be necessary to choose one from among them.
DQOs should be considered in choosing these methods to ensure that: a) the sample accurately
represents the portion of the environment to be characterized, b) the sample is of sufficient
volume to support the planned chemical analysis, and c) the sample remains stable during
shipping and handling.
(3) Describe the decontamination procedures and materials. Decontamination is primarily
applicable in situations of sample acquisition from solid, semi-solid, or liquid media, but it
should be addressed, if applicable, for continuous monitors as well. The investigator must
EPA QA/G-5 20 QA98
-------
consider the appropriateness of the decontamination procedures for the project at hand. For
example, if contaminants are present in the environmental matrix at the 1% level, it is probably
unnecessary to clean sampling equipment to parts-per-billion (ppb) levels. Conversely, if ppb-
level detection is required, rigorous decontamination or the use of disposable equipment is
required. Decontamination by-products must be disposed of according to EPA policies and the
applicable rules and regulations that would pertain to a particular situation, such as the
regulations of OSHA, the Nuclear Regulatory Commission (NRC), and State and local
governments.
B2.3 Identify Support Facilities for Sampling Methods
Support facilities vary widely in their analysis capabilities, from percentage-level accuracy to
ppb-level accuracy. The investigator must ascertain that the capabilities of the support facilities are
commensurate with the requirements of the sampling plan established in Step 7 of the DQO Process.
B2.4 Describe Sampling/Measurement System Failure Response and Corrective Action Process
This section should address issues of responsibility for the quality of the data, the methods for
making changes and corrections, the criteria for deciding on a new sample location, and how these
changes will be documented. This section should describe what will be done if there are serious flaws
with the implementation of the sampling methodology and how these flaws will be corrected. For
example, if part of the complete set of samples is found to be inadmissable, how replacement samples
will be obtained and how these new samples will be integrated into the total set of data should be
described.
B2.5 Describe Sampling Equipment, Preservation, and Holding Time Requirements
This section includes the requirements needed to prevent sample contamination (disposable
samplers or samplers capable of appropriate decontamination), the physical volume of the material to be
collected (the size of composite samples, core material, or the volume of water needed for analysis), the
protection of physical specimens to prevent contamination from outside sources, the temperature
preservation requirements, and the permissible holding times to ensure against degradation of sample
integrity.
B2.6 References
Publications useful in assisting the development of sampling methods include:
Solid and Hazardous Waste Sampling
U.S. Environmental Protection Agency. 1986. Test Methods for Evaluating SolidWaste (SW-846). 3rd Ed., Chapter 9.
U.S. Environmental Protection Agency. 1985. Characterization of Hazardous Waste Sites - A Methods Manual. Vol. I, Site
Investigations. EPA-600/4-84-075. Environmental Monitoring Systems Laboratory. Las Vegas, NV.
U.S. Environmental Protection Agency. 1984. Characterization of Hazardous Waste Sites - A Methods Manual. Vol. II,
Available Sampling Methods. EPA-600/4-84-076. Environmental Monitoring Systems Laboratory. Las Vegas, NV.
U.S. Environmental Protection Agency. 1987. A Compendium of Superfund Field Operations Methods. NTIS PB88-181557.
EPA/540/P-87/001. Washington, DC.
EPAQA/G-5 21 QA98
-------
Ambient Air Sampling
U.S. Environmental Protection Agency. 1994. Quality Assurance Handbook for Air Pollution Measurement Systems. Vol. I,
Principles. EPA 600/9-76-005. Section 1.4.8 and Appendix M.S.6.
U.S. Environmental Protection Agency. 1994. Quality Assurance Handbook for Air Pollution Measurement Systems. Vol. II,
EPA 600/R-94-038b. Sections 2.0.1 and 2.0.2 and individual methods.
U.S. Environmental Protection Agency. 1984. Compendium of Methods for the Determination of Toxic Organic Compounds in
Ambient Air. EPA/600-4-84-41. Environmental Monitoring Systems Laboratory. Research Triangle Park, NC.
Supplement: EPA-600-4-87-006. September 1986.
Source Testing (Air)
U.S. Environmental Protection Agency. 1994. Quality Assurance Handbook for Air Pollution Measurement Systems. Vol. Ill,
EPA 600/R-94-038c. Section 3.0 and individual methods.
Water/ Ground Water
U.S. Environmental Protection Agency. Handbook: Ground Water. Cincinnati, OH. EPA/625/6-87/016. March 1987.
U.S. Environmental Protection Agency. RCRA Ground Water Monitoring Technical Enforcement Guidance Document.
Washington, DC. 1986.
U.S. Environmental Protection Agency. Standard Methods for the Examination of Water and Waste-water. 16th ed.
Washington, DC. 1985.
Acid Precipitation
U.S. Environmental Protection Agency. 1994. Quality Assurance Handbook for Air Pollution Measurement Systems. Vol. V,
EPA 600/94-038e.
Meteorological Measurements
U.S. Environmental Protection Agency. 1989. Quality Assurance Handbook for Air Pollution Measurement Systems. Vol. IV,
EPA 600/4-90-003.
Radioactive Materials and Mixed Waste
U.S. Department of Energy. 1989. Radioactive-Hazardous Mixed Waste Sampling and Analysis: Addendum to SW-846.
Soils and Sediments
U.S. Environmental Protection Agency. 1985. Sediment Sampling Quality Assurance User's Guide. NTIS PB85-233542.
EPA/600/4-85/048. Environmental Monitoring Systems Laboratory. Las Vegas, NV.
U.S. Environmental Protection Agency. 1989. Soil Sampling Quality Assurance User's Guide. EPA/600/8-89/046.
Environmental Monitoring Systems Laboratory. Las Vegas, NV.
Earth, D.S., and T.H.Starks. 1985. Sediment Sampling Quality Assurance User's Guide. EPA/600-4-85/048. Prepared for
Environmental Monitoring and Support Laboratory. Las Vegas, NV.
Statistics, Geostatistics, and Sampling Theory
Myers, J.C. 1997. Geostatistical Error Measurement. New York: Van Nostrand Reinhold.
Pitard, F.F. 1989. Pierre Gy's Sampling Theory and Sampling Practice. Vollandll. Boca Raton, FL: CRC Press.
EPA QA/G-5 22 QA98
-------
Miscellaneous
American Chemical Society Joint Board/Council Committee on Environmental Improvement. 1990. Practical Guide for
Environmental Sampling and Analysis, Section II. Environmental Analysis. Washington, DC.
ASTM Committee D-34. 1986. Standard Practices for Sampling Wastes from Pipes and Other Point Discharges. Document
No. D34.01-001R7.
Keith, L. 1990. EPA's Sampling and Analysis Methods Database Manual. Austin, TX: Radian Corp.
Keith,L. 1991. Environmental Sampling and Analysis: A Practical Guide. Chelsea, MI: Lewis Publishers, Inc.
B3 SAMPLE HANDLING AND CUSTODY REQUIREMENTS
Describe the requirements and provisions for sample handling and custody in the field,
laboratory, and transport, taking into account the nature of the samples, the maximum
allowable sample holding times before extraction or analysis, and available shipping options
and schedules.
Include examples of sample labels, custody forms, and sample custody logs.
B3.1 Purpose/Background
This element of the QAPP should describe all procedures that are necessary for ensuring that:
(1) samples are collected, transferred, stored, and analyzed by authorized personnel;
(2) sample integrity is maintained during all phases of sample handling and analyses; and
(3) an accurate written record is maintained of sample handling and treatment from the time
of its collection through laboratory procedures to disposal.
Proper sample custody minimizes accidents by assigning responsibility for all stages of sample handling
and ensures that problems will be detected and documented if they occur. A sample is in custody if it is
in actual physical possession or it is in a secured area that is restricted to authorized personnel. The level
of custody necessary is dependent upon the project's DQOs. While enforcement actions necessitate
stringent custody procedures, custody in other types of situations (i.e., academic research) may be
primarily concerned only with the tracking of sample collection, handling, and analysis.
Sample custody procedures are necessary to prove that the sample data correspond to the sample
collected, if data are intended to be legally defensible in court as evidence. In a number of situations, a
complete, detailed, unbroken chain of custody will allow the documentation and data to substitute for the
physical evidence of the samples (which are often hazardous waste) in a civil courtroom. Some statutes
or criminal violations may still necessitate that the physical evidence of sample containers be presented
along with the custody and data documentation.
An outline of the scope of sample custody-starting from the planning of sample collection, field
sampling, sample analysis to sample disposal—should also be included. This discussion should further
stress the completion of sample custody procedures, which include the transfer of sample custody from
field personnel to lab, sample custody within the analytical lab during sample preparation and analysis,
and datastorage.
EPA QA/G-5 23 QA98
-------
B3.2 Sample Custody Procedure
The QAPP should discuss the sample custody procedure at a level commensurate with the
intended use of the data. This discussion should include the following:
(1) List the names and responsibilities of all sample custodians in the field and laboratories.
(2) Give a description and example of the sample numbering system.
(3) Define acceptable conditions and plans for maintaining sample integrity in the field prior
to and during shipment to the laboratory (e.g., proper temperature and preservatives).
(4) Give examples of forms and labels used to maintain sample custody and document
sample handling in the field and during shipping. An example of a sample log sheet is
given in Figure 5; an example sample label is given in Figure 6.
(5) Describe the method of sealing shipping containers with chain-of-custody seals. An
example of a seal is given in Figure 7.
(6) Describe procedures that will be used to maintain the chain of custody and document
sample handling during transfer from the field to the laboratory, within the laboratory,
and among contractors. An example of a chain-of-custody record is given in Figure 8.
(7) Provide for the archiving of all shipping documents and associated paperwork.
(8) Discuss procedures that will ensure sample security at all times.
(9) Describe procedures for within-laboratory chain-of-custody together with verification of
the printed name, signature, and initials of the personnel responsible for custody of
samples, extracts, or digests during analysis at the laboratory. Finally, document
disposal or consumption of samples should also be described. A chain-of-custody
checklist is included in Appendix C to aid in managing this element.
Minor documentation of chain-of-custody procedures is generally applicable when:
• Samples are generated and immediately tested within a facility or site; and
• Continuous rather than discrete or integrated samples are subjected to real- or near real-
time analysis (e.g., continuous monitoring).
The discussion should be as specific as possible about the details of sample storage, transportation, and
delivery to the receiving analytical facility.
EPA QA/G-5 24 QA98
-------
LU
<
Q
G SHEET NO.
O
LLJ
CL
2
<
OT
SURVEY
on
O
LL
Q
Oi
—
D
O
LJJ
W
LJJ
(fl
_l
<
Z
<
!
§
D)
52.
a:
LU
1
Q.
5
<
tn
LU
_l
Q.
s
<
tn
LL.
0
LU
Q.
3QINVAO
1ON3Hd
soiNvoyo aovai
gyaH
S3QIOIlS3d
nova
snviaiAi
asvayoaNvnio
Aiiaiayni
lAiyodnoo nvoad
lAiyodnoo nvioi
aynivyadiAiai
»AiiAiionaNoo
»Hd
oa
AiiNnv>inv
sanosa3QN3dsns
sanos nvioi
001
aoo
aoa
siN3iyinN
LU
<
Z
LU
W
LU
o:
Q.
y3NIVlNOO 3dAl
3i/\innoAnvioi
N3>1V13ndl/\IVS3l/\IIJ
STATION DESCRIPTION
z cc
O LU
p OQ
Bi
w z
O
£
Q
UJ
Z
<
UJ
d
1
0
£/) ^
£ S3
I !
•2: u
LU il
o:
Figure 5. An Example of a Sample Log Sheet
EPA QA/G-5
25
QA98
-------
(Name of Sampling Organization)
Sample Description:
Plant: Location:
Date:
Time:
Media: Station:
Sample Type: Preservative:
Sampled Bv:
Sample ID No.:
Lab No.
w
i_
CO
E
CD
cr
Figure 6. An Example of a Sample Label
9JT3Q
ivas
CUSTODY SEAL
o
3 Date
CD
•?
Signature
Figure 1. An Example of a Custody Seal
EPA QA/G-5
26
QA98
-------
STATION
NUMBER
STATION LOCATION
DATE
Relinquished by: (Signature)
Relinquished by: (signature)
Relinquished by: (signature)
Received by: (signature)
Received by: (Signature)
DATE/T
TIME
SAMPLERS (Signature)
SAMPLE TYPE
WATER
Comp
Grabx
AIR
SEQ
NO.
NO. OF
CONTAINERS
ANALYSIS
REQUIRED
Received by: (Signature)
Received by: (Signature)
Received by: (Signature)
Received by Mobile Laboratory for field
analysis: (Signature)
IME
Received for Laboratory by:
Method of Shipment:
DATE
DATE
DATE
DATE/
DATE
/TIME
/TIME
/TIME
TIME
/TIME
Distribution: Original - Accompany Shipment
1 Copy - Survey Coordinator Field Files
Figure 8. An Example of a Chain-of-Custody Record
EPA QA/G-5
27
QA98
-------
B4 ANALYTICAL METHODS REQUIREMENTS
Identify the analytical methods and equipment required, including sub-sampling or extraction
methods, laboratory decontamination procedures and materials (such as the case of hazardous
or radioactive samples), waste disposal requirements (if any), and specific performance
requirements for the method.
Identify analytical methods by number, date, and regulatory citation (as appropriate). If a
method allows the user to select from various options, then the method citations should state
exactly which options are being selected. For non-standard methods, such as unusual sample
matrices and situations, appropriate method performance study information is needed to
confirm the performance of the method for the particular matrix. If previous performance
studies are not available, they must be developed during the project and included as part of the
project results.
Address what to do when a failure in the analytical system occurs, who is responsible for
corrective action, and how the effectiveness of the corrective action shall be determined and
documented.
Specify the laboratory turnaround time needed, if important to the project schedule. Specify
whether a field sampling and/or laboratory analysis case narrative is required to provide a
complete description of any difficulties encountered during sampling or analysis.
B4.1 Purpose/Background
The choice of analytical methods will be influenced by the performance criteria, Data Quality
Objectives, and possible regulatory criteria. If appropriate, a citation of analytical procedures may be
sufficient if the analytical method is a complete SOP. For other methods, it may suffice to reference a
procedure (i.e., from Test Methods for Evaluating Solid Waste, SW-846) and further supplement it with
the particular options/variations being used by the lab, the detection limits actually achieved, the
calibration standards and concentrations used, etc. If the procedure is unique or an adaption of a
"standard" method, complete analytical and sample preparation procedures will need to be attached to
the QAPP.
Specific monitoring methods and requirements to demonstrate compliance traditionally were
specified in the applicable regulations and/or permits. However, this approach is being replaced by the
Performance-Based Measurement System (PBMS). PBMS is a process in which data quality needs,
mandates, or limitations of a program or project are specified and serve as a criterion for selecting
appropriate methods. The regulated body selects the most cost-effective methods that meet the criteria
specified in the PBMS. Under the PBMS framework, the performance of the method employed is
emphasized rather than the specific technique or procedure used in the analysis. Equally stressed in this
system is the requirement that the performance of the method be documented and certified by the
laboratory that appropriate QA/QC procedures have been conducted to verify the performance. PBMS
applies to physical, chemical, and biological techniques of analysis performed in the field as well as in the
laboratory. PBMS does not apply to the method-defined parameters.
EPA QA/G-5 28 QA98
-------
The QAPP should also address the issue of the quality of analytical data as indicated by the
data's ability to meet the QC acceptance criteria. This section should describe what should be done if the
calibration check samples exceed the control limits due to mechanical failure of the instrumentation, a
drift in the calibration curve occurs, or if a reagent blank indicates contamination. This section should
also indicate the authorities responsible for the quality of the data, the protocols for making changes and
implementing corrective actions, and the methods for reporting the data and its limitations.
Laboratory contamination from the processing of hazardous materials such as toxic or
radioactive samples for analysis and their ultimate disposal should be a considered during the planning
stages for selection of analysis methods. Safe handling requirements for project samples in the
laboratory with appropriate decontamination and waste disposal procedures should also be described.
B4.2 Subsampling
If subsampling is required, the procedures should be described in this QAPP element, and the
full text of the subsampling operating procedures should be appended to the QAPP. Because
subsampling may involve more than one stage, it is imperative that the procedures be documented fully
so that the results of the analysis can be evaluated properly.
B4.3 Preparation of the Samples
Preparation procedures should be described and standard methods cited and used where possible.
Step-by-step operating procedures for the preparation of the project samples should be listed in an
appendix. The sampling containers, methods of preservation, holding times, holding conditions, number
and types of all QA/QC samples to be collected, percent recovery, and names of the laboratories that will
perform the analyses need to be specifically referenced.
B4.4 Analytical Methods
The citation of an analytical method may not always be sufficient to fully characterize a method
because the analysis of a sample may require deviation from a standard method and selection from the
range of options in the method. The SOP for each analytical method should be cited or attached to the
QAPP, and all deviations or alternative selections should be detailed in the QAPP.
The matrix containing the subject analytes often dictates the sampling and analytical methods.
Gaseous analytes often must be concentrated on a trap in order to collect a measurable quantity. If the
matrix is a liquid or a solid, the analytes usually must be separated from it using various methods of
extraction. Sometimes the analyte is firmly linked by chemical bonds to other elements and must be
subjected to digestion methods to be freed for analysis.
Often the selected analytical methods may be presented conveniently in one or several tables
describing the matrix, the analytes to be measured, the analysis methods, the type, the precision/accuracy
data, the performance acceptance criteria, the calibration criteria, and etc. Appendix C contains a
checklist of many important components to consider when selecting analytical methods.
B4.5 References
Greenberg, A.E., L.S. Clescer, and A. D. Eaton, eds. 1992. Standard Methods for the Examination of Water and Waste-water.
18th ed. American Public Health Association. Water Environment Federation.
EPA QA/G-5 29 QA98
-------
U.S. Environmental Protection Agency. 1996. Quality Control: Variability in Protocols. EPA/600/9-91/034. Risk Reduction
Engineering Laboratory. U.S. EPA. Cincinnati, OH.
U.S. Environmental Protection Agency. Test Methods for Evaluating Solid Waste. SW-846. Chapter 2, "Choosing the
Correct Procedure."
B5 QUALITY CONTROL REQUIREMENTS
Identify required measurement QC checks for both the field and the laboratory. State the
frequency of analysis for each type of QC check, and the spike compounds sources and levels.
State or reference the required control limits for each QC check and corrective action required
when control limits are exceeded and how the effectiveness of the corrective action shall be
determined and documented.
Describe or reference the procedures to be used to calculate each of the QC statistics.
B5.1 Purpose/Background
QC is "the overall system of technical activities that measures the attributes and performance of
a process, item, or service against defined standards to verify that they meet the stated requirements
established by the customer." QC is both corrective and proactive in establishing techniques to prevent
the generation of unacceptable data, and so the policy for corrective action should be outlined. This
element will rely on information developed in section A7, "Quality Objectives and Criteria for
Measurement Data," which establishes measurement performance criteria.
B5.2 QC Procedures
This element documents any QC checks not defined in other QAPP elements and should
reference other elements that contain this information where possible. Most of the QC acceptance limits
of EPA methods are based on the results of interlaboratory studies. Because of improvements in
measurement methodology and continual improvement efforts in individual laboratories, these
acceptance limits may not be stringent enough for some projects. In some cases, acceptance limits are
based on intralaboratory studies (which often result in narrower acceptance limits than those based on
interlaboratory limits), and consultation with an expert may be necessary. Other elements of the QAPP
that contain related sampling and analytical QC requirements include:
Sampling Process Design (Bl), which identifies the planned field QC samples as well
as procedures for QC sample preparation and handling;
• Sampling Methods Requirements (B2), which includes requirements for determining if
the collected samples accurately represent the population of interest;
Sample Handling and Custody Requirements (B3), which discusses any QC devices
employed to ensure samples are not tampered with (e.g., custody seals) or subjected to
other unacceptable conditions during transport;
• Analytical Methods Requirements (B4), which includes information on the
subsampling methods and information on the preparation of QC samples in the sample
matrix (e.g., splits, spikes, and replicates); and
EPA QA/G-5 30 QA98
-------
• Instrument Calibration and Frequency (B7), which defines prescribed criteria for
triggering recalibration (e.g., failed calibration checks).
Table 1 lists QC checks often included in QAPPs. The need for the specific check depends on
the project objectives.
Table 1. Project Quality Control Checks
QC Check
Information Provided
Blanks
field blank
reagent blank
rinsate blank
method blank
transport and field handling bias
contaminated reagent
contaminated equipment
response of entire laboratory analytical system
Spikes
matrix spike
matrix spike replicate
analysis matrix spike
surrogate spike
analytical (preparation + analysis) bias
analytical bias and precision
instrumental bias
analytical bias
Calibration Check Samples
zero check
span check
mid-range check
calibration drift and memory effects
calibration drift and memory effects
calibration drift and memory effects
Replicates, splits, etc.
collocated samples
field replicates
field splits
laboratory splits
laboratory replicates
analysis replicates
sampling + measurement precision
precision of all steps after acquisition
shipping + interlaboratory precision
interlaboratory precision
analytical precision
instrument precision
Many QC checks result in measurement data that are used to compute statistical indicators of data
quality. For example, a series of dilute solutions may be measured repeatedly to produce an estimate of
the instrument detection limit. The formulas for calculating such Data Quality Indicators (DQIs) should
be provided or referenced in the text. This element should also prescribe any limits that define
acceptable data quality for these indicators (see also Appendix D, "Data Quality Indicators"). A QC
checklist should be used to discuss the relation of QC to the overall project objectives with respect to:
• the frequency and point in the measurement process in which the check sample is
introduced,
the traceability of the standards,
the matrix of the check sample,
the level or concentration of the analyte of interest,
the actions to be taken if a QC check identifies a failed or changed measurement system,
the formulas for estimating DQIs, and
• the procedures for documenting QC results, including control charts.
EPA QA/G-5
31
QA98
-------
Finally, this element should describe how the QC check data will be used to determine that
measurement performance is acceptable. This step can be accomplished by establishing QC "warning"
and "control" limits for the statistical data generated by the QC checks (see standard QC textbooks or
refer to EPA QA/G-5T for operational details).
Depending on the breadth of the potential audience for reviewing and implementing the QAPP, it
may be advantageous to separate the field QC from the laboratory QC requirements.
B6 INSTRUMENT/EQUIPMENT TESTING, INSPECTION, AND MAINTENANCE
REQUIREMENTS
Describe how inspections and acceptance testing of environmental sampling and measurement
systems and their components will be performed and documented.
Identify and discuss the procedure by which final acceptance will be performed by independent
personnel and/or by the EPA Project Officer.
Describe how deficiencies are to be resolved and when re-inspection will be performed.
Describe or reference how periodic preventive and corrective maintenance of measurement or
test equipment shall be performed. Identify the equipment and/or systems requiring periodic
maintenance. Discuss how the availability of critical spare parts, identified in the operating
guidance and/or design specifications of the systems, will be assured and maintained.
B6.1 Purpose/Background
The purpose of this element of the QAPP is to discuss the procedures used to verify that all
instruments and equipment are maintained in sound operating condition and are capable of operating at
acceptable performance levels.
B6.2 Testing, Inspection, and Maintenance
The procedures described should (1) reflect consideration of the possible effect of equipment
failure on overall data quality, including timely delivery of project results; (2) address any relevant site-
specific effects (e.g., environmental conditions); and (3) include procedures for assessing the equipment
status. This element should address the scheduling of routine calibration and maintenance activities, the
steps that will be taken to minimize instrument downtime, and the prescribed corrective action
procedures for addressing unacceptable inspection or assessment results. This element should also
include periodic maintenance procedures and describe the availability of spare parts and how an
inventory of these parts is monitored and maintained. The reader should be supplied with sufficient
information to review the adequacy of the instrument/equipment management program. Appending
SOPs containing this information to the QAPP and referencing the SOPs in the text are acceptable.
Inspection and testing procedures may employ reference materials, such as the National Institute
of Standards and Technology's (NIST's) Standard Reference Materials (SRMs), as well as QC standards
or an equipment certification program. The accuracy of calibration standards is important because all
data will be measured in reference to the standard used. The types of standards or special programs
should be noted in this element, including the inspection and acceptance testing criteria for all
EPA QA/G-5 32 QA98
-------
components. The acceptance limits for verifying the accuracy of all working standards against primary
grade standards should also be provided.
B7 INSTRUMENT CALIBRATION AND FREQUENCY
Identify all tools, gauges, instruments, and other sampling, measuring, and test equipment used
for data collection activities affecting quality that must be controlled and, at specified periods,
calibrated to maintain performance within specified limits.
Identify the certified equipment and/or standards used for calibration. Describe or reference
how calibration will be conducted using certified equipment and/or standards with known valid
relationships to nationally recognized performance standards. If no such nationally recognized
standards exist, document the basis for the calibration. Indicate how records of calibration
shall be maintained and be traceable to the instrument.
B7.1 Purpose/Background
This element of the QAPP concerns the calibration procedures that will be used for instrumental
analytical methods and other measurement methods that are used in environmental measurements. It is
necessary to distinguish between defining calibration as the checking of physical measurements against
accepted standards and as determining the relationship (function) of the response versus the
concentration. The American Chemical Society (ACS) limits the definition of the term calibration to the
checking of physical measurements against accepted standards, and uses the term standardization to
describe the determination of the response function.
B7.2 Identify the Instrumentation Requiring Calibration
The QAPP should identify any equipment or instrumentation that requires calibration to maintain
acceptable performance. While the primary focus of this element is on instruments of the measurement
system (sampling and measurement equipment), all methods require standardization to determine the
relationship between response and concentration.
B7.3 Document the Calibration Method that Will Be Used for Each Instrument
The QAPP must describe the calibration method for each instrument in enough detail for another
researcher to duplicate the calibration method. It may reference external documents such as EPA-
designated calibration procedures or SOPs providing that these documents can be easily obtained.
Nonstandard calibration methods or modified standard calibration methods should be fully documented
and justified.
Some instrumentation may be calibrated against other instrumentation or apparatus (e.g., NIST
thermometer), while other instruments are calibrated using standard materials traceable to national
reference standards. QAPP documentation for calibration apparatus and calibration standards are
addressed in B7.4 and B7.5.
Calibrations normally involve challenging the measurement system or a component of the
measurement system at a number of different levels over its operating range. The calibration may cover
a narrower range if accuracy in that range is critical, given the end use of the data. Single-point
EPA QA/G-5 33 QA98
-------
calibrations are of limited use, and two-point calibrations do not provide information on nonlinearity. If
single- or two-point calibrations are used for critical measurements, the potential shortcomings should be
carefully considered and discussed in the QAPP. Most EPA-approved analytical methods require
multipoint (three or more) calibrations that include zeros, or blanks, and higher levels so that unknowns
fall within the calibration range and are bracketed by calibration points. The number of calibration
points, the calibration range, and any replication (repeated measures at each level) should be given in the
QAPP.
The QAPP should describe how calibration data will be analyzed. The use of statistical QC
techniques to process data across multiple calibrations to detect gradual degradations in the measurement
system should be described. The QAPP should describe any corrective action that will be taken if
calibration (or calibration check) data fail to meet the acceptance criteria, including recalibration.
References to appended SOPs containing the calibration procedures are an acceptable alternative to
describing the calibration procedures within the text of the QAPP.
B7.4 Document the Calibration Apparatus
Some instruments are calibrated using calibration apparatus rather than calibration standards.
For example, an ozone generator is part of a system used to calibrate continuous ozone monitors.
Commercially available calibration apparatus should be listed together with the make (the manufacturer's
name), the model number, and the specific variable control settings that will be used during the
calibrations. A calibration apparatus that is not commercially available should be described in enough
detail for another researcher to duplicate the apparatus and follow the calibration procedure.
B7.5 Document the Calibration Standards
Most measurement systems are calibrated by processing materials that are of known and stable
composition. References describing these calibration standards should be included in the QAPP.
Calibration standards are normally traceable to national reference standards, and the traceability protocol
should be discussed. If the standards are not traceable, the QAPP must include a detailed description of
how the standards will be prepared. Any method used to verify the certified value of the standard
independently should be described.
B7.6 Document Calibration Frequency
The QAPP must describe how often each measurement method will be calibrated. It is desirable
that the calibration frequency be related to any known temporal variability (i.e., drift) of the
measurement system. The calibration procedure may involve less-frequent comprehensive calibrations
and more-frequent simple drift checks. The location of the record of calibration frequency and
maintenance should be referenced.
B7.7 References
American Chemical Society. 1980. "Calibration." Analytical Chemistry, Vol. 52, pps. 2,242-2,249.
Dieck, R.H. 1992. Measurement Uncertainty Methods and Applications. Research Triangle Park, NC: Instrument Society of
America.
Dux, J.P. 1986. Handbook of Quality Assurance for the Analytical Chemistry Laboratory. New York: VanNostrand
Reinhold.
EPA QA/G-5 34 QA98
-------
ILAC Task Force E. 1984. Guidelines for the Determination ofRecalibration Intervals of Testing Equipment Used in Testing
Laboratories. International Organization for Legal Metrology (OIML). International Document No. 10. 11 Rue
Twigot, Paris 95009, France.
Ku, H.H., ed. 1969. Precision Measurement and Calibration. Selected NBS Papers on Statistical Concepts and Procedures.
Special Publication 300. Vol.1. Gaithersburg, MD: National Bureau of Standards.
Liggett, W. 1986. "Tests of the Recalibration Period of a Drifting Instrument." In Oceans'86 Conference Record. Vol.3.
Monitoring Strategies Symposium. The Institute of Electrical and Electronics Engineers, Inc., Service Center.
Piscataway, NJ.
Pontius, P.E. 1974. Notes on the Fundamentals of Measurement as a Production Process. Publication No. NBSIR 74-545.
Gaithersburg, MD: National Bureau of Standards.
Taylor, J.T. 1987. Quality Assurance of Chemical Measurements. Boca Raton, FL: Lewis Publishers, Inc.
B8 INSPECTION/ACCEPTANCE REQUIREMENTS FOR SUPPLIES AND
CONSUMABLES
Describe how and by whom supplies and consumables shall be inspected and accepted for use in
the project. State acceptance criteria for such supplies and consumables.
B8.1 Purpose
The purpose of this element is to establish and document a system for inspecting and accepting
all supplies and consumables that may directly or indirectly affect the quality of the project or task. If
these requirements have been included under another section, it is sufficient to provide a reference.
B8.2 Identification of Critical Supplies and Consumables
Clearly identify and document all supplies and consumables that may directly or indirectly affect
the quality of the project or task. See Figures 9 and 10 for example documentation of
inspection/acceptance testing requirements. Typical examples include sample bottles, calibration gases,
reagents, hoses, materials for decontamination activities, deionized water, and potable water.
For each item identified, document the inspection or acceptance testing requirements or
specifications (e.g., concentration, purity, cell viability, activity, or source of procurement) in addition to
any requirements for certificates of purity or analysis.
B8.3 Establishing Acceptance Criteria
Acceptance criteria must be consistent with overall project technical and quality criteria (e.g.,
concentration must be within ± 2.5%, cell viability must be >90%). If special requirements are needed
for particular supplies or consumables, a clear agreement should be established with the supplier,
including the methods used for evaluation and the provisions for settling disparities.
B8.4 Inspection or Acceptance Testing Requirements and Procedures
Inspections or acceptance testing should be documented, including procedures to be followed,
individuals responsible, and frequency of evaluation. In addition, handling and storage conditions for
supplies and consumables should be documented.
EPA QA/G-5 35 QA98
-------
B8.5 Tracking and Quality Verification of Supplies and Consumables
Procedures should be established to ensure that inspections or acceptance testing of supplies and
consumables are adequately documented by permanent, dated, and signed records or logs that uniquely
identify the critical supplies or consumables, the date received, the date tested, the date to be retested (if
applicable), and the expiration date. These records should be kept by the responsible individual(s) (see
Figure 11 for an example log). In order to track supplies and consumables, labels with the information
on receipt and testing should be used.
These or similar procedures should be established to enable project personnel to (1) verify, prior
to use, that critical supplies and consumables meet specified project or task quality objectives; and
(2) ensure that supplies and consumables that have not been tested, have expired, or do not meet
acceptance criteria are not used for the project or task.
Unique identification no. (if not clearly shown)_
Date received
Date opened
Date tested (if performed)
Date to be retested (if applicable).
Expiration date
Figure 9. Example of a Record for Consumables
Critical
Supplies and
Consumables
Inspection/
Acceptance
Testing
Requirements
Acceptance
Criteria
Testing
Method
Frequency
Responsible
Individual
Handling/Storage
Conditions
Figure 10. Example of Inspection/Acceptance Testing Requirements
Critical Supplies
and Consumable
(Type, ID No.)
Date
Received
Meets Inspection/
Acceptance Criteria
(Y/N, Include Date)
Requires Retesting
(Y/N, If Yes, Include
Date)
Expiration
Date
Comments
Initials/Date
Figure 11. Example of a Log for Tracking Supplies and Consumables
EPA QA/G-5
36
QA98
-------
B9 DATA ACQUISITION REQUIREMENTS (NON-DIRECT MEASUREMENTS)
Identify any types of data needed for project implementation or decision making that are
obtained from non-measurement sources such as computer databases, programs, literature
files, and historical databases.
Define the acceptance criteria for the use of such data in the project and discuss any
limitations on the use of the data resulting from uncertainty in its quality.
Document the rationale for the original collection of data and indicate its relevance to this
project.
B9.1 Purpose/Background
This element of the QAPP should clearly identify the intended sources of previously collected
data and other information that will be used in this project. Information that is non-representative and
possibly biased and is used uncritically may lead to decision errors. The care and skepticism applied to
the generation of new data are also appropriate to the use of previously compiled data (for example, data
sources such as handbooks and computerized databases).
B9.2 Acquisition of Non-Direct Measurement Data
This element's criteria should be developed to support the objectives of element A7. Acceptance
criteria for each collection of data being considered for use in this project should be explicitly stated,
especially with respect to:
Representativeness. Were the data collected from a population that is sufficiently
similar to the population of interest and the population boundaries? How will potentially
confounding effects (for example, season, time of day, and cell type) be addressed so
that these effects do not unduly alter the summary information?
Bias. Are there characteristics of the data set that would shift the conclusions. For
example, has bias in analysis results been documented? Is there sufficient information to
estimate and correct bias?
• Precision. How is the spread in the results estimated? Does the estimate of variability
indicate that it is sufficiently small to meet the objectives of this project as stated in
element A7? See also Appendix D.
Qualifiers. Are the data evaluated in a manner that permits logical decisions on whether
or not the data are applicable to the current project? Is the system of qualifying or
flagging data adequately documented to allow the combination of data sets?
• Summarization. Is the data summarization process clear and sufficiently consistent
with the goals of this project? (See element D2 for further discussion.) Ideally,
observations and transformation equations are available so that their assumptions can be
evaluated against the objectives of the current project.
EPA QA/G-5 37 QA98
-------
This element should also include a discussion on limitations on the use of the data and the nature of the
uncertainty of the data.
BIO DATA MANAGEMENT
Describe the project data management scheme, tracing the path of the data from their
generation in the field or laboratory to their final use or storage. Describe or reference the
standard record-keeping procedures, document control system, and the approach used for
data storage and retrieval on electronic media.
Discuss the control mechanism for detecting and correcting errors and for preventing loss of
data during data reduction, data reporting, and data entry to forms, reports, and databases.
Provide examples of any forms or checklists to be used.
Identify and describe all data handling equipment and procedures to process, compile, and
analyze the data, including any required computer hardware and software. Address any
specific performance requirements and describe the procedures that will be followed to
demonstrate acceptability of the hardware/software configuration required.
Describe the process for assuring that applicable Agency information resource management
requirements and locational data requirements are satisfied. If other Agency data
management requirements are applicable, discuss how these requirements are addressed.
B10.1 Purpose/Background
This element should present an overview of all mathematical operations and analyses performed
on raw ("as-collected") data to change their form of expression, location, quantity, or dimensionality.
These operations include data recording, validation, transformation, transmittal, reduction, analysis,
management, storage, and retrieval. A diagram that illustrates the source(s) of the data, the processing
steps, the intermediate and final data files, and the reports produced may be helpful, particularly when
there are multiple data sources and data files. When appropriate, the data values should be subjected to
the same chain-of-custody requirements as outlined in element B3. Appendix G has further details.
B10.2 Data Recording
Any internal checks (including verification and validation checks) that will be used to ensure
data quality during data encoding in the data entry process should be identified together with the
mechanism for detailing and correcting recording errors. Examples of data entry forms and checklists
should be included.
B10.3 Data Validation
The details of the process of data validation and prespecified criteria should be documented in
this element of the QAPP. This element should address how the method, instrument, or system performs
the function it is intended to consistently, reliably, and accurately in generating the data. Part D of this
document addresses the overall project data validation, which is performed after the project has been
completed.
EPAQA/G-5 38 QA98
-------
B10.4 Data Transformation
Data transformation is the conversion of individual data point values into related values or
possibly symbols using conversion formulas (e.g., units conversion or logarithmic conversion) or a
system for replacement. The transformations can be reversible (e.g., as in the conversion of data points
using a formulas) or irreversible (e.g., when a symbol replaces actual values and the value is lost). The
procedures for all data transformations should be described and recorded in this element. The procedure
for converting calibration readings into an equation that will be applied to measurement readings should
be documented in the QAPP. Transformation and aberration of data for statistical analysis should be
outlined in element D3, "Reconciliation with Data Quality Objectives."
B10.5 Data Transmittal
Data transmittal occurs when data are transferred from one person or location to another or when
data are copied from one form to another. Some examples of data transmittal are copying raw data from
a notebook onto a data entry form for keying into a computer file and electronic transfer of data over a
telephone or computer network. The QAPP should describe each data transfer step and the procedures
that will be used to characterize data transmittal error rates and to minimize information loss in the
transmittal.
B10.6 Data Reduction
Data reduction includes all processes that change the number of data items. This process is
distinct from data transformation in that it entails an irreversible reduction in the size of the data set and
an associated loss of detail. For manual calculations, the QAPP should include an example in which
typical raw data are reduced. For automated data processing, the QAPP should clearly indicate how the
raw data are to be reduced with a well-defined audit trail, and reference to the specific software
documentation should be provided.
B10.7 Data Analysis
Data analysis sometimes involves comparing suitably reduced data with a conceptual model
(e.g., a dispersion model or an infectivity model). It frequently includes computation of summary
statistics, standard errors, confidence intervals, tests of hypotheses relative to model parameters, and
goodness-of-fit tests. This element should briefly outline the proposed methodology for data analysis
and a more detailed discussion should be included in the final report.
B10.8 Data Tracking
Data management includes tracking the status of data as they are collected, transmitted, and
processed. The QAPP should describe the established procedures for tracking the flow of data through
the data processing system.
B10.9 Data Storage and Retrieval
The QAPP should discuss data storage and retrieval including security and time of retention, and
it should document the complete control system. The QAPP should also discuss the performance
requirements of the data processing system, including provisions for the batch processing schedule and
the data storage facilities.
EPA QA/G-5 39 QA98
-------
EPA QA/G-5
40
QA98
-------
C ASSESSMENT/OVERSIGHT
Cl ASSESSMENTS AND RESPONSE ACTIONS
Identify the number, frequency, and type of assessment activities needed for this project.
List and describe the assessments to be used in the project. Discuss the information expected
and the success criteria for each assessment proposed. List the approximate schedule of
activities, identify potential organizations and participants. Describe how and to whom the
results of the assessments shall be reported.
Define the scope of authority of the assessors, including stop work orders. Define explicitly the
unsatisfactory conditions under which the assessors are authorized to act and provide an
approximate schedule for the assessments to be performed.
Discuss how response actions to non-conforming conditions shall be addressed and by whom.
Identify who is responsible for implementing the response action and describe how response
actions shall be verified and documented.
Cl.l Purpose/Background
During the planning process, many options for sampling design (see EPA QA/G-5S, Guidance
on Sampling Design to Support QAPPs), sample handling, sample cleanup and analysis, and data
reduction are evaluated and chosen for the project. In order to ensure that the data collection is
conducted as planned, a process of evaluation and validation is necessary. This element of the QAPP
describes the internal and external checks necessary to ensure that:
all elements of the QAPP are correctly implemented as prescribed,
• the quality of the data generated by implementation of the QAPP is adequate, and
corrective actions, when needed, are implemented in a timely manner and their
effectiveness is confirmed.
Although any external assessments that are planned should be described in the QAPP, the most
important part of this element is documenting all planned internal assessments. Generally, internal
assessments are initiated or performed by the internal QA Officer so the activities described in this
element should be related to the responsibilities of the QA Officer as discussed in Section A4.
C1.2 Assessment Activities and Project Planning
The following is a description of various types of assessment activities available to managers in
evaluating the effectiveness of environmental program implementation.
Cl.2.1 Assessment of the Subsidiary Organizations
A. Management Systems Review (MSR). A form of management assessment, this process is
a qualitative assessment of a data collection operation or organization to establish
whether the prevailing quality management structure, policies, practices, and procedures
are adequate for ensuring that the type and quality of data needed are obtained. The
EPAQA/G-5 41 QA98
-------
MSR is used to ensure that sufficient management controls are in place and carried out
by the organization to adequately plan, implement, and assess the results of the project.
See the Guidance for the Management Systems Review Process (EPA QA/G-3).
B. Readiness reviews. A readiness review is a technical check to determine if all
components of the project are in place so that work can commence on a specific phase.
Cl.2.2 Assessment of Project Activities
A. Surveillance. Surveillance is the continual or frequent monitoring of the status of a
project and the analysis of records to ensure that specified requirements are being
fulfilled.
B. Technical Systems Audit (TSA). A TSA is a thorough and systematic onsite qualitative
audit, where facilities, equipment, personnel, training, procedures, and record keeping
are examined for conformance to the QAPP. The TSA is a powerful audit tool with
broad coverage that may reveal weaknesses in the management structure, policy,
practices, or procedures. The TSA is ideally conducted after work has commenced, but
before it has progressed very far, thus giving opportunity for corrective action.
C. Performance Evaluation (PE). A PE is a type of audit in which the quantitative data
generated by the measurement system are obtained independently and compared with
routinely obtained data to evaluate the proficiency of an analyst or laboratory. "Blind"
PE samples are those whose identity is unknown to those operating the measurement
system. Blind PEs often produce better performance assessments because they are
handled routinely and are not given the special treatment that undisguised PEs
sometimes receive. The QAPP should list the PEs that are planned, identifying:
the constituents to be measured,
• the target concentration ranges,
the timing/schedule for PE sample analysis, and
• the aspect of measurement quality to be assessed (e.g., bias, precision,
and detection limit).
A number of EPA regulations and EPA-sanctioned methods require the successful
accomplishment of PEs before the results of the test can be considered valid. PE
materials are now available from commercial sources and a number of EPA Program
Offices coordinate various interlaboratory studies and laboratory proficiency programs.
Participation in these or in the National Voluntary Laboratory Accreditation Program
(NVLAP, run by NIST) should be mentioned in the QAPP.
D. Audit of Data Quality (ADQ). An ADQ reveals how the data were handled, what
judgments were made, and whether uncorrected mistakes were made. Performed prior to
producing a project's final report, ADQs can often identify the means to correct
systematic data reduction errors.
E. Peer review. Peer review is not a TSA, nor strictly an internal QA function, as it may
encompass non-QA aspects of a project and is primarily designed for scientific review.
Whether a planning team chooses ADQs or peer reviews depends upon the nature of the
EPA QA/G-5 42 QA98
-------
project, the intended use of the data, the policies established by the sponsor of the
project, and overall the conformance to the Program Office or Region's peer-review
policies and procedures. Reviewers are chosen who have technical expertise comparable
to the project's performers but who are independent of the project. ADQs and peer
reviews ensure that the project activities:
were technically adequate,
• were competently performed,
were properly documented,
• satisfied established technical requirements, and
satisfied established QA requirements.
In addition, peer reviews assess the assumptions, calculations, extrapolations, alternative
interpretations, methods, acceptance criteria, and conclusions documented in the
project's report. Any plans for peer review should conform with the Agency's peer-
review policy and guidance. The names, titles, and positions of the peer reviewers
should be included in the final QAPP, as should their report findings, the QAPP authors'
documented responses to their findings, and reference to where responses to peer-review
comments may be located, if necessary.
F. Data Quality Assessment (DQA). DQA involves the application of statistical tools to
determine whether the data meet the assumptions that the DQOs and data collection
design were developed under and whether the total error in the data is tolerable.
Guidance for the Data Quality Assessment Process (EPA QA/G-9) provides
nonmandatory guidance for planning, implementing, and evaluating retrospective
assessments of the quality of the results from environmental data operations.
C1.3 Documentation of Assessments
The following material describes what should be documented in a QAPP after consideration of
the above issues and types of assessments.
Cl.3.1 Number. Frequency, and Types of Assessments
Depending upon the nature of the project, there may be more than one assessment. A schedule
of the number, frequencies, and types of assessments required should be given.
Cl.3.2 Assessment Personnel
The QAPP should specify the individuals, or at least the specific organizational units, who will
perform the assessments. Internal audits are usually performed by personnel who work for the
organization performing the project work but who are organizationally independent of the management
of the project. External audits are performed by personnel of organizations not connected with the
project but who are technically qualified and who understand the QA requirements of the project.
Cl.3.3 Schedule of Assessment Activities
A schedule of audit activities, together with relevant criteria for assessment, should be given to
the extent that it is known in advance of project activities.
EPA QA/G-5 43 QA98
-------
Cl.3.4 Reporting and Resolution of Issues
Audits, peer reviews, and other assessments often reveal findings of practice or procedure that do
not conform to the written QAPP. Because these issues must be addressed in a timely manner, the
protocol for resolving them should be given here together with the proposed actions to ensure that the
corrective actions were performed effectively. The person to whom the concerns should be addressed,
the decision making hierarchy, the schedule and format for oral and written reports, and the
responsibility for corrective action should all be discussed in this element. It also should explicitly define
the unsatisfactory conditions upon which the assessors are authorized to act and list the project personnel
who should receive assessment reports.
C2 REPORTS TO MANAGEMENT
Identify the frequency and distribution of reports issued to inform management of the status
of the project; results of performance evaluations and systems audits; results of periodic data
quality assessments; and significant quality assurance problems and recommended solutions.
Identify the preparer and the recipients of the reports, and the specific actions management is
expected to take as a result of the reports.
C2.1 Purpose/Background
Effective communication between all personnel is an integral part of a quality system. Planned
reports provide a structure for apprising management of the project schedule, the deviations from
approved QA and test plans, the impact of these deviations on data quality, and the potential
uncertainties in decisions based on the data. Verbal communication on deviations from QA plans should
be noted in summary form in element Dl of the QAPP.
C2.2 Frequency, Content, and Distribution of Reports
The QAPP should indicate the frequency, content, and distribution of the reports so that
management may anticipate events and move to ameliorate potentially adverse results. An important
benefit of the status reports is the opportunity to alert the management of data quality problems, propose
viable solutions, and procure additional resources. If program assessment (including the evaluation of
the technical systems, the measurement of performance, and the assessment of data) is not conducted on
a continual basis, the integrity of the data generated in the program may not meet the quality
requirements. These audit reports, submitted in a timely manner, will provide an opportunity to
implement corrective actions when most appropriate.
C2.3 Identify Responsible Organizations
It is important that the QAPP identify the personnel responsible for preparing the reports,
evaluating their impact, and implementing follow-up actions. It is necessary to understand how any
changes made in one area or procedure may affect another part of the project. Furthermore, the
documentation for all changes should be maintained and included in the reports to management. At the
end of a project, a report documenting the Data Quality Assessment findings to management should be
prepared.
EPA QA/G-5 44 QA98
-------
D DATA VALIDATION AND USABILITY
Dl DATA REVIEW, VALIDATION, AND VERIFICATION REQUIREMENTS
State the criteria used to review and validate data.
Provide examples of any forms or checklists to be used.
Identify any project-specific calculations required.
Dl.l Purpose/Background
The purpose of this element is to state the criteria for deciding the degree to which each data
item has met its quality specifications as described in Group B. Investigators should estimate the
potential effect that each deviation from a QAPP may have on the usability of the associated data item,
its contribution to the quality of the reduced and analyzed data, and its effect on the decision.
The process of data verification requires confirmation by examination or provision of objective
evidence that the requirements of these specified QC acceptance criteria are met. In design and
development, verification concerns the process of examining the result of a given activity to determine
conformance to the stated requirements for that activity. For example, have the data been collected
according to a specified method and have the collected data been faithfully recorded and transmitted?
Do the data fulfill specified data format and metadata requirements. The process of data verification
effectively ensures the accuracy of data using validated methods and protocols and is often based on
comparison with reference standards.
The process of data validation requires confirmation by examination and provision of objective
evidence that the particular requirements for a specific intended use have been fulfilled. In design and
development, validation concerns the process of examining a product or result to determine conformance
to user needs. For example, have the data and assessment methodology passed a peer review to evaluate
the adequacy of their accuracy and precision in assessing progress towards meeting the specific
commitment articulated in the objective or subobjective. The method validation process effectively
develops the QC acceptance criteria or specific performance criteria.
Each of the following areas of discussion should be included in the QAPP elements. The
discussion applies to situations in which a sample is separated from its native environment and
transported to a laboratory for analysis and data generation. However, these principles can be adapted to
other situations (for example, in-situ analysis or laboratory research).
D1.2 Sampling Design
How closely a measurement represents the actual environment at a given time and location is a
complex issue that is considered during development of element B1. See Guidance on Sampling Designs
to Support QAPPs (EPA QA/G-5S). Acceptable tolerances for each critical sample coordinate and the
action to be taken if the tolerances are exceeded should be specified in element Bl.
Each sample should be checked for conformity to the specifications, including type and location
(spatial and temporal). By noting the deviations in sufficient detail, subsequent data users will be able to
EPA QA/G-5 45 QA98
-------
determine the data's usability under scenarios different from those included in project planning. The
strength of conclusions that can be drawn from data (see Guidance Document for Data Quality
Assessment, EPA QA/G-9) has a direct connection to the sampling design and deviations from that
design. Where auxiliary variables are included in the overall data collection effort (for example,
microbiological nutrient characteristics or process conditions), they should be included in this evaluation.
D1.3 Sample Collection Procedures
Details of how a sample is separated from its native time/space location are important for
properly interpreting the measurement results. Element B2 provides these details, which include
sampling and ancillary equipment and procedures (including equipment decontamination). Acceptable
departures (for example, alternate equipment) from the QAPP, and the action to be taken if the
requirements cannot be satisfied, should be specified for each critical aspect. Validation activities should
note potentially unacceptable departures from the QAPP. Comments from field surveillance on
deviations from written sampling plans also should be noted.
D1.4 Sample Handling
Details of how a sample is physically treated and handled during relocation from its original site
to the actual measurement site are extremely important. Correct interpretation of the subsequent
measurement results requires that deviations from element B3 of the QAPP and the actions taken to
minimize or control the changes, be detailed. Data collection activities should indicate events that occur
during sample handling that may affect the integrity of the samples.
At a minimum, investigators should evaluate the sample containers and the preservation methods
used and ensure that they are appropriate to the nature of the sample and the type of data generated from
the sample. Checks on the identity of the sample (e.g., proper labeling and chain-of-custody records) as
well as proper physical/chemical storage conditions (e.g., chain-of-custody and storage records) should
be made to ensure that the sample continues to be representative of its native environment as it moves
through the analytical process.
D1.5 Analytical Procedures
Each sample should be verified to ensure that the procedures used to generate the data (as
identified in element B4 of the QAPP) were implemented as specified. Acceptance criteria should be
developed for important components of the procedures, along with suitable codes for characterizing each
sample's deviation from the procedure. Data validation activities should determine how seriously a
sample deviated beyond the acceptable limit so that the potential effects of the deviation can be
evaluated during DQA.
D1.6 Quality Control
Element B5 of the QAPP specifies the QC checks that are to be performed during sample
collection, handling, and analysis. These include analyses of check standards, blanks, spikes, and
replicates, which provide indications of the quality of data being produced by specified components of
the measurement process. For each specified QC check, the procedure, acceptance criteria, and
corrective action (and changes) should be specified. Data validation should document the corrective
actions that were taken, which samples were affected, and the potential effect of the actions on the
validity of the data.
EPA QA/G-5 46 QA98
-------
D1.7 Calibration
Element B7 addresses the calibration of instruments and equipment and the information that
should be presented to ensure that the calibrations:
• were performed within an acceptable time prior to generation of measurement data;
• were performed in the proper sequence;
• included the proper number of calibration points;
• were performed using standards that "bracketed" the range of reported measurement
results (otherwise, results falling outside the calibration range are flagged as such); and
had acceptable linearity checks and other checks to ensure that the measurement system
was stable when the calibration was performed.
When calibration problems are identified, any data produced between the suspect calibration event and
any subsequent recalibration should be flagged to alert data users.
D1.8 Data Reduction and Processing
Checks on data integrity evaluate the accuracy of "raw" data and include the comparison of
important events and the duplicate rekeying of data to identify data entry errors.
Data reduction is an irreversible process that involves a loss of detail in the data and may involve
averaging across time (for example, hourly or daily averages) or space (for example, compositing results
from samples thought to be physically equivalent). Since this summarizing process produces few values
to represent a group of many data points, its validity should be well-documented in the QAPP. Potential
data anomalies can be investigated by simple statistical analyses (see Guidance for Data Quality
Assessment, EPA QA/G-9).
The information generation step involves the synthesis of the results of previous operations and
the construction of tables and charts suitable for use in reports. How information generation is checked,
the requirements for the outcome, and how deviations from the requirements will be treated, should be
addressed in this element.
D2 VALIDATION AND VERIFICATION METHODS
Describe the process to be used for validating and verifying data, including the chain of
custody for data throughout the life cycle of the project or task.
Discuss how issues shall be resolved and identify the authorities for resolving such issues.
Describe how the results are conveyed to the data users.
Precisely define and interpret how validation issues differ from verification issues for this
project.
EPA QA/G-5 47 QA98
-------
D2.1 Purpose/Background
The purpose of this element is to describe, in detail, the process for validating (determining if
data satisfy QAPP-defmed user requirements) and verifying (ensuring that conclusions can be correctly
drawn) project data. The amount of data validated is directly related to the DQOs developed for the
project. The percentage validated for the specific project together with its rationale should be outlined or
referenced. The QAPP should have a clear definition of what is implied by "verification" and
"validation."
D2.2 Describe the Process for Validating and Verifying Data
The individuals responsible for data validation together with the lines of authority should be
shown on an organizational chart and may be indicated in the chart in element A7. The chart should
indicate who is responsible for each activity of the overall validation and verification processes.
The data to be validated should be compared to "actual" events using the criteria documented in
the QAPP. The data validation procedure for all environmental measurements should be documented in
the SOPs for specific data validation. Verification and validation issues are discussed at length in
Guidance on Environmental Verification and Validation, (EPA QA/G-8).
D3 RECONCILIATION WITH DATA QUALITY OBJECTIVES
Describe how the results obtained from the project or task will be reconciled with the
requirements defined by the data user or decision maker.
Outline the proposed methods to analyze the data and determine possible anomalies or
departures from assumptions established in the planning phase of data collection.
Describe how issues will be resolved and discuss how limitations on the use of the data will be
reported to decision makers.
D3.1 Purpose/Background
The purpose of element D3 is to outline and specify, if possible, the acceptable methods for
evaluating the results obtained from the project. This element includes scientific and statistical
evaluations of data to determine if the data are of the right type, quantity, and quality to support their
intended use.
D3.2 Reconciling Results with DQOs
The DQA process has been developed for cases where formal DQOs have been established.
Guidance for Data Quality Assessment (EPA QA/G-9) focuses on evaluating data for fitness in decision
making and also provides many graphical and statistical tools.
DQA is a key part of the assessment phase of the data life cycle, as shown in Figure 1. As the
part of the assessment phase that follows data validation and verification, DQA determines how well the
validated data can support their intended use. If an approach other than DQA has been selected, an
outline of the proposed activities should be included.
EPA QA/G-5 48 QA98
-------
CHAPTER IV
QAPP REVISIONS AND RELATED GUIDANCE
QAPP REVISIONS
During the course of environmental data collection, it is possible that changes will occur and
revisions to the QAPP will have to be made. Any changes to the technical procedures should be
evaluated by the EPA QA Officer and Project Officer to determine if they significantly affect the
technical and quality objectives of the project. If so, the QAPP should be revised and reapproved, and a
revised copy should be sent to all the persons on the distribution list.
COMPARISON WITH PREVIOUS GUIDANCE (QAMS-005/80)
EPA's previous guidance for preparing QAPPs, Interim Guidelines and Specifications for
Preparing Quality Assurance Project Plans (QAMS-005/80), was released in December 1980. The
evolution of EPA programs, changing needs, and changes to quality management practices have
mandated the preparation of a new guidance. The QAPPs that will be generated based on this guidance
will be slightly different from those in the past because:
• New QAPP specifications are given in the R-5 requirements document.
• Additional guidance documents from the Agency including Guidance for the Data
Quality Objectives Process (EPA QA/G-4), and Guidance for Data Quality Assessment
(EPA QA/G-9), are available on important quality management practices. These
guidance documents show how the DQO Process, the QAPP, and the DQA Process link
together in a coherent way (see Appendix A for a crosswalk between the DQOs and the
QAPP).
• The new guidance includes flexibility in the requirements and reporting format.
However, if an element of the QAPP is not applicable to a particular project, the
rationale for not addressing the element should be included.
• The elements of the QAPP are now organized in an order that corresponds to the
customary planning, implementation, and assessment phases of a project. They have
been categorized into four groups for ease of implementation:
• Project Management,
Measurement/Data Acquisition,
• Assessment/Oversight, and
Data Validation and Usability.
• There are more elements identified than in the previous QAMS-005/80 guidance and this
encourages flexibility in construction of defensible QAPPs.
A comparison between the requirements of QAMS-005/80 and the R-5 document is presented in
Appendix A, "Crosswalk Between EPA QA/R-5 and QAMS-005/80."
EPA QA/G-5 49 QA98
-------
APPENDIX A
CROSSWALKS BETWEEN QUALITY ASSURANCE DOCUMENTS
This appendix consists of five sections. The first section describes the relationship between the
systems requirements developed in ANSI/ASQC E4-1994 and the Environmental Protection Agency
(EPA) Quality System requirements. The second section provides a crosswalk between the requirements
document for Quality Assurance Project Plans (QAPPs), EPA QA/R-5, EPA Requirements for Quality
Assurance Project Plans for Environmental Data Operations, and its predecessor document, QAMS
005/80, Interim Guidelines and Specifications for Preparing Quality Assurance Project Plans. The third
section provides a crosswalk between QA/R-5 and the elements of International Organization for
Standardization (ISO) 9000. The fourth section is a crosswalk between the requirements of the QAPP
and the steps of the Data Quality Objectives (DQOs) Process. The final section describes the Agency's
QA documents at the program and project levels.
AA1. RELATIONSHIP BETWEEN E4 AND EPA QUALITY SYSTEM
EPA Order 5360.1 establishes a mandatory Agency-wide Quality System that applies to all
organizations, both internal and external, performing work for EPA. (The authority for the requirements
defined by the Order are contained in the applicable regulations for extramural agreements.) These
organizations must ensure that data collected for the characterization of environmental processes and
conditions are of the appropriate type and quality for their intended use and that environmental
technologies are designed, constructed, and operated according to defined expectations. All EPA
Regional, Office, and Laboratory quality systems established in accordance with these requirements shall
comply with ANSI/ASQC E4-1994, Specifications and Guidelines for Quality Systems for
Environmental Data Collection and Environmental Technology Programs, which conforms generally to
ISO 9000. In addition, EPA has developed two documents S EPA QA/R-1, EPA Quality Systems
Requirements for Environmental Programs, and EPA QA/R-2, EPA Requirements for Quality
Management Plans S that specify the requirements for developing, documenting, implementing, and
assessing a Quality System. This appendix describes these three Agency documents (Order 5360.1, EPA
QA/R-1, and EPA QA/R-2) in order to define their relationships and roles in laying the foundation for
EPA's Quality System.
ANSI/ASQC E4-1994 provides the basis for the preparation of a quality system for an
organization's environmental programs. The document provides the requisite management and technical
area elements necessary for developing and implementing a quality system. The document first describes
the quality management elements that are generally common to environmental problems, regardless of
their technical scope. The document then discusses the specifications and guidelines that apply to
project-specific environmental activities involving the generation, collection, analysis, evaluation, and
reporting of environmental data. Finally, the document contains the minimum specifications and
guidelines that apply to the design, construction, and operation of environmental technology.
EPA QA/R-1 provides the details on EPA quality management requirements to organizations
conducting environmental programs. This document states that"... all EPA organizations and all
organizations performing work for EPA shall develop and establish Quality Systems, as appropriate, that
conform to the American National Standard ANSI/ASQC E4-1994, Specifications and Guidelines for
Quality Systems for Environmental Data Collection and Environmental Technology Programs, and its
additions and supplements from the American National Standards Institute (ANSI) and the American
Society for Quality Control (ASQC)." R-l applies to all EPA programs and organizations, unless
explicitly exempted, that produce, acquire, or use environmental data depending on the purposes for
EPAQA/G-5 A-l QA98
-------
which the data will be used. This document also applies to systems, facilities, processes, and methods
for pollution control, waste treatment, waste remediation, and waste packaging and storage. Essentially,
R-l formally describes how EPA Order 5360.1 applies to extramural organizations.
EPA Requirements for Quality Management Plans, EPA QA/R-2, discusses the development,
review, approval, and implementation of the Quality Management Plan (QMP). The QMP is a means of
documenting how an organization will plan, implement, and assess the effectiveness of the management
processes and structures (required under R-l) that relate to the Quality System. R-2 describes the
program elements that should be part of a QMP. These requirements match the quality management
elements described in ANSI/ASQC E4-1994 that are generally common to environmental projects.
These elements include the following: (1) management and organization, (2) quality system and
description, (3) personnel qualifications and training, (4) procurement of items and services, (5)
documents and records, (6) computer hardware and software, (7) planning, (8) implementation of work
processes, (9) assessment and response, and (10) quality improvement.
The procedures, roles, and responsibilities for QAPPs are addressed in the organization's QMP.
In essence, the QMP establishes the nature of the requirements for QAPPs for work done by or for that
organization.
EPA QA/G-5 A-2 QA98
-------
AA2. CROSSWALK BETWEEN EPA QA/R-5 AND QAMS-005/80
QAMS-005/80 ELEMENTS
1.0
2.0
3.0
4.0
5.0
6.0
7.0
8.0
9.0
10.0
11.0
12.0
13.0
14.0
15.0
16.0
Title Page with Provision for Approval
Signatures
Table of Contents
Project Description
Project Organization and
Responsibility
QA Objectives for Measurement
Data (PARCC)
Sampling Procedures
Sample Custody
Calibration Procedures and Frequency
Analytical Procedures
Data Reduction, Validation, and
Reporting
Internal Quality Control Checks and
Frequency
Performance and Systems
Preventive Maintenance
Specific Routine Procedures
Measurement Parameters Involved
Corrective Action
QA Reports to Management
QA/R-5 ELEMENTS
Al
A2
A5
A6
A4
A9
A7
Bl
B2
A8
B3
B7
B4
Dl
D2
B9
BIO
B5
Cl
B6
B8
D3
Cl
A3
C2
Title and Approval Sheet
Table of Contents
Problem Definition/Background
Project/Task Description
Project/Task Organization
Documentation and Records
Quality Objectives and Criteria for Measurement
Data
Sampling Process Design
Sampling Methods Requirements
Special Training Requirements or Certification
Sample Handling and Custody Requirements
Instrument Calibration and Frequency
Analytical Methods Requirements
Data Review, Validation, and Verification
Requirements
Validation and Verification Methods
Data Acquisition Requirements
Data Quality Management
Quality Control Requirements
Assessments and Response Actions
Instrument/Equipment Testing, Procedures and
Schedules Inspection, and Maintenance
Requirements
Inspection/ Acceptance Requirements for Supplies
and Consumables
Reconciliation with Data Used to Assess
PARCC for Quality Objectives Measurement
Assessments and Response Actions
Distribution List
Reports to Management
EPA QA/G-5
A-3
QA98
-------
AA3. CROSSWALK BETWEEN EPA QA/R-5 AND ISO 9000
EPA QA/R-5 Elements
Al
A2
A3
A4
A5
A6
A7
A8
A9
Bl
B2
B3
B4
B5
B6
B7
B8
B9
BIO
Cl
C2
Dl
D2
D3
Title and Approval Sheet
Table of Contents
Distribution List
Project/Task Organization
Problem Definition/Background
Project/Task Description
Quality Objectives and Criteria for
Measurement Data
Special Training Requirements/Certification
Documentation and Records
Sampling Process Design
Sampling Methods Requirements
Sample Handling and Custody Requirements
Analytical Methods Requirements
Quality Control Requirements
Instrument/Equipment Testing, Inspection, and
Maintenance Requirements
Instrument Calibration and Frequency
Inspection/ Acceptance Requirements for
Supplies and Consumables
Data Acquisition Requirements
Data Quality Management
Assessments and Response Actions
Reports to Management
Data Review, Validation, and Verification
Requirements
Validation and Verification Methods
Reconciliation with User Requirements
ISO 9000 Elements
N/A
N/A
N/A
4
Management Responsibility
N/A
N/A
5
5.2
Quality System Principles
Structure of the Quality System
N/A
N/A
8
10
16
10
11
13
Quality in Specification and Design
Quality of Production
Handling and Post-Production Functions
Quality of Production
Control of Production
Control of Measuring and Test Equipment
N/A
9
11.2
Quality in Procurement
Material Control and Traceability
N/A
N/A
5.4
14
15
5.3
6
11.7
12
Auditing the Quality System
Nonconformity
Corrective Action
Documentation of the Quality System
Economics - Quality Related Costs
Control of Verification Status
Verification Status
N/A
7
Quality in Marketing
EPA QA/G-5
A-4
QA98
-------
AA4.
CROSSWALK BETWEEN THE DQO PROCESS AND THE QAPP
Elements
Requirements
DQO Overlap
PROJECT MANAGEMENT
Al Title and Approval
Sheet
A2 Table of Contents
A3 Distribution List
A4 Project/Task
Organization
A5 Problem
Definition/Backgro
und
A6 Project/Task
Description
A7 Data Quality
Objectives for
Measurement Data
A8 Special Training
Requirements/
Certification
A9 Documentation
and Record
Title and approval sheet.
Document control format.
Distribution list for the QAPP revisions and final
guidance.
Identify individuals or organizations participating
in the project and discuss their roles,
responsibilities and organization.
1 ) State the specific problem to be solved or the
decision to be made.
2) Identify the decision maker and the principal
customer for the results.
1) Hypothesis test, 2) expected measurements, 3)
ARARs or other appropriate standards, 4)
assessment tools (technical audits), 5) work
schedule and required reports.
Decision(s), population parameter of interest,
action level, summary statistics and acceptable
limits on decision errors. Also, scope of the
project (domain or geographical locale).
Identify special training that personnel will need.
Itemize the information and records that must be
included in a data report package, including
report format and requirements for storage, etc.
N/A
N/A
List the members of the scoping team.
Step 1: State the Problem.
Step 1 : State the Problem requires
definition of the DQO scoping or
planning team, which includes the
decision maker, technical staff, data
users, etc. This step also requires the
specification of each member's role and
responsibilities.
Step 1: State the Problem/Step 2:
Identify the Decision requires a
description of the problem. It also
identifies the decision makers who
could use the data.
Step 1 : State the Problem/Step 2:
Identify the Decision requires a work
schedule. Step 3: Identify the Inputs
requires the ARARs or standards and
expected measurements. Step 6:
Specify Limits on Decision Errors.
Step 1: State the Problem, Step 2:
Identify the Decision, Step 4: Define
the Boundaries, Step 5: Develop a
Decision Rule, Step 6: Specify Limits
on Decision Errors.
Step 3: Identify the Inputs to the
Decision.
Step 3: Identify the Inputs to the
Decision, Step 7: Optimize the Design
for Obtaining Data.
MEASUREMENT/DATA ACQUISITION
Bl Sampling Process
Designs
(Experimental
Design)
B2 Sampling Methods
Requirements
B3 Sample Handling
and Custody
Requirements
B4 Analytical
Methods
Requirements
B5 Quality Control
Requirements
Outline the experimental design, including
sampling design and rationale, sampling
frequencies, matrices, and measurement
parameter of interest.
Sample collection method and approach.
Describe the provisions for sample labeling,
shipment, chain-of-custody forms, procedures for
transferring and maintaining custody of samples.
Identify analytical method(s) and equipment for
the study, including method performance
requirements.
Describe routine (real-time) QC procedures that
should be associated with each sampling and
measurement technique. List required QC checks
and corrective action procedures.
Step 5: Develop a Decision Rule, Step
7: Optimize the Design for Obtaining
Data.
Step 7: Optimize the Design for
Obtaining Data.
Step 3: Identify the Inputs to the
Decision.
Step 3: Identify the Inputs to the
Decision, Step 7: Optimize the Design
for Obtaining Data.
Step 3: Identify the Inputs to the
Decision.
EPA QA/G-5
A-5
QA98
-------
Elements
B6 Instrument/Equip
ment Testing
Inspection and
Maintenance
Requirements
B7 Instrument
Calibration and
Frequency
B8 Inspection/Accepta
nee Requirements
for Supplies and
Consumables
B9 Data Acquisition
Requirements
(Non-direct
Measurements)
BIO Data Management
Requirements
Discuss how inspection and acceptance testing,
including the use of QC samples, must be
performed to ensure their intended use as
specified by the design.
Identify tools, gauges and instruments, and other
sampling or measurement devices that need
calibration. Describe how the calibration should
be done.
Define how and by whom the sampling supplies
and other consumables will be accepted for use in
the project.
Define the criteria for the use of non-
measurement data such as data that come from
databases or literature.
Outline the data management scheme including
the path and storage of the data and the data
record-keeping system. Identify all data handling
equipment and procedures that will be used to
process, compile, and analyze the data.
DQO Overlap
Step 3: Identify the Inputs to the
Decision.
Step 3: Identify the Inputs to the
Decision.
N/A
Step 1 : State the Problem, Step 7:
Optimize the Design for Obtaining
Data.
Step 3: Identify the Inputs to the
Decision, Step 7: Optimize the Design
for Obtaining Data.
ASSESSMENT/OVERSIGHT
Cl Assessments and
Response Actions
C2 Reports to
Management
Describe the assessment activities needed for this
project. These may include DQA, PE, TSA,
MSR/PR/RR
Identify the frequency, content, and distribution
of reports issued to keep management informed.
Step 5: Develop a Decision Rule, Step
6: Specify Limits on Decision Errors.
N/A
DATA VALIDATION AND USABILITY
Dl Data Review,
Validation, and
Verification
Requirements
D2 Validation and
Verification
Methods
D3 Reconciliation
With Data Quality
Objectives
State the criteria used to accept or reject the data
based on quality.
Describe the process to be used for validating and
verifying data, including the chain-of-custody for
data throughout the lifetime of the project.
Describe how results will be evaluated to
determine if DQOs have been satisfied.
Step 7: Optimize the Design for
Obtaining Data.
Step 3: Identify the Inputs to the
Decision.
Step 7: Optimize the Design for
Obtaining Data.
EPA QA/G-5
A-6
QA98
-------
AA5. EPA QUALITY ASSURANCE DOCUMENTS
The Quality Assurance Division issues QA documents for use both internally (National Programs,
Centers, and Laboratories) and externally (state and local agencies, contractors, extramural agreement
holders, and nonprofit groups). The scopes of the documents span all aspects of QA and can be obtained
by writing QAD directly or by visiting the QAD Website:
http://es.epa.gov/ncerqa/qa/qa_docs.html
QAD documents fall into three categories: the EPA Quality Manual (for internal use);
Requirements documents (for external use, labeled 'R-xx'); and Guidance documents (for internal and
external use, labeled 'G-xx'). Requirements documents and the Quality Manual contain the Agency's
QA policies and Guidance documents contain nonmandatory guidance on how to achieve these QA
requirements.
Table Al shows the general numbering system for EPA's Quality System documents, and Table
A2 illustrates some specific documents available and under construction. The auxiliary letter on some of
the documents denotes specialized audiences or areas of interest. Figure Al shows the relationship
among the documents at the Policy and Program levels. Figure A2 demonstrates the sequence and
interrelationship of documents at the Program level.
Not all of the documents listed in Table A2 are available, as some are in various stages of
development and will not be finalized until late 1998. Consult the Website or contact QAD directly for
information on the current status and availability of all QAD documents.
Table AA1. Numbering System for EPA's Quality System Documents
1 = Quality System Policy and Quality Manual 6 = Standard Operating Procedures (SOPs)
2 = Quality Management Plans (QMPs) 7 = Technical Assessments (TAs)
3 = Management Systems Reviews (MSRs) 8 = Data Verification and Validation
4 = Data Quality Objectives (DQOs) 9 = Data Quality Assessment (DQA)
5 = Quality Assurance Project Plans (QAPPs) 10 = Training Issues
EPA QA/G-5 A-7 QA98
-------
Table AA2. Quality System Documents
Overview
QA/G-0
Program level
QA/R-1
QA/G-1
QA/R-2
QA/G-2
QA/G-2C
QA/G-2EA
QA/G-2F
QA/G-3
QA/G-10
Project level
QA/G-4
QA/G-4CS
QA/G-4D
QA/G-4HW
QA/G-4R
QA/R-5
QA/G-5
QA/G-5I
QA/G-5 S
QA/G-5T
QA/G-6
QA/G-7
QA/G-8
QA/G-9
QA/G-9D
EPA Quality System Description
EPA Quality Systems Requirements for Environmental Programs
Guidance for Developing Quality Systems for Environmental Data Operations
EPA Requirements for Quality Management Plans
Guidance for Preparing Quality Management Plans
Guide to Satisfying EPA Quality Assurance Requirements for Contracts
Guide to Implementing Quality Assurance in Extramural Agreements
Guide to Satisfying EPA Quality Assurance Requirements for Financial Assistance
Agreements
Guidance for the Management Systems Review Process
Guidance for Determining Quality Training Requirements for Environmental Data
Operations
Guidance for the Data Quality Objectives Process
The Data Quality Objectives Process: Case Studies
Data Quality Objectives Decision Errors Feasibility Trials (DEFT) Software
Guidance for the Data Quality Objectives Process for Hazardous Waste Sites
Guidance for the Data Quality Objectives for Researchers
EPA Requirements for Quality Assurance Project Plans
EPA Guidance for Quality Assurance Project Plans
Guidance for Data Quality Indicators
Guidance on Sampling Designs to Support Quality Assurance Project Plans
Guidance on Specialized Topics in Quality Assurance
Guidance for the Preparation of Standard Operating Procedures for Quality-Related
Operations
Guidance on Technical Assessments for Environmental Data Operations
Guidance on Environmental Data Verification and Validation
Guidance for Data Quality Assessment: Practical Methods for Data Analysis
Data Quality Evaluation Statistical Toolbox (DataQUEST).
EPA QA/G-5
A-8
QA98
-------
o
__i
o
Q_
Agency-
wide
Policies.
Requirements
and
procedures
EPA
Quality
Manual
Authority —
Internal
assessment
o
o
Q_
o
§
Quality
System
structure,
procedures,
standards
Quality
Management
Plan
(QMP)
(G-2)
(R-2)
Quality
Assurance
Annual
Report and
Work Plan
(QAARWP)
Ensures
resources for
implementation of
Quality System
procedures
Quality System
structure, procedures,
standards
Revisions to
QMP
Requirements,
procedures
Internal assessment of
Quality System
effectiveness
Assess-
ment of —
training
needs
Assessment
of Quality
System
effectiveness
Management
Systems Reviews
(MSRs)
(G-3)
Performance
measures
Organizational
responsibilities, Quality
System structure,
procedures, standards,
personnel
Ensure
adequacy of
knowledge,
skills
Assess conformity of
project components
to QMP
1
Performance
measures
Individual Projects
Figure AA1. Relationships Among EPA Quality System Documents at the Program Level
EPA QA/G-5 A-9 QA98
-------
PLANNING
IMPLEMENTATION
ASSESSMENT
INITIAL
PLANNING
DESIGN
IMPLEMENTATION
PLANNING
R-5
Requirements for
QA Project Plans
G-4/G-4R/G-4HW
Guidance on the
DQO Process
(decision making)
G-5
Guidance on QA
Project Plans
G-9
Guidance for Data
Quality Assessment
KEY
G-xx | EPA QA guidance document
^ primary outputs/inputs
^ background/reference information
Figure AA2. Relationship Among EPA Quality System Documents at the Project Level
EPAQA/G-5 A-10
QA98
-------
APPENDIX B
GLOSSARY OF QUALITY ASSURANCE AND RELATED TERMS
Acceptance criteria — Specified limits placed on characteristics of an item, process, or service defined
in requirements documents. (ASQC Definitions)
Accuracy — A measure of the closeness of an individual measurement or the average of a number of
measurements to the true value. Accuracy includes a combination of random error (precision) and
systematic error (bias) components that are due to sampling and analytical operations; the EPA
recommends using the terms "precision " and "bias ", rather than "accuracy," to convey the information
usually associated with accuracy. Refer to Appendix D, Data Quality Indicators for a more detailed
definition.
Activity — An all-inclusive term describing a specific set of operations of related tasks to be performed,
either serially or in parallel (e.g., research and development, field sampling, analytical operations,
equipment fabrication), that, in total, result in a product or service.
Assessment — The evaluation process used to measure the performance or effectiveness of a system and
its elements. As used here, assessment is an all-inclusive term used to denote any of the following: audit,
performance evaluation (PE), management systems review (MSR), peer review, inspection, or
surveillance.
Audit (quality) — A systematic and independent examination to determine whether quality activities
and related results comply with planned arrangements and whether these arrangements are implemented
effectively and are suitable to achieve objectives.
Audit of Data Quality (ADQ) — A qualitative and quantitative evaluation of the documentation and
procedures associated with environmental measurements to verify that the resulting data are of
acceptable quality.
Authenticate — The act of establishing an item as genuine, valid, or authoritative.
Bias — The systematic or persistent distortion of a measurement process, which causes errors in one
direction (i.e., the expected sample measurement is different from the sample's true value). Refer to
Appendix D, Data Quality Indicators, for a more detailed definition.
Blank — A sample subjected to the usual analytical or measurement process to establish a zero baseline
or background value. Sometimes used to adjust or correct routine analytical results. A sample that is
intended to contain none of the analytes of interest. A blank is used to detect contamination during
sample handling preparation and/or analysis.
Calibration — A comparison of a measurement standard, instrument, or item with a standard or
instrument of higher accuracy to detect and quantify inaccuracies and to report or eliminate those
inaccuracies by adjustments.
Calibration drift — The deviation in instrument response from a reference value over a period of time
before recalibration.
EPAQA/G-5 B-l QA98
-------
Certification — The process of testing and evaluation against specifications designed to document,
verify, and recognize the competence of a person, organization, or other entity to perform a function or
service, usually for a specified time.
Chain of custody — An unbroken trail of accountability that ensures the physical security of samples,
data, and records.
Characteristic — Any property or attribute of a datum, item, process, or service that is distinct,
describable, and/or measurable.
Check standard — A standard prepared independently of the calibration standards and analyzed exactly
like the samples. Check standard results are used to estimate analytical precision and to indicate the
presence of bias due to the calibration of the analytical system.
Collocated samples — Two or more portions collected at the same point in time and space so as to be
considered identical. These samples are also known as field replicates and should be identified as such.
Comparability — A measure of the confidence with which one data set or method can be compared to
another.
Completeness — A measure of the amount of valid data obtained from a measurement system compared
to the amount that was expected to be obtained under correct, normal conditions. Refer to Appendix D,
Data Quality Indicators, for a more detailed definition.
Confidence Interval — The numerical interval constructed around a point estimate of a population
parameter, combined with a probability statement (the confidence coefficient) linking it to the
population's true parameter value. If the same confidence interval construction technique and
assumptions are used to calculate future intervals, they will include the unknown population parameter
with the same specified probability.
Confidentiality procedure — A procedure used to protect confidential business information (including
proprietary data and personnel records) from unauthorized access.
Configuration — The functional, physical, and procedural characteristics of an item, experiment, or
document.
Conformance — An affirmative indication or judgment that a product or service has met the
requirements of the relevant specification, contract, or regulation; also, the state of meeting the
requirements.
Consensus standard — A standard established by a group representing a cross section of a particular
industry or trade, or a part thereof.
Contractor — Any organization or individual contracting to furnish services or items or to perform
work.
Corrective action — Any measures taken to rectify conditions adverse to quality and, where possible, to
preclude their recurrence.
EPA QA/G-5 B-2 QA98
-------
Data Quality Assessment (DQA) — The scientific and statistical evaluation of data to determine if data
obtained from environmental operations are of the right type, quality, and quantity to support their
intended use. The five steps of the DQA Process include: 1) reviewing the DQOs and sampling design,
2) conducting a preliminary data review, 3) selecting the statistical test, 4) verifying the assumptions of
the statistical test, and 5) drawing conclusions from the data.
Data Quality Indicators (DQIs) — The quantitative statistics and qualitative descriptors that are used to
interpret the degree of acceptability or utility of data to the user. The principal data quality indicators are
bias, precision, accuracy (bias is preferred), comparability, completeness, representativeness.
Data Quality Objectives (DQOs) — The qualitative and quantitative statements derived from the DQO
Process that clarify study's technical and quality objectives, define the appropriate type of data, and
specify tolerable levels of potential decision errors that will be used as the basis for establishing the
quality and quantity of data needed to support decisions.
Data Quality Objectives (DQO) Process — A systematic strategic planning tool based on the scientific
method that identifies and defines the type, quality, and quantity of data needed to satisfy a specified use.
DQOs are the qualitative and quantitative outputs from the DQO Process.
Data reduction — The process of transforming the number of data items by arithmetic or statistical
calculations, standard curves, and concentration factors, and collating them into a more useful form.
Data reduction is irreversible and generally results in a reduced data set and an associated loss of detail.
Data usability — The process of ensuring or determining whether the quality of the data produced meets
the intended use of the data.
Deficiency — An unauthorized deviation from acceptable procedures or practices, or a defect in an item.
Demonstrated capability — The capability to meet a procurement's technical and quality specifications
through evidence presented by the supplier to substantiate its claims and in a manner defined by the
customer.
Design — The specifications, drawings, design criteria, and performance requirements. Also, the result
of deliberate planning, analysis, mathematical manipulations, and design processes.
Design change — Any revision or alteration of the technical requirements defined by approved and
issued design output documents and approved and issued changes thereto.
Design review — A documented evaluation by a team, including personnel such as the responsible
designers, the client for whom the work or product is being designed, and a quality assurance (QA)
representative but excluding the original designers, to determine if a proposed design will meet the
established design criteria and perform as expected when implemented.
Detection Limit (DL) — A measure of the capability of an analytical method to distinguish samples that
do not contain a specific analyte from samples that contain low concentrations of the analyte; the lowest
concentration or amount of the target analyte that can be determined to be different from zero by a single
measurement at a stated level of probability. DLs are analyte- and matrix-specific and may be
laboratory-dependent.
EPA QA/G-5 B-3 QA98
-------
Distribution — 1) The appointment of an environmental contaminant at a point over time, over an area,
or within a volume; 2) a probability function (density function, mass function, or distribution function)
used to describe a set of observations (statistical sample) or a population from which the observations are
generated.
Document control — The policies and procedures used by an organization to ensure that its documents
and their revisions are proposed, reviewed, approved for release, inventoried, distributed, archived,
stored, and retrieved in accordance with the organization's requirements.
Duplicate samples — Two samples taken from and representative of the same population and carried
through all steps of the sampling and analytical procedures in an identical manner. Duplicate samples are
used to assess variance of the total method, including sampling and analysis. See also collocated sample.
Environmental conditions — The description of a physical medium (e.g., air, water, soil, sediment) or a
biological system expressed in terms of its physical, chemical, radiological, or biological characteristics.
Environmental data — Any parameters or pieces of information collected or produced from
measurements, analyses, or models of environmental processes, conditions, and effects of pollutants on
human health and the ecology, including results from laboratory analyses or from experimental systems
representing such processes and conditions.
Environmental data operations — Any work performed to obtain, use, or report information pertaining
to environmental processes and conditions.
Environmental monitoring — The process of measuring or collecting environmental data.
Environmental processes — Any manufactured or natural processes that produce discharges to, or that
impact, the ambient environment.
Environmental programs — An all-inclusive term pertaining to any work or activities involving the
environment, including but not limited to: characterization of environmental processes and conditions;
environmental monitoring; environmental research and development; the design, construction, and
operation of environmental technologies; and laboratory operations on environmental samples.
Environmental technology — An all-inclusive term used to describe pollution control devices and
systems, waste treatment processes and storage facilities, and site remediation technologies and their
components that may be utilized to remove pollutants or contaminants from, or to prevent them from
entering, the environment. Examples include wet scrubbers (air), soil washing (soil), granulated
activated carbon unit (water), and filtration (air, water). Usually, this term applies to hardware-based
systems; however, it can also apply to methods or techniques used for pollution prevention, pollutant
reduction, or containment of contamination to prevent further movement of the contaminants, such as
capping, solidification or vitrification, and biological treatment.
Estimate — A characteristic from the sample from which inferences on parameters can be made.
Evidentiary records — Any records identified as part of litigation and subject to restricted access,
custody, use, and disposal.
EPA QA/G-5 B-4 QA98
-------
Expedited change — An abbreviated method of revising a document at the work location where the
document is used when the normal change process would cause unnecessary or intolerable delay in the
work.
Field blank — A blank used to provide information about contaminants that may be introduced during
sample collection, storage, and transport. A clean sample, carried to the sampling site, exposed to
sampling conditions, returned to the laboratory, and treated as an environmental sample.
Field (matrix) spike — A sample prepared at the sampling point (i.e., in the field) by adding a known
mass of the target analyte to a specified amount of the sample. Field matrix spikes are used, for example,
to determine the effect of the sample preservation, shipment, storage, and preparation on analyte recovery
efficiency (the analytical bias).
Field split samples — Two or more representative portions taken from the same sample and submitted
for analysis to different laboratories to estimate interlaboratory precision.
Financial assistance — The process by which funds are provided by one organization (usually
governmental) to another organization for the purpose of performing work or furnishing services or
items. Financial assistance mechanisms include grants, cooperative agreements, and governmental
interagency agreements.
Finding — An assessment conclusion that identifies a condition having a significant effect on an item or
activity. An assessment finding may be positive or negative, and is normally accompanied by specific
examples of the observed condition.
Goodness-of-fit test — The application of the chi square distribution in comparing the frequency
distribution of a statistic observed in a sample with the expected frequency distribution based on some
theoretical model.
Grade — The category or rank given to entities having the same functional use but different
requirements for quality.
Graded approach — The process of basing the level of application of managerial controls applied to an
item or work according to the intended use of the results and the degree of confidence needed in the
quality of the results. (See also Data Quality Objectives (DQO) Process.)
Guidance — A suggested practice that is not mandatory, intended as an aid or example in complying
with a standard or requirement.
Guideline — A suggested practice that is not mandatory in programs intended to comply with a standard.
Hazardous waste — Any waste material that satisfies the definition of hazardous waste given in 40 CFR
261, "Identification and Listing of Hazardous Waste."
Holding time — The period of time a sample may be stored prior to its required analysis. While
exceeding the holding time does not necessarily negate the veracity of analytical results, it causes the
qualifying or "flagging" of any data not meeting all of the specified acceptance criteria.
Identification error — The misidentification of an analyte. In this error type, the contaminant of
concern is unidentified and the measured concentration is incorrectly assigned to another contaminant.
EPA QA/G-5 B-5 QA98
-------
Independent assessment — An assessment performed by a qualified individual, group, or organization
that is not a part of the organization directly performing and accountable for the work being assessed.
Inspection — The examination or measurement of an item or activity to verify conformance to specific
requirements.
Internal standard — A standard added to a test portion of a sample in a known amount and carried
through the entire determination procedure as a reference for calibrating and controlling the precision
and bias of the applied analytical method.
Laboratory split samples — Two or more representative portions taken from the same sample and
analyzed by different laboratories to estimate the interlaboratory precision or variability and the data
comparability.
Limit of quantitation — The minimum concentration of an analyte or category of analytes in a specific
matrix that can be identified and quantified above the method detection limit and within specified limits
of precision and bias during routine analytical operating conditions.
Management — Those individuals directly responsible and accountable for planning, implementing, and
assessing work.
Management system — A structured, nontechnical system describing the policies, objectives,
principles, organizational authority, responsibilities, accountability, and implementation plan of an
organization for conducting work and producing items and services.
Management Systems Review (MSR) — The qualitative assessment of a data collection operation
and/or organization(s) to establish whether the prevailing quality management structure, policies,
practices, and procedures are adequate for ensuring that the type and quality of data needed are obtained.
Matrix spike — A sample prepared by adding a known mass of a target analyte to a specified amount of
matrix sample for which an independent estimate of the target analyte concentration is available. Spiked
samples are used, for example, to determine the effect of the matrix on a method's recovery efficiency.
Mean (arithmetic) — The sum of all the values of a set of measurements divided by the number of
values in the set; a measure of central tendency.
Mean squared error — A statistical term for variance added to the square of the bias.
Measurement and Testing Equipment (M&TE) — Tools, gauges, instruments, sampling devices, or
systems used to calibrate, measure, test, or inspect in order to control or acquire data to verify
conformance to specified requirements.
Memory effects error — The effect that a relatively high concentration sample has on the measurement
of a lower concentration sample of the same analyte when the higher concentration sample precedes the
lower concentration sample in the same analytical instrument.
Method — A body of procedures and techniques for performing an activity (e.g., sampling, chemical
analysis, quantification), systematically presented in the order in which they are to be executed.
EPA QA/G-5 B-6 QA98
-------
Method blank — A blank prepared to represent the sample matrix as closely as possible and analyzed
exactly like the calibration standards, samples, and quality control (QC) samples. Results of method
blanks provide an estimate of the within-batch variability of the blank response and an indication of bias
introduced by the analytical procedure.
Mid-range check — A standard used to establish whether the middle of a measurement method's
calibrated range is still within specifications.
Mixed waste — A hazardous waste material as defined by 40 CFR 261 Resource Conservation and
Recovery Act (RCRA) and mixed with radioactive waste subject to the requirements of the Atomic
Energy Act.
Must — When used in a sentence, a term denoting a requirement that has to be met.
Nonconformance — A deficiency in a characteristic, documentation, or procedure that renders the
quality of an item or activity unacceptable or indeterminate; nonfulfillment of a specified requirement.
Objective evidence — Any documented statement of fact, other information, or record, either
quantitative or qualitative, pertaining to the quality of an item or activity, based on observations,
measurements, or tests that can be verified.
Observation — An assessment conclusion that identifies a condition (either positive or negative) that
does not represent a significant impact on an item or activity. An observation may identify a condition
that has not yet caused a degradation of quality.
Organization — A company, corporation, firm, enterprise, or institution, or part thereof, whether
incorporated or not, public or private, that has its own functions and administration.
Organization structure — The responsibilities, authorities, and relationships, arranged in a pattern,
through which an organization performs its functions.
Outlier — An extreme observation that is shown to have a low probability of belonging to a specified
data population.
Parameter — A quantity, usually unknown, such as a mean or a standard deviation characterizing a
population. Commonly misused for "variable," "characteristic," or "property."
Peer review — A documented critical review of work generally beyond the state of the art or
characterized by the existence of potential uncertainty. Conducted by qualified individuals (or an
organization) who are independent of those who performed the work but collectively equivalent in
technical expertise (i.e., peers) to those who performed the original work. Peer reviews are conducted to
ensure that activities are technically adequate, competently performed, properly documented, and satisfy
established technical and quality requirements. An in-depth assessment of the assumptions, calculations,
extrapolations, alternate interpretations, methodology, acceptance criteria, and conclusions pertaining to
specific work and of the documentation that supports them. Peer reviews provide an evaluation of a
subject where quantitative methods of analysis or measures of success are unavailable or undefined, such
as in research and development.
EPA QA/G-5 B-7 QA98
-------
Performance Evaluation (PE) — A type of audit in which the quantitative data generated in a
measurement system are obtained independently and compared with routinely obtained data to evaluate
the proficiency of an analyst or laboratory.
Pollution prevention — An organized, comprehensive effort to systematically reduce or eliminate
pollutants or contaminants prior to their generation or their release or discharge into the environment.
Precision — A measure of mutual agreement among individual measurements of the same property,
usually under prescribed similar conditions expressed generally in terms of the standard deviation. Refer
to Appendix D, Data Quality Indicators, for a more detailed definition.
Procedure — A specified way to perform an activity.
Process — A set of interrelated resources and activities that transforms inputs into outputs. Examples of
processes include analysis, design, data collection, operation, fabrication, and calculation.
Project — An organized set of activities within a program.
Qualified data — Any data that have been modified or adjusted as part of statistical or mathematical
evaluation, data validation, or data verification operations.
Qualified services — An indication that suppliers providing services have been evaluated and
determined to meet the technical and quality requirements of the client as provided by approved
procurement documents and demonstrated by the supplier to the client's satisfaction.
Quality — The totality of features and characteristics of a product or service that bears on its ability to
meet the stated or implied needs and expectations of the user.
Quality Assurance (QA) — An integrated system of management activities involving planning,
implementation, assessment, reporting, and quality improvement to ensure that a process, item, or service
is of the type and quality needed and expected by the client.
Quality Assurance Program Description/Plan — See quality management plan.
Quality Assurance Project Plan (QAPP) — A formal document describing in comprehensive detail the
necessary quality assurance (QA), quality control (QC), and other technical activities that must be
implemented to ensure that the results of the work performed will satisfy the stated performance criteria.
The QAPP components are divided into four classes: 1) Project Management, 2) Measurement/Data
Acquisition, 3) Assessment/Oversight, and 4) Data Validation and Usability. Requirements for preparing
QAPPs can be found in EPA QA/R-5.
Quality Control (QC) — The overall system of technical activities that measures the attributes and
performance of a process, item, or service against defined standards to verify that they meet the stated
requirements established by the customer; operational techniques and activities that are used to fulfill
requirements for quality. The system of activities and checks used to ensure that measurement systems
are maintained within prescribed limits, providing protection against "out of control" conditions and
ensuring the results are of acceptable quality.
Quality control (QC) sample — An uncontaminated sample matrix spiked with known amounts of
analytes from a source independent of the calibration standards. Generally used to establish intra-
EPA QA/G-5 B-8 QA98
-------
laboratory or analyst-specific precision and bias or to assess the performance of all or a portion of the
measurement system.
Quality improvement — A management program for improving the quality of operations. Such
management programs generally entail a formal mechanism for encouraging worker recommendations
with timely management evaluation and feedback or implementation.
Quality management — That aspect of the overall management system of the organization that
determines and implements the quality policy. Quality management includes strategic planning,
allocation of resources, and other systematic activities (e.g., planning, implementation, and assessment)
pertaining to the quality system.
Quality Management Plan (QMP) — A formal document that describes the quality system in terms of
the organization's structure, the functional responsibilities of management and staff, the lines of
authority, and the required interfaces for those planning, implementing, and assessing all activities
conducted.
Quality system — A structured and documented management system describing the policies, objectives,
principles, organizational authority, responsibilities, accountability, and implementation plan of an
organization for ensuring quality in its work processes, products (items), and services. The quality
system provides the framework for planning, implementing, and assessing work performed by the
organization and for carrying out required quality assurance (QA) and quality control (QC).
Radioactive waste — Waste material containing, or contaminated by, radionuclides, subject to the
requirements of the Atomic Energy Act.
Readiness review — A systematic, documented review of the readiness for the start-up or continued use
of a facility, process, or activity. Readiness reviews are typically conducted before proceeding beyond
project milestones and prior to initiation of a major phase of work.
Record (quality) — A document that furnishes objective evidence of the quality of items or activities
and that has been verified and authenticated as technically complete and correct. Records may include
photographs, drawings, magnetic tape, and other data recording media.
Recovery — The act of determining whether or not the methodology measures all of the analyte
contained in a sample. Refer to Appendix D, Data Quality Indicators, for a more detailed definition.
Remediation — The process of reducing the concentration of a contaminant (or contaminants) in air,
water, or soil media to a level that poses an acceptable risk to human health.
Repeatability — The degree of agreement between independent test results produced by the same
analyst, using the same test method and equipment on random aliquots of the same sample within a short
time period.
Reporting limit — The lowest concentration or amount of the target analyte required to be reported
from a data collection project. Reporting limits are generally greater than detection limits and are usually
not associated with a probability level.
EPA QA/G-5 B-9 QA98
-------
Representativeness — A measure of the degree to which data accurately and precisely represent a
characteristic of a population, a parameter variation at a sampling point, a process condition, or an
environmental condition. See also Appendix D, Data Quality Indicators.
Reproducibility — The precision, usually expressed as variance, that measures the variability among the
results of measurements of the same sample at different laboratories.
Requirement — A formal statement of a need and the expected manner in which it is to be met.
Research (applied) — A process, the objective of which is to gain the knowledge or understanding
necessary for determining the means by which a recognized and specific need may be met.
Research (basic) — A process, the objective of which is to gain fuller knowledge or understanding of
the fundamental aspects of phenomena and of observable facts without specific applications toward
processes or products in mind.
Research development/demonstration — The systematic use of the knowledge and understanding
gained from research and directed toward the production of useful materials, devices, systems, or
methods, including prototypes and processes.
Round-robin study — A method validation study involving a predetermined number of laboratories or
analysts, all analyzing the same sample(s) by the same method. In a round-robin study, all results are
compared and used to develop summary statistics such as interlaboratory precision and method bias or
recovery efficiency.
Ruggedness study — The carefully ordered testing of an analytical method while making slight
variations in test conditions (as might be expected in routine use) to determine how such variations affect
test results. If a variation affects the results significantly, the method restrictions are tightened to
minimize this variability.
Scientific method — The principles and processes regarded as necessary for scientific investigation,
including rules for concept or hypothesis formulation, conduct of experiments, and validation of
hypotheses by analysis of observations.
Self-assessment — The assessments of work conducted by individuals, groups, or organizations directly
responsible for overseeing and/or performing the work.
Sensitivity — the capability of a method or instrument to discriminate between measurement responses
representing different levels of a variable of interest. Refer to Appendix D, Data Quality Indicators, for a
more detailed definition.
Service — The result generated by activities at the interface between the supplier and the customer, and
the supplier internal activities to meet customer needs. Such activities in environmental programs
include design, inspection, laboratory and/or field analysis, repair, and installation.
Shall — A term denoting a requirement that is mandatory whenever the criterion for conformance with
the specification permits no deviation. This term does not prohibit the use of alternative approaches or
methods for implementing the specification so long as the requirement is fulfilled.
EPAQA/G-5 B-10 QA98
-------
Significant condition — Any state, status, incident, or situation of an environmental process or
condition, or environmental technology in which the work being performed will be adversely affected
sufficiently to require corrective action to satisfy quality objectives or specifications and safety
requirements.
Software life cycle — The period of time that starts when a software product is conceived and ends
when the software product is no longer available for routine use. The software life cycle typically
includes a requirement phase, a design phase, an implementation phase, a test phase, an installation and
check-out phase, an operation and maintenance phase, and sometimes a retirement phase.
Source reduction — Any practice that reduces the quantity of hazardous substances, contaminants, or
pollutants.
Span check — A standard used to establish that a measurement method is not deviating from its
calibrated range.
Specification — A document stating requirements and referring to or including drawings or other
relevant documents. Specifications should indicate the means and criteria for determining conformance.
Spike — A substance that is added to an environmental sample to increase the concentration of target
analytes by known amounts; used to assess measurement accuracy (spike recovery). Spike duplicates
are used to assess measurement precision.
Split samples — Two or more representative portions taken from one sample in the field or in the
laboratory and analyzed by different analysts or laboratories. Split samples are quality control (QC)
samples that are used to assess analytical variability and comparability.
Standard deviation — A measure of the dispersion or imprecision of a sample or population distribution
expressed as the positive square root of the variance and has the same unit of measurement as the mean.
Standard Operating Procedure (SOP) — A written document that details the method for an operation,
analysis, or action with thoroughly prescribed techniques and steps and that is officially approved as the
method for performing certain routine or repetitive tasks.
Supplier — Any individual or organization furnishing items or services or performing work according to
a procurement document or a financial assistance agreement. An all-inclusive term used in place of any
of the following: vendor, seller, contractor, subcontractor, fabricator, or consultant.
Surrogate spike or analyte — A pure substance with properties that mimic the analyte of interest. It is
unlikely to be found in environmental samples and is added to them to establish that the analytical
meyhod has been performed properly.
Surveillance (quality) — Continual or frequent monitoring and verification of the status of an entity and
the analysis of records to ensure that specified requirements are being fulfilled.
Technical review — A documented critical review of work that has been performed within the state of
the art. The review is accomplished by one or more qualified reviewers who are independent of those
who performed the work but are collectively equivalent in technical expertise to those who performed the
original work. The review is an in-depth analysis and evaluation of documents, activities, material, data,
EPAQA/G-5 B-ll QA98
-------
or items that require technical verification or validation for applicability, correctness, adequacy,
completeness, and assurance that established requirements have been satisfied.
Technical Systems Audit (TSA) — A thorough, systematic, on-site qualitative audit of facilities,
equipment, personnel, training, procedures, record keeping, data validation, data management, and
reporting aspects of a system.
Traceability — The ability to trace the history, application, or location of an entity by means of recorded
identifications. In a calibration sense, traceability relates measuring equipment to national or
international standards, primary standards, basic physical constants or properties, or reference materials.
In a data collection sense, it relates calculations and data generated throughout the project back to the
requirements for the quality of the project.
Trip blank — A clean sample of a matrix that is taken to the sampling site and transported to the
laboratory for analysis without having been exposed to sampling procedures.
Validation — Confirmation by examination and provision of objective evidence that the particular
requirements for a specific intended use have been fulfilled. In design and development, validation
concerns the process of examining a product or result to determine conformance to user needs. See also
Appendix G, Data Management.
Variance (statistical) — A measure or dispersion of a sample or population distribution.
Verification — Confirmation by examination and provision of objective evidence that specified
requirements have been fulfilled. In design and development, verification concerns the process of
examining a result of a given activity to determine conformance to the stated requirements for that
activity.
EPAQA/G-5 B-12 QA98
-------
APPENDIX C
CHECKLISTS USEFUL IN QUALITY ASSURANCE REVIEW
This appendix contains three checklists:
AC.l Sample Handling, Preparation, and Analysis Checklist
AC.2 QAPP Review Checklist
AC.3 Chain-of-Custody Checklist
These three checklists were developed as tools for quality assurance (QA) managers to screen for
completeness of documentation. This appendix was not intended to be used or adapted for auditing
purposes. The items listed on the checklists are not ranked or identified to indicate which items are
trivial and which are of major importance. When using these checklists, it is extremely important to
ensure that a mechanism be established for assessing and addressing important comments or violations
during the data assessment (e.g., Data Quality Assessment [DQA]) stage.
AC1. SAMPLE HANDLING, PREPARATION, AND ANALYSIS CHECKLIST
This checklist covers most of the appropriate elements performed during the analysis of
environmental samples. Functions not appropriate for a specific analysis should be annotated.
Information on the collection and handling of samples should be completely documented to
allow the details of sample collection and handling to be re-created. All information should be entered
in ink at the time the information was generated in a permanently bound logbook. Errors should not be
erased or crossed-out but corrected by putting a line through the erroneous information and by entering,
initialing, and dating the correct information. Blank spaces should have an obliterating line drawn
through to prevent addition of information. Each set of information should have an identifying printed
name, signature, and initials.
Sample Handling
• Field Logs Documentation of events occurring during field sampling to
identify individual field samples.
• Sample Labels Links individual samples with the field log and the chain-of-
custody record.
• Chain-of-Custody Records Documentation of exchange and transportation of samples from
the field to final analysis.
• Sample Receipt Log Documentation of receipt of the laboratory or organization of
the entire set of individual samples for analysis.
Sample Preparation and Analysis
• Sample Preparation Log Documents the preparation of samples for a specific method.
• Sample Analysis Log Records information on the analysis of analytical results.
• Instrument Run Log Records analyses of calibration standards, field samples, and
quality control (QC) samples.
Chemical Standards
• Chemical Standard Receipt Log Records receipt of analytical standards and chemicals.
• Standards/Reagent Preparation Log Records of the preparation of internal standards, reagents,
spiking solutions, surrogate solutions, and reference materials.
EPAQA/G-5 C-l QA98
-------
AC.l SAMPLE HANDLING, REPORTING, AND ANALYSIS CHECKLIST
Field Loss
ELEMENT
Project name/ID and location
Sampling personnel
Geological observations including map
Atmospheric conditions
Field measurements
Sample dates, times, and locations
Sample identifications present
Sample matrix identified
Sample descriptions (e.g., odors and colors)
Number of samples taken per location
Sampling method/equipment
Description of any QC samples
Any deviations from the sampling plan
Difficulties in sampling or unusual circumstances
COMMENT
Sample Labels
ELEMENT
Sample ID
Date and time of collection
Sampler's signature
Characteristic or parameter investigated
Preservative used
COMMENT
Chain of Custody Records
ELEMENT
Project name/ID and location
Sample custodian signatures verified and on file
Date and time of each transfer
Carrier ID number
Integrity of shipping container and seals verified
Standard Operating Procedures (SOPs) for receipt on file
Samples stored in same area
Holding time protocol verified
SOPs for sample preservation on file
Identification of proposed analytical method verified
Proposed analytical method documentation verified
QA Plan for proposed analytical method on file
COMMENT
EPA QA/G-5
C-2
QA98
-------
AC.l SAMPLE HANDLING, REPORTING, AND ANALYSIS CHECKLIST (CONTINUED)
Sample Receipt Los
ELEMENT
Date and time of receipt
Sample collection date
Client sample ID
Number of samples
Sample matrices
Requested analysis, including method number(s)
Signature of the sample custodian or designee
Sampling kit code (if applicable)
Sampling condition
Chain-of-custody violations and identities
COMMENT
SAMPLE PREPARATION AND ANALYSIS
Sample Preparation Loss
ELEMENT
Parameter/analyte of investigation
Method number
Date and time of preparation
Analyst's initials or signature
Initial sample volume or weight
Final sample volume
Concentration and amount of spiking solutions used
QC samples included with the sample batch
ID for reagents, standards, and spiking solutions used
COMMENT
Sample Analysis Loss
ELEMENT
Parameter analyte of investigation
Method number/reference
Date and time of analysis
Analyst's initials or signature
Laboratory sample ID
Sample aliquot
Dilution factors and final sample volumes (if applicable)
Absorbance values, peak heights, or initial concentrations reading
Final analyte concentration
Calibration data (if applicable)
Correlation coefficient (including parameters)
Calculations of key quantities available
Comments on interferences or unusual observations
QC information, including percent recovery
COMMENT
EPA QA/G-5
C-3
QA98
-------
AC.l SAMPLE HANDLING, REPORTING, AND ANALYSIS CHECKLIST (CONTINUED)
Instrument Run Loss
ELEMENT
Name/type of instrument
Instrument manufacturer and model number
Serial number
Date received and date placed in service
Instrument ID assigned by the laboratory (if used)
Service contract information, including service representative details
Description of each maintenance or repair activity performed
Date and time when of each maintenance or repair activity
Initials of maintenance or repair technicians
COMMENT
CHEMICAL STANDARDS
Chemical/Standard Receipt Loss
ELEMENT
Laboratory control number
Date of receipt
Initials or signature of person receiving chemical
Chemical name and catalog number
Vendor name and log number
Concentration or purity of standard
Expiration date
COMMENT
Standards/Reagent Preparation Los
ELEMENT
Date of preparation
Initials of analyst preparing the standard solution or reagent
Concentration or purity of standard or reagent
Volume or weight of the stock solution or neat materials
Final volume of the solution being prepared
Laboratory ID/control number assigned to the new solution
Name of standard reagent
Standardization of reagents, titrants, etc. (if applicable)
Expiration date
COMMENT
References
Roserance, A. and L. Kibler. 1994. "Generating Defensible Data," Environmental Testing and Analysis. May/June.
Roserance, A. and L. Kibler. 1996. "Documentation and Record Keeping Guidelines." In Proceedings of the 12th
Annual Waste Testing and Quality Assurance Symposium. July.
EPA QA/G-5
C-4
QA98
-------
AC.2 QAPP REVIEW CHECKLIST
ELEMENT
Al. Title and Approval Sheet
Title
Organization's name
Dated signature of project manager
Dated signature of quality assurance officer
Other signatures, as needed
A2. Table of Contents
A3. Distribution List
A4. Project/Task Organization
Identifies key individuals, with their responsibilities (data users, decision-
makers, project QA manager, subcontractors, etc.)
Organization chart shows lines of authority and reporting responsibilities
A5. Problem Definition/Background
Clearly states problem or decision to be resolved
Provides historical and background information
A6. Project/Task Description
Lists measurements to be made
Cites applicable technical, regulatory, or program-specific quality standards,
criteria, or objectives
Notes special personnel or equipment requirements
Provides work schedule
Notes required project and QA records/reports
A7. Quality Objectives and Criteria for Measurement Data
States project objectives and limits, both qualitatively and quantitatively
States and characterizes measurement quality objectives as to applicable
action levels or criteria
A8. Special Training Requirements/Certification Listed
States how provided, documented, and assured
A9. Documentation and Records
Lists information and records to be included in data report (e.g., raw data,
field logs, results of QC checks, problems encountered)
States requested lab turnaround time
Gives retention time and location for records and reports
Bl. Sampling Process Design (Experimental Design)
States the following:
Type and number of samples required
Sampling design and rationale
Sampling locations and frequency
Sample matrices
COMMENTS
EPA QA/G-5
C-5
QA98
-------
AC.2 QAPP REVIEW CHECKLIST (CONTINUED)
ELEMENT
Classification of each measurement parameter as either critical or needed for
information only
Appropriate validation study information, for nonstandard situations
B2. Sampling Methods Requirements
Identifies sample collection procedures and methods
Lists equipment needs
Identifies support facilities
Identifies individuals responsible for corrective action
Describes process for preparation and decontamination of sampling
equipment
Describes selection and preparation of sample containers and sample volumes
Describes preservation methods and maximum holding times
B3. Sample Handling and Custody Requirements
Notes sample handling requirements
Notes chain-of -custody procedures, if required
B4. Analytical Methods Requirements
Identifies analytical methods to be followed (with all options) and required
equipment
Provides validation information for nonstandard methods
Identifies individuals responsible for corrective action
Specifies needed laboratory turnaround time
B5. Quality Control Requirements
Identifies QC procedures and frequency for each sampling, analysis, or
measurement technique, as well as associated acceptance criteria and
corrective action
References procedures used to calculate QC statistics including precision and
bias/accuracy
B6. Instrument/Equipment Testing, Inspection, and Maintenance Requirements
Identifies acceptance testing of sampling and measurement systems
Describes equipment preventive and corrective maintenance
Notes availability and location of spare parts
B7. Instrument Calibration and Frequency
Identifies equipment needing calibration and frequency for such calibration
Notes required calibration standards and/or equipment
Cites calibration records and manner traceable to equipment
B8. Inspection/Acceptance Requirements for Supplies and Consumables
States acceptance criteria for supplies and consumables
Notes responsible individuals
B9. Data Acquisition Requirements for Nondirect Measurements
COMMENTS
EPA QA/G-5
C-6
QA98
-------
AC.2 QAPP REVIEW CHECKLIST (CONTINUED)
ELEMENT
Identifies type of data needed from nonmeasurement sources (e.g., computer
databases and literature files), along with acceptance criteria for their use
Describes any limitations of such data
Documents rationale for original collection of data and its relevance to this
project
BIO. Data Management
Describes standard record-keeping and data storage and retrieval requirements
Checklists or standard forms attached to QAPP
Describes data handling equipment and procedures used to process, compile,
and analyze data (e.g., required computer hardware and software)
Describes process for assuring that applicable Office of Information Resource
Management requirements are satisfied
Cl. Assessments and Response Actions
Lists required number, frequency and type of assessments, with approximate
dates and names of responsible personnel (assessments include but are not
limited to peer reviews, management systems reviews, technical systems
audits, performance evaluations, and audits of data quality)
Identifies individuals responsible for corrective actions
C2. Reports to Management
Identifies frequency and distribution of reports for:
Project status
Results of performance evaluations and audits
Results of periodic data quality assessments
Any significant QA problems
Preparers and recipients of reports
Dl. Data Review, Validation, and Verification
States criteria for accepting, rejecting, or qualifying data
Includes project-specific calculations or algorithms
D2. Validation and Verification Methods
Describes process for data validation and verification
Identifies issue resolution procedure and responsible individuals
Identifies method for conveying these results to data users
D3. Reconciliation with User Requirements
Describes process for reconciling project results with DQOs and reporting
limitations on use of data
COMMENTS
References
Personal Communication, Margo Hunt, EPA Region II, February, 1996.
Personal Communication, Robert Dona, EPA Region VII, November, 1997.
EPA QA/G-5
C-7
QA98
-------
AC.3 CHAIN-OF-CUSTODY CHECKLIST
Item
1 . Is a sample custodian designated?
If yes, name of sample custodian.
2. Are the sample custodian's procedures and responsibilities
documented?
If yes, where are these documented?
3. Are written Standard Operating Procedures (SOPs) developed
for receipt of samples?
If yes, where are the SOPs documented (laboratory manual,
written instructions, etc.)?
4. Is the receipt of chain-of-custody record(s) with samples being
documented?
If yes, where is this documented?
5. Is the nonreceipt of chain-of-custody record(s) with samples
being documented?
If yes, where is this documented?
6. Is the integrity of the shipping container(s) being documented
(custody seal(s) intact, container locked, or sealed properly,
etc.)?
If yes, where is security documented?
7. Is the lack of integrity of the shipping container(s) being
documented (i.e., evidence of tampering, custody seals broken
or damaged, locks unlocked or missing, etc.)?
If yes, where is nonsecurity documented?
8. Is agreement between chain-of-custody records and sample
tags being verified and documented?
If yes, state source of verification and location of
documentation.
9. Are sample tag numbers recorded by the sample custodian?
If yes, where are they recorded?
10. Are written SOPs developed for sample storage?
If yes, where are the SOPs documented (laboratory manual,
written instructions, etc.)?
1 1 . Are samples stored in a secure area?
If yes, where and how are they stored?
12. Is sample identification maintained?
If yes, how?
13. Is sample extract (or inorganics concentrate) identification
maintained?
If yes, how?
14. Are samples that require preservation stored in such a way as
to maintain their preservation?
If yes, how are the samples stored?
Y
N
Comment
EPA QA/G-5
QA98
-------
AC.3 CHAIN-OF-CUSTODY CHECKLIST (CONTINUED)
Item
15. Based upon sample records examined to determine holding
times, are sample holding time limitations being satisfied?
Sample records used to determine holding times:
16. Are written SOPs developed for sampling handling and
tracking?
If yes, where are the SOPs documented (laboratory manual,
written instructions, etc.)?
17. Do laboratory records indicate personnel receiving and
transferring samples in the laboratory?
If yes, what laboratory records document this?
18. Does each instrument used for sample analysis (GC, GC/MS,
AA, etc.) have an instrument log?
If no, which instruments do not?
19. Are analytical methods documented and available to the
analysts?
If yes, where are these documented?
20. Are QA procedures documented and available to the analysts?
If yes, where are these documented?
21. Are written SOPs developed for compiling and maintaining
sample document files?
If yes, where are the SOPs documented (laboratory manual,
written instructions, etc.)?
22. Are sample documents filed by case number?
If no, how are documents filed?
23. Are sample document files inventoried?
24. Are documents in the case files consecutively numbered
according to the file inventories?
25. Are documents in the case files stored in a secure area?
If yes, where and how are they stored?
26. Has the laboratory received any confidential documents?
27. Are confidential documents segregated from other laboratory
documents?
If no, how are they filed?
28. Are confidential documents stored in a secure manner?
If yes, where and how are they stored?
29. Was a debriefing held with laboratory personnel after the audit
was completed?
30. Were any recommendations made to laboratory personnel
during the debriefing?
Y
N
Comment
EPA QA/G-5
C-9
QA98
-------
APPENDIX D
DATA QUALITY INDICATORS
INTRODUCTION
Data Quality Indicators (DQIs) are qualitative and quantitative descriptors used in interpreting
the degree of acceptability or utility of data. The principal DQIs are precision, bias, representativeness,
comparability, and completeness. Secondary DQIs include sensitivity, recovery, memory effects, limit of
quantitation, repeatability, and reproducibility. Establishing acceptance criteria for the DQIs sets
quantitative goals for the quality of data generated in the analytical measurement process. DQIs may be
expressed for entire measurement systems, but it is customary to allow DQIs to be applied only to
laboratory measurement processes. The issues of design and sampling errors, the most influential
components of variability, are discussed separately in EPA QA/G-5S, Guidance on Sampling Designs to
Support QAPPs.
Of the five principal DQIs, precision and bias are the quantitative measures, representativeness
and comparability are qualitative, and completeness is a combination of both quantitative and qualitative
measures.
The five principal DQIs are also referred to by the acronym PARCC, with the "A" in PARCC
referring to accuracy instead of bias. This inconsistency results because some analysts believe accuracy
and bias are synonymous, and PARCC is a more convenient acronym than PBRCC. Accuracy comprises
both random error (precision) and systematic error (bias), and these indicators are discussed separately in
this appendix. DQIs are discussed at length in EPA QA/G-5I, Guidance on Data Quality Indicators.
ADI. PRINCIPAL DQIs: PARCC
AD1.1 PARCC: Precision
Precision is a measure of agreement among replicate measurements of the same property, under
prescribed similar conditions. This agreement is calculated as either the range (R) or as the standard
deviation (s). It may also be expressed as a percentage of the mean of the measurements, such as relative
range (RR) (for duplicates) or relative standard deviation (RSD).
For analytical procedures, precision may be specified as either intralaboratory (within a
laboratory) or interlaboratory (between laboratories) precision. Intralaboratory precision estimates
represent the agreement expected when a single laboratory uses the same method to make repeated
measurements of the same sample. Interlaboratory precision refers to the agreement expected when two
or more laboratories analyze the same or identical samples with the same method. Intralaboratory
precision is more commonly reported; however, where available, both intralaboratory and interlaboratory
precision are listed in the data compilation.
When possible, a sample subdivided in the field and preserved separately is used to assess the
variability of sample handling, preservation, and storage along with the variability of the analysis process.
When collocated samples are collected, processed, and analyzed by the same organization,
intralaboratory precision information on sample acquisition, handling, shipping, storage, preparation, and
analysis is obtained. Both samples can be carried through the steps in the measurement process together
EPAQA/G-5 D-l QA98
-------
to provide an estimate of short-term precision. Likewise, the two samples, if separated and processed at
different times or by different people and/or analyzed using different instruments, provide an estimate of
long-term precision.
AD1.2 PARCC: Bias
Bias is the systematic or persistent distortion of a measurement process that causes errors in one
direction. Bias assessments for environmental measurements are made using personnel, equipment, and
spiking materials or reference materials as independent as possible from those used in the calibration of
the measurement system. When possible, bias assessments should be based on analysis of spiked
samples rather than reference materials so that the effect of the matrix on recovery is incorporated into
the assessment. A documented spiking protocol and consistency in following that protocol are important
to obtaining meaningful data quality estimates. Spikes should be added at different concentration levels
to cover the range of expected sample concentrations. For some measurement systems (e.g., continuous
analyzers used to measure pollutants in ambient air), spiking samples may not be practical, so
assessments should be made using appropriate blind reference materials.
For certain multianalyte methods, bias assessments may be complicated by interferences among
multiple analytes, which prevents all of the analytes from being spiked into a single sample. For such
methods, lower spiking frequencies can be employed for analytes that are seldom or never found. The
use of spiked surrogate compounds for multianalyte gas chromatography/ mass spectrometry (GC/MS)
procedures, while not ideal, may be the best available procedure for assessment of bias.
AD1.3 PARCC: Accuracy
Accuracy is a measure of the closeness of an individual measurement or the average of a number
of measurements to the true value. Accuracy includes a combination of random error (precision) and
systematic error (bias) components that result from sampling and analytical operations.
Accuracy is determined by analyzing a reference material of known pollutant concentration or by
reanalyzing a sample to which a material of known concentration or amount of pollutant has been added.
Accuracy is usually expressed either as a percent recovery (P) or as a percent bias (P - 100).
Determination of accuracy always includes the effects of variability (precision); therefore, accuracy is
used as a combination of bias and precision. The combination is known statistically as mean square
error.
Mean square error (MSB) is the quantitative term for overall quality of individual measurements
or estimators. To be accurate, data must be both precise and unbiased. Using the analogy of archery, to
be accurate, one must have one's arrows land close together and, on average, at the spot where they are
aimed. That is, the arrows must all land near the bull's-eye (see Figure AD. 1).
Mean square error is the sum of the variance plus the square of the bias. (The bias is squared to
eliminate concern over whether the bias is positive or negative.) Frequently, it is impossible to quantify
all of the components of the mean square error—especially the biases—but it is important to attempt to
quantify the magnitude of such potential biases, often by comparison with auxiliary data.
AD1.4 PARCC: Representativeness
Representativeness is a measure of the degree to which data accurately and precisely represent a
characteristic of a population parameter at a sampling point or for a process condition or environmental
EPA QA/G-5 D-2 QA98
-------
(a) High bias + low precision = low accuracy
(b) Low bias + low precision = low accuracy
(c) High bias + high precision = low accuracy
(d) Low bias + high precision = high accuracy
Figure ADI. Measurement Bias and Random Measurement Uncertainties:
Shots at a Target
condition. Representativeness is a qualitative term that should be evaluated to determine whether in situ
and other measurements are made and physical samples collected in such a manner that the resulting data
appropriately reflect the media and phenomenon measured or studied.
AD1.5 PARCC: Comparability
Comparability is the qualitative term that expresses the confidence that two data sets can
contribute to a common analysis and interpolation. Comparability must be carefully evaluated to
establish whether two data sets can be considered equivalent in regard to the measurement of a specific
variable or groups of variables. In a laboratory analysis, the term comparability focuses on method type
comparison, holding times, stability issues, and aspects of overall analytical quantitation.
There are a number of issues that can make two data sets comparable, and the presence of each
of the following items enhances their comparability:
• two data sets should contain the same set of variables of interest;
• units in which these variables were measured should be convertible to a common metric;
• similar analytic procedures and quality assurance should be used to collect data for both
data sets;
• time of measurements of certain characteristics (variables) should be similar for both
data sets;
EPA QA/G-5
D-3
QA98
-------
• measuring devices used for both data sets should have approximately similar detection
levels;
• rules for excluding certain types of observations from both samples should be similar;
• samples within data sets should be selected in a similar manner;
• sampling frames from which the samples were selected should be similar; and
• number of observations in both data sets should be of the same order or magnitude.
These characteristics vary in importance depending on the final use of the data. The closer two
data sets are with regard to these characteristics, the more appropriate it will be to compare them. Large
differences between characteristics may be of only minor importance, depending on the decision that is
to be made from the data.
Comparability is very important when conducting meta-analysis, which combines the results of
numerous studies to identify commonalities that are then hypothesized to hold over a range of
experimental conditions. Meta-analysis can be very misleading if the studies being evaluated are not
truly comparable. Without proper consideration of comparability, the findings of the meta-analysis may
be due to an artifact of methodological differences among the studies rather than due to differences in
experimentally controlled conditions. The use of expert opinion to classify the importance of differences
in characteristics among data sets is invaluable.
AD1.6 PARCC: Completeness
Completeness is a measure of the amount of valid data obtained from a measurement system,
expressed as a percentage of the number of valid measurements that should have been collected (i.e.,
measurements that were planned to be collected).
Completeness is not intended to be a measure of representativeness; that is, it does not describe
how closely the measured results reflect the actual concentration or distribution of the pollutant in the
media sampled. A project could produce 100% data completeness (i.e., all samples planned were
actually collected and found to be valid), but the results may not be representative of the pollutant
concentration actually present.
Alternatively, there could be only 70% data completeness (30% lost or found invalid), but, due to
the nature of the sample design, the results could still be representative of the target population and yield
valid estimates. Lack of completeness is a vital concern with stratified sampling. Substantial incomplete
sampling of one or more strata can seriously compromise the validity of conclusions from the study. In
other situations (for example, simple random sampling of a relatively homogeneous medium), lack of
completeness results only in a loss of statistical power. The degree to which lack of completeness affects
the outcome of the study is a function of many variables ranging from deficiencies in the number of field
samples acquired to failure to analyze as many replications as deemed necessary by the QAPP and
DQOs. The intensity of effect due to incompleteness of data is sometimes best expressed as a qualitative
measure and not just as a quantitative percentage.
Completeness can have an effect on the DQO parameters. Lack of completeness may require
reconsideration of the limits for the false negative and positive error rates because insufficient
completeness will decrease the power of the statistical test.
The following four situations demonstrate the importance of considering the planned use of the
data when determining the completeness of a study. The purpose of the study is to determine whether the
average concentration of dioxin in surface soil is no more than 1.0 ppb. The DQOs specified that the
EPA QA/G-5 D-4 QA98
-------
sample average should estimate the true average concentration to within ±0.30 ppb with 95 %
confidence. The resulting sampling design called for 30 samples to be drawn according to a simple
random sampling scheme. The results were as follows:
Study result Completeness Outcome
1. 1.5 ppb ± 0.28 ppb 97% satisfies DQOs and study purpose
2. 500 ppb ± 0.28 ppb 87% satisfies DQOs and study purpose
3. 1.5 ppb ±0.60 ppb 93% doesn't satisfy either
4. 500 ppb ±0.60 ppb 67% fails DQOs but meets study purpose
For all but the third situation, the data that were collected completely achieved their purpose,
meeting data quality requirements originally set out, or providing a conclusive answer to the study
question. The degree of incompleteness did not affect some situations (situations 2 and 4) but may have
been a prime cause for situation 3 to fail the DQO requirements. Expert opinion would then be required
to ascertain if further samples for situation 3 would be necessary in order to meet the established DQOs.
Several factors may result in lack of completeness: (1) the DQOs may have been based on poor
assumptions, (2) the survey design may have been poorly implemented, or (3) the design may have
proven impossible to carry out given resource limitations. Lack of completeness should always be
investigated, and the lessons learned from conducting the study should be incorporated into the planning
of future studies.
AD2. OTHER DATA QUALITY INDICATORS
AD2.1 Sensitivity
Sensitivity is the capability of a method or instrument to discriminate between measurement
responses representing different levels of a variable of interest. Sensitivity is determined from the value
of the standard deviation at the concentration level of interest. It represents the minimum difference in
concentration that can be distinguished between two samples with a high degree of confidence.
AD2.2 Recovery
Recovery is an indicator of bias in a measurement. This is best evaluated by the measurement of
reference materials or other samples of known composition. In the absence of reference materials, spikes
or surrogates may be added to the sample matrix. The recovery is often stated as the percentage
measured with respect to what was added. Complete recovery (100%) is the ultimate goal. At a
minimum, recoveries should be constant and should not differ significantly from an acceptable value.
This means that control charts or some other means should be used for verification. Significantly low
recoveries should be pointed out, and any corrections made for recovery should be stated explicitly.
AD2.3 Memory Effects
A memory effect occurs when a relatively high-concentration sample influences the measurement
of a lower concentration sample of the same analyte when the higher concentration sample precedes the
lower concentration sample in the same analytical instrument. This represents a fault in an analytical
measurement system that reduces accuracy.
EPA QA/G-5 D-5 QA98
-------
AD2.4 Limit of Quantitation
The limit of quantitation is the minimum concentration of an analyte or category of analytes in a
specific matrix that can be identified and quantified above the method detection limit and within
specified limits of precision and bias during routine analytical operating conditions.
AD2.5 Repeatability
Repeatability is the degree of agreement between independent test results produced by the same
analyst using the same test method and equipment on random aliquots of the same sample within a short
time period.
AD2.6 Reproducibility
Reproducibility is the precision that measures the variability among the results of measurements
of the same sample at different laboratories. It is usually expressed as a variance and low values of
variance indicate a high degree of reproducibility.
AD2.7 DQIs and the QAPP
At a minimum, the following DQIs should be addressed in the QAPP: accuracy and/or bias,
precision, completeness, comparability, and representativeness. Accuracy (or bias), precision,
completeness, and comparability should be addressed in Section A7.3, Specifying Measurement
Performance Criteria. Refer to that section of the G-5 text for a discussion of the information to present
and a suggested format. Representativeness should be discussed in Sections B4.2 (Subsampling) and Bl
(Sampling Design).
Table ADI. Principal Types of Error
Types of Error
Sources of Error
Random Error
(precision; "P" in PARCC)
Natural variability in the population from which the sample is
taken.
Measurement system variability, introduced at each step of
sample handling and measurement processes.
Systematic Error
(accuracy/bias; "A" in PARCC)
Interferences that are present in sample matrix.
Loss (or addition) of contaminants during sample collection and
handling.
Loss (or addition) of contaminants during sample preparation
and analysis.
Calibration error or drift in the response function estimated by
the calibration curve.
EPA QA/G-5
D-6
QA98
-------
Lack of representativeness
("R" in PARCC)
Sample is not representative of the population, which often
occurs in judgmental sampling because not all the units of the
population have equal or known selection probabilities.
Sample collection method does not extract the material from its
natural setting in a way that accurately captures the desired
qualities to be measured.
Subsample (taken from a sample for chemical analysis) is not
representative of the sample, which occurs because the sample
is not homogeneous and the subsample is taken from the most
readily available portion of the sample. Consequently, other
parts of the sample had less chance of being selected for analysis.
Lack of comparability
("C" in PARCC)
Failure to use similar data collection methods, analytical
procedures, and QA protocols.
Failure to measure the same parameters over different data sets.
Lack of completeness
("C" in PARCC)
Lack of completeness sometimes caused by loss of a sample,
loss of data, or inability to collect the planned number of
samples.
Incompleteness also occurs when data are discarded because
they are of unknown or unacceptable quality.
AD2.8 References
American Society for Quality Control. 1996. Definitions of Environmental Quality Assurance Terms.
Milwaukee, WI: ASQC Press.
Gilbert, R.O. 1987. Statistical Methods for Environmental Pollution Monitoring. New York: Van
Nostrand.
Ott, W.R. 1985. Environmental Statistics and Data Analysis. Boca Raton, FL: Lewis Publishers Inc.
Taylor, J.K. and T.W. Stanley, eds. 1985. Quality Assurance for Environmental Measurements.
Philadelphia, PA: American Society for Testing and Materials.
Taylor, J.K. 1987. Quality Assurance of Chemical Measurements. Chelsea, MI: Lewis Publishers Inc.
U.S. Environmental Protection Agency. 1984. Chapter 5. Calculation of Precision, Bias, and Method
Detection Limit for Chemical and Physical Measurements.
U.S. Environmental Protection Agency. 1994. AEERL Quality Assurance Procedures Manual for
Contractors and Financial Assistance Recipients.
EPA QA/G-5
D-7
QA98
-------
U.S. Environmental Protection Agency. 1994. EPA Requirements for Quality Management Plans. EPA
QA/R-2, Draft Interim Final. August.
Youden, W.J. 1967. Journal of the Association of Official Analytical Chemists. Vol. 50. p. 1007.
EPA QA/G-5 D-8 QA98
-------
APPENDIX E
QUALITY CONTROL TERMS
AE1. QUALITY CONTROL OPERATIONS
Quality control (QC) plays an increasingly important role in environmental studies, especially
when those studies are conducted to decide how to address an environmental problem. To minimize the
chance of making an incorrect decision, data of adequate quality must be collected. The purpose of QC
is to ensure that measurement and other data-producing systems operate within defined performance
limits as specified in planning. QC programs can both lower the chances of making an incorrect decision
and help the data user understand the level of uncertainty that surrounds the decision. QC operations
help identify where error is occurring, what the magnitude of that error is, and how that error might
impact the decision-making process. This appendix provides a brief overview of this complex topic. It
surveys the different types of QC samples that can be applied to environmental studies and evaluates
how they are currently deployed as specified by EPA methods and regulations.
AE1.1 General Objectives
The most important QC questions a project manager should consider are:
• What are the QC requirements for the methods to be used in the project?
• What types of problems in environmental measurement systems do these requirements
enable the Agency to detect?
Addressing these questions should provide the manager with the background needed for defining
a uniform, minimum set of QC requirements for any environmental data collection activity.
Understanding existing QC requirements for environmental data generation activities provides a
framework for considering what set of QC requirements should be considered "core" requirements
irrespective of the end use of the data.
While it is difficult to define a standard of data quality regardless of its intended use, core QC
requirements can be established that will enable one to provide data of known quality in accordance with
the Agency's QA program. This program requires that all environmental data collection efforts gather
information on bias, variability, and sample contamination. These error types are incurred throughout
the data generation process, including all sampling and analytical activities (i.e., sample collection,
handling, transport, and preparation; sample analysis; and subsampling). The principal issue centers on
what level of detail in the error structure should QC operations be capable of revealing, given that it is
impractical to explore every known potential source of error.
AE1.2 Background
Many of the essential elements of a Quality Assurance Project Plan (QAPP) apply directly to
sampling and analytical activities and include: Quality assurance (QA) objectives for measurement data
specified in terms of Data Quality Indicators (precision, accuracy, bias, representativeness and
comparability); sampling procedures; sample custody; calibration procedures and frequency; analytical
procedures; internal QC checks and frequency; performance and system audits and frequency; and
specific routine procedures that should be used to assess both data precision and the completeness of the
specific measurement parameters involved.
EPAQA/G-5 E-l QA98
-------
AE1.3 Definitions and Terminology
In order to ensure that managers have a uniform perspective of QC requirements, it is necessary
to discuss some basic terminology and definitions. QC and QA, total study error and its components,
types of QC operations, and Good Laboratory Practices (GLPs) will be discussed here. Specific
definitions of these terms and others are provided in Appendix B, Glossary of Quality Assurance and
Monitoring Terms, while Table E. 1 summarizes the results of a study on how these terms are defined and
used in EPA and non-EPA literature. Five commonly available sources are discussed in Table E.I:
Appendix B in EPA QA/G-5; American Society for Quality Control (1996); van Ee, Blume, and Starks
(1989); Taylor (1987); and Keith (1988).
AE1.1.3 Quality Control vs. Quality Assurance
All of the cited literature provides somewhat similar definitions for both QA and QC. QC
activities are designed to control the quality of a product so that it meets the user's needs. QA includes
QC as one of the activities needed to ensure that the product meets defined standards of quality.
These two terms have been defined in slightly different ways by other authors, but all are in
agreement that QC is a component of QA. Many authors define QC as "those laboratory operations
whose objective is to ensure that the data generated by the laboratory are of known accuracy to some
stated, quantitative degree of probability" (Dux 1986). The objective of QC is not to eliminate or
minimize errors but to measure or estimate what they are in the system as it exists. The same authors
then define QA as the ability to prove that the quality of the data is as reported. QA relies heavily on
documentation, including documentation of implemented QC procedures, accountability, traceability,
and precautions to protect raw data.
AE1.3.2 PC Samples
Table E.I offers a broad survey of commonly used QC terms, including the definitions of QC
sample types that span the measurement process. The authors cited in Table E.I define different sample
types in varied ways; however, the definitions are not contradictory.
AE1.3.3 Good Laboratory Practices
The Food and Drug Administration (FDA) promulgated the first version of the Good Laboratory
Practices (GLPs) in 1978. The EPA enacted similar guidance requirements in 1983 for Resource
Conservation Recovery Act (RCRA) and Federal Insecticide, Fungicide, and Rodenticide Act (FIFRA)
compliance. The FIFRA GLPs were revised in 1988. Though much of the content relates to laboratory
animal science, many requirements are relevant to the analytical chemist. The GLP standards for FIFRA
(40 Code of Federal Regulations [CFR] Part 160) and the Toxic Substances Control Act (TSCA) (40
CFR 792) are similar (Dux 1986). Selected topics of FIFRA subparts A through K appear below.
Subpart A General Provisions.
Subpart B Organization and Personnel. Includes QA unit.
Subpart C Facilities. Includes: facilities for handling test, control, and reference
substances; laboratory operations areas; and specimen and data storage
facilities.
Subpart D Equipment. Includes: maintenance and calibration of equipment.
Subpart E Testing Facilities Operation. Includes: standard operation procedures
(SOPs); reagents and solutions.
EPA QA/G-5 E-2 QA98
-------
Subpart F Test, Control, and Reference Substances. Includes: characterization and
handling; mixtures of substances with carriers.
Subpart G Protocol for and Conduct of a Study.
Subpart H Reserved.
Subpart I Reserved.
Subpart J Records and Reports. Includes: reporting of study results; storage and
retrieval of records and data; and retention of records.
GLPs are defined similarly by the Agency and by Taylor (1987) as an acceptable way to perform
some basic laboratory operation or activity that is known or believed to influence the quality of its
outputs.
AE2. QUALITY CONTROL REQUIREMENTS IN EXISTING PROGRAMS
To identify QC requirements for this section, standard EPA method references, such as SW-846,
and the CFR were reviewed together with information on non-EPA methods identified through a
computerized literature search. Within the EPA literature, some of the major programs were reviewed,
including the Drinking Water, Air and the Contract Laboratory Program (CLP). Different types of
methods, such as gas chromatography (GC), atomic absorption (AA), and inductively coupled plasma
(ICP), and different media were included in this process, but it was not intended to be exhaustive.
AE2.1 Summary of QC Requirements by Program and Method
Table AE.2 presents the frequency of QC requirements for different selected programs and Table
AE.3 presents information for methods. In cases where different programs use dissimilar terms for
similar QC samples, the table uses the term from the program or method.
AE2.2 Comparing Various QC Requirements
AE2.2.1 QC Requirements for Program Offices
Table AE.2 shows that QC requirements vary considerably and are established by the Program
Office responsible for the data collection activity. Ambient air monitoring methods (Office of Air
Quality Planning and Standards [OAQPS]) require periodic analysis of standards for assessment of
accuracy (combination of imprecision and bias) for manual methods, and analysis of collocated samples
for the assessment of imprecision. Prevention of Significant Deterioration (PSD) and State and Local
Air Monitoring Stations (SLAMS) make a unique distinction in defining two terms: precision checks and
accuracy checks. These checks entail essentially the same QC requirements, but they are performed by
different parties; the accuracy check is essentially an external audit, while the precision check is an
internal QC operation. It should be noted that some water methods require additional QC operations for
GC/MS than for other methods (e.g., tuning, isotopic dilution).
In general, the wet chemistry analytical methods (the toxicity characteristic leaching procedure
[TCLP] being a preparation method) require periodic analysis of blanks and calibration standards. Most
require analysis of matrix spikes and replicate samples, the exceptions being the 200 Series (no spikes or
replicates) and the 600 series (GC/MS require no replicates).
While the QC operations for the PSD and SLAMS methods appear minimal, these monitoring
programs require active QA programs that include procedures for zero/span checks. (The zero check
may be considered a blank sample, while the span check may be considered a calibration check sample.)
EPA QA/G-5 E-3 QA98
-------
The Program Office Quality Assurance Officer (QAO) or representative should have details on
specific QC requirements.
AE2.2.2 Organized by Type of Potential Problem
Table AE.3 lists the QC requirements of various EPA measurement methods and presents the
required frequencies for different kinds of QC operations. The table is divided into four sections, one for
each general type of QC problem:
• Contamination: This occurs when the analyte of interest or an interferant is introduced
through any of a number of sources, including contaminated sample equipment,
containers, and reagents. The contaminant can be the analyte of interest or another
chemical that interferes with the measurement of the analyte or causes loss or generation
of the analyte.
• Calibration Drift: This is a nonrandom change in the measurement system over time,
such as a (systematic) change in instrument response over time. It is often detectable by
periodic remeasurement of a standard.
• Bias: This can be regarded as a systematic error caused by contamination and calibration
drift and also by numerous other causes, such as extraction efficiency by the solvent,
matrix effect, and losses during shipping and handling.
• Imprecision: This is a random error, observed as different results from repeated
measurements of the same or identical samples.
For internal consistency, the names of QC operations used in Table AE.3 are those given in the specific
reference methods.
AE2.3 Using QC Data
The relationships between monitoring design specifications and the final use of the data
described above incorporate two significant assumptions: (1) laboratory measurements, through the use
of internal standards or other adjustments that are integral to the analytical protocol, are unbiased; and
(2) the variance structure of these measurements does not change over time. Bias enters as a
consequence of under-recovery of the contaminant of interest during the sample preparation stage of the
analytical protocol and as undetected drift in calibration parameters. The variance of measurements also
may change over time due to unintentional changes in the way samples are prepared and/or to
degradation of the electromechanical instrumentation used to analyze the samples. QC samples are
intended to detect bias and variability changes and should be specified in the QAPP.
QC samples that address bias are calibration check standards (CCSs) and spiked samples
(performance check samples [PCSs]). CCSs typically consist of reagent water samples spiked with the
concentrations used to develop the calibration curve. Measurements obtained by analyzing these
samples, which reflect the existing calibration relationship, are compared to the actual concentrations
that were added to the samples. If the difference exceeds a prespecified calibration test limit, the
measurement system is considered "out of control" and the calibration function is re-estimated.
EPA QA/G-5 E-4 QA98
-------
Detecting a change in calibration parameters is a statistical decision problem in detecting a
material change in the calibration function. In many QC programs, CCSs typically are analyzed at the
beginning and end of each shift and after any other QC sample has detected a failure. By definition,
significant change in the calibration parameters leads to biased measurements of field samples. This can
be detected through use of statistical tests.
A spiked sample typically has the same matrix characteristics found in field samples, but it has
been spiked (as soon after the sample is taken as is practical) with a known concentration of the target
contaminant. Because spiked samples are intended to detect recovery changes, they are processed
through the same preparation steps as field samples, and the spiked sample measurement is used to form
an estimate of recovery. Significant changes lead to the conclusion that measurements of field samples
are biased.
The second of the two monitoring program assumptions identified at the beginning of this
section is a constant variance structure for monitoring data over time. Measurements from split (or
duplicate) field samples provide a check on this variance assumption. Changes in measurement
variability, for example a uniform increase in the standard deviation or changes in the way variability
depends on concentration, have a direct impact on subsequent investigations.
AE2.4 Classifying QC Samples: Control versus Assessment
QC programs are designed foremost to detect a measurement process entering an "out of control"
state so that corrective measures can be initiated. QC samples used in this way are performing a control
function. Each of the three types of QC samples previously discussed, CCSs, spiked samples, and split
(or duplicate) samples, may be used for control. In addition, spiked samples and split samples also may
be used to estimate measurement bias and variability. QC samples that also can be used to estimate
measurement parameters are sometimes referred to as quality assessment samples. These should not be
confused with the much larger Data Quality Assessment Process; see also EPA QA/G-9, Guidance for
Data Quality Assessment.
QC samples that are used for control must be analyzed and reported soon after they are obtained
if their intervention potential is to be realized. Among the three types of QC samples discussed above,
CCSs are the most likely to be effective for control purposes. Spiked samples and split samples
generally are not effective for control purposes, in part because they are analyzed "blind" and therefore
the results cannot be reviewed immediately. Spiked samples and split samples, however, may be used
for control if consecutive batches of similar field samples are being analyzed.
Spiked samples and split samples can be effective quality assessment samples. For example,
spiked samples may be used to indicate the presence of bias. The estimate is applied as a bias correcting
adjustment to individual measurements or to batches of measurements before the measurements are used
in compliance tests. The adjustment improves the test by eliminating bias. However, the variance of the
adjusted estimate used in the test is greater than the variance of the unadjusted estimate.
Split (or duplicate) samples also can be used as quality assessment samples, but their application
in the monitoring program is not as constructive as the application of spiked samples. Split samples lead
to an estimate of the measurement replication component of variability. (The variance of a measurement
has, at a minimum, a sampling component and a measurement replication component, which is
sometimes referred to as measurement error. If the sampling design involves stratification, the variance
will include additional components.) If the estimate based on split samples suggests a measurement
replication standard deviation larger than the value assumed in establishing the original sampling design,
a loss in efficiency will result.
EPA QA/G-5 E-5 QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Blank sample
A clean sample or a sample of matrix
processed so as to measure artifacts in the
measurement (sampling and analysis)
process.
Blanks provide a measure of various
cross-contamination sources, background
levels in reagents, decontamination
efficiency, and other potential error that
can be introduced from sources other
than the sample. A rinsate blank
(decontamination sample) measures any
chemical that may have been on the
sampling and sample preparation tools
after the decontamination process is
completed.
The measured value obtained when a
specified component of a sample is not
present during measurement. Measured
value/signal for the component is
believed to be due to artifacts; it should
be deducted from a measured value to
give a net value due to the component
contained in a sample. The blank
measurement must be made to make the
correction process valid.
Samples expected to have negligible or
unmeasurable amounts of the substance
of interest. They are necessary for
determining some of the uncertainty due
to random errors. Three kinds required
for proper quality assurance: equipment
blanks, field blanks, and sampling
blanks.
Blind sample
A subsample submitted for analysis with
a composition and identity known to the
submitter but unknown to the analyst.
Used to test analyst or laboratory
proficiency in execution of the
measurement process.
Single-Blind Samples: Field Rinsate
Blanks, Preparation Rinsate Blank, Trip
Blank
A sample submitted for analysis whose
composition is known to the submitter
but unknown to the analyst. One way to
test the proficiency of a measurement
process.
Calibration
standard
A substance or reference material used to
calibrate an instrument, (calibration
check standard, reference standard,
quality control check sample)
In physical calibration, an artifact
measured periodically, the results of
which typically are plotted on a control
chart to evaluate the measurement
process.
Or quality control calibration standard
(CCS). In most laboratory procedures, a
solution containing the analyte of
interest at a low but measurable
concentration. Standard deviation of the
CCSs is a measure of instrument
precision unless the CCS is analyzed as
a sample, in which case it is a measure
of method precision.
Checks sample
Example: ICP Interference Check Sample -
Part A contains potential interfering
analytes. Part B contains both the
analytes of interest and the target
analytes. Part A and B are analyzed
separately to determine the potential for
interferences.
Check standard
A substance or reference material
obtained from a source independent from
the source of the calibration standard;
used to prepare check samples, (control
standard")
Laboratory control standards are
certified standards, generally supplied by
an outside source. They are used to
ensure that the accuracy of the analysis
is in control.
EPA QA/G-5
E-6
QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Double blind
samples
Samples that can not be distinguished
from routine samples by analytical
laboratory. Examples: Field Evaluation
Samples, Low Level Field Evaluation
Samples, External Laboratory Evaluation
Samples, Low Level External Laboratory
Evaluation Samples, Field Matrix Spike,
Field Duplicate, Field Split
A sample known by the submitter but
submitted to an analyst so that neither its
composition nor its identification as a
check sample are known to the analyst.
Duplicate
measurement
A second measurement made on the same
(or identical) sample of material to assist
in the evaluation of measurement
Duplicate
sample
Two samples taken from and
representative of the same population and
carried through all steps of the sampling
and analytical procedures in an identical
manner. Used to assess variance of the
total method including sampling and
analysis.
Field duplicate - an additional sample
taken near the routine field sample to
determine total within-batch
measurement variability.
Analytical laboratory duplicate - a
subsample of a routine sample analyzed
by the same method. Used to determine
method precision. It is non-blind so it
can only be used by the analyst in
internal control, not an unbiased estimate
of analytical precision.
A second sample randomly selected from
a population of interest to assist in the
evaluation of sample variance.
Error
The difference between a computed,
observed, or measured value or condition
and the true, specified, or theoretical
value or condition.
Difference between the true or expected
value and the measured value of a
quantity or parameter.
Field blank
Used to estimate incidental or accidental
contamination of a sample during the
collection procedure. One should be
allowed per sampling team per day per
collection apparatus. Examples include
matched-matrix blank, sampling media
or trip blank, equipment blank.
EPA QA/G-5
E-7
QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Good
Laboratory
Practices
(GLPs)
Either general guidelines or formal
regulations for performing basic
laboratory operations or activities that are
known or believed to influence the
quality and integrity of the results.
An acceptable way to perform some basic
operation or activity in a laboratory that
is known or believed to influence the
quality of its outputs. GLPs ordinarily
are essentially independent of the
measurement techniques used.
Instrument
blank
Also called system blank. Used to
establish baseline response of an
analytical system in the absence of a
sample. Not a simulated sample but a
measure of instrument or system
background response.
Method blank
One of the most important in any
process. DDI water processed through
analytical procedure as a normal sample.
After use to determine the lower limit of
detection, a reagent blank is analyzed for
each 20 samples and whenever a new
batch of reagents is used.
Non-blind
sample
QC samples with a concentration and
origin known to the analytical laboratory.
Examples: Laboratory Control Sample,
Pre-digest Spike, Post-digest Spike,
Analytical Laboratory Duplicate, Initial
Calibration Verification and Continuing
Calibration Verification Solutions, Initial
Calibration Blank and Continuing
Calibration Blank Solution, CRDL
Standard for ICP and AA, Linear Range
Verification Check Standard, ICP
Interference Check Sample.
Performance
Evaluation
(PE)
A type of audit in which the quantitative
data generated in a measurement system
are obtained independently and compared
with routinely obtained data to evaluate
the proficiency of an analyst or
laboratory.
rPefmed in EPA OA/G-5. App. B1
EPA QA/G-5
E-8
QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Quality
assessment
Assessment is the evaluation of
environmental data to determine if they
meet the quality criteria required for a
specific application.
The overall system of activities that
provides an objective measure of the
quality of data produced.
The overall system of activities whose
purpose is to provide assurance that the
quality control activities are done
effectively. It involves a continuing
evaluation of performance of the
production system and the quality of the
products produced.
Quality
assessment
sample (QAS)
Those samples that allow statements to
be made concerning the quality of the
measurement system. Allow assessment
and control of data quality to assure that
it meets original objectives. Three
categories: double-blind, single-blind,
and non-blind.
Quality
assurance
(QA)
An integrated system of activities
involving planning, quality control,
quality assessment, reporting and quality
improvement to ensure that a product or
service meets defined standards of
quality with a stated level of confidence.
A system of activities whose purpose is
to provide to the producer or user of a
product or service the assurance that it
meets defined standards of quality. It
consists of two separate, but related
activities, quality control and quality
assessment.
Same as van Ee.
Quality
control (QC)
The overall system of technical activities
whose purpose is to measure and control
the quality of a product or service so that
it meets the needs of users. The aim is to
provide quality that is satisfactory,
adequate, dependable, and economical.
The overall system of activities whose
purpose is to control the quality of the
measurement data so that they meet the
needs of the user.
The overall system of activities whose
purpose is to control the quality of a
product or service so that it meets the
needs of users. The aim is to provide
quality that is satisfactory, adequate
dependable, and economic.
Quality
control sample
An uncontaminated sample matrix spiked
with known amounts of analytes from a
source independent from the calibration
standards. Generally used to establish
intralaboratory or analyst specific
precision and bias or to assess
performance of all or part of the
measurement system. (Laboratory control
sample)
rPefmed in EPA OA/G-5. App. B1
A sample of well-characterized soil,
whose analyte concentrations are known
to the laboratory. Used for internal
laboratory control. Also called QC audit
sample.
A material of known composition that is
analyzed concurrently with test samples
to evaluate a measurement process.
Used in quality control procedures to
determine whether or not the analytical
procedures are in control.
EPA QA/G-5
E-9
QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Reagent blank
A sample consisting of reagent(s),
without the target analyte or sample
matrix, introduced into analytical
procedure at the appropriate point and
carried through all subsequent steps to
determine the contribution of the reagents
in the absence of matrix and the involved
analytical steps to error in the observed
value (analytical blank, laboratory blank).
(Defined in EPA QA/G-5, App. B)
Also called method blank. Used to
detect and quantitate contamination
introduced during sample preparation
and analysis. Contains all reagents used
in sample preparation and analysis and is
carried through the complete analytical
procedure.
Reference
material
A material or substance, one or more
properties of which are sufficiently well
established to be used for the calibration
of an apparatus, the assessment of a
measurement method, or for the
assignment of values to materials.
Sample
preparation
blank
Required when methods like stirring,
mixing, blending, or subsampling are
used to prepare a sample prior to
analysis. One should be prepared per 20
samples processed.
Sampling
equipment
blank
Used to determine types of contaminants
introduced through contact with
sampling equipment; also to verify the
effectiveness of cleaning procedures.
Prepared by collecting water or solvents
used to rinse sampling equipment.
Solvent blank
Used to detect and quantitate solvent
impurities; the calibration standard
corresponds to zero analyte
concentration. Consists only of solvent
used to dilute the sample.
EPA QA/G-5
E-10
QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Spiked sample
A sample prepared by adding a known
mass of target analyte to a specified
amount of matrix sample for which an
independent estimate of target analyte
concentration is available. Spiked
samples are used, for example, to
determine the effect of the matrix on a
method's recovery efficiency (matrix
spike).
A sample prepared by adding a known
amount of reference chemical to one of a
pair of split samples. Comparing the
results of the analysis of a spiked
member to that of the non-spiked
member of the split measures spike
recovery and provides a measure of the
analytical bias.
Field matrix spike - a routine sample
spiked with the contaminant of interest in
the field.
Matrix control or field spike -for sample
matrices where a complex mixture (e.g.
sediments, sludges) may interfere with
analysis, a field spike may be required to
estimate the magnitude of those
interferences. Losses from transport,
storage treatment, and analysis can be
assessed by adding a known amount of
the analyte of interest to the sample in
the field.
Split sample
Two or more representative portions
taken from a sample or subsample and
analyzed by different analysts or
laboratories. Split samples are used to
replicate the measurement of the
variable(s) of interest.
Samples can provide: a measure of
within-sample variability; spiking
materials to test recovery; and a measure
of analytical and extraction errors.
Where the sample is split determines the
components of variance that are
measured. Field split - a sample is
homogenized and spilt into two samples
of theoretically equal concentration at the
sampling site. Indicate within-batch
measurement error. Also called
replicates.
A replicate portion or subsample of a
total sample obtained in such a manner
that is not believed to differ significantly
from other portions of the same sample.
Total
measurement
error
The sum of all the errors that occur from
the taking of the sample through the
reporting of results; the difference
between the reported result and the true
value of the population that was to have
been sampled.
Transport
blank
Used to estimate sample contamination
from the container and preservative
during transport and storage of the
sample. One should be allowed per day
per type of sample.
EPA QA/G-5
E-ll
QA98
-------
Table AE1. Comparison of QC Terms
Terms
ASQC, Definitions of Environmental
Quality Assurance Terms
or
EPA QA/G-5 Appendix B
van Ee, Blume, and Starks
A Rationale for the Assessment of Errors
in the Sampling of Soils
John Keenan Taylor
Quality Assurance of Chemical
Measurements
Lawrence H. Keith, ed.
Principles of Environmental Sampling
Trip blank
A clean sample of matrix that is carried
to the sampling site and transported to
the laboratory for analysis without having
been exposed to sampling procedures.
(Defined in EPA QA/G-5, App. B)
Used when volatile organics are sampled.
Consists of actual sample containers
filled with ASTM Type II water, kept
with routine samples throughout
sampling event, packaged for shipment
with routine samples and sent with each
shipping container to the laboratory.
Used to determine the presence or
absence of contamination during
shipment.
A type of field blank also called
sampling media blank. To detect
contamination associated with the
sampling media such as filters, traps, and
sample bottles. Consists of sampling
media used for sample collection.
EPA QA/G-5
E-12
QA98
-------
Table AE2. QC Requirements for Programs
Potential
Problems:
QC
Samples to
Identify
Potential
Problems:
CLP
Organics:
1991
Statement of
Work,
Exhibit E
Contamination
Blanks
Volatiles
Semi-
volatiles
Pesticides/
Aroclor
A method
blank once
every 12
hours.
A method
blank with
every batch.
Instrument
blank at start
of analyses
and every 12
hours.
Method
blank with
each case,
14 days, or
batch.
Sulfur
blanks are
sometimes
required.
Calibration
Drift
Calibration
Check
Samples
Continuing
calibration
standard every
12 hours. BFB
analysis once
every 12 hours.
DFTPP analysis
once every 12
hours.
Continuing
calibration
standard every
12 hours.
Performance
evaluation
mixture to
bracket 12-hour
periods.
Bias
Spike
Matrix spike
with every
case, batch, 20
samples, or 14
days.
Matrix spike
with every
case, batch, 20
samples, or 14
days.
Matrix spike
with every 20
samples.
Standard
3 system monitoring
compounds added to
every sample.
8 surrogates spiked into
each sample.
2 surrogates added to
each sample.
Imprecision
Replicate
Matrix spike
duplicate with
every case,
batch, 20
samples, or 14
days.
Matrix spike
duplicate with
every case,
batch, 20
samples, or 14
days.
Matrix spike
duplicate with
every 20
samples.
Collocated
Other
EPA QA/G-5
E-13
QA98
-------
Table AE2. QC Requirements for Programs
Potential
Problems:
QC
Samples to
Identify
Potential
Problems:
CLP
Inorganics:
1991
Statement
of Work,
Exhibit E
PSD
40CFR
Part 58
Appendix B
Contamination
Blanks
Initial calibration blank; then
continuing calibration blank
10% or every 2 hours.
Preparation blank with every
batch.
Calibration
Drift
Calibration
Check
Samples
Initial
calibration
verification
standard; then
continuing
calibration
verification
10% or every 2
hours.
Bias
Spike
1 spike for
every batch.
Method of
standard
additions for
AA if spikes
indicate
problem.
Standard
Interference check
sample for ICP
2 x /8 hours.
Laboratory control
sample with each
batch.
For SO2, NO2, O3, and
CO, response check I/
sampling quarter. For
TSP and lead, sample
flow check
I/sampling quarter.
For lead, check with
audit strips I/quarter.
Imprecision
Replicate
1 duplicate/
batch. For AA,
duplicate
injections.
Collocated
For TSP and
lead, collocated
sample I/week
or every 3rd
day for
continuous
sampling.
Other
For S02, N02,
O3, and CO,
precision
check once
every 2
weeks.
EPA QA/G-5
E-14
-------
Table AE2. QC Requirements for Programs
Potential
Problems:
QC
Samples to
Identify
Potential
Problems:
SLAMS
40CFR
Part 58
Appendix A
A Rationale
for the
Assessment
of Errors in
the
Sampling of
Soils, by
van Ee,
Blume, and
Starks
Contamination
Blanks
Preparation rinsate blanks and
field rinsate blanks discussed,
but no frequency given.
Calibration
Drift
Calibration
Check
Samples
Bias
Spike
Standard
For automated SO2,
N02, 03, and CO
response check for at
least 1 analyzer (25%
of all) each quarter.
For manual SO2 and
NO2, analyze audit
standard solution each
day samples are
analyzed (at least
2x/quarter). For TSP,
PM10, and lead,
sample flow rate
check at least 1
analyzer/quarter (25%
of all analyzers). For
lead, check with audit
strips I/quarter.
At least 2 1 pairs of
field evaluation
samples. At least 20
pairs of external
laboratory evaluation
samples if estimating
components of
variance is important.
Imprecision
Replicate
At least 20
pairs or 10
triples of field
duplicates. At
least 20 pairs of
preparation
splits if
estimating
variance is
important.
Collocated
For manual
methods,
including lead,
collocated
sample I/week.
Other
For automated
S02, N02, 03,
and CO,
precision
check once
every 2
weeks.
EPA QA/G-5
E-15
QA98
-------
Table AE3. QC Requirements for Methods
Potential
Problems:
QC Samples to
Identify Potential
Problems:
SW-846 Method
7000 (Proposed
Update I)
Atomic Absorption
SW-846 Method
8000 (Proposed
Update I) Gas
Chromatography
503.1 Volatile
Aromatic and
Unsaturated Organic
Compounds in
Water by Purge and
Trap GC (from
PB89-220461)
200 Atomic
Absorption Methods
(from EPA-600-4-
79-020)
624-Purgeables
40 CFR Part 136,
Appendix A
Contamination
Blanks
Reagent blank as
part of daily
calibration.
Reagent blank
before sample
analysis and for
each batch of up
to 20 samples.
Laboratory
reagent blank with
each batch. Field
reagent blank with
each set of field
samples.
Reagent blank at
least daily.
Reagent water
blank daily.
Calibration Drift
Calibration
Check Samples
Mid-range standard
analyzed every 10
samples.
A daily calibration
sample analyzed.
Calibration verified
daily with 1 or more
calibration standards.
Daily checks at least
with reagent blank and
1 standard.
Verification with an
additional standard
every 20 samples.
Analyze BFB every
day analyses are
performed.
Bias
Spike
1 spiked matrix
sample analyzed
every 20 samples or
analytical batch.
Method of standard
additions required
for difficult matrices.
1 matrix spike for
each batch of up to
20 samples.
Laboratory-fortified
blank with each
batch or 20 samples.
Spike a minimum of
5% of samples.
Standard
QC check sample
required, but
frequency not
specified.
QC sample
analyzed at least
quarterly.
Analysis of an
unknown
performance
sample at least
once per year.
Surrogate
standards used
with all samples.
Analyze QC
check samples as
5% of analyses.
Imprecision
Replicate
1 replicate sample
every 20 samples or
analytical batch; 1
spiked replicate
sample for each
matrix type.
1 replicate or matrix
spike replicate for
each analytical batch
of up to 20 samples.
Samples collected in
duplicate.
Laboratory- fortified
blanks analyzed in
duplicate at least
quarterly.
Collocated
Other
EPA QA/G-5
E-16
QA98
-------
Table AE3. QC Requirements for Methods
Potential
Problems:
QC Samples to
Identify Potential
Problems:
1624-Volatile
Organic Compounds
by Isotope Dilution
GC/MS
40 CFR Part 136,
Appendix A
TCLP-Fed. Reg.,
Vol 55, No. 126
Friday, June 29,
1990
SW-846 Method
6010 (Proposed
Update I)
Inductively Coupled
Plasma Atomic
Emission
Spectroscopy
Contamination
Blanks
Blanks analyzed
initially and with
each sample lot.
1 blank for every
20 extractions.
At least 1 reagent
blank with every
sample batch.
Calibration Drift
Calibration
Check Samples
Aqueous standard with
BFB, internal
standards, and
pollutants is analyzed
daily. A standard used
to compare syringe
injection with purge
and trap.
Verify calibration
every 10 samples and
at the end of the
analytical run with a
blank and standard.
Bias
Spike
All samples spiked
with labeled
compounds.
1 matrix spike for
each waste type and
for each batch.
Spiked replicate
samples analyzed at
a frequency of 20%.
Standard
An interference
check sample
analyzed at the
beginning and end
of each run or 8-
hour shift.
Imprecision
Replicate
8 aliquots of the
aqueous
performance
standard analyzed
initially.
1 replicate with
every batch or 20
samples. Also spiked
replicates analyzed,
as discussed under
"Spikes."
Collocated
Other
EPA QA/G-5
E-17
QA98
-------
AE2.4 References
American Society for Quality Control. Environmental Restoration Committee. Terms and Definitions Task Group.
1996. Definitions of Environmental Quality Assurance Terms. Milwaukee, WI: Quality Press.
American Society for Quality Control. Chemical Process Industries Division. 1987. Quality Assurance for the
Chemical Process Industries, a Manual of Good Practices. Washington, DC.
Dux, James P. 1986. Handbook of Quality Assurance for the Analytical Chemistry Laboratory.
Federal Insecticide, Fungicide andRodenticide Act (FIFRA). 1989. Good Laboratory Practices Standards. Final
Rule. Federal Register, vol. 54, no. 158, August.
Good Laboratory Practices: An Agrochemical Perspective. 1987. Division of Agrochemicals, 194th Meeting of the
American Chemical Society.
Grant, E.L. and R.S. Leavenworth. 1988. Statistical Quality Control, 6th Edition. New York: McGraw-Hill.
Griffith, Gary K. 1996. Statistical Process Control Methods for Long and Short Runs, 2nd Edition. Milwaukee,
WI: ASCQ Quality Press.
Hayes, Glenn E. and Harry G. Romig. 1988. Modern Quality Control. Revised Edition. Encino, CA.
Juran, J.M. and Frank M. Gryna. 1993. Quality Planning and Analysis, 3rd Edition. New York: McGraw-Hill.
Keith, Lawrence H., ed. 1988. Principles of Environmental Sampling. Washington, DC: American Chemical
Society Press.
Taylor, John Keenan. 1987. Quality Assurance of Chemical Measurements. Chelsea, MI: Lewis Publishers, Inc.
van Ee, J. Jeffrey, Louis J. Blume, and Thomas H. Starks. 1989. A Rationale for the Assessment of Errors in
Sampling of Soils. EPA/600/X-89/203.
EPAQA/G-5 E-18 QA98
-------
APPENDIX F
SOFTWARE FOR THE DEVELOPMENT AND PREPARATION OF A QUALITY
ASSURANCE PROJECT PLAN
This appendix contains three sections:
AF1. an overview of the potential need for software in QAPP preparation,
AF2. information on existing software, and
AF3. information on software availability and sources.
The information presented in this appendix on various types of software that may be useful in
constructing a QAPP is only a subset of what is available to the QA Manager. Mention of certain
products or software does not constitute endorsement, only that some potentially useful material can be
obtained from those products.
AF1. OVERVIEW OF POTENTIAL NEED FOR SOFTWARE IN QAPP PREPARATION
The general structure of a QAPP can be adapted easily for an organization's needs by automating
some of the components of the QAPP. Several commercial and governmental organizations have
produced software to facilitate this automation. The software needs are categorized under the four
classes of QAPP elements. Within each category is an explanation of the general functions of the
software that could prove useful in preparing, reviewing, or implementing a QAPP. In addition, the
QAPP elements to which the software applies are listed.
AF1.1 Class A: Project Management
This type of software can be used to produce planning documentation and preparation of the
QAPP document. In addition, this type of software can be used to produce other project documentation
such as Standard Operating Procedures (SOPs), Quality Management Plans (QMPs), and Data Quality
Objectives (DQOs) reports.
GENERAL SOFTWARE FUNCTIONS
Provides the user guidance on what to address in each QAPP element and
serves as a template for the production of the QAPP document.
Generates flowcharts to assist in preparing project organization charts and in
illustrating processes that occur in the project, such as sample collection and
analysis or data management.
Identifies training or certification required for personnel in given program
areas.
Provides applicable regulatory standards (e.g., action or clean-up levels) for the
various program areas (e.g., air, water, and solid waste).
Provides guidance on implementing the DQO Process.
QAPP
ELEMENTS
All elements
A4, BIO
A8
A6
A5, A6, A7
EPA QA/G-5
F-l
QA98
-------
AF1.2 Class B: Measurement and Data Acquisition
This type of software can be used to assist in the design of a sampling plan. In addition, this
software can provide information on analytical methods and sample collection and handling.
GENERAL SOFTWARE FUNCTIONS
Assists in the development of sampling designs that will meet specified
DQOs. The software should handle a variety of general design types with
and without compositing, such as simple random sampling, grid sampling,
and stratified sampling.
Provides information on analytical procedures and sampling methods for
various contaminants and media. This software provides QC data for the
analytical method (method detection limit [MDL], precision, and bias),
references to standard methods, and SOPs (where calibration and
maintenance information could be found).
Assists in tracking samples and assisting with documenting sample handling
and custody.
Integrates QC design and sampling design to meet DQOs and facilitate Data
Quality Assessment (DQA).
QAPP ELEMENTS
Bl
B2, B4, B5, B6, B7
B3
Bl, B5, BIO
AF1.3 Class C: Assessment and Oversight
This software can assist in assessment and oversight activities.
GENERAL SOFTWARE FUNCTIONS
Produces checklists, checklist templates, or logic diagrams (such as problem
diagnostics) for Technical Systems Audits (TSAs), Management Systems
Reviews (MSRs), and Audits of Data Quality (ADQs).
Perform DQA and facilitates corrective actions during the implementation
phase as preliminary or field screening data become available.
QAPP ELEMENTS
Cl
Cl, C2
AF1.4 Class D: Data Validation and Usability
This software assists in validating data and assessing its usability.
GENERAL SOFTWARE FUNCTIONS
Assists in performing data validation and usability.
Assists in performing data quality assessment.
QAPP ELEMENTS
D2
D3
EPA QA/G-5
F-2
QA98
-------
AF2. EXISTING SOFTWARE
This information is summarized as a list of identified software; a more detailed description of
each item is found in Section AF3. A variety of commercial software packages are available to assist in
statistical analysis, laboratory QC, and related activities, but this appendix focuses on software used
specifically by those preparing, implementing, and reviewing QAPPs. See Table AF.l for a summary of
the software described below.
AF2.1 Template Software
Several applications have been implemented in word-processing software that provide guidance
on how to complete each QAPP element and a template for the discussion portion. Four examples of
these applications are:
• Quality Integrated Work Plan Template (QIWP) (Section AF3, No. 2)
• QAPP Template (Section AF3, No. 3)
• Region 5 QAPP Template (Section AF3, No. 4)
A more sophisticated application, Quality Assurance Sampling Plan for Environmental Response
(QASPER), was identified that combines a template with links to a variety of lists that provide the user
response options (Section AF3, No. 1).
AF2.2 Flowcharting Software
Various flowcharting software is commercially available. One example found in QA/QC
literature is allCLEAR III (Section AF3, No. 5). Other more sophisticated packages link the flowchart
diagrams to active databases or simulation modeling capabilities.
AF2.3 Regulatory Standards Software
This software provides regulatory limits under the various statutes for a wide variety of
contaminants:
• Environmental Monitoring Methods Index (EMMI) (Section AF3, No. 6)
• Clean-Up Criteria for Contaminated Soil and Groundwater (an example of a commercially
available product) (Section AF3, No. 8)
AF2.4 Sampling Design Software
A variety of software has been developed to assist in the creation of sampling designs:
• Decision Error Feasibility Trials (DEFT) (Section AF3, No. 9)
• GeoEAS (Section AF3, No. 10)
• ELIPGRID-PC (Section AF3, No. 11)
• DQOPro (Section AF3, No. 12)
In addition, there are many statistical packages that support sampling design.
EPA QA/G-5 F-3 QA98
-------
AF2.5 Analytical Methods Software
This software provides information on method detection limits (MDLs) and method summaries
for a wide variety of analytical methods:
• EMMI (Section AF3, No. 6)
• EPA's Sampling and Analysis Methods Database (Section AF3, No. 7)
AF2.6 Data Validation Software
The Research Data Management and Quality Control System (RDMQ) (Section AF3, No. 13) is
a data management system that allows for the verification, flagging, and interpretation of data.
AF2.7 Data Quality Assessment Software
Several software packages have been developed to perform data quality assessment tasks.
Examples of this software include:
• DataQUEST (Section AF3, No. 14)
• ASSESS (Section AF3, No. 15)
• RRELSTAT (Section AF3, No. 16)
Note that most commercially available statistical packages (not listed above) perform a variety of
DQA tasks.
AF2.8 QAPP Review
QATRACK (Section AF3, No. 17) is used to track QAPPs undergoing the review process.
laoie AVI. aonware Avaiiaoie to ivieet i^vrr ueveio
SOFTWARE NEED
PROJECT MANAGEMENT
Template guidance
Flowcharting
Regulatory standards
MEASUREMENT AND DATA
ACQUISITION
Sample design
Analytical and sampling procedures
Integrating QC design and sampling design
to meet DQOs and facilitate DQA.
QAPP
ELEMENTS
All elements
A4, BIO
A6
Bl
B2, B4, B5, B6,
B7
B1,B5, BIO
)ineiu i>eeas
EXISTING SOFTWARE
QASPER, QWIP, QAPP Template
allCLEAR III
EMMI, Clean-Up Criteria for Contaminated
Soil and Groundwater
DEFT, GeoEAS, ELIPGRID-PC, DRUMs,
DQOPro, miscellaneous statistical packages
EMMI, EPA's Sampling and Analysis Database
DQOPro
EPA QA/G-5
F-4
QA98
-------
SOFTWARE NEED
ASSESSMENT AND OVERSIGHT
Data Quality Assessment
DATA VALIDATION AND USABILITY
Data validation
Data Quality Assessment
QAPP
ELEMENTS
Cl, C2
D2
D3
EXISTING SOFTWARE
DataQUEST, ASSESS, RRELSTAT
RDMQ
DataQUEST, ASSESS, RRELSTAT,
miscellaneous statistical packages
AF3. SOFTWARE AVAILABILITY AND SOURCES
The wide variety of existing software has potential to meet the needs identified for preparing
QAPPs. As illustrated in Table AF.l, at least one example of a software tool was identified that could
potentially be applied to aspects of QAPP preparation or implementation for all but three of the need
areas. The capabilities of the existing software should match the QAPP needs, as most of the software
was developed for use with a QAPP or for environmental data collection or analysis. Software not
designed for these uses could be modified or used to form the basis of an application that is more tailored
to QAPP preparation or implementation.
AF3.1 Quality Assurance Sampling Plan for Environmental Response (QASPER), Version 4.0
QASPER allows the creation and editing of a Quality Assurance sampling plan for
environmental response. The plan template consists of 11 sections: (1) title page, (2) site background,
(3) data use objectives, (4) sampling design, (5) sampling and analysis, (6) SOPs, (7) QA requirements,
(8) data validation, (9) deliverables, (10) project organization and responsibilities, and (11) attachments.
While preparing the plan, the user may enter the required information or select from the options provided
in a variety of "picklists." The picklists cover topics such as holding times, methods, preservatives, and
sampling approaches. The user may add or delete options from the picklists. QASPER also provides
various utility functions such as backing up, restoring, exporting, and importing a plan. Output may be
directed to a file or a printer. Contact: EPA, (732) 906-6921, Quality Assurance Sampling Plan for
Environmental Response (QASPER Version 4.0 User's Guide; latest version is QASPER Version 4.1,
January 1995.
AF3.2 Quality Integrated Work Plan (QIWP) Template for R&D and Monitoring Projects
The QIWP template is a tool designed to assist with planning, managing, and implementing a
specific monitoring or R&D project. The QIWP template is formatted with comment boxes that provide
guidance on the information to provide in each section. When activated, the text in the comment boxes
will appear on screen; however, they will not appear in a printout. An asterisk indicates where the user
should begin entering the discussion for each section. The QIWP document control format is already set
up in the template header. When a particular element is considered not applicable, the rationale for that
decision must be stated in response to that element. Once the user is satisfied with the information
entered under all elements of the template, the resulting printout is the combined project work plan and
QA plan. In addition, a printout of the QIWP template, prior to entering project related information, can
be used as a checklist for planning and review purposes. Other software packages available are the
QWIP Template for Model Development Projects and the QWIP Template for Model Application
Projects. Contact: EPA, (919) 541-3779 and North American Research Strategy for Tropospheric Ozone
(NARSTO) homepage.
EPA QA/G-5
F-5
QA98
-------
AF3.3 Region 2 QAPP Template
This package contains an annotated template containing instructions for completing each section
of the QAPP. The users are also instructed where to insert their discussions within the template. After
completing the QAPP, the italicized instructions are not printed, leaving only the preparer's discussion.
In addition, a table of contents is automatically generated. The template describes the information that
should be provided under the main topics of project management, measurement/data acquisition, data,
assessment/oversight, and references. The project management section covers the introduction, goals of
the project, organization of the project participants and of QA, and DQOs. The measurement/data
acquisition section discusses the topics to address to describe the statistical research design and
sampling. This section also covers the elements related to sample analysis: description of the instrument,
calibration, QC, consumables, and preventative maintenance. The data section provides for a discussion
of the data management procedures. The assessment/oversight section covers audits and QA reports.
The next section is a list of references. Finally, six tables are provided as examples for displaying
information on the following topics: (1) measurement quality criteria; (2) sample collection, handling,
and preservation; (3) instrument data and interferences; (4) instrument calibration, (5) QC checks; and
(6) preventive maintenance. Contact: EPA, (401) 782-3163, or (503) 754-4670.
AF3.4 Region 5 QAPP Template
This software consists of two model documents (one for Superfund sites and one for RCRA
sites) that describe the preparation of a QAPP in a series of elements. Each element contains two types
of information: (1) content requirements that are presented as smaller text and (2) structural guidance
that is presented as larger text and headed by the appropriate section number. This information is
intended to show to the QAPP preparer the requirements that must be described in each element and the
level of detail that is typically needed to gain Region 5 approval. Example text is provided that should be
deleted and replaced with the specific site information.
A TSCA Model Plan template is also available that attempts to be a comprehensive guide to all
the data gathering activities for Fiscal Year 94 Title IV grantees. In this template, headers are provided
in "background" format, and text that may apply to specific situations is in an italic font. Open spaces
indicate where the preparer's input is required. Contact: EPA, (312) 886-6234.
AF3.5 allCLEARIII
This software enables the creation of simple process diagrams, organizational charts, or decision
trees. It also creates diagrams from text outlines, spreadsheets, and database information. Contact:
American Society for Quality Control Quality Press, Publications Catalogue, (800) 248-1946.
AF3.6 Environmental Monitoring Methods Index (EMMI)
This software consists of an analytical methods database containing more than 4,200 analytes,
3,400 analytical and biological methods, and 47 regulatory and nonregulatory lists. EMMI cross-
references analytes, methods, and lists and has information about related laws, organizations, and other
chemical databases. This information does not include measurement method performance such as
precision and bias. Contact: DynCorp Environmental Technical Support, (703) 519-1222.
EPA QA/G-5 F-6 QA98
-------
AF3.7 EPA's Sampling and Analysis Methods Database, 2nd Edition
This software has a menu-driven program allowing the user to search a database of 178 EPA-
approved analytical methods with more than 1,300 method and analyte summaries. The database covers
industrial chemicals, pesticides, herbicides, dioxins, and PCBs and focuses on water, soil matrices, and
quality parameters. The software generates reports that are stand-alone documents that can be browsed,
printed, or copied to files. Each report contains information for initial method selection such as
applicable matrices, analytical interferences and elimination recommendations, sampling and
preservation requirements, MDLs, and precision, accuracy, and applicable concentration ranges.
Contact: Radian Corporation, (512) 454-4797.
AF3.8 Clean-Up Criteria for Contaminated Soil and Groundwater, 2nd edition
This software consists of a one-volume document and diskette summarizing cleanup criteria
developed by EPA, all 50 State regulatory agencies, and select countries outside the United States.
Contact: ASTM Publications Catalogue, (610) 832-9585, http://www.astm.org.
AF3.9 Decision Error Feasibility Trials (DEFT)
This package allows quick generation of cost information about several simple sampling designs
based on the DQO constraints. The DQO constraints can be evaluated to determine their appropriateness
and feasibility before the sampling and analysis design is finalized.
This software supports the Guidance for the Data Quality Objectives Process, EPA QA/G-4, that
provides general guidance to organizations on developing data quality criteria and performance
specifications for decision-making. The Data Quality Objectives Decision Error Feasibility Trials
(DEFT) User's Guide, contains detailed instructions on how to use DEFT software and provides
background information on the sampling designs that the software uses. Contact: EPA, (202) 564-6830.
AF3.10 GeoEAS
Geostatistical Environmental Assessment Software (GeoEAS) is a collection of interactive
software tools for performing two-dimensional geostatistical analyses of spatially distributed data.
Programs are provided for data file management, data transformations, univariate statistics, variogram
analysis, cross-validation, kriging, contour mapping, post plots, and line/scatter plots. Users may alter
parameters and re-calculate results or reproduce graphs, providing a "what if analysis capability.
This software and a user's guide can be downloaded through the Office of Research and
Development (ORD) World Wide Web site at http://www.epa.gov/ORD or
http://www.epa.gov/ORD/nerl.htm. Contact: GEO-EAS 1.2.1 User's Guide, EPA/600/8-91/008, April,
1991, EPA, (702)798-2248.
AF3.11 ELIPGRID-PC
ELIPGRID-PC calculates the probabilities related to hitting a single hot spot. The user has the
following options: (1) calculating the probability of detecting a hot spot of given size and shape when
using a specified grid, (2) calculating the grid size required to find a hot spot of given size and shape with
specified confidence, (3) calculating the size of the smallest hot spot likely to be hit with a specified
sampling grid, (4) calculating a grid size based on fixed sampling cost, and (5) displaying a graph of the
EPA QA/G-5 F-7 QA98
-------
probability of hitting a hot spot versus sampling costs. Contact: ELIPGEJD-PC: UPGRADED
VERSION, Oak Ridge National Laboratory/TM-13103, (970) 248-6259.
AF3.12 DQOPro
This software consists of a series of three computer programs that calculate the number of
samples needed to meet specific DQOs. DQOPro provides answers for three objectives: (1) determining
the rate at which an event occurs, (2) determining an estimate of an average within a tolerable error, and
(3) determining the sampling grid necessary to detect "hot-spots." Contact: Radian International, (512)
454.4797.
AF3.13 Research Data Management and Quality Control System (RDMQ)
This software is a data management system that allows for the verification, flagging, and
interpretation of data. PvDMQ is a menu-driven application with facilities for loading data, applying QC
checks, viewing and changing data, producing tabular and graphical reports, and exporting data in ASCII
files. PvDMQ provides a shell environment that allows the user to perform these tasks in a structured
manner. Contact: Environment Canada, (416) 639-5722, or EPA, (919) 541-2408.
AF3.14 DataQUEST
This tool is designed to provide a quick and easy way for managers and analysts to perform
baseline Data Quality Assessment. The goal of the system is to allow those not familiar with standard
statistical packages to review data and verify assumptions that are important in implementing the DQA
Process. This software supports the Guidance for Data Quality Assessment, EPA QA/G-9, that
demonstrates the use of the DQA Process in evaluating environmental data sets. Contact: EPA, (202)
564-6830.
AF3.15 ASSESS I.Ola
This software tool was designed to calculate variances for quality assessment samples in a
measurement process. The software performs the following functions: (1) transforming the entire data
set, (2) producing scatter plots of the data, (3) displaying error bar graphs that demonstrate the variance,
and (4) generating reports of the results and header information. Contact: EPA, (702) 798-2367.
AF3.16 QATRACK
This Microsoft Access software provides a database that tracks QAPPs requiring approval. Data
are entered into QATRACK during the assistance agreement start-up stage, as soon as the QA manager
reviews and signs the agreement. Users can edit the data, query the database to perform data reviews,
and archive files once the QAPP is approved. Contact: EPA, (919) 541-2408.
EPA QA/G-5 F-8 QA98
-------
APPENDIX G
ISSUES IN DATA MANAGEMENT
AG1. INTRODUCTION
EPA QA/G-5 provides guidance on many different operations that involve generating, collecting,
manipulating, and interpreting environmental data. These activities include field sampling, sample
handling and storage, laboratory analysis, modeling, data storage and retrieval, and Data Quality
Assessment. All these activities generate data or require data to be manipulated in some way, usually
with the aid of a computerized data management tool such as a database, spreadsheet, computer model, or
statistical program.
This appendix expands the guidance currently provided in EPA QA/G-5, Section BIO, Data
Management. Guidance is provided on Quality Assurance (QA) considerations and planning for the
development, implementation, and testing of computer-based tools that perform the data management
aspects of the overall environmental project described in the Quality Assurance Project Plan (QAPP).
These data management aspects include data storage, data acquisition, data transformations, data
reduction, modeling, and other data management tasks associated with environmental data collection
projects. This guidance can be used for applications developed in-house or for those developed using
commercial software. It can be used for systems of different sizes, from individual spreadsheet
applications to large integrated systems. The amount of planning and documentation involved are
tailored according to the use of the data and the size and complexity of the application.
This appendix incorporates into EPA QA/G-5 the QA elements of guidance from the EPA Office
of Information Resources Management (OIRM) and applicable industry standards, such as those of the
Institute of Electronic and Electrical Engineers, relating to development of information and data
management systems. Because data and information system development projects differ widely in many
different respects, this appendix does not attempt to address the low-level details of planning,
implementation and assessment nor does it provide step-by-step procedures to follow when developing a
data management system. These details are left to other EPA guidance documents (See Section AG2.4),
national consensus standards, and the best judgement of the personnel on each project.
AG2. REGULATORY AND POLICY FRAMEWORK
This section provides a brief overview of the legislation, policies, standards and guidelines most
applicable to the development of EPA data management and information systems. Sections AG2.1 and
AG2.2 of this overview are intended to provide the QAPP preparer (specifically the preparer of the data
management section) with a general understanding of the relevant agency-level policies, Sections AG2.3
and AG2.4 provide a reference for the major guidance documents containing more specific and detailed
information on development of data management systems.
AG2.1 Legislation
The following is a summary of the major legislative policies that pertain to information
technology and the development of data management systems. The two most relevant pieces of legislation
are:
(1) the Paperwork Reduction Act (PRA) of 1980 (P.L. 96-511) as amended in 1986 (P.L. 99-500)
and 1995 (P.L. 104-13), and
EPA QA/G-5 G-l QA98
-------
(2) the Clinger-Cohen Act of 1996 (P.L.-104-208). (Note that the Clinger-Cohen Act is the
amended title for the Information Technology Management Reform Act and the Federal
Acquisition Reform Act of 1996 (P.L. 104-106)).
The overall purpose of the PRA is to reduce paperwork and enhance the economy and efficiency
of the government and private sector by improving Federal information policy development and
implementation. The PRA establishes a broad mandate for executive agencies to perform their
information activities in an efficient, effective, and economical manner. The 1995 amendments
established several broad objectives for improving the management of Federal information resources.
These objectives include maximizing the utility of information, improving the quality and use of
information to strengthen decision making, and establishing uniform resource management policies.
The Clinger-Cohen Act (CCA) sets forth requirements for the Office of Management and Budget
(OMB) and the individual executive agencies. OMB responsibilities include promoting and improving
the acquisition, use, and disposal of information technology by the Federal Government to improve the
productivity, efficiency, and effectiveness of Federal programs. In addition, the CCA requires each
agency to design and implement a process for maximizing the value and assessing and managing the risks
of information technology acquisitions. The CCA also requires each agency to utilize the same
performance- and results-based management practices as encouraged by OMB.
AG2.2 Policy Circulars and Executive Orders
Circular A-130 implements OMB authority under the PRA and sets forth the policy that applies
to the information activities of all the executive agencies. The policies include requirements for
information management planning as well as information systems and information technology
management. Part of the information management policy is that agencies, when creating or collecting
data, need to plan from the outset how to perform the following data management functions: (1) data
processing and transmission, (2) data end use and integrity protection, (3) data access, (4) data
dissemination, (5) data storage and retrieval, and (6) data disposal. In addition, these planning activities
need to be documented. The information systems and information technology management policies
describe an information system life cycle that is defined as the phases through which an information
system passes. These phases are typically characterized as initiation, development, operation, and
termination. However, no specific number of phases is set, and the life cycle management techniques
that agencies use may vary depending on the complexity and risk inherent in the project. In addition, the
division between the phases of the system life cycle may not be distinct.
Current implementation of the CCA comes through Executive Order 13011, which outlines the
executive agencies. The agencies are to strengthen the quality of decisions about the use of information
resources to meet mission needs and establish mission-based performance measures for information
systems. In addition, to establish agency-wide and project-level management structures and processes
responsible and accountable for managing, selecting, controlling, and evaluating investments in
information systems.
G2.3 Federal Information Processing Standards
The National Institute of Standards and Technology (NIST) develops standards for Federal
computer systems. NIST issues these standards and guidelines as Federal Information Processing
Standards (FIPS) for government-wide use. NIST develops FIPS when there are compelling Federal
government requirements (such as for security and interoperability) and there are no acceptable industry
standards or solutions. FIPS publications include standards, guidelines, and program information
EPA QA/G-5 G-2 QA98
-------
documents in the following seven subject areas: (1) general publications, (2) hardware standards and
guidelines, (3) software standards and guidelines, (4) data standards and guidelines, (5) computer
security standards and guidelines, (6) operations standards and guidelines, and (7) telecommunications
standards. Additional information about FIPS, including ordering information and a list and description
of the individual documents, is available online using the World Wide Web (WWW) at the following
Uniform Resource Locator (URL) address: http//www.nist.gov/itl/div879/pubs/.
AG2.4 EPA Guidance
EPA's Office of Information Resources Management (OIRM), which has the primary functional
responsibility for Information Resources Management (IRM) policy development and overall
management of EPA's IRM program, has published several IRM guidance documents. The Information
Resources Management Policy Manual 2100 establishes a policy framework for managing information
resources in the Agency. The document is intended to provide a structure for the implementation of
legislation concerning the management of Federal information resources such as the PRA. Also, the
manual establishes the authorities and responsibilities under which the OIRM will function. The Policy
Manual consists of twenty chapters that cover subjects such as software management, information
security, system life cycle management, and information and data management. The Policy Manual can
be obtained online using the WWW at the following URL address: http://www.epa.gov/irmpoli8/.
The System Design and Development Guidance document provides a framework that Agency
managers can use to document a problem and justify the need for an information-system-based solution.
The document also provides guidance for identifying solutions to specified problems and for information
system development. The guidance consists of three volumes (A, B, and C). Volume A provides a
method for documenting the need for an information system and developing an initial system concept
that describes the inputs, outputs, and processes of the proposed system. Volume B provides guidance
for developing design options that satisfy the initial system concept developed in Volume A. Volume B
also gives guidance for selecting the most cost-effective solution. Volume C describes the system-design
and development process (and the required associated documentation) and outlines a software
management plan that is used to ensure the quality of EPA software design, development,
implementation, and maintenance efforts. This document can be obtained online using the WWW at the
following URL address: http://www.epa.gov/irmpoli8/.
Additional EPA guidance documents pertaining to information system development, operations,
and maintenance are listed in Section G4, References. Up-to-date OIRM documents can be obtained
online using the WWW at the following URL address: http://www.epa.gov/irmpoli8/.
Another source of guidance is EPA Quality Assurance Division's (QAD) Development
Management System Template. The template includes a description of the roles of management in
planning for the development of data management systems. The responsible project officer or
contracting officer representative outlines a management scheme based upon the planning and
documentation activities that satisfy OIRM policy or an organization's Quality Management Plan. The
project manager works with the quality assurance manager to identify the tasks, work products, and
management procedures for the project.
AG3. QA PLANNING FOR INFORMATION SYSTEMS
Data generated or managed by an information system must be defensible and appropriate to their
final use or the conclusions to be drawn from the data. To help ensure that data will be defensible,
EPA QA/G-5 G-3 QA98
-------
project teams should include adequate QA planning in the development of data management or other
information systems. There are three elements to QA planning for data management:
• Needs Analysis—identifying applicable qualitative and quantitative requirements and
establishing corresponding quality goals.
• Planning and Implementing—implementing an appropriate planning and management
framework for achieving these goals.
• Verification—testing and auditing to determine that the established goals are being met.
AG3.1 Quality Assurance Needs Analysis
The type and magnitude of the QA effort needed in developing a new information system
depends on the qualitative and quantitative criteria that the data must meet and on the complexity and
magnitude of the project. Other specific concerns such as security and system performance also help
define the QA program requirements. Only by establishing the ultimate needs and objectives for data
quality in the early planning stages can appropriate decisions be made to guide the system development
process to a successful conclusion.
AG3.1.1 Quantitative and Qualitative Criteria
Considerations similar to those in the Data Quality Objectives (DQO) framework can be used to
identify and define the general criteria that computer-processed data must meet. For example, very high
standards must be set for information systems that generate or manage data supporting Congressional
testimony, for developing new laws and regulations, for litigation, or for real-time health and safety
protection. More modest levels of defensibility and rigor are required for data used for technology
assessment or "proof of principle," where no litigation or regulatory actions are expected. Still lower
levels of defensibility apply to basic exploratory research requiring extremely fast turn-around, or high
flexibility and adaptability. In this case, the work may have to be replicated under tighter controls or the
results carefully reviewed prior to publication. By analyzing the end-use needs, appropriate criteria can
be established to guide the information system development process.
More detailed criteria can also be developed to address the specific ways in which computer-
generated or computer-processed results can be in error. The following are some specific questions to be
asked when quantitative or qualitative objectives are being defined:
• What is the required level of accuracy/uncertainty for numerical approximations?
• Are the correct data elements being used in calculations (e.g., the correct "cell" in a
spreadsheet)?
• Have the appropriate statistical models, mathematical formulas, etc. been chosen?
• What "chain-of-custody" requirements pertain to the data and results?
AG3.1.2 Project Scope. Magnitude, and Complexity Criteria
Software and systems development projects vary widely in scope and magnitude. Application of
effective management controls (including the QA program) are critical for successful performance on
large projects. Risks associated with large, complex projects commonly include overruns and schedule
delays. The integrity of results can also be compromised by rushing to complete an overdue project.
Table Gl summarizes risks as a function of project size or scope.
EPA QA/G-5 G-4 QA98
-------
Table AG1. Project Scope and Risks
PROJECT SCOPE
Large Project (information system development
is a major component)
Medium Size Project (including projects in
which an information system is not the major
component)
Small Projects (including projects in which
computer-related development is a minor
component)
Projects with ad hoc software development and
data management practices (no QA program)
POTENTIAL RISKS
• Major budget overruns
• Schedule slippage
• Unusable system or data
• Public relations problems
• Budget overrun
• Schedule slippage
• Uncertain data quality
• Lack of confidence in data
• Lack of data traceability
• Schedule slippage
• Lack of confidence in data
• Inefficient use of time and resources
EPA OIRM's Chapter 17, System Life Cycle Management, in Information Resources
Management Policy Manual, provides a similar rationale for categorizing information systems. Four
system types are defined based on the significance of the risk assessment for the Information System.
Major factors included in this risk assessment are the importance of the data, the cost of the system, and
the organizational scope of the system. For the purposes of a management review, OIRM defines
Information Systems using the following classes:
• A Major Agency System is a system that is mission critical for multiple AAships or
Regions or Agency Core Financial System or has a life cycle cost greater than $25
million or $5 million annually.
• A Major AAship or Regional System is a system that is mission critical for one AAship
or Regional Office or has a life cycle cost greater than $10 million or $1 million
annually.
• A Significant Program Office System is a system that is mission critical in one Program
Office or has a life cycle cost greater than $2 million or $100,000 annually.
• A Local Office or Individual Use System is a system for local office or individual user or
costs less than $100,000 annually for one project.
AG3.1.3 Other Quality Issues
While the issues discussed in the preceding two sections are of key importance in determining
the necessary level of the QA effort, there are many individual quality issues that should not be
overlooked in defining the requirements for a particular project. These issues should be addressed in
project planning, implementation, and testing. Some commonly encountered issues are discussed in the
following text.
AG3.1.3.1 Security Issues. There are many different types of threats to data security and
communications. Common concerns include viruses, hackers, and interception of e-mail. If these
concerns apply for a particular system, the following issues should be addressed during system planning.
Tests and audits may be planned to assess system vulnerability. Some of the management and QA
techniques that can be employed in this assessment include:
EPA QA/G-5
G-5
QA98
-------
• reviewing the project requirements documentation to ensure that security issues are
included among project requirements;
• reviewing the testing documents to ensure that security features are adequately and
thoroughly tested; and
• planning audits to be conducted by security personnel outside the immediate project
team.
AG3.1.3.2 Communication Issues. Most business computers are extensively interconnected through the
Internet, agency networks, or local networks. Computer communications is a rapidly changing area of
technology. Consequently, communications software and hardware are frequently the source of
problems to developers and users. Some communication issues that might be addressed in system
planning, design, and testing include the following:
• adequately defining the communication interfaces;
• thoroughly testing the communications hardware and software, including "stress testing"
under high load and adverse conditions; and
• conducting a beta test that encompasses users with a variety of different hardware and
communications connections.
AG3.1.3.3 Software Installation Issues. Many software packages are being developed and distributed by
the Agency to run on the individual user's personal computer. Many of these use auto-installation
routines that copy files into various directories and modify system initialization and registry files.
Planning the necessary systems requirements should address the following considerations:
• testing on as many different platforms as possible including various combinations of
processors, memory sizes, video controllers, and printers (Beta Testing can be extremely
helpful for this);
• including an "uninstall" program that not only deletes files, but also properly removes
entries in the initialization and registry files; and
• ensuring that both the "setup" and "uninstall" routines are thoroughly tested and
debugged before release.
AG3.1.3.4 Response Time Issues. A frequently overlooked aspect of computerized systems is the
impact of of system load and the resulting effect on response time. Response time is important not only
for real-time data acquisition and control systems, but also for interactive user interfaces. It is a good
idea to establish quantitative objectives for response time performance for all interactive and real-time
systems. These goals must be explicit and testable. A typical specification might be that the user should
not wait longer than x seconds for a response after submitting a request to the program.
AG3.1.3.5 Compliance with EPA and other Federal Policies and Regulations. Since individual
managers and scientists may not track information systems regulations and policy, requirements should
be determined at project inception. Some of the more important policies have been summarized in
Section AG2 of this appendix. Many of the policies and guidances are aimed at ensuring individual
project success, while others are intended to foster Agency-wide goals, including consistency of
hardware and software platforms, purchasing economies, and security. For example, EPA's Acquisition
Regulation requires Agency contractors to collect and review OIRM's most recent policies by
downloading the most current documents available online at OIRM's WWW Site.
EPA QA/G-5 G-6 QA98
-------
AG3.2 System Development Planning
Proper planning, execution, and QA protocols are vital to the success of projects involving
information systems development, software development, or computer data processing. The project
management team should work closely with the responsible QA staff to implement a program that best
suits the needs of the individual project. A few of the issues to be addressed include the level of
documentation required, schedule, personnel assignments, and change control. The following section
describes a commonly used planning framework and associated documentation that is based on the
widely recognized software- or system-development life cycle.
AG3.2.1 System Development Life Cycle
Software and information system development projects tend to evolve in distinct phases.
Recognition of this fact can be helpful in planning and managing a new project. Table G2 outlines eight
commonly recognized stages in the system development life cycle, along with typical activities and
documentation for each stage. This approach can be modified to meet the needs of individual projects.
Table AG2. Software Development Life Cycle
LIFE CYCLE
STAGE
TYPICAL ACTIVITIES
DOCUMENTATION
Needs Assessment
and High- Level
Requirements
Definition
Assessment of needs and requirements
through literature search, interviews with
users and other experts.
• Needs Assessment
Documentation (e.g., QA
Project Plan)
• Requirements Document
Detailed
Requirements
Analysis
Listing of all inputs, outputs, actions,
computations, etc. that the system is to
perform.
Listing of ancillary needs such as
security, user interface requirements.
Design team meetings.
Detailed Requirements
Document, including
Performance, Security, User
Interface Requirements etc.
System Development
Standards
System Design
Translation of requirements into a design
to be implemented.
Design Document(s)
including Technical Design
(algorithms, etc.),
Software/Systems Design
Implementation
Controls
Coding and configuration control.
Design/implementation team meetings.
In-line comments
Change control
documentation
Testing, Verification,
and Validation
Verification that the system meets
requirements.
Verification that the design has been
correctly implemented.
Beta Testing (users outside team).
Acceptance Testing (for final acceptance
of a contracted product).
Implement necessary corrective actions.
• Test Plan
• Test Result Documentation
• Corrective Action
Documentation
• Beta Test Comments
• Acceptance Test Results
EPA QA/G-5
G-7
QA98
-------
LIFE CYCLE
STAGE
Installation and
Training
Operations,
Maintenance, and
User Support
System Retirement
and Archival
TYPICAL ACTIVITIES
Installing data management system and
training users.
Use of the system or data requires usage
instructions and maintenance resources.
Information on how data or software can
be retrieved if needed.
DOCUMENTATION
• Installation Documentation
• User's Guide
• User's Guide
• Maintenance Manual or
Programmer's Manual
• Project files
• Final Report
AG3.2.2 Planning Documentation
Individual project and QA managers should tailor documentation to meet the specific needs of
their project. References in Section AG4 such as EPA System Design and Development Guidance and
Chapter 17, System Life Cycle Management, in Information Resources Management Policy Manual
describe in more detail the various types of documentation related to the system life cycle planning
phases. The following list describes in more detail some of the planning documentation listed in Table
AG2:
• Requirements Documentation—The high-level requirements document gives an
overview of the functions an information system must perform. Detailed requirements
documents define all critical functions that the completed information system must
support. Performance goals derived from analysis of the project's DQOs should be
included among the requirements. In addition, frequently overlooked issues such as
those described in Section AG3.1.3 should be addressed. Requirements documentation
should be reviewed by the end-user, if possible, to ensure that critical functions and other
requirements have not been overlooked.
• Design Documentation—Design documents are used to plan and describe the structure of
the computer program. These are particularly important in multi-programmer projects in
which modules written by different individuals must interact. Even in small or single-
programmer projects, a formal design document can be useful for communication and
for later reference.
• Coding Standards or SOPs—These may apply to a single project, an entire organizational
Branch, or other functional group. Uniform standards for code formats, subroutine
calling conventions, and in-line documentation can significantly improve the
maintainability of software.
• Testing Plans—Testing, which is discussed in Section AG3.3, must be planned in
advance and must address all original requirements and performance goals. Specific
procedures for the corrective action and retesting process should be described in QA
planning documents and implemented in the Testing Plan.
• Data Dictionary—A data dictionary can be useful to developers, users, and maintenance
programmers who may need to modify the system later. The data dictionary is often
EPA QA/G-5
G-8
QA98
-------
developed before code is written as part of the design process. The dictionary should be
updated as necessary when new elements are added to the data structure. A data
dictionary need not be a separately written document. For example, the record definition
files required for many database systems can serve this purpose, provided that they are
available in a form that is readily accessible to the user or maintenance programmer.
• User's Manual—The user's manual can often borrow heavily from the requirements
document because all of the software's functions should be specified there. The scope of
the user's manual should take into account such issues as the level and sophistication of
the intended user and the complexity of the interface. Online help can also be used to
serve this function.
• Maintenance Manual—The maintenance manual's purpose is to explain a program's logic
and organization for the maintenance programmer. This manual should also contain
crucial references documenting algorithms, numerical methods, and assumptions.
Instructions on how to rebuild the system from source code must be included. The
maintenance manual will often borrow heavily from the design manual.
• Source Code—It is usually not necessary to print the source code in hard copy form
unless needed for a specific purpose. However, it is very important to archive computer-
readable copies of source code according to the policies of each Office, Region, National
Center, or Laboratory.
AG3.3 Audits and Testing
As with any project involving generation or handling of environmental data, audits can be used
to verify that goals and objectives are being met. Audits of the Information System development
process, audits of security, and data verification audits may be particularly helpful when conducted by
personnel outside the immediate project team. Security audits by someone with expertise in this field
can be valuable when data confidentiality and prevention of tampering are important issues. Data
verification audits can be conducted using a known data set. Such a data set might be developed by an
end-user or an outside expert to verify that the information system produces the expected results.
Testing procedures and criteria need not be specified in detail by the QA Project Plan (or
equivalent document); however, the general extent and approach to testing should be described. QA
planning documents for developing a new information system should generally provide the following
project elements:
• a list of planned test documentation to be written;
• a description of the types of testing that will be conducted;
• a schedule for testing and audits; and
• a section on corrective actions.
The purpose of testing is not simply to detect errors but also to verify that the completed software
meets user requirements. In designing any test, the "correct" or "acceptable" outputs should be known in
advance, if possible. Testing should be planned in an orderly, structured way and documented. A phased
approach to testing, which is often employed in larger scale information system development projects,
might employ a sequence of testing procedures such as those presented in Sections AG3.3.1 through
AG3.3.5.
EPA QA/G-5 G-9 QA98
-------
AG3.3.1 Individual Module Tests
Individual module tests are applied to individual functions. For sequential programming
languages, such as FORTRAN, BASIC, or C, individual modules might include functions and
subroutines. For other types of software (e.g., spreadsheets), defining a functional module is more
problematic, because the software may not be designed in a modular way. However, well-planned design
strategies, such as compartmentalized design, can ease the testing effort.
AG3.3.2 Integration Tests
Integration tests are done to check the interfaces between modules and to detect unanticipated
interactions between them. Integration testing should be done in a hierarchical way, increasing the
number of modules tested and the subsystem complexity as testing proceeds. Each level of subsystem
integration should ideally correspond to a unified subset of system functions such as the "user interface."
Because all the elements may not be present, it may be necessary to develop test data sets or
hardware/software test beds to conduct the tests effectively.
When problems are encountered at any level of integration or system testing, it is necessary to
track the errors back to their origin, which may be any phase of the project. When the original reason for
the problem is identified, all affected modules and subsystems should be corrected and retested as
described in the next section.
AG3.3.3 Regression Testing
After a system module has been modified, all testing performed on the original version of the
module should be repeated, including all integration tests that include the module. This reduces the
chance that any new "bugs" introduced by the changes will go undetected while modifying the code to
correct an existing problem. Spreadsheets may be particularly difficult to test thoroughly after changes
have been made because their data dependencies are often difficult to trace. In such cases, it may be
useful to have a suite of tests that can be run whenever a change is made to verify that other functions are
not affected.
AG3.3.4 System Testing
Testing the full system is the ultimate level of integration testing and should be done in a realistic
simulation of the end-user's operational environment. If a detailed requirements document was written,
each requirement should be tested systematically. It is often helpful for a representative end-user to
participate in the system test to verify that all requirements have been implemented as intended.
Elements of the special tests described in Section AG3.3.5 can be incorporated into the system test.
For some projects, the in-house system test may be the final stage of testing. For larger or more
critical projects, formal acceptance tests or beta testing would follow. The system test should exercise all
functions possible, and the data sets used to demonstrate the software should be as realistic as possible.
AG3.3.5 Other Special Testing
AG3.3.5.1 Stress Testing should be included in the system-level testing whenever a system might be
load-sensitive (e.g., real-time data acquisition and control systems). The stress test should attempt to
simulate the maximum input, output, and computational load expected during peak usage. The specific
rates of input, output, and processing for which the system is designed are important criteria in the
EPAQA/G-5 G-10 QA98
-------
original requirements specification. The maximum load is a key quality indicator and should have been
specified early in planning. The load can be defined quantitatively using criteria such as the frequency of
inputs and outputs or the number of computations or disk accesses per unit of time. Developing an
artificial test bed to supply the necessary inputs may be necessary. The test bed can consist of hardware,
software, or a combination of the two that presents the system with realistic inputs to be processed. The
project team can write programs to carry out this testing, or automated tools may be available
commercially. Test data sets may be necessary if the software needs external inputs to run.
AG3.3.5.2 Acceptance Testing refers to contractually required testing that must be done before
acceptance by the customer and final payment. Specific procedures and the criteria for passing the
acceptance test should be listed before the test is done. A stress test is a recommended part of the
acceptance test, along with thorough evaluation of the user interface.
AG3.3.5.3 Beta Testing refers to a system-level verification in which copies of the software are
distributed outside the project group. In beta testing, the users typically do not have a supplied testing
protocol to follow; instead, they use the software as they would normally and record any anomalies
encountered. Users report these observations to the developers, who address the problems before release
of the final version.
AG3.3.5.4 Spreadsheet testing is particularly difficult because of an inherent lack of readability and
structure. One of the best ways to test spreadsheets is to challenge them with known data, although this
can be very time-consuming. Another approach is to independently recede some or all of the spreadsheet
and compare results. Software packages for spreadsheet analysis exist, but their usefulness for testing
must be evaluated on a case-by-case basis.
AG3.4 Examples
The following examples present three different data management projects: a computer model, a
spreadsheet, and a local-area-network-distributed database. Some of the QA, management, and testing
issues peculiar to each type of project are discussed.
AG3.4.1 Model Development
Mathematical models are widely used in the environmental sciences. Modeling is necessary
when the complexity of a particular situation makes a simple solution impossible, as when many different
processes are closely coupled and occur simultaneously. Some models are used to generate data that may
be used for planning and regulatory purposes.
A high level of mathematical and scientific expertise is required to develop and test the
algorithms used to represent the different physical processes. This expertise is often a scarce and
valuable resource. Consequently, a team approach may be used under which the senior scientific staff
concentrates on developing, testing, and documenting the "core" algorithms, while support staff take care
of other duties on the development project, including developing the user interface, communications,
coding, and documentation. Quality Assurance planning for developing a new model should include the
following:
• The staffing section of the QAPP should state the relevant qualifications for the key
scientific personnel. The need for peer review of novel algorithms should be addressed
if new research developments are to be incorporated in the model. Guidance documents
on conducting peer review of models are referenced in Section AG4.5. The topics
EPAQA/G-5 G-ll QA98
-------
addressed include verification testing, model code documentation, and review of the
conceptual and mathematical performance of a model.
• The end use of the data produced will dictate how exhaustively the models must be
tested and the types of demonstrations that should be done before release. A regulatory
model should be compared with existing regulatory models using identical or similar
data sets. Environmental models such as air dispersion models can be compared with
actual field data. However, care should be taken in evaluating discrepancies between
model results and the field data, because differences between monitoring data and model
results can arise from a variety of sources.
• Capabilities and needs of the end-users will dictate how much effort is spent developing
and testing the user interface and on providing user documentation and online help
functions. User interface issues should be addressed in the requirements definition, and
these functions should be tested exhaustively. Beta testing results should be reviewed
carefully to identify problems with the user interface.
• It may be possible to develop specific objectives for parameters such as bias and
precision by modeling cases that have known and accurate results. This is usually
possible only in relatively simple cases, since new models are usually developed to
expand beyond the capabilities of currently available models.
AG3.4.2 Spreadsheet for Data Processing in an Ongoing Project
Spreadsheets have replaced hand calculators for many simple applications but can sometimes
grow to approach the complexity of a database management system. Spreadsheets developed on an ad
hoc basis are usually not tested in any systematic way and may not be archived with project data.
Consequently, there can be little accountability for the correctness of calculations, even when those
results are used for sensitive applications such as regulatory reporting. This lack of testing and
verification can present significant risks. The following QA guidelines are suggested for spreadsheets
developed or used in support of projects involving environmental data:
• QA or other project planning documents should indicate all data processing tasks to be
done using spreadsheets. The origin of any spreadsheets obtained from outside the
project group should be documented.
• Spreadsheets should be developed by personnel with the appropriate education and
training. Personnel who maintain or use the spreadsheet should also have appropriate
qualifications and training.
• Documentation should be provided for correct use and maintenance of the spreadsheet.
• Data quality audits for projects processing environmental data should examine all
spreadsheets used to produce reportable data for the project. Questions such as the
following should be asked during the audit:
Have all critical calculations performed by the spreadsheet been verified (i.e.,
has the spreadsheet been tested)? Is there a record of validation including the
date and the specific inputs and outputs?
EPAQA/G-5 G-12 QA98
-------
Have significant changes been made to the spreadsheet since the last time its
output was validated?
Are users properly trained in the use of the spreadsheet? Do they have sufficient
reference material? An interview with users other than the spreadsheet
developer may be helpful in determining this.
What provisions are there for quality control of manual data input? As with any
other type of manual data entry situation, duplicate key entry or similar means of
quality control should be used when entering large data sets.
Does the spreadsheet incorporate complex table lookup functions or macros?
These features significantly complicate spreadsheets and can make their detailed
operation virtually impossible to understand fully. In such cases, the auditor
should review the reasonableness of outputs produced by the spreadsheet using a
known data set.
• Provisions should be made for archiving the spreadsheet in a format that is usable in case
data have to be reprocessed. The time window in which data may have to be reprocessed
should be considered. Some spreadsheets (as well as other types of computer software)
can sometimes remain in use long after the original project has ended, and
documentation must be provided so that the spreadsheet's functions can be understood at
a later time.
AG3.4.3 A Local-Area-Network-Distributed Database Application
Communication software is complex and is evolving rapidly. This leads to fundamental concerns
in the areas of security, privacy, and accountability. The following example, based on a real system now
in use for reporting and distributing environmental data, will illustrate some of the QA considerations
relevant to a relatively simple distributed application:
The data base application resides on a centralized server with PC- or workstation-based clients
accessing the data over a local area network (LAN). Users can also communicate with the server
using dial-up access or via the Internet. By relying on a commercially available communications
product, the system has been developed by existing project personnel, none of whom have formal
training in computer science. The database programming was done using a popular,
commercially available data base development product. Individual project team members and
some outside users can log on remotely and are able to add and modify data, query the data base,
and generate reports.
Management and QA planning for this project should address the normal concerns of ensuring
that the system is acquired and installed within budget and on-schedule, and that calculations and reports
are correct. QA concerns specific to this system include the following:
AG3.3.4.1 Security. There are many potential security vulnerabilities, and planners should identify as
many of these as possible and state explicitly how they will be prevented. Specific tests should be
conducted that address the security features of the system. Some specific methods for addressing
security vulnerabilities include the following:
• Using separate passwords for user log on, for remote dial-in, and for access to sensitive
portions of the database.
• Restricting downloads of files that could contain viruses and performing regular virus
checks on all machines on the network. Viruses are easily transmitted over the Internet,
and can spread rapidly over LANs. Viruses represent both an operational and a security
EPAQA/G-5 G-13 QA98
-------
risk. Recent viruses have infected wordprocessor macros. Because word processing
files are frequently interchanged via the WWW and e-mail, even such "nonexecutable"
files can pose a danger.
AG3.4.3.2 Privacy. Since this system may contain records of proprietary business data and voluntarily
submitted emissions information, the records must be kept private. The means for ensuring that privacy
is protected was a fundamental requirement for the system. Regular QA reviews are done to verify that
established privacy-related procedures are being followed. These include encrypting the identifying
records and restricting use only to personnel with special password-protected access.
AG3.4.3.3 Personnel Qualifications and Training. Although it is common for technical or clerical
personnel to develop small information systems using currently available "User-friendly" software and
systems environments, this practice can represent a significant risk to a larger project. On the project
described in the example, the qualifications of project staff had been carefully evaluated with respect to
experience with similar information-systems development projects of comparable magnitude. A key
person with this experience was identified and was made the lead programmer/developer. The QA
Officer also had significant computer background and was able to provide additional support during
project implementation. A number of books, utility programs, and other aids were purchased for the
project.
AG3.4.3.4 Data Defensibility and Traceability. With many different users having read/write access to a
common data set, assurance of data integrity was a concern. If absolute traceability of each data item had
been required, an Audit Trail, which records each transaction and includes the date, time, and person
responsible for the change, would be a fundamental part of the requirement. However, an audit trail was
not deemed necessary for this particular project. Backup copies of the data base are being maintained on
a weekly basis and are archived. This serves the dual purpose of providing a backup as well as to trace
any data tampering that might occur. Backups have proved valuable in this relatively open environment
when a user inadvertently overwrites or deletes files. Occasional internal audits are performed to detect
any unexplained changes in the data set overtime.
AG4. REFERENCES
This section provides references that were used in developing this appendix along with
documents that provide more detailed coverage of the topics listed below.
AG4.1 General
U.S. Environmental Protection Agency. 1995. Air Pollution Prevention and Control Division Quality
Assurance Procedures Manual, Appendix G Quality Assurance Planning for Software and Data
Management Projects. Revision 1. Research Triangle Park, NC.
U.S. Environmental Protection Agency. 1996. EPA Guidance for Quality Assurance Project Plans, EPA
QA/G-5. Washington, DC.
AG4.2 Legislation
Clinger-Cohen Act of 1996 (P.L.-104-208). (Note that the Clinger-Cohen Act is the amended title for
the Information Technology Management Reform Act and the Federal Acquisition Reform Act
of 1996 (P.L. 104-106).
Information Technology Management Reform Act of 1996.
EPA QA/G-5 G-14 QA98
-------
Paperwork Reduction Act of 1980 (P.L. 96-511) as amended in 1986 (P.L. 99-500) and 1995
(P.L. 104-13).
AG4.3 Executive Orders and Policy Directives
Executive Order, Federal Information Technology, July 17, 1996.
Office of Management and Budget Circular Number A-130, Management of Federal Information
Resources, February, 1996.
AG4.4 System Development, Operations and Maintenance Guidance Documents
Institute of Electrical and Electronics Engineers. 1994. Software Engineering. Piscataway, NJ.
U.S. Environmental Protection Agency. 1987. Data Standards for the Electronic Transmission of
Laboratory Measurement Results. EPA Directive Number 2180.2. Washington, DC.
U.S. Environmental Protection Agency. 1993. EPA Information Security Manual. EPA Directive
Number 2195. Washington, DC.
U.S. Environmental Protection Agency. 1993. EPA System Design and Development Guidance. EPA
Directive Number 2182. Washington, DC.
U.S. Environmental Protection Agency. 1993. Hardware and Software Standards. Washington, DC.
U.S. Environmental Protection Agency. 1993. Operations and Maintenance Manual. EPA Directive
Number 2181. Washington, DC.
U.S. Environmental Protection Agency. 1994. EPA Information Paper, Distributed System Management,
Draft for Comments. EPA 722/003. Washington, DC.
U.S. Environmental Protection Agency. 1995. Information Resources Management Policy Manual. EPA
Directive Number 2100. Washington, DC.
U.S. Environmental Protection Agency. 1995. Information Technology Architecture Road Map. EPA
612/002A. Washington, DC.
AG4.5 Modeling
U.S. Environmental Protection Agency. 1994. Guidance for Conducting External Peer Review of
Environmental Regulatory Models. EPA 100-B-94-001.
U.S. Environmental Protection Agency. 1994. Report of the Agency Task Force on Environmental
Regulatory Modeling. EPA 500-R-94-001.
EPAQA/G-5 G-15 QA98
-------
FEDERAL EMERGENCY MANAGEMENT AGENCY FEMA 265/JULY 1995
MANAGING FLOODPLAIN DEVELOPMENT
IN
APPROXIMA TE ZONE A AREAS
A GUIDE FOR OBTAINING AND DEVELOPING
BASE (100-YEAR) FLOOD ELEVATIONS
APRIL 1995
-------
FOREWORD
This guide was developed for use by community officials, property
owners, developers, surveyors, and engineers who may need to
determine Base (100-year) Flood Elevations (BFEs) in special flood
hazard areas designated as approximate Zone A on the Federal
Emergency Management Agency's Flood Insurance Rate Maps published
as part of the National Flood Insurance Program. One of the
primary goals of this document is to provide a means of determining
BFEs at a minimal cost.
The guidance provided herein is primarily intended for use in
riverine and lake areas where flow conditions are fairly uniform,
and do not involve unusual flow regimes (rapidly varying flow, two-
dimensional flow, supercritical flow, hydraulic jumps, etc.).
This guide is not to be used for areas that experience alluvial fan
flooding or areas that contain characteristics of alluvial fan
flooding. In addition, this guide is not to be used in Zone V
(velocity) areas or coastal Zone A areas that are subject to
flooding due to storm surge from hurricanes and other coastal
storms. Furthermore, guidance on determining regulatory floodways
is not provided in this guide.
Notes on the .PDF version of the Zone A Manual
Appendices 8 and 9 (hand calculations) were not included in the
.PDF (Internet) version of this Manual. The information was not in
a text format. The scanned images of Appendices 8 & 9 would have
made the file size of this document much larger. To keep the file
size down they were omitted.
-------
TABLE OF CONTENTS
Page
I. INTRODUCTION 1-1
II. NATIONAL FLOOD INSURANCE PROGRAM BACKGROUND II-1
III. APPLICABLE NATIONAL FLOOD INSURANCE PROGRAM FLOODPLAIN
MANAGEMENT REQUIREMENTS IN APPROXIMATE ZONE A AREAS ..111-1
Requirements for Obtaining Base (100-year)
Flood Elevation Data III-l
Requirements for Developing Base (100-year)
Flood Elevation Data Ill-2
Use of Draft or Preliminary Flood Insurance
Study Data Ill-7
Advantages of Developing Base (100-year)
Flood Elevation Data Ill-8
IV. OBTAINING EXISTING BASE (100-YEAR) FLOOD ELEVATIONS ...IV-1
Federal Emergency Management Agency IV-1
Other Federal Agencies IV-3
Other State and Local Agencies IV-4
V. DEVELOPING BASE (100-YEAR) FLOOD ELEVATIONS V-l
Simplified Methods V-l
Contour Interpolation V-2
Data Extrapolation V-7
Detailed Methods V-ll
Topography V-ll
Existing Topographic Maps V-ll
Datum Requirements for Field Surveys V-12
Number of Cross Sections Required V-13
Proper Location of Cross Sections V-13
Hydrology V-l5
Discharge-Drainage Area Relationships V-16
Regression Equations V-19
TR-55 V-20
Rational Formula V-20
Other Hydrograph Methods V-21
Hydraulics V-22
Normal Depth V-23
Critical Depth V-26
Step-Backwater Analysis V-28
Hydraulic Structures V-28
VI. OBTAINING LETTERS OF MAP CHANGE VI -1
11
-------
TABLE OF CONTENTS (continued)
FIGURES
Figure 1 - Flood Hazard Boundary Map II-2
Figure 2 - Flood Insurance Rate Map II-3
Figure 3 - Proposed 76-Lot Subdivision Ill-2
Figure 4 - Proposed 6.7-Acre Subdivision Ill-3
Figure 5 - Proposed 76-Lot Subdivision III-4
Figure 6 - Proposed 5.6-Acre Subdivision III-4
Figure 7 - Proposed 6.7-Acre Subdivision Ill-5
Figure 8 - Contour Interpolation Method -
Riverine Flooding Example 1 V-4
Figure 9 - Contour Interpolation Method -
Riverine Flooding Example 2 V-5
Figure 10 - Contour Interpolation Method -
Lacustrine Flooding Example 3 V-6
Figure 11 - Data Extrapolation Method - Profile V-8
Figure 12 - Data Extrapolation Method - Plan View V-8
Figure 13 - Data Extrapolation Method - Profile V-9
Figure 14 - Data Extrapolation Method - Plan View V-9
Figure 15 - Data Extrapolation Method - Profile V-10
Figure 16 - Cross Section Orientation V-14
Figure 17 - Locate Cross Sections at Points of Flood
Discharge Changes V-14
Figure 18 - Cross Section Locations at Structures V-15
Figure 19 - Wendy Run Drainage Basin V-18
Figure 20 - Discharge-Drainage Area Plot V-18
Figure 21 - 100-Year Discharge Estimates for Site A
and Site B V-19
Figure 22 - Channel Bank Stations V-25
Figure 23 - Weir Flow - Embankment Profile is Not
Horizontal V-30
Figure 24 - Weir Flow Over Road V-31
111
-------
TABLE OF CONTENTS (continued)
Page
APPENDICES
Appendix 1 - Glossary of Floodplain Analysis Terms Al-1
Appendix 2 - Flood Insurance Study Data Request Form A2-1
Appendix 3 - Federal Emergency Management Agency Offices
and Other Federal and State Agencies A3-1
Appendix 4 - State Hydrology Reports A4-1
Appendix 5 - Manning1s "n" Values A5-1
Appendix 6 - QUICK-2 Computer Program Manual A6-1
Appendix 7 - Hydraulic Computer Manuals A7-1
Appendix 8 - Normal Depth Hand Calculations A8-1
Appendix 9 - Weir Flow Hand Calculations A9-1
Appendix 10 - Worksheet A10-1
IV
-------
Guide for Approximate Zone A Areas Introduction
I. INTRODUCTION
This guide is primarily intended to assist local community
officials in administering and enforcing the floodplain
management requirements of the National Flood Insurance
Program (NFIP). This document provides guidance for
determining Base (100-year) Flood Elevations (BFEs) in special
flood hazard areas that have been identified and designated as
approximate Zone A on a community's NFIP maps. Zone A
identifies an approximately studied special flood hazard area
for which no BFEs have been provided. Although BFEs are not
provided, the community is still responsible for ensuring that
new development within approximate Zone A areas is constructed
using methods that will minimize flood damages. This often
requires obtaining or calculating BFEs at a development site.
Developers, property owners, engineers, surveyors, and others
at the local level who may be required to develop BFEs for use
in approximate Zone A areas should also find this guide
useful. Included in this guide are methodologies that can be
used to develop BFEs, which can be used to determine the
elevation or floodproofing requirements for buildings. The
detailed methodologies described in this guide can also be
used to develop the BFE information necessary to obtain a
Letter of Map Amendment or a Letter of Map Revision Based on
Fill from the Federal Emergency Management Agency (FEMA) to
remove a legally defined property or structure from a special
flood hazard area. In addition, Letter of Map Revision
requestors may use the detailed methods in this document to
develop the BFE information that must be submitted to FEMA to
demonstrate that an area will not be flooded during the 100-
year flood.
1-1
-------
Guide For Approximate Zone A Areas NFIP Background
II. NATIONAL FLOOD INSURANCE PROGRAM BACKGROUND
In 1968, the United States Congress passed the National Flood
Insurance Act, which created the NFIP. Congress recognized
that the success of this program required that community
participation be widespread, that studies be conducted to
accurately assess the flood risk within each participating
flood-prone community, and that insurance premium rates be
established based on the risks involved and accepted actuarial
principles. To meet these objectives, the 1968 Act called
for: 1) the identification and publication of information
within five years for all floodplain areas that have special
flood hazards; and 2) the establishment of flood-risk zones in
all such areas to be completed over a 15-year period following
the passage of the act.
Within the first year of NFIP operation, it became evident
that the time required to complete the detailed flood
insurance studies would markedly delay implementation in many
flood-prone communities. As a result, an interim means for
more rapid community participation in the NFIP had to be
provided. The Housing and Urban Development Act of 1969
expanded participation by authorizing an Emergency Program
under which insurance coverage could be provided at non-
actuarial, federally-subsidized rates in limited amounts
during the period prior to completion of a community's flood
insurance study.
Until engineering studies could be conducted for these
communities, Flood Hazard Boundary Maps, such as the one shown
in Figure 1, "Flood Hazard Boundary Map," which delineated the
boundaries of the community's special flood hazard areas, were
prepared using available data or approximate methods. The
Flood Hazard Boundary Maps identified, on an approximate
basis, the areas within a community subject to inundation by
the 100-year flood (i.e., Zone A). The 100-year flood has a
one-percent chance of being equalled or exceeded in any given
year. The Flood Hazard Boundary Map was intended to assist
communities in managing floodplain development, and insurance
agents and property owners in identifying those areas where
the purchase of flood insurance was advisable.
The Flood Disaster Protection Act of 1973, which also amended
the 1968 Act, required that flood-prone communities be
notified of their flood hazards to encourage program
participation. This notification was accomplished through the
publication of Flood Hazard Boundary Maps for all communities
that were identified as containing flood hazard areas. In
addition, the 1973 Act required the purchase of flood
insurance by property owners who were being assisted by
II-l
-------
Guide For Approximate Zone A Areas
NFIP Background
Figure 1 - Flood Hazard Boundary Map
Federal programs, or by Federally supervised, regulated, or
insured agencies or institutions, in the acquisition or
improvement of land or facilities located, or to be located,
in special flood hazard areas. This act also severely limited
Federal financial assistance in the flood hazard areas of
communities which did not join the NFIP.
The initial Flood Hazard Boundary Maps for communities
identified as having flood hazards were prepared using
available floodplain data contained in reports developed by a
variety of Federal, State, and local sources. For those
communities that had no available flood information,
approximate hydrologic and hydraulic methods or historical
flood data were used to determine the extent of the special
flood hazard areas.
Flood Insurance Studies that used detailed hydrologic and
hydraulic analyses to develop BFEs and designate floodways and
risk zones were subsequently developed for most NFIP
communities. The results of a Flood Insurance Study were
II-2
-------
Guide For Approximate Zone A Areas
NFIP Background
issued to the community in the form of a Flood Insurance Rate
Map (FIRM), such as the one shown in Figure 2, "Flood
Insurance Rate Map," and, in most cases, a Flood Boundary and
Floodway Map and a Flood Insurance Study report. Once more
detailed risk data were provided, the community could enter
the Regular Program whereby more comprehensive floodplain
management requirements were imposed and higher amounts of
insurance could be purchased by owners of structures.
LIMIT OF
DETAILED STUDY
\ "CORPORATE
Figure 2 - Flood Insurance Rate Map
As early as 1976, FEMA recognized that some communities did
not require a detailed Flood Insurance Study because there
were few existing buildings in the floodplain and minimal
development pressure. Shortly thereafter, FEMA began
utilizing a special conversion process whereby communities
11-3
-------
Guide For Approximate Zone A Areas NFIP Background
were converted to the Regular Program without a Flood
Insurance Study. Consequently, these communities were
converted using FIRMs in which all of the special flood hazard
areas were designated as approximate Zone A, without BFEs.
Although over 10,000 communities have now been provided
detailed Flood Insurance Studies and issued FIRMs that include
BFEs, many floodplains are still designated as approximate
Zone A without BFEs. Due to the costs of developing detailed
risk data, areas not subject to development pressure are
studied using approximate methodologies and continue to be
shown on the FIRM as approximate Zone A areas. FEMA only
provides BFEs for the floodplains of those flooding sources
that are currently subject to development pressure or are
projected at the initiation of a Flood Insurance Study or
Flood Insurance Study restudy to be subject to development
pressure during the immediate future. Generally, a planning
period of approximately five years is used. Even in these
cases, BFEs are provided on a priority basis due to funding
constraints. The community plays a major part in the
determination of the level of detail required in the study of
selected streams. As a result, most communities will have
FIRMs that include special flood hazard areas for flooding
sources that have been studied in detail with BFEs and special
flood hazard areas for flooding sources that have been studied
using approximate methods, and have been designated as
approximate Zone A.
II-4
-------
Guide For Approximate Zone A Areas Floodplain Management
III. APPLICABLE NATIONAL FLOOD INSURANCE PROGRAM
FLOODPLAIN MANAGEMENT REQUIREMENTS IN APPROXIMATE
ZONE A AREAS
The primary requirement for community participation in the
NFIP is the adoption and enforcement of floodplain management
regulations that meet the minimum standards of the NFIP
regulations in Title 44 of the Code of Federal Regulations
(CFR) Section 60.3. These minimum standards vary depending on
the type of flood risk data provided to the community by FEMA.
The intent of floodplain management regulations is to minimize
the potential for flood damages to new construction and to
avoid aggravating existing flood hazard conditions that could
increase potential flood damages to existing structures. To
protect structures in riverine and lacustrine areas, the NFIP
regulations require that the lowest floor (including basement)
of all new construction and substantial improvements of
residential structures be elevated to or above the BFE. New
or substantially improved non-residential structures in
riverine areas must either be elevated or floodproofed (made
watertight) to or above the BFE.
Requirements for Obtaining BFE Data
In areas designated as approximate Zone A, where BFEs have not
been provided by FEMA, communities must apply the provisions
of Paragraph 60.3(b) of the NFIP regulations. Subparagraph
60.3(b)(4) requires that communities:
Obtain, review and reasonably utilize any
base flood elevation and floodway data
available from a Federal, State, or other
source... [ 44 CFR 60.3 (b)(4) ]
Section IV describes the sources from which BFE data may be
obtained. These data are to be used as criteria for requiring
that new construction, substantial improvements, and other
development within all approximate Zone A areas meet the
applicable requirements in Paragraphs 60.3 (c) and (d) of the
NFIP regulations, including the requirement that structures
have their lowest floors elevated to or above the BFE (or
floodproofed to or above the BFE for non-residential
structures) . These data should be used as long as they
reasonably reflect flooding conditions expected during the
base (100-year) flood, are not known to be scientifically or
technically incorrect, and represent the best data available.
Communities should consider formally adopting these data by
reference as part of their floodplain management regulations.
III-l
-------
Guide For Approximate Zone A Areas
Floodplain Management
Requirements for Developing BFE Data
Under Subparagraph 60.3(b)(3) of the NFIP regulations,
communities must also:
Require that all new subdivision proposals
and other proposed development (including
proposals for manufactured home parks and
subdivisions) greater than 50 lots or 5
acres, whichever is the lesser, include
within such proposals base flood elevation
data; [ 44 CFR 60.3 (b)(3) ]
This means that any subdivision which meets this threshold must
be evaluated to determine if the subdivision proposal is
affected by an approximate Zone A area and whether BFE data are
required. BFE data are required for the affected lots in the
subdivisions shown in Figure 3, "Proposed 76-Lot Subdivision,"
and Figure 4, "Proposed 6.7-Acre Subdivision." Figure 3 clearly
shows a 76-lot subdivision with several lots affected by an
approximate Zone A area. The subdivision depicted in Figure 4
is only 12 lots, but because the subdivision is greater than 5
acres and clearly shows buildable sites affected by an
approximate Zone A area, BFE data are required.
ZONE A
Figure 3 - Proposed 76-Lot Subdivision
III-2
-------
Guide For Approximate Zone A Areas
Floodplain Management
ZONE A
Figure 4 - Proposed 6.7-Acre Subdivision
Communities are encouraged to address the flood hazards at the
earliest stages of subdivision planning rather than at the
actual placement of individual structures. If a community can
work with the developer and others when land is being
subdivided, many long-term floodplain management benefits can be
achieved, particularly if the floodplain is avoided altogether.
In Figure 5, "Proposed 76-Lot Subdivision," the entire
approximate Zone A area is to be dedicated as open space. If
the planned subdivision shows the floodplain is contained
entirely within an open space lot, it may not be necessary to
conduct a detailed engineering analysis to develop BFE data.
Also, it may not be necessary to develop detailed BFE data in
large-lot subdivisions or single-lot subdivisions that are
within the thresholds under Subparagraph 60.3 (b) (3) of the NFIP
regulations when the actual building sites are clearly outside
of the Zone A area. In Figure 6, "Proposed 5.6-Acre
Subdivision," it is evident from the topographic features of
this 5.6-acre subdivision that the building sites would be
clearly out of the floodplain since the proposal indicates a
steep grade between the approximate Zone A area and the building
sites which are located on natural high ground.
III-3
-------
Guide For Approximate Zone A Areas
Floodplain Management
ZONE A
Open Space Lot
Figure 5 - Proposed 76-Lot Subdivision
Building j
Site
ZONE A
Figure 6 - Proposed 5.6-Acre Subdivision
III-4
-------
Guide For Approximate Zone A Areas
Floodplain Management
If the grade between the actual building sites and the
approximate Zone A area of the proposed subdivision is
relatively gradual, asshown in Figure 7, "Proposed 6.7-Acre
Subdivision," the floodplain could extend beyond what is shown
on the Flood Insurance Rate Map. It is very likely that flooding
could affect the building sites.In this case, an analysis should
be conducted to determine the location of the 100-year
floodplain and the BFE.
280
ZONE A
275
Figure 7 - Proposed 6.7 Acre Subdivision
For developments that exceed the thresholds identified in NFIP
regulations Subparagraph 60.3 (b) (3), BFEs must be either
obtained from other sources or developed using detailed
methodologies comparable to those contained in a Flood Insurance
Study. Section V describes some of the detailed methodologies
available that can be used to develop BFE data when none are
available from the sources listed in Section IV.
If the size of the new subdivision or other proposed development
falls below the thresholds specified in NFIP regulations
Subparagraph 60.3 (b) (3) and no BFE data are available from the
sources listed in Section IV, the community must still apply, at
a minimum, the requirements of Subparagraph 60.3(a)(3) to
proposed structures or Subparagraph 60.3 (a) (4) to subdivisions
and other developments within approximate Zone A areas. These
paragraphs require that permit officials:
III-5
-------
Guide For Approximate Zone A Areas Floodplain Management
Review all permit applications to determine whether
proposed building sites will be reasonably safe from
flooding. If a proposed building site is in a flood-
prone area, all new construction and substantial
improvements shall (i) be designed (or modified) and
adequately anchored to prevent floatation, collapse, or
lateral movement..., (ii) be constructed with materials
resistant to flood damage, (iii) be constructed by
methods and practices that minimize flood damages, and
(iv) be constructed with electrical, heating,
ventilation, plumbing, and other service facilities that
are designed and/or located so as to prevent water from
entering or accumulating within the components during
conditions of flooding. [44 CFR 60.3(a)(3)]
Review subdivision proposals ... including manufactured
home parks or subdivisions ... to assure that (i) all
such proposals are consistent with the need to minimize
flood damage within the flood-prone area, (ii) all
public utilities and facilities ... are located and
constructed to minimize or eliminate flood damage, and
(iii) adequate drainage is provided to reduce exposure
to flood hazards; [44 CFR 60.3(a)(4)]
One way that communities can ensure that building sites will be
reasonably safe from flooding for proposed developments that
fall below the thresholds in Subparagraph 60.3 (b) (3) is to use
the simplified methods outlined in Section V for estimating a
BFE. Another approach to ensure that a building site is
reasonably safe from flooding is to require the structure to be
elevated above the highest adjacent grade by a specified number
of feet based on the local official's knowledge of flood
conditions in the area. In the absense of available BFE data
from other sources, the community may require the permit
applicant to elevate the structure two or more feet above the
highest adjacant grade which qualifies the structure for reduced
flood insurance rates. Elevation of the structure to four feet
above the highest adjacant grade will enable the structure to
qualify for substantially reduced flood insurance rates.
However, some states and communities require that BFE data be
developed for all subdivisions and/or floodplain development
within approximate Zone A areas, not just those subdivisions
III-6
-------
Guide For Approximate Zone A Areas Floodplain Management
which meet the 50-lot or 5-acre threshold. A community may, at
its discretion, require the use of detailed methods for such
development. While this requirement is more restrictive than
NFIP minimum requirements, the NFIP regulations specifically
recognize and encourage states and communities to adopt and
enforce more restrictive floodplain management regulations in
those instances where the state or community believes that it is
in the best interest of its citizens.
No matter what the size of the subdivision or other development
proposal, requests to revise or amend effective Flood Insurance
Study information through the procedures outlined in Part 65 and
Part 70 of the NFIP regulations must be based on detailed
methodologies presented in Section V or other methodologies
comparable to those found in a Flood Insurance Study. The
analysis used to develop the BFEs must be certified by a
registered professional engineer or licensed land surveyor, as
appropriate, if the BFEs are to be used to revise or amend an
NFIP map.
Use of Draft or Preliminary Flood Insurance Study Data
The data from a draft or preliminary flood insurance study
constitutes "available data" under Subparagraph 60.3(b)(4).
Communities must reasonably utilize the draft or preliminary
flood insurance study data under the section of their ordinance
that requires the use of other base flood data when detailed BFE
data has not been published in a flood insurance study.
Communities are given discretion in using draft or preliminary
flood insurance study data only to the extent that the technical
or scientific validity of the proposed flood elevation data is
questioned. If a community decides not to use the draft or
preliminary flood insurance data in a FIS because it is
questioning the data through a valid appeal, the community must
still assure that buildings are constructed using methods and
practices that minimize flood damages in accordance with the
requirements under Subparagraphs 60.3(a)(3) and (4).
When all appeals have been resolved and a notice of final flood
elevations has been provided by FEMA, communities are required
to use the data from the flood insurance study for regulating
floodplain development in accordance with Subparagraph
60.3(b)(4) since the data represents the best data available.
Communities must regulate floodplain development using the flood
insurance study data under Subparagraph 60.3(b)(4) until such
time as the community has adopted the effective FIRM and flood
insurance study.
III-7
-------
Guide For Approximate Zone A Areas Floodplain Management
Advantages of Developing BFE Data
While the NFIP regulations do not require that communities
develop BFE data in approximate Zone A areas when proposed
development is below the thresholds in NFIP regulations
Subparagraph 60.3 (b) (3), there are significant advantages and
financial benefits for communities and individual property
owners that develop BFE data. These advantages and benefits
include:
• protecting structures up to the BFE will minimize and
reduce future flood losses, resulting in long-term
savings to the individual, the community, and the
National Flood Insurance Fund;
• flood insurance policies in approximate Zone A areas
that are rated using a BFE will often qualify for
significantly lower insurance rates than policies that
are rated without a BFE. The difference in flood
insurance premiums could be substantial;
• less burden will be placed on the permit official
because he or she can require protection to a specified
elevation. Without a BFE, the permit official must make
judgements as to what constitutes "reasonably safe from
flooding" and "constructed with materials and practices
that minimize flood damages";
• the NFIP's Community Rating System establishes flood
insurance premium discounts of up to 45 percent for
policy holders within communities that have a
floodplain management program that exceeds NFIP minimum
requirements. Sizable Community Rating System credits
are available for Community Rating System communities
that develop BFEs for areas designated as approximate
Zone A on their Flood Hazard Boundary Map or FIRM, or
that require site-specific engineering analyses for
development proposals; and
• by specifying a BFE in an approximate Zone A area, a
building or property can, in some circumstances, be
removed from the floodplain by issuance of a Letter of
Map Amendment or Letter of Map Revision in accordance
with Part 65 and Part 70 of the NFIP regulations. While
these procedures eliminate the requirement that flood
insurance be purchased as a condition of obtaining a
loan from a Federally insured or regulated lender, a
lending institution may, at its discretion, require the
purchase of flood insurance.
III-8
-------
Guide For Approximate Zone A Areas Obtaining BFEs
IV. OBTAINING EXISTING BASE (100-YEAR) FLOOD ELEVATIONS
The NFIP Regulations at 44 CFR 60.3 require that structures be
elevated or floodproofed (non-residential structures only) to
provide protection from flood damage. A BFE must be established
before such flood protection measures can be used. There are a
variety of computational methods that can be employed to
determine BFEs. However, these methods can be costly. Before
computational methods are used, every attempt should be made to
obtain information, in the form of floodplain studies or
computations, from Federal, State, or local agencies. Data
obtained from these agencies may be adequate to determine BFEs
with little or no additional research, computation, or cost.
Local officials who obtain BFE data should maintain the
information for future reference. Local officials should also
consider making a search for BFE data for the entire community.
By doing so, the local officials may not have to conduct a
search each time a floodplain development permit is requested.
If the data reasonably reflect flooding conditions, a community
should consider adopting the information into its floodplain
management ordinance.
Provided below are a list of agencies that can be contacted to
determine if any BFE data have already been developed. When
obtained, these data should be evaluated to ensure that they
reasonably reflect flooding conditions expected at the site
during the 100-year flood, are scientifically or technically
correct, and represent the best data available.
Three major sources of existing data are highlighted in this
section: FEMA, other Federal agencies, and State and local
agencies.
FEMA
FEMA's technical evaluation contractors maintain libraries that
contain technical and administrative data developed during the
preparation of Flood Insurance Studies, as well as the resulting
Flood Insurance Study reports and NFIP maps. FEMA can be
contacted to determine whether or not sufficient information
exists in the back-up data to calculate BFEs. For some flooding
sources that are designated as approximate Zone A, FEMA may have
detailed flooding information that has not yet been incorporated
into the community's Flood Insurance Study. FEMA can be
contacted to obtain this information where it exists.
IV-1
-------
Guide For Approximate Zone A Areas Obtaining BFEs
FEMA regularly conducts restudies of flood hazards in an effort
to keep Flood Insurance Studies accurate and up-to-date. As
part of these restudies, detailed BFE data for approximate Zone
A areas may be developed. During the time that elapses between
FEMA obtaining restudy data and the incorporation of BFE data
areas into a revised Flood Insurance Study, a community may
reasonably use the BFE data from the restudy in approximate Zone
A areas in accordance with Subparagraph 60.3(b)(4).
In addition, flooding sources restudied by FEMA may often impact
several communities. FEMA may be unable to immediately update
the Flood Insurance Study for every community impacted due to
funding constraints. Therefore, BFEs may have been developed
for streams within a community that have not yet been
incorporated into the Flood Insurance Study.
It is also possible that a previous request to revise flood
hazards along a stream or lake may be on file with FEMA, and
that BFEs, which may be applicable to other areas of the same
stream or lake, may have been computed for that request.
FEMA data should be sought when trying to obtain or determine
BFEs for an approximate Zone A area, so that if BFEs have
already been determined for an approximate Zone A area, then
other BFE determinations in the same area can be based on the
same methodology. However, if it is determined that a more
scientifically or technically accurate determination than that
which is available in FEMA's back-up data is warranted, then a
more detailed methodology, such as those described in Section V,
should be utilized.
Data requests should be directed to the appropriate FEMA
technical evaluation contractor at the address listed on the
following page:
IV-2
-------
Guide For Approximate Zone A Areas Obtaining BFEs
FEMA Regions I-V
(States East of the Mississippi
River and Minnesota)
Flood Insurance Information Specialist
c/o Dewberry & Davis
2953 Prosperity Avenue
Fairfax, Virginia 22031
FAX: (703) 876-0073
Phone: (703) 876-0148
FEMA Regions VI-X
(States West of the
Mississippi River)
FEMA Project Library
c/o Michael Baker, Jr., Inc.
3601 Eisenhower Avenue
Suite 600
Alexandria, Virginia 22304
FAX: (703) 960-9125
Phone: (703) 960-8800
An instruction sheet entitled Flood Insurance Study (FIS) Data
Requests is provided in Appendix 2. This sheet contains
pertinent information and instructions for requesting Flood
Insurance Study data.
A fee is charged for locating, retrieving, copying, and mailing
Flood Insurance Study back-up data based on the cost of
materials and a standard hourly rate for time spent to fill the
request. FEMA will inform the requestor if the requested data
are available and of the required fee. The requestor should
allow two to four weeks for the request to be processed.
Other Federal Agencies
Information regarding BFEs may be obtained from other Federal
agencies involved in floodplain management. A fee may be
required to obtain some of the products or services available
through these agencies. The following is a list of some of the
Federal Agencies involved in floodplain management and the
information, which may be useful in obtaining and determining
BFEs, that they produce.
IV-3
-------
Guide For Approximate Zone A Areas
Obtaining BFEs
AGENCY
U.S. Army Corps
of Engineers
U.S. Department of the
Interior, Geological
Survey
U.S. Department of
Agriculture, Natural
Resources Conservation
Service (NRCS)
U.S. Department of
Transportation, Federal
Highway Administration
U.S. Department of Commerce
National Technical
Information, Service
Tennessee Valley Authority
Other State and Local Agencies
PRODUCT
Floodplain Information
Reports
Technical Manuals
Computer Programs
Computational Assistance
Topographic Maps
Water Resource
Investigations
Technical Bulletins
Water Supply Papers
Computer Programs
Watershed Studies
Technical Releases
Computer Programs
Floodplain Studies
Design Manuals
Computer Programs
Design Manuals
Computer Programs
Floodplain Studies
If back-up data from Federal agencies are unavailable or are not
useful, information regarding BFEs may be obtained from State or
local agencies involved in floodplain management. On the
following page is a list of State and local agencies involved in
floodplain management that may be contacted to obtain BFE
information. Again, fees may be applicable for this
information.
For example, some state agencies, such as a Department of
Natural Resources or a Geological Survey, may conduct floodplain
studies using state funds. In some states, these agencies may
maintain a repository for flood data. The NFIP state
coordinating agency can also be contacted. A list of the State
coordinating agencies is provided in Appendix 3. Other state
agencies, such as a Department of Transportation, do engineering
for specific types of projects, such as road and bridge
construction, in which BFE data may have been developed for
these projects. In general, when calling these agencies, the
IV-4
-------
Guide For Approximate Zone A Areas Obtaining BFEs
caller should ask for a copy of any back-up data (reports,
computations, computer models, maps) associated with a
FLOODPLAIN STUDY or DRAINAGE STUDY, for the area of the
particular stream of interest. In addition, some state
agencies, such as a Department of Natural Resources, may
maintain historic lake level data.
The local public works department or the local transportation
department may have developed flood data in designing sewer and
storm drainage systems and local roads. For example, plans for
a sanitary sewer line which runs parallel to the stream and area
of interest may have 100-year flood elevations on the profile of
the sanitary sewer. Also, if there are culverts or bridges
which cross the same stream within 1,000 feet of the area of
interest, there may be hydrologic and hydraulic information
pertaining to the 100-year flood discharge and elevation which
may be pertinent to the site. Finally, if there are any nearby
residential or commercial developments along the same stream,
the development (site) plans for these projects may also include
information about the 100-year flood.
Other possible sources of data include regional organizations,
such as Flood Control Districts, Levee Improvement Districts,
Watershed Districts, or Soil and Water Conservation Districts.
These organizations may have developed flood profiles for
smaller streams or for reaches of streams impacted by flood
control or drainage projects.
State Agencies;
Departments of Environmental Conservation
Departments of Environmental Protection
Departments of Floodplain Management
Departments of Natural Resources
Departments of Transportation
Departments of Water Resources
Geological Survey
Local or Regional Agencies;
Flood Control Districts
Levee Improvement Districts
Local Planning Commissions
Local Public Works Departments
Municipal Utility Districts
River Basin Commissions
Water Control Boards
A partial list of Federal and State agencies is provided in
Appendix 3.
IV-5
-------
Guide For Approximate Zone A Areas Developing BFEs
V. DEVELOPING BASE (100-YEAR) FLOOD ELEVATIONS
If sufficient BFE information cannot be obtained from the
sources described in Section IV, then the community should
consider conducting, or requiring the applicant to conduct, a
site specific engineering analysis to determine a BFE. This
section describes several simplified and detailed methods for
estimating or developing BFE data, and provides guidance for
using them.
As noted in Section III, a detailed method is required under
Subparagraph 60.3(b)(3) of the NFIP regulations for proposed
development greater than 50 lots or 5 acres, whichever is the
lesser. If the BFEs developed will be used to revise or amend
NFIP maps, they must be developed using the detailed
methodologies described in this section or other methods
comparable to those in a Flood Insurance Study.
If no BFE data are available and the proposed development is
below the thresholds specified in Subparagraph 60.3(b)(3) of the
NFIP regulations, the simplified methods for estimating BFEs
described in the following section may be used. Simplified
methods are less expensive and less time consuming than the
detailed methods described later in this section. However,
communities have the discretion to determine which method should
be used when a proposed development is below the aforementioned
thresholds.
Simplified Methods
There are situations in which a simplified approach for
estimating the BFE may yield an acceptable level of accuracy.
For simplified methods to be used, very specific conditions must
be met as discussed below.
Simplified methods are appropriate for floodplain management
purposes only. These methods may be used for the purpose of
meeting the requirements of NFIP regulations Subparagraphs
60.3(a)(3) and 60.3(a)(4) for developments, such as isolated
small subdivisions in rural areas which are below the threshold
in Subparagraph 60.3(b)(3), or single lots. Subparagraphs
60.3(a)(3) and 60.3(a)(4) require the community to determine
whether proposed building sites are reasonably safe from
flooding and ensuring that subdivision proposals are consistent
with the need to minimize flood damage within flood-prone areas.
V-l
-------
Guide For Approximate Zone A Areas Developing BFEs
Simplified methods may not be used by the community to complete
an Elevation Certificate used for flood insurance rating.
Communities must use the detailed methodologies described in
this section or other methods comparable to those in a Flood
Insurance Study for completing the Elevation Certificate. A
flood insurance policy for a structure for which a simplified
method is used may be rated without an elevation certificate.
However, the flood insurance rate may be higher than if the
structure is rated using an Elevation Certificate.
Contour Interpolation
Contour interpolation involves superimposing approximate Zone A
boundaries onto a topographic map in order to estimate a BFE.
BFEs obtained by this method can only be assumed to be as
accurate as one-half of the contour interval of the topographic
map that is used. Therefore, the smaller the contour interval
of the topographic map, the higher the accuracy of the BFE
determined from the map. The procedures for using this method
are outlined below. Steps 1 through 5 are the same for both
riverine and lacustrine (lake) flooding sources.
Step 1 - Obtain a topographic map showing the site being
analyzed
Step 2 - Reduce or enlarge the FIRM or topographic map as
necessary so that the two are at the same scale
Step 3 - Superimpose the approximate Zone A (100-year)
floodplain boundary from the FIRM onto the topographic
map
Step 4 - Determine if this method is within the acceptable
accuracy limits. The floodplain boundary must
generally conform with the contour lines along the
flooding source in question. The difference between
the water-surface elevations determined on the right
overbank and the left overbank must be within one-half
of the map contour interval. For lacustrine flooding
sources, the difference between the highest and lowest
determined water-surface elevations around the
flooding source must be within one-half of the map
contour interval. Otherwise, this method is not
acceptable.
Step 5 - If the method is acceptable, then determine the BFE.
Detailed guidance for determining the BFE is provided
below.
V-2
-------
Guide For Approximate Zone A Areas Developing BFEs
Determining BFEs for Riverine flooding:
On each side of the stream in the vicinity of the site,
determine the ground elevation at which the superimposed Zone A
boundary lies by interpolating between two contour lines. Add
one-half of the map contour interval to the lower of the two
interpolated elevations. This is the approximate BFE for the
site (be sure to perform this method at each structure
location).
By adding one-half of the contour interval to the lowest
interpolated water-surface elevation, two things are achieved:
1) the final BFE is within one-half of the contour interval of
both interpolated water-surface elevations and, therefore, is
still within the acceptable tolerance of the topographic map
(generally regarded as ± one-half of the map contour interval);
2) it is a conservative estimate of the BFE. If the BFE
determined under this procedure seems too high, then a detailed
analysis may be performed to justify lowering it.
Example 1
Using a county topographic map with a contour interval of 5
feet, the approximate Zone A boundary crosses contour
elevations on the left and right bank at 323 and 325 feet,
respectively, as shown in Figure 8, "Contour Interpolation
Method - Riverine Flooding Example 1." The difference
between these two water-surface elevations is 2 feet, which
is less than one-half of the contour interval or 2.5 feet.
Therefore, this method is acceptable for use on this
portion of the stream. Add 323 feet (lowest interpolated
water-surface elevation) plus 2.5 feet (one-half of the
contour interval), which yields a BFE of 325.5.
V-3
-------
Guide For Approximate Zone A Areas
Developing BFEs
Proposed Structur
s m / i j-- ->.
Figure 8 - Contour Interpolation Method
Riverine Flooding Example 1
Example 2
Using a U.S. Geological Survey quadrangle map with a
contour interval of 10 feet, the approximate Zone A
boundary crosses contour elevations on the left and right
bank of 422 and 430 feet, respectively, as shown in Figure
9, "Contour Interpolation Method - Riverine Flooding
Example 2." The difference between these two water-surface
elevations is 8 feet, which is greater than one-half of the
contour interval or 5 feet. Therefore, this method is not
acceptable for use on this portion of the stream, and
another method must be used to determine the BFE.
V-4
-------
Guide For Approximate Zone A Areas
Developing BFEs
Proposed
Structure
Figure 9 - Contour Interpolation Method -
Riverine Flooding Example 2
Determining BFEs for Lacustrine (Lake) flooding:
Determine the contour elevations that the approximate Zone A
boundary crosses (i.e. the BFE) around the perimeter of the lake
or ponding area. Assuming that the highest and lowest
determined water-surface elevations are within one-half of the
map contour interval of each other, add one-half of the map
contour interval to the lowest water-surface elevation to
determine the BFE for the site.
Example 3
Using a U.S. Geological Survey quadrangle map with a
contour interval of 10 feet, the approximate Zone A
boundary crosses low and high determined water-surface
elevations along the perimeter of the ponding area of 280
and 283 feet, respectively, as shown in Figure 10, "Contour
Interpolation Method - Lacustrine Flooding Example 3." The
difference between these two water-surface elevations is 3
feet, which is less than one-half of the contour interval
or 5 feet. Therefore, this method is acceptable for use on
this ponding area. Add 280 feet (lowest water-surface
elevation) and 5 feet (one-half of the contour interval),
which yields a BFE of 285 feet.
V-5
-------
Guide For Approximate Zone A Areas
Developing BFEs
. ,,^. \
roposed Structure
V * ti 1***~ "V \ 1
Tadma Lake
r
Figure 10 - Contour Interpolation Method -
Lacustrine Flooding Example 3
V-6
-------
Guide For Approximate Zone A Areas Developing BFEs
Data Extrapolation
If a site is within 500 feet upstream of a stream reach for
which a 100-year flood profile has been computed by detailed
methods, and the floodplain and channel bottom slope
characteristics are relatively similar to the downstream
reaches, data extrapolation may be used to determine the BFE.
The stream in the vicinity of the site, however, must be free
of backwater effects from downstream hydraulic structures.
The procedure for using this method is outlined below.
Step 1 - Determine the location of the site on the flood
profile for the detailed study stream
Step 2 - Extrapolate the last segment of the 100-year flood
profile that has a constant water-surface slope to
the location of the site. The BFE at the site can
then be obtained directly from the profile
Figures 11-12 on the following pages depict situations (i.e.,
properties "Y" and "Z"), in which the data extrapolation
method may and may not be used. Figures 13-14 depict a
situation in which the data extrapolation method may not be
used because the highway may have an effect on the 100-year
water-surface elevations. If the 100-year flood profile
changes just prior to the limit of detailed study, as shown in
Figure 15, the data extrapolation method should not be used.
V-7
-------
Guide For Approximate Zone A Areas
Developing BFEs
40
ft 3b
CD
i—
UJ
111
LL
z 30
O
t
UJ
ti 25
20
8
= ,:
00
. ^ *
: = "•
« I
1000
.-•
• K
=
?
T
•
; =
— ^. •-
a » •-
1200
_-=•
- = -
•• *
-___
--y,
f**
r c
C)r _
,^_
^r ~
c i
• •
• i
.r~
Sr=
rr "
0
0' »-
¥
Z
4
ft
4
rr
n
9
4
1
LEGEND
500-YEAR FLOOD
100-YEAR FLOOD
2KSWSSS STREAM BED
40
35
30
25
1400 1600 1800 2000 2200
STREAM DISTANCE IN FEET ABOVE THE CORPORATE LIMITS
Figure 11 - Data Extrapolation Method - Profile
1"=400'
Figure 12 - Data Extrapolation
Method - Plan View
-Property Y is approximately 370' upstream of the limit of
detailed study (as measured along the streamline). Using
the profile below, we can extrapolate the 100-year flood
profile to determine that the BFE for property Y is equal to
33'.
-Property Z is approximately 700' upstream of the limit of
detailed study (as measured along the streamline), and is
therefore beyond the limit of data extrapolation.
V-8
-------
Guide For Approximate Zone A Areas
Developing BFEs
100
s 95
§
LU
UJ
LL
z 90
ELEVATION I
§ £
_ »*
**
**
«•
•*
«•
^
ff
*
^
• —
?
S
35
=
T)
•i
W
=
•- •
9 •
• «
=
•-
*ri
a
•
K
S a
-* — 500' — >-
C
r
\L
f
c
c
!
V
•
J
\
^
b
3
t,
•
,)
E
?
I
i
t
i
).
r
1
^^S^N
LEGEND
500-YEAR FLOOD
100-YEAR FLOOD
STREAM BED
1500
100
95
90
85
2000 2500 3000 3500 4000 4500
STREAM DISTANCE IN FEET ABOVE THE CORPORATE LIMITS
5000
1"=1000'
Figure 13 - Data Extrapolation Method - Profile
ZONE A
Figure 14 -
Data Extrapolation
Method - Plan View
-State Route 27 may have an effect on the 100-year
water-surface elevations. Therefore, data
extrapolation should not be used to obtain a BFE
for property R.
V-9
-------
Guide For Approximate Zone A Areas
Developing BFEs
£.K3
s 21°
CD
ill
ill
u_
z 205
O
LU
m 200
lac:
**•
P*»
r~
^
**
**
•*
ff
**
f*
**
^
ff
**
^
^
,
+*
r*
ff
ff
*
**
"
ff
f
-.
ff-
•*
*"
f*
f*
fl
II
^*
ff
if,
"•
^
rf
^
f
tfl
*•
r*
ff
*
„
ff
*•
,*•
^
^
"
**
**
^
^
^
J
^
^*
^*
«*
^
J
f
•**
•*
•*
r
ff
*
^
*
f
•
I
(
^
•
f
i
<
K
•
^
1
)
i
[
•
^
>
•
ml*
LEGEND
500-YEAR FLOOD
100-YEAR FLOOD
STREAM BED
215
210
205
200
46 48 50 52 54 56 58
STREAM DISTANCE IN HUNDREDS OF FEET ABOVE THE CORPORATE LIMITS
Figure 15 - Data Extrapolation Method - Profile
60
V-10
-------
Guide For Approximate Zone A Areas Developing BFEs
Detailed Methods
Three essential factors must be determined either by hand
calculations or by computer model to determine a BFE by detailed
methods. These factors are: 1) floodplain geometry
(topography); 2) flood discharge and/or volume (hydrology); and
3) flood height (hydraulics).
Topography involves the measurement of the geometry of a cross
section(s) of the floodplain, which includes horizontal and
vertical coordinates. The vertical coordinate, or elevation, is
related to a vertical datum, such as the National Geodetic
Vertical Datum of 1929 or North American Vertical Datum of 1988.
The horizontal coordinate, or station, is measured from a
reference point along the cross section to establish actual
ground points.
Hydrology for the particular location along a stream involves
the determination of the peak rate of stream flow [usually
measured in cubic feet per second (cfs)] that will occur during
a flood (for purposes of determining the BFE, the 100-year
flood). When determining lake or pond elevations, a 100-year
flood hydrograph is required to determine the BFE.
Hydraulics involves the determination of the water-surface
elevation that will occur during a flood (for purposes of
determining the BFE, the 100-year flood), the selection of a
method to relate the flood discharge to a flood depth, and the
selection of Manning's roughness coefficients or "n" values.
These "n" values vary depending on the type of materials; degree
of irregularity; variation of shape, obstructions, and
vegetation; and degree of meandering related to the channel and
the floodplain of a stream.
The following sections discuss various methods for determining
the topography, hydrology, and hydraulics for a particular
location in order to determine a BFE.
Topography
Existing Topographic Maps
Before initiating field surveys, determine if there is existing
detailed topographic mapping that can be used to generate cross-
section data. To adequately describe a floodplain and for use
with a hydraulic method to calculate a BFE, topographic map
scales and contour intervals must be the same as, or more
detailed than, those used to prepare the
V-ll
-------
Guide For Approximate Zone A Areas Developing BFEs
community's Flood Insurance Study. The greater the level of
detail on the topographic map, the more accurate the BFE
determination. If the community does not have a Flood Insurance
Study, an existing topographic survey should, at a minimum, be
as detailed as the U.S. Geological Survey quadrangle map for the
area. Regardless of the level of detail of the existing
topographic map used, it is suggested that the geometry of the
actual stream channel be obtained by a site visit if the cross
sections are to be used for hydraulic analyses.
Datum Requirements for Field Surveys
If a greater level of detail is desired than is available from
existing topographic mapping, then a field survey should be
performed. If it is necessary to establish a BFE for insurance
purposes or to meet the requirements of 60.3 of the NFIP
Regulations, the survey must be referenced to the same datum
that is used to produce the FIRM, which is usually the National
Geodetic Vertical Datum of 1929 or the North American Vertical
Datum of 1988. Reference marks giving elevations to this datum
are given in the published Flood Insurance Studies. If the
reference marks cannot be located in the field, or are simply
too far away, additional reference mark information may be
obtained from the State's U.S. Geological Survey or
Transportation office. Local surveyors are generally familiar
with nearby reference marks. In approximate Zone A areas, if it
is not economically feasible to reference survey information to
a known reference mark, an assumed datum may be used, provided
that the BFE, structure, and lot elevations are referenced to
the same assumed datum; however, data developed using such an
assumed datum may not be sufficient to revise a FIRM. All
surveys must be certified by a registered professional engineer
or land surveyor.
If the sole purpose of determining relative flood heights is to
meet the requirements set forth in Section 60.3 (a) of the NFIP
regulations, any assumed datum may be used. In this instance, a
depth of flooding can be established at a particular location
without having to reference it to a datum (i.e., National
Geodetic Vertical Datum). However, in order for an insurable
structure to be eligible for a lower insurance rate based on the
BFE, the survey may need to be referenced to the same datum that
was used for the FIRM (i.e., National Geodetic Vertical Datum or
North American Vertical Datum).
V-12
-------
Guide For Approximate Zone A Areas Developing BFEs
Number of Cross Sections Required
If the determination of the BFE is for only one lot, one cross
section is required across the 100-year floodplain through the
property in question. For large parcels and multi-lot
subdivisions, at least one cross section is required at each end
of the parcel or subdivision. Additional cross sections must be
added if the difference in the computed 100-year water-surface
elevations at the two cross sections is more than one foot and
the distance between the cross sections is greater than 500
feet.
Proper Location of Cross Sections
The following guidelines should be used to determine the proper
location for cross sections:
Flow Path: Cross sections must be oriented perpendicular
to the anticipated flow path of the 100-year flood, as
shown in Figure 16, "Cross Section Orientation."
Channel Characteristics: Cross sections should be located
where changes in channel characteristics, such as slope,
shape, and roughness, occur.
Discharge: Cross sections should be located at points
along a stream where changes in flood discharge occur, such
as upstream of tributaries, as shown in Figure 17, "Locate
Cross Sections at Points of Flood Discharge Changes."
Structures: A minimum of two cross sections are required
to compute a BFE at or near a structure, such as a bridge
or dam. If the floodplain configurations upstream and
downstream of the structure are similar, two cross sections
may be used. One cross section should represent the
structure profile including the profile of the road or
embankment. When obtaining the structure profile in the
field, measurements of the structure opening, if there is
one, and any piers should also be obtained. The other
cross section should represent the natural valley cross
section downstream of the structure and should not include
any part of the structure or embankment. The natural
valley cross section should be located at a distance equal
to the width of the structure opening, W, measured from the
downstream foot of the embankment or wing walls, as shown
in Figure 18, "Cross Section Locations at Structures."
V-13
-------
Guide For Approximate Zone A Areas
Developing BFEs
Figure 16 - Cross Section Orientation
Drainage
area
boundary
Contours
Figure 17 - Locate Cross Sections at Points of Flood Discharge
Changes
V-14
-------
Guide For Approximate Zone A Areas
Developing BFEs
Roadway
Embankment
Roadway
Embankment
Figure 18 - Cross Section Locations at Structures
If the floodplain configurations upstream and downstream of
the structure are different and the structure is a bridge,
an additional cross section should be used upstream of the
structure. The cross section should be located at a
distance equal to the width of the structure opening
upstream of the structure as measured from the foot of the
embankment or wing walls.
The stations and elevations for cross section ground points
outside of the stream channel may be obtained from a
topographic map. The size of the structure opening, piers,
and channel geometry, however, should be obtained by field
survey.
Hydrology
There are a number of methodologies that may be used to develop flood
discharges for approximate Zone A areas. The methods discussed below
were selected because they are fairly simple to use, require
information that is easily obtainable, and provide reasonable
discharge estimates for streams where more detailed hydrologic
analyses have not been performed. These methods, which have been
ordered based on ease of use and expected level of accuracy, include
discharge-drainage area relationships, regression equations, the
NRCS TR-55 graphical peak discharge and tabular hydrograph methods,
V-15
-------
Guide For Approximate Zone A Areas Developing BFEs
and the rational formula. Other hydrograph methods will also be
noted but not described in detail due to their complexity.
Discharge-Drainage Area Relationships
This method is suggested for approximate Zone A areas because it
is straightforward and the only data needed are drainage areas
and corresponding 100-year flood discharges. These data can be
obtained from the Summary of Discharges table in a Flood
Insurance Study report.
The relationship between drainage area and discharge is non-
linear in most cases; therefore, the drainage areas and
corresponding 100-year flood discharges from the Flood Insurance
Study should be plotted on log-log paper as shown in Figure 20
from the example which begins on the following page. The
streams plotted may have varying drainage areas; however, other
watershed characteristics should be similar. A straight line
should be drawn through the plotted points as shown in Figure
21. The 100-year flood discharge for a particular location can
then be estimated based on the drainage area at the location as
shown in Figure 21 from the example.
Limitations - If the relationship of plotted points cannot be
approximated by a straight line, then this method should not be
used. In addition, this method is not appropriate when the
stream along which the site is located is regulated by dams,
detention ponds, canals, or other flow control structures or
diversions.
V-16
-------
Guide For Approximate Zone A Areas
Developing BFEs
EXAMPLE: DISCHARGE-DRAINAGE AREA RELATIONSHIPS
The following is a Summary of Discharges table from a Flood
Insurance Study report.
TABLE 1 - SUMMARY OF DISCHARGES
FLOODING SOURCE
AND LOCATION
DRAINAGE AREA
(sa. miles)
PINE CREEK
At confluence with
Saddle River 20.39
At Calvin Street 16.3
At Caitlin Avenue 14 . 9
ROCK RUN
Downstream of confluence
of Ramsey Brook 12.6
Upstream of confluence
of Ramsey Brook 10.1
GOOSE CREEK
Downstream of confluence
of Valentine Brook 9 .1
Upstream of confluence
of Valentine Brook 6 . 2
COON CREEK
Downstream of confluence
of Allendale Brook 14.3
Upstream of confluence
of Allendale Brook 12.9
PEAK DISCHARGES (cfs)
10-YEAR 50-YEAR 100-YEAR 500-YEAR
2,220
1, 907
1,860
4, 165
3,617
3,285
5,310
4,612
4,090
9,010
7,300
6,570
1,640
1,390
1,285
965
1, 805
1,670
2,895
2,455
2,270
1,700
3,185
2, 950
3,605
3,055
2,825
2,120
3, 965
3,670
5,795
4, 910
4,540
3,405
6,370
5, 900
Assume that Wendy Run is a stream within the same community as
the streams listed in the table, and that the Wendy Run drainage
basin, shown in Figure 19, has similar characteristics to the
stream basins from the table. First, plot the drainage areas
and corresponding 100-year discharges on log-log paper as shown
in Figure 20 on the following page. Then draw a straight line
through the plotted points as shown in Figure 21.
At Property A, the drainage area for Wendy Run is 8.5 square
miles. Using the drainage area curve created from the Flood
Insurance Study Summary of Discharges table, the 100-year
discharge at Property A is estimated to be 2,750 cfs, as shown
on Figure 21. At Property B, with a drainage area of 12.0
square miles, an estimated 100-year discharge of 3,600 is
obtained, as shown on Figure 21.
V-17
-------
Guide For Approximate Zone A Areas
Developing BFEs
Drainage
area
boundary
Contours
Figure 19 - Wendy Run Drainage Basin
10000
T3
o
o
CD
Q.
ft
0)
U-
o
ID
O
o
gi
CO
jr
o
co
OS
CD
Q.
1000
!0
Drainage Area (Square Miles)
Figure 20 - Discharge-Drainage Area Plot
V-18
-------
Guide For Approximate Zone A Areas
Developing BFEs
Discharge-Drainage Area Curve
T3
O
U
-------
Guide For Approximate Zone A Areas Developing BFEs
The general form of these regression equations is:
Q = K * Ax * By * Cz
where: Q = discharge (cfs)
K = regression equation constant
A,B,
and C = watershed variables
X,Y,
and Z = exponents
Watershed variables may include parameters such as drainage area
(in square miles), stream slope (in feet/mile), and impervious
area (in percent).
Limitations - Care must be taken when using these publications
because restrictions generally apply when the watershed is
heavily urbanized (i.e., high percentage of impervious land), or
where the runoff is regulated by the use of dams, detention
ponds, canals and other flow diversions. Other restrictions
based on the physical parameters of the watershed, such as
drainage area or stream slope, may also apply. Limitations of
these equations are normally stated in each report and should be
examined closely.
TR-55
The NRCS TR-55 "Urban Hydrology for Small Watersheds" contains
two methods for computing flood discharges: the Graphical Peak
Discharge method and the Tabular Hydrograph method. TR-55 is
straightforward in its approach and method of computation. TR-
55 takes into account the effects of urbanization, rainfall
distribution, soil types and conditions, ground cover types, and
other watershed characteristics. A method for estimating the
effects of storage on peak flood discharges is also included in
TR-55.
Limitations - In general, TR-55 should not be used in areas
where flow is divided between closed storm sewer systems and
overland conveyance areas, or where drainage areas exceed 2,000
acres. More specific limitations for using TR-55 are contained
in Chapters 2 through 6 of the NRCS TR-55 manual.
Rational Formula
This method estimates peak discharge rates for small watershed
areas not covered by regression equations and for areas where
the NRCS TR-55 method is not applicable. The Rational Formula
V-20
-------
Guide For Approximate Zone A Areas Developing BFEs
is based on the drainage area, rainfall intensity, watershed
time of concentration, and a runoff coefficient. The
generalized equation is:
Q = C * I * A
where: Q = discharge (cfs)
C = runoff coefficient
I = rainfall intensity (inches/hour)
A = drainage area (acres)
The runoff coefficient, C, varies with soil type, land use, and
terrain slope and can be obtained from text books on hydrology.
The intensity of rainfall, I, is determined based on the total
rainfall for a selected exceedence probability and a duration
equal to the time of concentration for the watershed. The time
of concentration for the watershed can be computed using the
method described in the NRCS TR-55 manual or methods described
in hydrology text books. For approximate Zone A areas, the
exceedence probability is equal to 1 percent (100-year storm
frequency). The 1 percent exceedence probability total rainfall
(100-year rainfall) for the computed duration can be obtained
from Technical Paper No. 40, Hydro 35, and precipitation-
frequency atlases published by the National Weather Service.
Dividing the total rainfall by the computed duration will yield
the intensity of rainfall.
Limitations - This method must not be used where the runoff is
regulated by the use of dams, detention ponds, canals and other
flow diversions. Also, this method is not recommended for
drainage areas greater than 200 acres, but can be used with
caution for drainage areas up to 640 acres (one square mile).
Other Hydrograph Methods
There are numerous other methods that can be used to determine
flood discharges based on rainfall-runoff relationships. The
following hydrograph methods are described in detail within
their respective technical reports and, therefore, will not be
described in detail within this guide. These methodologies in
general are good for any size watershed, and most of the methods
include computations that take into consideration areas where
the runoff is regulated by the use of dams, detention ponds,
canals and other flow diversions. These methods are recommended
for determining BFEs for ponds or lakes that are designated as
approximate Zone A. Besides TR-55, two of the more widely used
hydrograph methods are the NRCS' TR-20 and the U.S. Army Corps
of Engineers' HEC-1 computer programs.
V-21
-------
Guide For Approximate Zone A Areas Developing BFEs
TR-20 and HEC-1 provide a very detailed calculation of discharge
through the generation, addition, and routing of runoff
hydrographs. The effect on peak flood discharges due to dams,
road crossings, and large floodplain storage areas is more
accurately assessed with these programs. These models require
experience on the part of the user if they are to produce
realistic determinations of peak discharge.
Limitations - The limitations of these methods are thoroughly
described in their manuals. Because these methods involve many
variables and assumptions, the potential for error is great.
The users of these models must be thoroughly versed in the
limitations and assumptions of the computational methods
contained in these models. As with any synthetic model
depicting rainfall-runoff relationships, extreme care needs to
be taken to ensure that the results of the model are reasonable.
It is highly recommended that the discharges produced by these
hydrograph methods be compared to discharges produced by another
hydrologic method of equal accuracy or by calibrating the model
to an actual storm event.
Hydraulics
There are various hydraulic methods that may be used to
determine BFEs along riverine flooding sources. The appropriate
method to use depends on flow conditions and the size of the
area that is being analyzed. For developments of equal to or
less than 50 lots or 5 acres, the normal depth method, which is
described in greater detail below, is usually adequate for
determining BFEs. After normal depth has been computed, flow
conditions should be analyzed. If flow is classified as
subcritical (i.e., normal depth is greater than critical depth),
normal depth is used as the BFE. If flow is classified as
supercritical (i.e., normal depth is less than critical depth),
then critical depth is used as the BFE for natural channels.
For engineered channels, supercritical (normal) depth may be
used for the BFE, provided that the backwater from the normal
depth of the downstream cross section is considered properly.
If more than one cross section is required, step-backwater
computations should be used to determine BFEs along riverine
flooding sources.
The procedures for computing normal depth, critical depth, and
step-backwater by hand are outlined below. As an alternative to
hand calculations, the QUICK-2 computer program may be used.
V-22
-------
Guide For Approximate Zone A Areas Developing BFEs
QUICK-2 is a user-friendly computer program developed by FEMA
that may be used for normal depth, critical depth, or step-
backwater computations for regular or irregular shaped cross
sections. To aid the users of this guide in computing BFEs, the
QUICK-2 computer program and user's manual are located in
Appendix 6. The user's manual contains a tutorial section which
leads a new user through the calculation process using "real
life" examples. For those not using the QUICK-2 program, the
following sections on Normal Depth and Critical Depth illustrate
how to compute these depths by hand (see Appendix 8 for an
example of a Normal Depth hand calculation).
Normal Depth
Normal depth is the depth expected for a stream when the flow is
uniform, steady, one-dimensional, and is not affected by
downstream obstructions or flow changes. For uniform flow, the
channel bottom slope, water-surface slope, and energy slope are
parallel and are, therefore, equal. For normal depth
computations, the flow is considered steady because the
discharge is assumed to be constant; therefore, the depth of
flow does not change during the time interval under
consideration.
Normal depth calculations (also called the "slope/area method")
compute BFEs at a cross section. The standard formula for
determining normal depth at a cross section is Manning's
formula. The standard Manning's equation is:
Q = 1.486 x A x (R6") x S5 / n
where: Q = discharge (cfs)
A = cross section area (ft2)
R = hydraulic radius (ft) = A/WP
WP = wetted perimeter (ft)
S = energy slope (ft/ft)
n = Manning's roughness coefficient
The cross section area refers to the area below the water-
surface elevation, and the wetted perimeter refers to the length
of the ground surface along the cross section below the water-
surface elevation. The channel bottom slope is used in lieu of
the energy slope.
As noted earlier, Manning's "n" values vary depending on the
physical features of the stream channel and the channel
overbanks. The results of normal depth calculations can differ
V-23
-------
Guide For Approximate Zone A Areas Developing BFEs
significantly depending on the Manning's "n" values used;
therefore, care should be taken to ensure that the Manning's "n"
values selected accurately reflect conditions at the site being
analyzed. Manning's "n" values should be selected based on
field inspection, field photographs, and topographic mapping. A
list of accepted Manning's "n" values has been included in
Appendix 5. Various methods for computing normal depth are
described below.
Computer Programs for Computing Normal Depth
In addition to QUICK-2, the following Federal Government
computer programs have the capability to perform normal depth
computations:
Computer Program Agency
HEC-2 U.S. Army Corps of Engineers
HEC-RAS U.S. Army Corps of Engineers
WSPRO U.S. Geological Survey
WSP2 NRCS
SFD FEMA
PSUPRO FEMA
Please note that HEC-RAS is still being tested and had not yet
been released to the general public when this guide was
published. Furthermore, FEMA has not yet approved the model for
requests to revise NFIP maps. Please contact our Headquarters
office to determine the current status of HEC-RAS.
In addition to the above-referenced programs, there are other
engineering computer programs and models, which perform normal
depth calculations, that are available through various
commercial vendors. References for the hydraulic computer
programs listed above are in Appendix 7.
Normal Depth Hand Calculations
If a computer is not available, it is possible to perform hand
computations to calculate normal depth for the 100-year flood at
a cross section by following steps 1-11 listed below.
Step 1 - Obtain a topographic map or conduct a field survey to
obtain a cross section at the site where normal depth
should be determined. If a topographic map is used,
the channel geometry should be obtained from
measurements taken in the field. The cross section
should be oriented perpendicular to the expected 100-
year floodplain.
V-24
-------
Guide For Approximate Zone A Areas Developing BFEs
Step 2 - Compute the 100-year discharge by applying one of the
methods described in the hydrology section of this
guide.
Step 3 - Plot the cross section on graph paper with the
stations and the corresponding elevations. (The
stations and elevations are obtained from the
topographic map and/or from field survey).
Step 4 - Select the left and right channel bank stations. The
channel bank stations are those stations where the
ground slope becomes flatter moving away from the
channel bottom as shown in Figure 22, "Channel Bank
Stations." Photographs taken in the field and the
contours on the topographic maps are also helpful when
defining the channel bank stations. Do not place the
channel bank stations at the bottom of the channel.
Figure 22 - Channel Bank Stations
Step 5 - Select appropriate Manning's roughness coefficients
for the left overbank, channel, and right overbank
from the "n" values given in Appendix 5. These values
should be determined by reviewing the field
photographs and visiting the site.
Step 6 - Compute the cross section area, wetted perimeter,
hydraulic radius, and conveyance for each segment
(i.e., left overbank, channel, and right overbank) for
at least three elevations. The conveyance, K, of a
segment is given as:
K = (1.486/n) x A x R
where: A = cross section area (ft2
R = hydraulic radius (ft)
WP = wetted perimeter (ft)
R = A/WP
V-25
-------
Guide For Approximate Zone A Areas Developing BFEs
Step 7 - Compute the channel bottom slope, S, from the
topographic map or from field survey.
Step 8 - Compute the discharge, Q, for each segment of the
cross section at each elevation by multiplying K by
S°'5.
Step 9 - Add the discharges from each segment at the same
elevation to obtain the total discharge.
Step 10 - Plot the total discharges and the corresponding
elevations on graph paper.
Step 11 - The BFE can be determined from this graph for the 100-
year flood discharge computed in Step 2.
An example of a normal depth hand calculation is included in
Appendix 8.
Critical Depth
After computing normal depth, the type of flow should be
checked. If the velocity head from the normal depth computation
is equal to or more than half the hydraulic depth, the flow is
supercritical and the critical depth should be used to establish
the BFE. The velocity head, HV, for an irregular cross section
is computed using the following equation:
HV = aV2/2g
where: a = velocity coefficient
V = mean velocity = QT / A,, (fps)
QT = total discharge (cfs)
A,, = total flow area (ft2)
g = acceleration due to gravity = 32.2 ft/sec2
The velocity coefficient, a, is determined using the following
equation:
F (K/A) + (K3/A2) +
V-26
-------
Guide For Approximate Zone A Areas Developing BFEs
where: K^, Kc, K^, K^ = conveyance for left overbank,
channel, right overbank, and
total conveyance, respectively
A^, Ac, \, A,, = flow area for left overbank,
channel, right overbank, and
total flow area, respectively
Hydraulic depth, h, is computed by using the following
relationship:
h = A, / T
where: T = top water-surface width at the normal depth
A,, = Total Flow Area
If the velocity head is greater than or equal to one-half the
hydraulic depth, the flow is supercritical.
For prismatic channels, the following equation can be used to
determine the critical depth:
Q! = A! or Q = VgA3 / T
g T
For a series (3 or more) of water-surface elevations, compute
the corresponding total area, A, water-surface topwidth, T, and
the critical discharge, Q, using Q = VgA3 / T. Compute the
value of right hand side of the above equation. Plot the water-
surface elevations and the corresponding discharge values on
graph paper. The critical water-surface elevation and,
therefore, critical depth, can be determined from this graph for
a range of discharge values.
For rectangular channels, critical depth can be computed
directly from the above equation and is expressed in the
following relationship:
Dc = { Q / (5.67 T) }
The energy is minimum at the critical depth. For irregular
cross sections, critical depth is determined from the
relationship between the water-surface elevation and the energy.
The energy is computed by adding the water-surface elevation and
the corresponding velocity head (or energy grade elevation).
For irregular cross sections, the velocity coefficient, a (a),
must be considered when computing velocity head (HV) . Several
water-surface elevations should be assumed and corresponding
energy grade elevations computed. These values are then plotted
V-27
-------
Guide For Approximate Zone A Areas Developing BFEs
on a graph of water-surface elevation versus energy grade
elevation. The critical water-surface elevation and, therefore,
critical depth, can be determined from this graph where the
energy (i.e., energy grade elevation) is minimum.
Step-Backwater Analysis
Step-backwater computations are based on the principle of
conservation of energy, which states that the energy at the
upstream cross section is equal to the energy at the downstream
cross section plus the losses between the two cross sections.
The losses considered in the step-backwater analysis are the
friction loss and the transition loss.
The equations and the procedure used in the step-backwater
analysis are explained in the QUICK-2 user's manual in Appendix
6. Although hand computations can be done to perform the step-
backwater analysis, it is advisable to use the QUICK-2 program
or other Federally approved programs for ease of computation.
The QUICK-2 program currently does not model the effects of
bridges or culverts or supercritical flow.
The QUICK-2 program uses the default friction slope method,
which is the average conveyance method, from the HEC-2 program
to compute friction losses. For transition losses, a
contraction coefficient of 0.1 and an expansion coefficient of
0.3 should be used in the computations.
The reach lengths between the two cross sections for the left
overbank, channel, and right overbank are required for step-
backwater computations. The distance for the left overbank
should be measured between the center of the floodplains of the
left overbank at each cross section. The same is true for the
right overbank. The channel distance should be measured along
the streambed, and therefore will account for the meandering of
the stream channel.
In general, starting water-surface elevations are obtained from
normal depth computations (slope/area method) at the first cross
section. If there is a structure downstream of the study area,
the backwater effects of the structure must be considered in
determining the starting water-surface elevation. If there is a
known 100-year water-surface elevation at the downstream end of
the study area, that water-surface elevation should be used as
the starting water-surface elevation.
V-28
-------
Guide For Approximate Zone A Areas Developing BFEs
Hydraulic Structures
As stated earlier, normal depth is the depth expected for a
stream when the flow is uniform, steady, one-dimensional, and is
not affected by downstream obstructions or flow changes.
However, there are situations in which a physical structure
located downstream of a particular site will cause an
obstruction or alteration of the flow, resulting in a flood
depth at the site higher than the normal depth. The discussion
below describes the appropriate methods for determining BFEs for
reaches that include hydraulic structures.
Hydraulic structures that are common in approximate Zone A areas
include road and railroad crossings, including embankments,
dams, bridges and culverts, and canal crossings. The flow over
the road, railroad, embankment, dam or canal can be described as
weir flow. Weir flow can be calculated by hand or by computer
program in order to determine the BFE. When flow passes through
a bridge or culvert, the BFE can be determined through the use
of nomographs or computer programs. The BFE at a structure
where flow travels through a bridge or culvert and over the
crossing can be determined by nomographs, but is more easily
determined with a computer program.
Weir Flow
Determination of the water-surface elevation for weir flow
requires at least two cross sections. The first cross section
represents the natural valley section downstream of the
structure, and the second cross section represents the road
profile and the opening of the structure (refer to Figure 18,
"Cross Section Locations at Structures." If the approach
velocity head is to be considered, then a third cross section is
required that represents the natural valley section upstream of
the structure. In most situations, however, the velocity head
can be assumed to be negligible, and a third cross section is
not necessary.
The water-surface elevation downstream of the structure should
be determined by using normal depth computations at the first
cross section, provided there are no structures further
downstream that can create backwater effects (refer to the
methods for determining normal depth described previously).
The second cross section, which represents the profile along the
top of the structure including the road or the embankment,
should be used to determine the weir length for use in the
equation for weir flow, as shown on the following page.
V-29
-------
Guide For Approximate Zone A Areas
Developing BFEs
Q =
x C x L x H
where: Q = discharge (cfs)
@ = submergence factor
C = weir coefficient, varies from 2.6 to 3.0 and can
be obtained from hydraulic text books
L = weir length (ft)
H = available head (ft), measured from top of weir to
the selected energy grade elevation
Several values for H should be selected and the corresponding
discharge computed until the total weir flow is larger than the
100-year flood discharge. Plot the discharges and the
corresponding energy grade elevations on graph paper. The 100-
year flood energy grade elevation can be determined from this
graph. For an approximate analysis, the computed energy
gradient elevation can be considered the BFE.
If the structure profile is not horizontal, as shown in Figure
23, "Weir Flow - Embankment Profile is Not Horizontal," several
structure segments should be used and an average energy depth,
H, for that segment should be determined for use in the above
equation for selected energy grade elevations. The sum of the
weir flow from each segment will then be equal to the total weir
flow for the selected energy grade elevation.
ROAD OR EMBANKMENT PROFILE
ENERGY GRADE LINE
L(3) - I - L(4)
Q = (@ CL(1) HAVG(1)3/2) + (@ CL(2) HAVG(2)3/2) + (@ CL(3) HAVG(3)3/2
(@ CL(4) HAVG(4)3/2)
Figure 23 - Weir Flow - Embankment Profile is Not Horizontal
V-30
-------
Guide For Approximate Zone A Areas
Developing BFEs
If the downstream water-surface elevation is higher than the
minimum road elevation, a submergence factor may be considered
in the weir flow computation. The submergence factor is
dependent upon the D/H ratio, where D is the downstream depth of
water above the road and H is the upstream energy grade depth
above the road, as shown in Figure 24, "Weir Flow Over Road."
The submergence factor must be considered when the D/H ratio is
more than 0.79. For a non-horizontal road profile, the D/H
ratio must be computed for each road segment. The submergence
factor, @, can be determined from the curve in "Hydraulics of
Bridge Waterways" (Reference 1, Figure 24) and some typical
values are given in the table below.
Figure 24 - Weir Flow Over Road
D/H
0.998
0.992
0.986
0.976
0.962
@
0.30
0.40
0.50
0.60
0.70
D/H
0.944
0.932
0.915
0.889
0.700
@
0.80
0.85
0.90
0.95
1.00
Other procedures used in Federal agency backwater computer programs
can also be used to determine the submergence factor.
A third cross section may be used to determine a more accurate water-
surface elevation upstream of the structure. This may be done by
assuming water-surface elevations and calculating the corresponding
velocity heads (HV) until an assumed water-surface elevation plus its
velocity head at that elevation equal the same energy gradient
elevation obtained from the weir flow equation. The velocity head,
HV, can be calculated using the following equation:
HV = a (Q/A)2/2g
V-31
-------
Guide For Approximate Zone A Areas Developing BFEs
where: a = velocity coefficient
Q = 100-year flood discharge (cfs)
A = cross section area (ft2) at the assumed
water-surface elevation
g = Acceleration due to gravity = 32.2 ft/sec2
An example of a weir flow computation is included in Appendix 9.
Flow Through Structures
Culverts
At least two cross sections are required to determine the water-
surface elevation upstream of a culvert. The first cross section
should represent the natural valley cross section downstream of the
culvert, and the second cross section should represent the top of the
embankment profile and the opening of the structure (refer to Figure
13, "Cross Section Locations at Structures"). The size, type,
length, and upstream and downstream invert elevations of the culvert
should be obtained by field survey. The wing wall angle and the
entrance opening configuration, such as sharp edge or rounded edge,
should also be determined from a field survey. The Federal Highway
Administration publication "Hydraulic Design of Highway Culverts"
(Reference 2) should be referenced before going to the field so that
all the necessary information for culvert flow computations can be
collected during one field survey. Water-surface elevations upstream
of the culvert can then be computed using the nomographs contained in
the above-mentioned publication and the procedures outlined below.
The first cross section should be used to determine the normal depth
downstream of the culvert, which will be considered as the tailwater
(refer to section on normal depth computations).
Two computations are required to determine the headwater when using
Federal Highway Administration nomographs. One computation is for
inlet control, and the other computation is for outlet control. The
headwater elevations from the two computations are then compared.
The higher of the two should be selected as the upstream headwater
elevation. If this headwater elevation is higher than the top of
embankment profile, weir flow will occur. Perform at least three
weir flow computations for headwater elevations between the headwater
that assumes that all the flow is culvert flow (the first
computation) and the minimum top of embankment elevation. For each
selected headwater elevation, compute the culvert flow using Federal
Highway Administration nomographs. Combine the weir flow and culvert
flow for each selected headwater elevation and plot on graph paper.
The BFE for the 100-year flood discharge can then be obtained from
this graph.
V-32
-------
Guide For Approximate Zone A Areas Developing BFEs
If the site in question is not located immediately upstream of a
structure, a normal depth should be computed at the site. The 100-
year water-surf ace elevation at the site should be the higher of the
two elevations from the culvert computation and the normal depth
computation.
Federal Highway Administration nomographs predict only the energy
grade elevation upstream of the culverts. In most applications, the
velocity head is assumed to be negligible and, therefore, the energy
grade elevation approximates the actual water-surface elevation. If
a more accurate water-surface elevation is desired, a hydraulic
computer model, such as HEC-2, should be used to determine the BFE.
The procedure outlined in the weir flow section to compute a water-
surface elevation that corresponds to a certain energy grade
elevation may also be used to determine a BFE upstream of a culvert.
Bridges
Although hand computations can be performed by following the
procedures for bridge routines in Federal agency computer models, it
is recommended that the water-surface elevation upstream of bridges
be determined using a computer model. The number of cross sections
required at the structure depends upon the type of bridge routine
used. Four cross sections are required if the special bridge routine
in the HEC-2 program is used, and six cross sections are required if
the normal bridge routine in the HEC-2 program is used. Three cross
sections are required if the bridge routines in the WSPRO program and
the WSP2 program are used. A step-backwater analysis is also
required to compute the water-surface elevations with these bridge
routines. The following programs are recommended to compute the
water-surface elevation upstream of a bridge:
Computer Program Agency
HEC-2 U.S. Army Corps of Engineers
*HEC-RAS U.S. Army Corps of Engineers
WSPRO U.S. Geological Survey
WSP2 NRCS
*Not available for general use when this guide was published;
please contact our Headquarters office for current status.
V-33
-------
Guide For Approximate Zone A Areas Developing BFEs
REFERENCE
I. U.S. Department of Transportation, Federal Highway
Administration, Hydraulics of Bridge Waterways, Washington,
B.C., March 1978.
2. U.S. Department of Transportation, Federal Highway
Administration, Hydraulic Design of Highway Culverts,
Washington, B.C., September 1985.
V-34
-------
Guide For Approximate Zone A Areas Letters of Map Change
VI. OBTAINING LETTERS OF MAP CHANGE
Once detailed methods have been applied to develop BFE data, these
data may be suitable for revising an NFIP map via a Letter of Map
Correction. On October 1, 1992, FEMA implemented the use of
detailed application and certification forms for requesting
revisions to NFIP maps. Therefore, if a map revision is requested,
the appropriate forms should be submitted.
FEMA has implemented a procedure to recover costs associated with
reviewing and processing requests for modifications to published
flood information and maps. Specific information about these fees
is presented in the application and certification forms.
These forms, along with other useful documents pertaining to the
NFIP, may be obtained from our technical evaluation contractors at
the addresses listed below:
FEMA Regions I-V
Dewberry & Davis
Management Engineering and
Technical Services Division
8401 Arlington Boulevard
Fairfax, Virginia 22031
FAX: (703) 876-0073
FEMA Regions VI-X
Michael Baker, Jr., Inc.
3601 Eisenhower Avenue
Suite 600
Alexandria, Virginia 22304
FAX: (703) 960-9125
This information is also available through the FEMA Regional
Offices listed in Appendix 3.
To provide additional assistance to those who develop BFE data, a
worksheet that synopsizes the procedures detailed in this guide is
found in Appendix 10.
VI-1
-------
Appendix 1
Glossary of Floodplain Analysis Terms
1-Percent Annual Chance Flood: the flood that has a one-percent chance of being
equaled or exceeded on the average in any given year; equivalent to the 100-year
flood.
100-Year Flood: the flood that is equaled or exceeded once in 100 years on the
average; equivalent to the one percent annual chance flood.
Alluvial Stream: a stream that has formed its channel by the process of
aggradation. The sediment in the stream is similar to the material in the bed
and banks.
Base Flood: the flood having a one percent chance of being equalled or exceeded
in any given year (the 100-year flood).
Base Flood Elevation (BFE): the water-surface elevation associated with the base
flood.
Conveyance: a measure of the carrying capacity of the channel section. Flow
(Discharge (Q) ) is directly proportional to conveyance (K) . The proportional
factor is the square root of the energy slope; expressed as Q = K * SK.
Cross Section: a vertical profile of the ground surface taken perpendicular to
the direction of flood flow. The profile is defined by coordinates of ground
elevation and horizontal distance (station).
Discharge: a measure of flow volume per unit of time. In hydrology, units of
flow are usually cubic feet per second (cfs).
Exceedence Frequency: the frequency that a flood of a certain discharge will be
equaled or exceeded in any given year; equal to the inverse of the recurrence
interval.
Flood: (a) a general and temporary condition of partial or complete inundation of
normally dry land areas from: (1) the overflow of inland or tidal waters; (2) the
unusual and rapid accumulation or runoff of surface waters from any source;
(3) mudslides (i.e., mudflows), which are proximately caused by flooding as
defined in (a)(2) above and are akin to a river of liquid and flowing mud on the
surfaces of normally dry land areas, as when earth is carried by a current of
water and deposited along the path of the current. (b) The collapse or
subsidence of land along the shore of a lake or other body of water as a result
of erosion or undermining caused by waves or currents of water exceeding
anticipated cyclical levels or suddenly caused by an unusually high water level
in a natural body of water, accompanied by a severe storm, or by an unanticipated
force of nature, such as flash flood or abnormal tidal surge, or by some
similarly unusual and unforeseeable event, which results in flooding as defined
in (a) (1) above.
Flood Crest: the maximum height of a flood, usually measured as an elevation or
depth.
Flood Hazard: the potential for inundation that involves the risk to life,
health, property, and natural floodplain values.
Al-1
-------
Appendix 1 - continued
Glossary of Floodplain Analysis Terms
Floodplain: any land area, such as the lowland and relatively flat areas
adjoining inland and coastal waters, susceptible to being inundated by water from
any source.
Floodway: the channel of a river or other watercourse and the adjacent land areas
that must be reserved in order to discharge the base flood without cumulatively
increasing the water-surface elevation more than a designated height. The base
flood is defined as the one-percent chance flood and the designated height is
usually one foot above the base flood elevation; however, this height may vary
(but is not more than one foot) depending on what the State has adopted.
Floodway Fringe: the area between the floodway boundary and the 100-year
floodplain boundary.
Flow: equivalent to discharge.
Flow Area: the cross section (see discharge) area of the floodplain below a given
water-surface elevation.
Hazardous Flow: conditions that exist when the product of the depth of flow and
its corresponding velocity are greater than ten (10) . For example a flow depth
of 3 feet and a flow velocity of 4 feet per second (3x4 = 12) would be
considered hazardous flow.
Hydraulic Depth: an average depth computed as the Flow Area divided by the top
width of the floodplain for a given water-surface elevation.
Lacustrine Flooding: Flooding produced by a lake or pond.
Peak Discharge: the maximum instantaneous discharge of a flood at a given
location.
Recurrence Interval: the average interval of time required for a flood of a
specific discharge to occur or be exceeded; equal to the inverse of the
exceedence frequency.
Riverine Flooding: Flooding produced by a river or stream.
Shallow Flooding: a designated AO, AH, or VO zone on a community's Flood
Insurance Rate Map with a one percent or greater annual chance of flooding to an
average depth of one to three feet where a clearly defined channel does not
exist, the path of flooding is unpredictable, and velocity flow may be evident.
Such flooding is characterized by ponding or sheet flow.
Slope (Energy): the rate of energy loss of a watercourse.
Slope (Ground): the change in vertical ground elevation over a horizontal
distance, usually based on the change in the vertical elevation of the stream
bottom.
Steady Flow: state of flow where the depth of flow does not change with time.
Subcritical Flow: state of flow where the gravitational forces are more
pronounced than the inertial forces. The flow tends to have a low velocity. In
general, in this flow regime, the hydraulic depth is more than twice the velocity
head.
Supercritical Flow: state of flow where the inertial forces become dominant. The
flow tends to have a high velocity. In general, in this flow regime, the
velocity head is equal to or more than half the hydraulic depth.
Al-2
-------
Appendix 1 - continued
Glossary of Floodplain Analysis Terms
Unsteady Flow: state of flow where the depth of flow changes with time.
Uniform Flow: depth is constant over channel length, and the channel shape, slope
and boundary roughness are constant over the channel length.
Varied Flow: depth of flow changes along the channel length.
Gradually Varied Flow: depth of flow changes gradually over the channel length.
Rapidly Varied Flow: depth changes abruptly over a short channel length.
Velocity: a rate of movement (i.e., distance divided by time). For water, the
rate is expressed in feet per second. Because water in a channel does not all
move at the same velocity at every point, an average value is used to described
flow velocity. This average velocity equals the discharge divided by the flow
area (Q/A).
Velocity Head: the kinetic energy term (a V2 / 2g) , in the total energy of flow.
The velocity coefficient (a) is used to adjust for the distribution of velocity
in a cross section of differing roughness.
Al-3
-------
Appendix 2
Flood Insurance Study Data Request Form
FLOOD INSURANCE STUDY (FIS) DATA REQUESTS
Requests for FIS data should be made in writing to:
Regions I-V Regions VI-X
Flood Insurance Information Specialist FEMA Project Library
c/o Dewberry & Davis c/o Michael Baker, Jr., Inc.
2953 Prosperity Avenue 3601 Eisenhower Avenue
Fairfax, Virginia 22031 Suite 600
FAX: (703) 876-0073 Alexandria, Virginia 22304
FAX: (703) 960-9125
The following information should be included in your written request:
• Complete community name (including county)
• Community Identification Number
• Name(s) of flooding source(s) and specific location(s) for which
data are needed
• Specific data needed:
HEC-2 input and output files
Topographic data
etc.
• Effective date of FIRM/FBFM for which data are requested (enclose an
annotated copy of FIRM/FBFM if available identifying area of interest)
• Agreement to pay costs associated with processing the request
• Fee limit after which authorization is needed
• Contact person's name, address, and phone number
The average request takes approximately 2 to 4 weeks to fill and may cost between
$100 to $200.
You will be contacted after we have determined if the data are available and the
cost to fill the request has been determined.
Do not include payment with your request letter.
Checks or money orders should be made payable to the National Flood Insurance
Program and sent to:
Federal Emergency Management Agency
Fee Collection System
P.O. Box 398
Merrifield, Virginia 22116
Data will be released upon receipt of payment.
A2-1
-------
Appendix 3
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Federal Emergency Management Agency Offices
HEADQUARTERS
500 C Street, SW
Washington, D.C. 20472
(202) 646-3680
FAX: (202) 646-4596
REGION I
(Connecticut, Maine, Massachusetts, New Hampshire, Rhode Island & Vermont)
J.W. McCormack Post Office & Courthouse Building, Room 462
Boston, MA 02109
(617) 223-9561
FAX: (617) 223-9574
REGION II
(New Jersey, New York, Puerto Rico & Virgin Islands)
26 Federal Plaza, Room 1349
New York, NY 10278
(212) 225-7200
FAX: (212) 225-7262
REGION III
(Delaware, District of Columbia, Maryland, Pennsylvania, Virginia & West
Virginia)
Liberty Square Building, Second Floor
105 South Seventh Street
Philadelphia, PA 19106
(215) 931-5512
FAX: (215) 931-5501
REGION IV
(Alabama, Florida, Georgia, Kentucky, Mississippi, North Carolina, South Carolina
& Tennessee)
1371 Peachtree Street, N.E., Suite 700
Atlanta, GA 30309
(404) 853-4400
FAX: (404) 853-4440
REGION V
(Illinois, Indiana, Michigan, Minnesota, Ohio & Wisconsin)
175 West Jackson Boulevard
Fourth Floor
Chicago, IL 60604
(312) 408-5552
FAX: (312) 408-5551
A3-1
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Federal Emergency Management Agency Offices (continued)
REGION VI
(Arkansas, Louisiana, New Mexico, Oklahoma & Texas)
Federal Regional Center
800 North Loop 288
Denton, TX 76201-3698
(817) 898-5165
FAX: (817) 898-5195
REGION VII
(Iowa, Kansas, Missouri & Nebraska)
Federal Office Building, Room 300
911 Walnut Street
Kansas City, MO 64106
(816) 283-7002
FAX: (816) 283-7018
REGION VIII
(Colorado, Montana, North Dakota, South Dakota, Utah & Wyoming)
Denver Federal Center, Bldg. 710
P.O. Box 25267
Denver, CO 80225-0267
(303) 235-4830
FAX: (303) 235-4849
REGION IX
(Arizona, California, Hawaii & Nevada)
Presidio of San Francisco
Building 105
San Francisco, CA 94129
(415) 923-7100
FAX: (415) 923-7147
REGION X
(Alaska, Idaho, Oregon & Washington)
Federal Regional Center
130 - 228th Street, SW
Bothell, WA 98021-9796
(206) 487-4678
FAX: (206) 487-4613
A3-2
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
ALABAMA
Alabama Department of Economics
and Community Affairs
State Planning Division
401 Adams Avenue
Montgomery, AL 36103
(205) 242-5442
U.S. Geological Survey
District Chief
Water Resources Division
520 19th Avenue
Tuscaloosa, AL 35401
(205) 752-8104
U.S. Department of Agriculture
Natural Resources Conservation
Service
665 Opelika Rd.
P.O. Box 311
Auburn, AL 36830
(205) 887-4506
NFIP State Coordinator
Mr. Gene Anderson
Director, Alabama Department
of Economic and Community Affairs
P.O. Box 5690
401 Adams Avenue
Montgomery, AL 36103-5690
(205) 242-5499
ALASKA
Alaska Department of
Community and Regional Affairs
Municipal and Regional
Assistance Division
333 West 4th Avenue, Suite 220
Anchorage, AK 99501
(907) 269-4500
U.S. Geological Survey
District Chief
Water Resources Division
4230 University Drive, Suite 201
Anchorage, AK 99508-4138
(907) 786-7100
U.S. Department of Agriculture
Natural Resources Conservation
Service
949 East 36th Avenue
Suite 400
Anchorage, AK 99504
(907) 271-2424
NFIP State Coordinator
Mr. Bob Walsh
Municipal and Regional
Assistance Division
333 West 4th Avenue, Suite 220
Anchorage, AK 99501
(907) 269-4500
ARIZONA
Arizona Department of Water Resources
15 South 15th Avenue
Phoenix, AZ 85004
(602) 242-1553
U.S. Geological Survey
District Chief
Water Resources Division
375 South Euclid
Tucson, AZ 85719
(602) 670-6671
U.S. Department of Agriculture
Natural Resources Conservation
Service
3008 Federal Building
230 N. 1st Avenue
Phoenix, AZ 85025
(602) 261-6711
NFIP State Coordinator
Ms. Elizabeth A. Rieke
Director, Arizona Department
of Water Resources
15 South 15th Avenue
Phoenix, AZ 85007
(602) 542-1540
ARKANSAS
Arkansas Soil and Water
Conservation Commission
1 Capitol Mall, Suite 2D
Little Rock, AR 72201
(501) 371-1611
A3-3
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Geological Survey
Water Resources Division
401 Hardin Road
Little Rock, AR 72211
(501) 228-3600
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Office Bldg.
700 West Capitol
P.O. Box 2323
Little Rock, AR 72203
(501) 324-6335
NFIP State Coordinator
Mr. Randy Young
Director
Arkansas Soil & Water
Conservation Commission
101 East Capitol
Little Rock, AR 72201
(501) 682-1611
CALIFORNIA
California Department of
Water Resources
P.O. Box 942836
Sacramento, CA 94236-0001
(916) 653-5791
U.S. Geological Survey
District Chief
Water Resources Division
Federal Building, Room W-2233
2800 Cottage Way
Sacramento, CA 95825
(916) 978-4633
U.S. Department of Agriculture
Natural Resources Conservation
Service
2121 C 2nd Street
Davis, CA 95616
(916) 757-8200
NFIP State Coordinator
Mr. David Kennedy, Director
California Department of
Water Resources
P.O. Box 942836
Sacramento, CA 94236-0001
(916) 653-7007
COLORADO
Urban Drainage and Flood Control
District
2480 West 26th Avenue
Suite 156B
Denver, CO 80211
Colorado Water Conservation Board
State Centennial Building, Room 721
1313 Sherman Street
Denver, CO 80203
(303) 866-3441
U.S. Geological Survey
District Chief
Water Resources Division
Denver Federal Center, Building 53
Box 25046 (Mail Stop 415)
Lakewood, CO 80225-0046
(303) 236-4882
U.S. Department of Agriculture
Natural Resources Conservation
Service
655 Parfait Street
Room E200C
Lakewood, CO 80215
(303) 236-2886
NFIP State Coordinator
Mr. Daries C. Lile, P.E.
Director, Colorado Water
Conservation Board
State Centennial Building
1313 Sherman Street
Denver, CO 80203
(303) 866-3441
CONNECTICUT
State Department of
Environmental Protection
79 Elm Street, 3rd Floor
Hartford, CT 06106
(203) 424-3706
U.S. Geological Survey
Hydrologist-in-Charge
Connecticut Office
Water Resources Division
Abraham A. Ribicoff Federal
Building, Room 525
450 Main Street
Hartford, CT 06103
(203) 240-3060
A3-4
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Department of Agriculture
Natural Resources Conservation
Service
16 Professional Park Road
Storrs, CT 06268
(203) 429-9361
NFIP State Coordinator
Mr. Timothy Keeney, Commissioner
State Department of Environmental
Protection
165 Capitol Avenue
State Office Building
Hartford, CT 06106
(203) 566-2110
DELAWARE
Department of Natural Resources and
Environmental Control
Division of Soil and Water
Conservation
P.O. Box 1401
89 Kings Highway
Dover, DE 19903
(302) 739-4403
U.S. Geological Survey
Hydrologist-in-Charge
Delaware Office
Water Resources Division
Federal Building, Room 1201
300 South New Street
Dover, DE 19904
(302) 734-2506
U.S. Department of Agriculture
Natural Resources Conservation
Service
3500 South DuPont Highway
Dover, DE 19901
(302) 697-6176
NFIP State Coordinator
Mr. John A. Hughes, Director
Delaware Department of Natural &
Environmental Control
Richardson and Robbins Building
P.O. Box 1401
Dover, DE 19903
(301) 736-4411
DISTRICT OF COLUMBIA
Department of Consumer
Regulatory Affairs
614 H Street Northwest
Washington, DC 20001
(202) 727-7170
U.S. Geological Survey
District Chief
Water Resources Division
208 Carroll Building
8600 La Salle Road
Towson, MD 21286
(410) 828-1535
NFIP State Coordinator
Mr. Donald G. Murray, Director
Department of Consumer Regulatory
Affairs
Office of the Director
614 H Street, NW., Suite 1120
Washington, D.C. 20001
(202) 727-7170
FLORIDA
Department of Community Affairs
East Howard Building
2740 Centerview Drive
Tallahassee, FL 32399-2100
(904) 488-8466
U.S. Geological Survey
District Chief
Water Resources Division
227 North Bronough Street,
Suite 3015
Tallahassee, FL 32301
(904) 942-9500
U.S. Department of Agriculture
Natural Resources Conservation
Service
P.O. Box 141510
Gainesville, FL 32614
(904) 338-9500
NFIP State Coordinator
Ms. Linda Lomis Shelley, Secretary
Florida Department of Community
Affairs
2740 Centerview Drive
Tallahassee, FL 32399-2100
(904) 488-8466
A3-5
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
GEORGIA
Department of Natural Resources
Environmental Protection Division
Floyd Towers East, Suite 1252
205 Butler Street Southeast
Atlanta, GA 30334
(404) 656-4713
U.S. Geological Survey
District Chief
Water Resources Division
3039 Amwiler Road, Suite 130
Atlanta, GA 30360
(404) 447-9803
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building
355 East Hancock Avenue
P.O. Box 832
Athens, GA 30613
(404) 546-2272
NFIP State Coordinator
Mr. Joe D. Tanner, Commissioner
Georgia Department of Natural
Resources
205 Butler Street, S.E.
Floyd Towers East, Suite 1252
Atlanta, GA 30334
(404) 656-3500
HAWAII
Hawaii Board of Land and
Natural Resources
1151 Punchbowl Road, Room 220
Honolulu, HI 96813
(808) 587-0446
U.S. Geological Survey
District Chief
Water Resources Division
677 Ala Moana Boulevard, Suite 415
Honolulu, HI 96813
(808) 522-8290
U.S. Department of Agriculture
Natural Resources Conservation
Service
300 Ala Moana Boulevard
P.O. Box 50004
Honolulu, HI 96850
(808) 546-3165
NFIP State Coordinator
Mr. William W. Paty, Chairperson
Commission on Water Resource
Management and Board of Land and
Natural Resources
State of Hawaii
P.O. Box 621
Honolulu, HI 96809
(808) 587-0401
IDAHO
Department of Water Resources
State House
1301 North Orchard Street
Boise, ID 83706
(208) 327-7900
U.S. Geological Survey
District Chief
Water Resources Division
230 Collins Road
Boise, ID 83702
(208) 387-1300
U.S. Department of Agriculture
Natural Resources Conservation
Service
3244 Elder Street
Room 124
Boise, ID 83705
(208) 334-1601
NFIP State Coordinator
Mr. R. Keith Higginson, Director
Idaho Department of Water Resources
1301 N. Orchard
Boise, ID 83706
(208) 327-7900
ILLINOIS
Illinois Department of Transportation
Local Flood Plains Programs
310 South Michigan, Room 1606
Chicago, IL 60604
(312) 793-3123
U.S. Geological Survey
District Chief
Water Resources Division
Champaign County Bank Plaza
102 East Main Street
Fourth Floor
Urbana, IL 61801
(217) 398-5353
A3-6
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building
2110 West Park Court
Suite C
Champaign, IL 61821
(217)398-5212
NFIP State Coordinator
Mr. Michael Lene, Secretary
Illinois Department of Transportation
2300 S. Dirksen Parkway
Springfield, IL 62764
(217) 728-5597
INDIANA
Department of Natural Resources
608 State Office Building W-256
402 West Washington Street
Indianapolis, IN 46204-2748
(317) 232-4020
U.S. Geological Survey
District Chief
Water Resources Division
5957 Lakeside Boulevard
Indianapolis, IN 46278
(317) 290-3333
U.S. Department of Agriculture
Natural Resources Conservation
Service
6013 Lakeside Boulevard
Indianapolis, IN 46275
(317) 290-3030
NFIP State Coordinator
Mr. James B. Ridenour, Director
Indiana Department of Natural
Resources
608 State Office Building
Indianapolis, IN 46204
(317) 232-4020
IOWA
Iowa Department of Natural Resources
Wallace State Office Building
Des Moines, IA 50319-0034
(515) 281-5385
U.S. Geological Survey
District Chief
Water Resources Division
P.O. Box 1230
Iowa City, IA 52244-1230
(Street Address:
Federal Building, Room 269
400 South Clinton Street)
(319) 337-4191
U.S. Department of Agriculture
Natural Resources Conservation
Service
Wallace Building
Des Moines, IA 50319
(515) 284-5851
NFIP State Coordinator
Mr. Larry Wilson, Director
Iowa Department of Natural Resources
Wallace State Office Building
Des Moines, IA 50319
(515) 281-5385
KANSAS
Division of Water Resources
Kansas State Board of Agriculture
901 South Kansas Avenue, 2nd Floor
Topeka, KS 66612-1283
(913) 296-3717
U.S. Geological Survey
District Chief
Water Resources Division
4821 Quail Crest Place
Lawrence, KS 66049
(913) 842-9901
U.S. Department of Agriculture
Natural Resources Conservation
Service
P.O. Box 600
760 South Broadway
Salina, KS 67401
(913) 823-4500
NFIP State Coordinator
Mr. David L. Pope, P.E.
Chief Engineer & Director
Kansas State Board of Agriculture
Division of Water Resources
901 S. Kansas, 2nd Floor
Topeka, KS 66612-1283
(913) 296-3717
A3-7
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
KENTUCKY
Kentucky Department
of Natural Resources
Division of Water
Fort Boone Plaza
14 Reilly Road
Frankfort, KY 40601
(502) 564-3410
U.S. Geological Survey
District Chief
Water Resources Division
2301 Bradley Avenue
Louisville, KY 40217
(502) 582-5241
U.S. Department of Agriculture
Natural Resources Conservation
Service
771 Corporate Drive, Suite 110
Lexington, KY 40503
(606) 224-7350
FTS 355-2749
NFIP State Coordinator
Mr. Jack Wilson, Director
Kentucky Division of Water
18 Reilly Road
Fort Boone Plaza
Frankfort, KY 40601
(502) 564-3410
LOUISIANA
Louisiana Department of
Urban and Community Affairs
P.O. Box 94455, Capitol Station
Baton Rouge, LA 70804
(504) 342-9794
U.S. Geological Survey
District Chief
Water Resources Division
P.O. Box 66492
Baton Rouge, LA 70896
(Street Address:
6554 Florida Boulevard
Baton Rouge, LA 70806)
(504) 389-0281
U.S. Department of Agriculture
Natural Resources Conservation
Service
3636 Government Street
Alexandria, LA 71301
(318) 487-8094
NFIP State Coordinator
General Jude W. P. Patlin, Secretary
Louisiana Department of
Transportation & Development
P.O. Box 94245
Baton Rouge, LA 70804-9245
(504) 379-1200
MAINE
Maine State Planning Office
State House Station 38
184 State Street
Augusta, ME 04333
(207) 287-3261
U.S. Geological Survey
Hydrologist-in-Charge
Maine Office
Water Resources Division
26 Ganneston Drive
Augusta, ME 04330
(207) 622-8208
U.S. Department of Agriculture
Natural Resources Conservation
Service
USDA Building
University of Maine
5 Godfrey Drive
Orono, ME 04473
(207) 866-7241
NFIP State Coordinator
Mr. Michael W. Aube, Commissioner
Department of Economic and
Community Development
State House Station 59
State Street
Augusta, ME 04333
(207) 287-2656
MARYLAND
Maryland State Resources
Administration
Tawes State Office Building, D-2
501 Taylor Avenue
Annapolis, MD 21401
(410) 974-3041
U.S. Geological Survey
District Chief
Water Resources Division
208 Carroll Building
8600 La Salle Road
Towson, MD 21286
(410) 828-1535
A3-S
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Department of Agriculture
Natural Resources Conservation
Service
339 Busch's Frontage Road
Suite 301
Annapolis, MD 21401-5534
(410) 757-0861
NFIP State Coordinator
Ms. Catherine Pieper Stevenson
Director, Maryland Water Resources
Administration
Tawes State Office Building D-2
Annapolis, MD 21401
(301) 974-3896
MASSACHUSETTS
Massachusetts Water
Resources Commission
State Office Building
100 Cambridge Street
Boston, MA 02202
(617) 727-3267
U.S. Geological Survey
District Chief
Water Resources Division
28 Lord Road
Marlborough, MA 01752
(508) 485-6360
U.S. Department of Agriculture
Natural Resources Conservation
Service
451 West Street
Amherst, MA 01002
(413) 253-4350
NFIP State Coordinator
Mr. Peter C. Webber, Commissioner
Massachusetts Department of
Environmental Management
State Office Building
100 Cambridge Street
Boston, MA 02202
(617) 727-3180 x600
MICHIGAN
Engineering Water
Management Commission
Michigan Department of
Natural Resources
P.O. Box 30028
Lansing, MI 48909
(517) 373-3930
U.S. Geological Survey
District Chief
Water Resources Division
6520 Mercantile Way, Suite 5
Lansing, MI 48910
(517) 887-8903
U.S. Department of Agriculture
Natural Resources Conservation
Service
Room 101
1405 S. Harrison Road
East Lansing, MI 48823
(517) 337-6701
NFIP State Coordinator
Mr. Roland Harms, Director
Michigan Department of Natural
Resources
Land and Water Management Division
P.O. Box 30028
Lansing, MI 48909
(517) 373-3930
MINNESOTA
Flood Plains/Shoreline
Management Section
Division of Waters
Department of Natural Resources
500 Lafayette Road, Box 30
St. Paul, MN 55515-4032
(612) 297-2405
U.S. Geological Survey
District Chief
Water Resources Division
2280 Woodale Road
Moundsville, MN 55112
(612) 783-3100
U.S. Department of Agriculture
Natural Resources Conservation
Service
600 Farm Credit Building
375 Jackson Street
St. Paul, MN 55101
(612) 290-3675
NFIP State Coordinator
Mr. Ronald Nargang, Director
Minnesota Department of Natural
Resources
Division of Water
500 LaFayette Road, Box 32
St. Paul, MN 55515-0432
(612) 296-4800
A3-9
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
MISSISSIPPI
Mississippi Research and
Development Center
3825 Ridgewood Road
Jackson, MI 39211
(601) 982-6376
U.S. Geological Survey
District Chief
Water Resources Division
Federal Office Building, Suite 710
100 West Capitol Street
Jackson, MS 39269
(601) 965-4600
U.S. Department of Agriculture
Natural Resources Conservation
Service
100 W. Capitol
Suite 1321
Federal Building
Jackson, MS 39269
(601) 969-5205
NFIP State Coordinator
Mr. J. E. Maher, Director
Mississippi Emergency Management
Agency
1410 Riverside Drive
P.O. Box 4501
Jackson, MS 39216
(601) 352-9100
MISSOURI
Department of Natural Resources
P.O. Box 176
205 Jefferson Street
Jefferson City, MO 65102
(314) 751-4422
U.S. Geological Survey
District Chief
Water Resources Division
1400 Independence Road,
(Mail Stop) 200
Rolla, MO 65401
(314) 341-0824
U.S. Department of Agriculture
Natural Resources Conservation
Service
601 Business Loop
70 West Parkdale Center, Suite 250
Columbia, MO 65202
(314) 876-0903
NFIP State Coordinator
Director
Missouri Department of Natural
Resources
101 N. Jefferson Street
P.O. Box 176
Jefferson City, MO 65102
(314) 751-4422
MONTANA
Montana Department of Natural
Resources and Conservation
1520 East Sixth Avenue
Helena, MT 59620
(406) 444-6646
U.S. Geological Survey
Federal Building, Room 428
Drawer 10076
301 South Park Avenue
Helena, MT 59626-0076
(406) 449-5302
U.S. Department of Agriculture
Natural Resources Conservation
Service
10 E. Babcock
Room 443
Bozeman, MT 59715
(406) 587-6811
NFIP State Coordinator
Mr. Mark Simonich, Director
Montana Department of Natural
Resources and Conservation
1520 East 6th Ave.
Helena, MT 59620
(406) 444-6699
NEBRASKA
Nebraska Natural Resources
Commission
P.O. Box 94876
Lincoln, NE 68509-4876
(402) 471-2081
U.S. Geological Survey
District Chief
Water Resources Division
Federal Building, Room 406
100 Centennial Mall North
Lincoln, NE 68508
(402) 437-5082
A3-10
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building, Rm. 345
U.S. Courthouse
100 Centennial Mall, North
P.O. Box 82502
Lincoln, NE 68508-3866
(402) 437-5300
NFIP State Coordinator
Mr. Dayle Williamson, Director
Nebraska Natural Resources
Commission
P.O. Box 94876
Lincoln, NE 68509
(402) 471-2081
NEVADA
Division of Emergency Management
State of Nevada
Capitol Complex
Carson City, NV 89710
(702) 885-4240
U.S. Geological Survey
Hydrologist-in-Charge
Nevada Office
Water Resources Division
Federal Building, Room 224
705 North Plaza Street
Carson City, NV 89701
(702) 882-1388
U.S. Department of Agriculture
Natural Resources Conservation
Service
5301 Longway Lane
Building F, Suite 201
Reno, NV 89511
(702) 784-5863
NFIP State Coordinator
Mr. David McNinch
Nevada Division of Emergency
Management
2525 S. Carson
Capitol Complex
Carson City, NV 89710
(702) 885-4240
NEW HAMPSHIRE
Office of Emergency Management
State Office Park South
107 Pleasant Street
Concord, NH 03301
(603) 271-2231
U.S. Geological Survey
Hydrologist-in-Charge
New Hampshire Office
Water Resources Division
525 Clinton Street, RFD 2
Bow, NH 03304
(603) 225-4681
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building
2 Madbury Road
Durham, NH 03824
(603) 868-7581
NFIP State Coordinator
Col. George L. Iverson, Director
Governor's Office of
Emergency Management
State Office Park South
107 Pleasant Street
Concord, NH 03301
(603) 271-2231
NEW JERSEY
New Jersey Department of
Environmental Protection and Energy
Flood Plain Management Section
CN 419
Trenton, NJ 08625-0419
(609) 292-2296
New Jersey Department of
Environmental Protection and Energy
Division of Natural and Historic
Resources Engineering and
Construction
Element
Floodplain Management Section
Station Plaza 5
501 East State Street, 1st Floor
Trenton, New Jersey 08625-0419
(609) 292-2296
A3-11
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Geological Survey
District Chief
Water Resources Division
Mountain View Office Park,
Suite 206
810 Bear Tavern Road
West Trenton, NJ 08628
(609) 771-0065
U.S. Department of Agriculture
Natural Resources Conservation
Service
1370 Hamilton Street
Somerset, NJ 08873
(908) 725-3848
NFIP State Coordinator
Mr. Scott A. Weiner, Commissioner
New Jersey Department of
Environmental Protection and Energy
CN 402
Trenton, NJ 08625
(609) 292-2885
NEW MEXICO
New Mexico State Engineer's Office
Bataan Memorial Building
P.O. Box 25102
Santa Fe, NM 87504-5102
(505) 827-6091
U.S. Geological Survey
District Chief
Water Resources Division
4501 Indian School Road, NE
Suite 200
Albuquerque, NM 87110
(505) 262-5300
U.S. Department of Agriculture
Natural Resources Conservation
Service
P.O. Box 2007
517 Gold Avenue, SW., Rm. 301
Albuquerque, NM 87102
(505) 766-3277
NFIP State Coordinator
Mr. Keith Lough
Office of Emergency Planning
and Coordination
Department of Public Safety
P.O. Box 1628
Santa Fe, NM 87503
(505) 827-6091
NEW YORK
Flood Protection Bureau
New York Department of
Environmental Conservation
50 Wolf Road
Albany, NY 12233-3507
(518) 457-3157
U.S. Geological Survey
District Chief
Water Resources Division
445 Broadway, Room 343
Albany, NY 12201
(518) 472-3107
U.S. Department of Agriculture
Natural Resources Conservation
Service
441 South Salina Street
5th floor, Suite 354
Syracuse, NY 13202
(315) 477-6508
FTS 950-5521
NFIP State Coordinator
Mr. James F. Kelly, Director
Flood Protection Bureau
New York State Department of
Environmental Conservation
50 Wolf Road, Room 330
Albany, NY 12233-3507
(518) 457-3157
NORTH CAROLINA
North Carolina Department of
Crime Control and Public Safety
Division of Emergency Management
116 West Jones Street
Raleigh, NC 27603
(919) 733-3867
U.S. Geological Survey
District Chief
Water Resources Division
P.O. Box 30728
3916 Sunset Ridge Road
Raleigh, NC 27622
(919) 856-4510
U.S. Department of Agriculture
Natural Resources Conservation
Service
4405 Bland Avenue
Suite 205
Raleigh, NC 27609
(919) 790-2888
A3-12
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
NFIP State Coordinator
Mr. Joseph F. Myers, Director
North Carolina Division of
Emergency Management
116 West Jones Street
Raleigh, NC 27603
(919) 733-3867
NORTH DAKOTA
State Water Commission
900 East Boulevard
Bismarck, ND 58505
(701) 224-2750
U.S. Geological Survey
District Chief
Water Resources Division
821 East Interstate Avenue
Bismarck, ND 58501
(701) 250-4601
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building, Rm. 270
Rosser Ave. & Third St.
P.O. Box 1458
Bismarck, ND 58502
(701) 250-4435
NFIP State Coordinator
Mr. David A. Sprycnzynatyk
State Engineer
North Dakota State Water Commission
900 E. Boulevard
Bismark, ND 58505
(701) 224-4940
OHIO
Ohio Department of Natural Resources
Flood Plain Planning Unit
Division of Water
1939 Fountain Square
Columbus, OH 43224
(614) 265-6750
U.S. Geological Survey
District Chief
Water Resources Division
975 West Third Avenue
Columbus, OH 43212
(614) 469-5553
U.S. Department of Agriculture
Natural Resources Conservation
Service
Room 522
Federal Building
200 North High Street
Columbus, OH 43215
(614) 469-6962
NFIP State Coordinator
Mrs. Frances S. Buchholzer, Director
Ohio Department of Natural
Resources
Fountain Square
Columbus, OH 43224
(614) 264-6875
OKLAHOMA
Oklahoma Water Resources Board
600 North Harvey Avenue
P.O. Box 150
Oklahoma City, OK 73101-0150
(405) 231-2500
U.S. Geological Survey
District Chief
Water Resources Division
202 NW Sixty Sixth, Building 7
Oklahoma City, OK 73116
(405) 843-7570
U.S. Department of Agriculture
Natural Resources Conservation
Service
100 USDA
Suite 203
Stillwater, OK 74074
(405) 742-1200
NFIP State Coordinator
Mrs. Patricia P. Eaton
Executive Director
Oklahoma Water Resources Board
600 N. Harvey
Oklahoma City, OK 73101
(405) 231-2500
OREGON
Department of Land Conservation
and Development
1175 Court Street Northeast
Salem, OR 97310
(503) 373-0050
A3-13
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Geological Survey
Hydrologist-in-Charge
Oregon Office
Water Resources Division
847 Northeast 19th Avenue,
Suite 300
Portland, OR 97323
(503) 251-3200
U.S. Department of Agriculture
Natural Resources Conservation
Service
2115 SE Morrison
Portland, OR 97214
(503) 231-2270
NFIP State Coordinator
Mr. Richard Benner
Oregon Department of Land
Conservation and Development
1175 Court Street, N.E.
Salem, OR 97310
(503) 378-4928
PENNSYLVANIA
Department of Community Affairs
317 Forum Building
Harrisburg, PA 17120
(717) 787-7160
U.S. Geological Survey
District Chief
Water Resources Division
840 Market Street
Harrisburg, PA 17043-1586
(717) 730-6900
Ms. Karen A. Miller, Secretary
Pennsylvania Department of
Community Affairs
P.O. Box 155
317 Forum Building
Harrisburg, PA 17120
(717) 787-7160
U.S. Department of Agriculture
Natural Resources Conservation
Service
One Credit Union Place
Suite 340
Harrisburg, PA 17110-2993
NFIP State Coordinator
Federal Building
U.S. Courthouse
805 985
Federal Square Station
Harrisburg, PA 17108
(717) 782-2202
FTS 590-2202
PUERTO RICO
Puerto Rico Planning Board
1492 Ponce De Leon Avenue, Suite 417
Santurce, Puerto Rico 00907
(809) 729-6920
U.S. Geological Survey
District Chief
Water Resources Division
GPO Box 4424
San Juan, PR 00936
(Street Address:
GSA Center, Building 652
Highway 28, Pueblo Viejo)
(809) 783-4660
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building, Rm. 639
Chardon Avenue
GPO Box 4868
San Juan, PR 00936
(809) 753-4206
NFIP State Coordinator
Ms. Norma N. Burgos, President
Puerto Rico Planning Board
P.O. Box 41119
San Juan, PR 00940-9985
(809) 727-4444
RHODE ISLAND
Statewide Planning Program
Rhode Island Office of State Planning
1 Capitol Hill
Providence, RI 02908
(401) 277-2656
U.S. Geological Survey
Hydrologist-in-Charge
Rhode Island Office
Water Resources Division
275 Promanade Street, Suite 150
Providence, RI 02908
(401) 331-9050
A3-14
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
U.S. Department of Agriculture
Natural Resources Conservation
Service
40 Quaker Lane, Suite 46
West Warwick, RI 02886
(401) 828-1300
NFIP State Coordinator
Mr. Daniel W. Varin
Associate Director
Department of Transportation
Office of Systems Planning
1 Capitol Hill
Providence, RI 02908-5872
(401) 277-6578
SOUTH CAROLINA
South Carolina Water and Natural
Resources Commission
1201 Main Street, Suite 1100
Columbia, SC 29201
(803) 737-0800
U.S. Geological Survey
District Chief
Water Resources Division
Stevenson Center, Suite 129
720 Gracern Road
Columbia, SC 29210-7651
(803) 750-6100
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Bldg., Rm. 950
1835 Assembly St.
Columbia, SC 29201
(803) 765-5681
NFIP State Coordinator
Mr. Danny Johnson, Director
Surface Water Division
South Carolina Water Resources
Commission
1201 Main Street, Suite 1100
Columbia, SC 29201
(803) 737-0800
SOUTH DAKOTA
Disaster Assistance Programs
Emergency and Management Services
500 East Capitol
Pierre, SD 57501
(605) 773-3231
U.S. Geological Survey
District Chief
Water Resources Division
Federal Building, Room 317
200 Fourth Street Southwest
Huron, SD 57350-2469
(605) 353-7176
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building, Rm. 203
200 4th Street, SW
Huron, SD 57350
(605) 353-1092
NFIP State Coordinator
Mr. Gary N. Whitney, Director
South Dakota Department of
Military and Veteran Affairs
Division of Emergency and
Disaster Services
500 E. Capitol
Pierre, SD 57501
(605) 773-3231
TENNESSEE
Tennessee Department of Economic
and Community Development
Division of Community Development
320 Sixth Avenue North, Sixth Floor
Nashville, TN 37243-0405
(615) 741-1888
U.S. Geological Survey
District Chief
Water Resources Division
810 Broadway, Suite 500
Nashville, TN 37203
(615) 736-5424
U.S. Department of Agriculture
Natural Resources Conservation
Service
U.S. Courthouse, Rm. 675
801 Broadway Street
Nashville, TN 37203
(615) 736-5471
FTS 852-5471
A3-15
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
NFIP State Coordinator
Mr. Michael McGuire
Assistant Commissioner
Tennessee Department of Economic and
Community Development
320 Sixth Avenue
North Nashville, TN 37219-5408
(615) 741-2211
TEXAS
Texas Natural Resource
Conservation Commission
P.O. Box 13087
Capitol Station
Austin, TX 78711-3087
(512) 239-1000
U.S. Geological Survey
District Chief
Water Resources Division
8011 Cameron Road
Austin, TX 78754
(512) 873-3000
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Bldg.
101 S. Main Street
Temple, TX 76501
(817) 774-1214
NFIP State Coordinator
Mr. Jesus Galza
Executive Director
Texas Water Commission
P.O. Box 13087
Capitol Station
Austin, TX 78711-3087
(512) 463-7791
UTAH
Office of Comprehensive
Emergency Management
State Office Building, Room 1110
Salt Lake City, UT 84114
(801) 538-3400
U.S. Geological Survey
District Chief
Water Resources Division
2363 Foothill Drive
Salt Lake City, UT 84109
(801) 467-7970
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Building
125 S. State Street
P.O. Box 11350
Salt Lake City, UT 84147
(801) 524-5068
NFIP State Coordinator
Director
Ms. Lorayne Frank,
Department of Public Safety
Division of Comprehensive Emergency
Management
State Office Building, Room 1110
450 North Main
Salt Lake City, UT 84114
(801) 538-3400
VERMONT
Agency of Natural Resources
Department of Environmental
Conservation
Water Quality Division
103 South Main Street - ION
Waterbury, VT 05671-0408
(802) 241-3777
U.S. Geological Survey
District Chief
Water Resources Division
P.O. Box 628
Montpelier, VT 05602
(802) 828-4479
U.S. Department of Agriculture
Natural Resources Conservation
Service
69 Union Street
Winooski, VT 05404
(802) 951-6795
NFIP State Coordinator
Mr. Chuck Clarde, Secretary
Agency of Natural Resources
Center Building
103 South Main Street
Waterbury, VT 05671-0301
(802) 244-7347
A3-16
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
VIRGIN ISLANDS
Virgin Islands of the U.S.
Virgin Islands Planning Department
and Natural Resources
Charlotte Amalie
Nisky Center, Suite 231
St. Thomas, VI 00802
(809) 774-3320
U.S. Geological Survey
District Chief
Water Resources Division
GPO Box 4424
San Juan, PR 00936
(Street Address:
GSA Center, Building 652
Highway 28, Pueblo Viejo)
(809) 783-4660
NFIP State Coordinator
Mr. Roy E. Adams, Commissioner
Virgin Islands Department of
Planning and Natural Resources
Suite 231, Nisky Center
Charlotte Amalie
St. Thomas, VI 00802
(809) 774-3320
VIRGINIA
Virginia State Department of
Environmental Quality
4900 Cot Road
Glen Allen, VA 23060
(804) 527-5000
U.S. Geological Survey
Hydrologist-in-Charge
Virginia Office
Water Resources Division
3600 West Broad Street
Room 606
Richmond, VA 23230
(804) 771-2427
U.S. Department of Agriculture
Natural Resources Conservation
Service
1606 Santa Rosa Road
Suite 209
Richmond, VA 23229
(804) 287-1689
NFIP State Coordinator
Mr. Roland B. Geddes, Director
Department of Conservation and
Historic Resources
203 Governor Street, Suite 206
Richmond, VA 23219
(804) 786-4356
WASHINGTON
Department of Ecology
P.O. Box 47600
Olympia, WA 98504-7600
(206) 407-6000
U.S. Geological Survey
District Chief
Water Resources Division
1201 Pacific Avenue, Suite 600
Tacoma, WA 98402
(206) 593-6510
U.S. Department of Agriculture
Natural Resources Conservation
Service
316 Boone Avenue
Suite 456
Spokane, WA 99201
(509) 353-2336
NFIP State Coordinator
Mr. Chuck Clark
Washington Department of Ecology
P.O. Box 47600
Olympia, WA 98504-7600
(206) 459-6168
WEST VIRGINIA
West Virginia Office of
Emergency Services
Room EB-80, Capitol Building
Charleston, WV 25305
(304) 348-5380
U.S. Geological Survey
District Chief
Water Resources Division
11 Dunbar Street
Charleston, WV 25301
(304) 347-5130
U.S. Department of Agriculture
Natural Resources Conservation
Service
75 High Street, Rm. 301
Morgantown, WV 26505
(304) 291-4151
A3-17
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
Other Federal and State Agencies
NFIP State Coordinator
Mr. Carl Bradford, Director
West Virginia Office of
Emergency Services
Room EBI-80
Capitol Building
Charleston, WV 25305
(304) 348-5380
WISCONSIN
Department of Natural Resources
Dam Safety/Floodplain
Management Section
P.O. Box 7921
Madison, WI 53707
(608) 266-2621
U.S. Geological Survey
District Chief
Water Resources Center
University of Wisconsin/Madison
1975 Willard Drive
Madison, WI 53706-4042
(608) 262-3577
U.S. Department of Agriculture
Natural Resources Conservation
Service
6515 Watts Road
Suite 200
Madison, WI 53719
(608) 264-5341
NFIP State Coordinator
Mr. Carroll D. Besandy, Secretary
Wisconsin Department of Natural
Resources
P.O. Box 7921
Madison, WI 53707
(608) 266-2121
WYOMING
Wyoming Emergency Management Agency
P.O. Box 1709
Cheyenne, WY 82003-1709
(307) 777-4900
U.S. Geological Survey
District Chief
Water Resources Division
P.O. Box 1125
Cheyenne, WY 82003
(Street Address:
2617 East Lincoln Way
Cheyenne, WY 82001
(307) 772-2153
U.S. Department of Agriculture
Natural Resources Conservation
Service
Federal Office Building
100 East "B" Street
Casper, WY 82601
(307) 261-5231
NFIP State Coordinator
Mr. Joe Daly, Coordinator
Wyoming Emergency Management
Agency
P.O. Box 1709
Cheyenne, WY 82003
(307) 777-7566
A3-18
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
U.S. Army Corps of Engineers
U.S. Army Corps of Engineers
Headquarters
20 Massachusetts Ave., NW
Washington, D.C. 20314-1000
Attn: CECW-PF
202/272-0169
U.S. Army Corps of Engineers
Lower Miss. Valley Division
P.O. Box 80
Vicksburg, MS 39181-0080
Attn: CELMV-PD-CM
601/634-5827
U.S. Army Corps of Engineers
Memphis District
167 North Main Street, B-202
Memphis, TN 38103-1894
Attn: CELMM-PD-M
901/544-3968
U.S. Army Corps of Engineers
New Orleans District
P.O. Box 60267
New Orleans, LA 70160-0267
Attn: CELMN-PD-FG
504/865-1121
U.S. Army Corps of Engineers
St. Louis District
1222 Spruce Street
St. Louis, MO
63103-2833
Attn: CELMS-PD-M
314/331-8483
U.S. Army Corps of Engineers
Vicksburg District
2101 North Frontage Road
Vicksburg, MS 39180-0060
Attn: CELMK-PD-FS
601/631-5416
U.S. Army Corps of Engineers
Missouri River Division
12565 West Center Road
Omaha, NE 68104-3869
Attn: CEMRD-PD-F
402/221-7273
U.S. Army Corps of Engineers
Kansas City District
700 Federal Building Kansas
City, MO 64106-2896
Attn: CEMRK-PD-P
816/426-3674
U.S. Army Corps of Engineers
Omaha District
215 North 17th Street
Omaha, NE 68102-4978
Attn: CEMRO-PD-F
402/221-4596
U.S. Army Corps of Engineers
North Atlantic Division
90 Church Street
New York, NY 10007-2979
Attn: CENAD-PL-FP
212/264-7482
U.S. Army Corps ot Engineers
Baltimore District
Supervisor of Baltimore Harbor
P.O. Box 1715
Baltimore, MD 21201-1715
Attn: CENAB-PL-B
410/962-7608
U.S. Army Corps of Engineers
New York District, Planning Division,
Floodplain Management Section
26 Federal Plaza
New York, NY 10278
Attn: CENAN-PL-FP
212/264-8870
U.S. Army Corps of Engineers
Norfolk District
Supervisor of Norfolk Harbor
803 Front Street
Norfolk, VA 23510-1096
Attn: CENAO-PL-FP
804/441-7779
U.S. Army Corps of Engineers
Philadelphia District
U.S. Customs House
2nd & Chestnut Streets
Philadelphia, PA 19106-2991
Attn: CENAP-PL-F
215/656-6516
U.S. Army Corps of Engineers
North Central Division
111 North Canal Street, 14th Floor
Chicago, IL 60606
Attn: CENCD-PD-FP
312/353-1279
A3-19
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
U.S. Army Corps of Engineers
U.S. Army Corps of Engineers
Buffalo District, Planning Division,
Floodplain Management Section
1776 Niagara Street
Buffalo, NY 14207-3199
Attn: CENCB-PD-FP
716/879-4104
U.S. Army Corps of Engineers
Chicago District
111 North Canal Street
14th Floor
Chicago, IL 60606
Attn: CENCC-PD-R
312/353-6400
U.S. Army Corps of Engineers
Detroit District
477 Michigan Avenue
Detroit, MI 48226
Attn: CENCE-PD-PF
313/226-6773
U.S. Army Corps of Engineers
Rock Island District
P.O. Box 2004
Clock Tower Building
Rock Island, IL 61204-2004
Attn: CENCR-PD-F
309/788-4750
U.S. Army Corps of Engineers
St. Paul District
190 Phipps Street East
St. Paul, MN 55101-1638
Attn: CENCS-PD-FS
612/290-5200
U.S. Army Corps of Engineers
New England Division
424 Trapelo Road
Waltham, MA 02254-9149
Attn: CENED-PL-B
617/647-8111
U.S. Army Corps of Engineers
North Pacific Division
333 Southwest 1st Avenue
Portland, OR 97204
Attn: CENPD-PL-FS
503/326-6021
U.S. Army Corps of Engineers
Alaska District
P.O. Box 898
Anchorage, AK 99506-0898
Attn: CENPA-EN-PL-FP
907/753-2504
U.S. Army Corps of Engineers
Portland District
P.O. Box 2946
Portland, OR 97208-2946
Attn: CENPP-PL-CF
503/326-6411
U.S. Army Corps of Engineers
Seattle District
P.O. Box 3755
Seattle, WA 98124-2255
Attn: CENPS-EN-HH
206/764-3660
U.S. Army Corps of Engineers
Walla Walla District
Bldg. 602 City-County Airport
Walla Walla, WA 99362-9265
Attn: CENPW-PL-FP
509/522-6589
U.S. Army Corps of Engineers
Ohio River Division
P.O. Box 59
Louisville, KY 40201-0059
Attn: CEORD-PD-J
502/582-5782
U.S. Army Corps of Engineers
Huntington District
502 8th Street
Huntington,WV 25701-2070
Attn: CEORH-PD-S
304/529-5644
U.S. Army Corps of Engineers
Louisville District
P.O. Box 59
Louisville, KY 40201-0059
Attn: CEORL-PD-S
502/582-5742
U.S. Army Corps of Engineers
Nashville District
P.O. Box 1070
Nashville, TN 37202-1070
Attn: CEORN-ED-P
615/736-5055
U.S. Army Corps of Engineers
Pittsburgh District
William S. Moorehead Fed. Bldg.
1000 Liberty Avenue
Pittsburgh, PA 15222-4186
Attn: CEORP-PD-J
412/644-6924
A3-20
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
U.S. Army Corps of Engineers
U.S. Army Corps of Engineers
Pacific Ocean Division
Ft. Shatter, HI 96858-5440
Attn: CEPOD-ED-PH
808/438-7009
U.S. Army Corps of Engineers
Charleston District, P.O. Box 919
Charleston, SC 29402-0919
Attn: CESAC-EN-PH
803/727-4263
U.S. Army Corps of Engineers
South Atlantic Division
611 South Cobb Drive
Marietta, GA 30060
Attn: CESAD-PD-A
404/421-5296
U.S. Army Corps of Engineers
Jacksonville District
P.O. Box 4970
Jacksonville, FL 32232-0019
Attn: CESAJ-PD-FP
904/232-2234
U.S. Army Corps of Engineers
Mobile District
P.O. Box 2288
Mobile, AL 36628-0001
Attn: CESAM-PD-P
205/694-3879
U.S. Army Corps of Engineers
Savannah District
P.O. Box 889
Savannah, GA 31402-0889
Attn: CESAS-PD-F
912/652-5822
U.S. Army Corps of Engineers
Wilmington District
P.O. Box 1890
Wilmington, NC 28402-1890
Attn: CESAW-PD-F
910/251-4822
U.S. Army Corps of Engineers
South Pacific Division, Room 720
630 Sansome Street
San Francisco, CA 94111-2206
Attn: CESPD-PD-P
415/705-2427
U.S. Army Corps of Engineers
Los Angeles District
P.O. Box 2711
Los Angeles, CA 90053-2325
Attn: CESPL-PD-WF
213/894-5450
U.S. Army Corps of Engineers
Sacramento District
1325 G Street
Sacramento, CA 95814-4794
Attn: CESPK-PD-F
916/557-6700
U.S. Army Corps of Engineers
San Francisco District
211 Main Street
San Francisco, CA 9410S-1905
Attn: CESPN-PE-W
415/744-3029
U.S. Army Corps of Engineers
Southwestern Division
1114 Commerce Street
Dallas, TX 75242-0216
Attn: CESWD-PL-M
214/767-2310
U.S. Army Corps of Engineers
Albuquerque District
P.O. Box 1580
Albuquerque, NM 87103-1580
Attn: CESWA-ED-PH
505/766-2635
U.S. Army Corps of Engineers
Fort Worth District
P.O. Box 17300
Fort Worth, TX 76102-0300
Attn: CESWF-PL-F
817/334-3207
U.S. Army Corps of Engineers
Galveston District
P.O. Box 1229
Galveston, TX 77553-1229
Attn: CESWG-PL-P
409/766-3023
U.S. Army Corps of Engineers
Little Rock District
P.O. Box 867
Little Rock, AR 72203-0867
Attn: CESWL-PL-F
501/378-5611
U.S. Army Corps of Engineers
Tulsa District
P.O. Box 61
Tulsa, OK 74121 0061
Attn: CESWT-PL-GF
918/581-7896
A3-21
-------
Appendix 3 - continued
Federal Emergency Management Agency Offices
and Other Federal and State Agencies
River Basin Commissions
Delaware River Basin Commission
25 State Police Drive
Box 7360
West Trenton, NJ 08628
609/883-9500
Susquehanna River Basin Commission
1721 North Front Street
Harrisburg, PA
717/238-0422
A3-22
-------
Appendix 4
State Hydrology Reports
ALABAMA
Olin, D.A. and Bingham, R.H., 1982, Synthesized flood frequency of urban streams
in Alabama: U.S. Geological Survey Water-Resources Investigations 82-683.
Olin, D.A., 1984, Magnitude and frequency of floods in Alabama: U.S. Geological
Survey Water-Resources Investigations 84-4191.
ALASKA
Lamke, R.D., 1978, Flood characteristics of Alaskan streams: U.S. Geological
Survey Water-Resources Investigations 78-129.
ARIZONA
Eychaner, J.H., 1984, Estimation of magnitude and frequency of floods in Pima
County, Arizona, with comparisons of alternative methods: U.S. Geological Survey
Water-Resources Investigations 84-4142.
ARKANSAS
Neely, B.L., Jr., 1986, Magnitude and frequency of floods in Arkansas: U.S.
Geological Survey Water-Resources Investigations 86-4335.
CALIFORNIA
Waananen, A.O., and Crippen, J.R., 1977, Magnitude and frequency of floods in
California: U.S. Geological Survey Water-Resources Investigations 77-21 (PB-272
510/AS).
COLORADO
Hedman, E.R., Moore, D.O. and Livingston, R.K., 1972, Selected streamflow
characteristics as related to channel geometry of perennial streams in Colorado:
U.S. Geological Survey Open-File Report.
Kircher, J.E., Choquette, A.F., and Richter, B.D., 1985, Estimation of natural
streamflow characteristics in Western Colorado: U.S. Geological Survey Water-
Resources Investigations 85-4086.
Livingston, R.K., 1980, Rainfall-runoff modeling and preliminary regional flood
characteristics of small rural watersheds in the Arkansas River Basin in
Colorado: U.S. Geological Survey Water-Resources Investigations 80-112 (NTIS).
Livingston, R.K., and Minges, D.R., 1987, Techniques for estimating regional
flood characteristics of small rural watersheds in the plains regions of eastern
Colorado: U.S. Geological Survey Water-Resources Investigations 87-4094.
McCain, J.R., and Jarrett, R.D., 1976, Manual for estimating flood
characteristics of natural flow streams in Colorado: Colorado Water Conservation
Board, Technical Manual No. 1.
A4-1
-------
Appendix 4 - continued
State Hydrology Reports
CONNECTICUT
Weiss, L.A., 1975, Flood flow formula for urbanized and non-urbanized areas of
Connecticut: Watershed Management Symposium of ASCE Irrigation and Drainage
Division, August 11-13, 1975, pp. 658-675.
DELAWARE
Simmons, R.H., and Carpenter, D.H., 1978, Technique for estimating the magnitude
and frequency of floods in Delaware: U.S. Geological Survey Water-Resources
Investigations Open-File Report 78-93.
DISTRICT OF COLUMBIA
None listed
FLORIDA
Bridges, W.C., 1982, Technique for estimating the magnitude and frequency of
floods on natural-flow streams in Florida: U.S. Geological Survey Water-Resources
Investigations Open-File Report 82-4012.
Franklin, M.A., 1984, Magnitude and frequency of floods from urban streams in
Leon County, Florida: U.S. Geological Survey Water-Resources Investigations
84-4004.
Lopez, M.A., and Woodham, W.M., 1982, Magnitude and frequency of floods on small
urban watersheds in the Tampa Bay area, west-central Florida: U.S. Geological
Survey Water-Resources Investigations 82-42.
GEORGIA
Inman, E.J., 1983, Flood-frequency relations for urban streams in metropolitan
Atlanta, Georgia: U.S. Geological Survey Water-Resources Investigations
83-4203.
Price, M., 1979, Floods in Georgia, magnitude and frequency: U.S. Geological
Survey Water-Resources Investigations 78-137 (PB-80 146 244).
HAWAII
Matsuoka, I., 1978, Flow characteristics of streams in Tutuila, American Somoa:
U.S. Geological Survey Open-File Report 78-103.
Nakahara, R.H., 1980, An analysis of the magnitude and frequency of floods on
Oahu, Hawaii: U.S. Geological Survey Water-Resources Investigations 80-45
(PB-81 109 902) .
A4-2
-------
Appendix 4 - continued
State Hydrology Reports
IDAHO
Harenberg, W.A., 1980, Using channel geometry to estimate flood flows at ungaged
sites in Idaho: U.S. Geological Survey Water-Resources Investigations 80-32
(PB-81 153 736) .
Kjelstrom, L.C., and Moffatt, R.L., 1981, Method of estimating flood-frequency
parameters for streams in Idaho: U.S. Geological Survey Open-File Report 81-909.
Thomas, C.A., Harenburg, W.A., and Anderson, J.M., 1973, Magnitude and frequency
of floods in small drainage basins in Idaho: U.S. Geological Survey Water-
Resources Investigations 7-73 (PB-222 409) .
ILLINOIS
Allen, H.E., Jr., and Bejcek, R.M., 1979, Effects of urbanization on the
magnitude and frequency of floods in northeastern Illinois: U.S. Geological
Survey Water-Resources Investigations 79-36 (PB-299 065/AS).
Curtis, G.W., 1987, Technique for estimating flood-peak discharges and
frequencies on rural streams in Illinois: U.S. Geological Survey Water-Resources
Investigations 87-4207.
INDIANA
Glatfelter, D.R., 1984, Technique for estimating the magnitude and frequency of
floods in Indiana: U.S. Geological Survey Water-Resources Investigations 84-4134.
IOWA
Lara, 0., 1978, Effects of urban development on the flood flow characteristics of
Walnut Creek basin, Des Moines metropolitan area, Iowa: U.S. Geological Survey
Water-Resources Investigations 78-11 (PB-284 093/AS).
KANSAS
Clement, R.W., 1987, Floods in Kansas and techniques for estimating their
magnitude and frequency: U.S. Geological Survey Water-Resources Investigations
87-4008.
Hedman, E.R., Kastner, W.M., and Hejl, H.R., 1974, Selected streamflow
characteristics as related to active-channel geometry of streams in Kansas:
Kansas Water Resources Board Technical Report No. 10.
Peek, C.O., and Jordan, P.R., 1978, Determination of peak discharge from rainfall
relations for urbanized basins, Wichita, Kansas: U.S. Geological Survey Open-File
Report 78-974.
KENTUCKY
Choquette, A.F., 1987, Regionalization of peak discharges for streams in
Kentucky: U.S. Geological Survey Water-Resources Investigations 87-4029.
A4-3
-------
Appendix 4 - continued
State Hydrology Reports
LOUISIANA
Lee, F.N., 1985, Floods in Louisiana, Magnitude and Frequency, Fourth Edition:
Department of Transportation and Development, Water Resources Technical Report
No. 36.
Lowe, A.S., 1979, Magnitude and frequency of floods in small watersheds in
Louisiana: Louisiana Department of Transportation and Development, Office of
Highways, Research Study No. 65-2H.
MAINE
Morrill, R.A., 1975, A technique for estimating the magnitude and frequency of
floods in Maine: U.S. Geological Survey Open-File Report.
MARYLAND
Carpenter, D.H., 1980, Technique for estimating magnitude and frequency of floods
in Maryland: U.S. Geological Survey Water-Resources Investigations Open-File
Report 80-1016.
MASSACHUSETTS
Wandle, S.W., 1983, Estimating peak discharges and frequencies of small rural
streams in Massachusetts: U.S. Geological Survey Water-Supply Paper 2214.
MICHIGAN
Holtschlag, D.J., and Croskey, H.M., 1984, Statistical models for estimating flow
characteristics of Michigan streams: U.S. Geological Survey Water-Resources
Investigations 84-4270.
MINNESOTA
Jacques, J.E., and Lorenz, D.L., 1987, Techniques for estimating the magnitude
and frequency of floods in Minnesota: U.S. Geological Survey Water-Resources
Investigations 87-4170.
MISSISSIPPI
Colson, B.E., and Hudson, J.W., 1976, Flood frequency of Mississippi streams:
Mississippi State Highway Department.
MISSOURI
Becker, L.D., 1986, Techniques for estimating flood-peak discharges for urban
streams in Missouri: U.S. Geological Survey Water-Resources Investigations Report
86-4322.
Hauth, L.D., 1974, A technique for estimating the magnitude and frequency of
Missouri floods: U.S. Geological Survey Open-File Report.
A4-4
-------
Appendix 4 - continued
State Hydrology Reports
MISSOURI continued
Spencer, D.W., and Alexander, T.W., 1978, Techniques for estimating the magnitude
and frequency of floods in St. Louis County, Missouri: U.S. Geological Survey
Water-Resources Investigations 78-139 (PB-298 245/AS).
MONTANA
Omang, R.J., 1983, Mean annual runoff and peak flow estimates based on channel
geometry of streams in southeastern Montana: U.S. Geological Survey Water-
Resources Investigations Report 82-4092.
Omang, R.J., Parrett, C., and Hull, J.A., 1986, Methods of estimating magnitude
and frequency of floods in Montana based on data through 1983: U.S. Geological
Survey Water-Resources Investigations Report 86-4027.
Parrett, C., 1983, Mean annual runoff and peak flow estimates based on channel
geometry of streams in northeastern and western Montana: U.S. Geological Survey
Water-Resources Investigations Report 83-4046.
Parrett, C., Hull, J.A., and Omang, R.J., 1987, Revised techniques for estimating
peak discharges from channel width in Montana: U.S. Geological Survey Water-
Resources Investigations 87-4121.
NEBRASKA
Beckman, E.W., 1976, Magnitude and frequency of floods in Nebraska: U.S.
Geological Survey Water-Resources Investigations 76-109 (PB-260 842/AS).
NEVADA
Moore, D.O., 1974, Estimating flood discharges in Nevada using channel-geometry
measurements: Nevada State Highway Department Hydrologic Report No. 1.
Moore, D.O., 1976, Estimating peak discharges from small drainages in Nevada
according to basin areas within elevation zones: Nevada State Highway Department
Hydrologic Report No. 3.
NEW HAMPSHIRE
LeBlanc, D.R., 1978, Progress report on hydrologic investigations of small
drainage areas in New Hampshire--Preliminary relations for estimating peak
discharges on rural, unregulated streams: U.S. Geological Survey Water-Resources
Investigations 78-47 (PB-284 127/AS).
NEW JERSEY
Stankowski, S.J., 1974, Magnitude and frequency of floods in New Jersey with
effects of urbanization: New Jersey Department of Environmental Protection
Special Report 38.
Velnick, Anthony J. and Laskowski, Stanley L. , 1979, Technique for estimating
depth of 100-year flood in New Jersey: Open-File Report 79-419.
A4-5
-------
Appendix 4 - continued
State Hydrology Reports
NEW MEXICO
Hejl, H.R., Jr., 1984, Use of selected basin characteristics to estimate mean
annual runoff and peak discharges for ungaged streams in drainage basins
containing strippable coal resources, northwestern New Mexico: U.S. Geological
Survey Water-Resources Investigations 84-4264.
Scott, A.G., and Kunkler, J.L., 1976, Flood discharges of streams in New Mexico
as related to channel geometry: U.S. Geological Survey Open-File Report.
Waltmeyer, S.D., 1986, Techniques for estimating flood-flow frequency for
unregulated streams in New Mexico: U.S. Geological Survey Water-Resources
Investigations 86-4104.
NEW YORK
Lomia, Richard, 1991, Regionalization of flood discharges for rural, unregulated
streams in New York, excluding Long Island: U.S. Geological Survey Water-
Resources Investigations Report 90-4197.
NORTH CAROLINA
Gunter, H.C., Mason, R.R., and Stamey, T.C., 1987, Magnitude and frequency of
floods in rural and urban basins of North Carolina: U.S. Geological Survey Water-
Resources Investigations 87-4096.
Martens, L.S., 1968, Flood inundation and effects of urbanization in metropolitan
Charlotte, North Carolina: U.S. Geological Survey Water-Supply Paper 1591-C.
Putnam, A.L., 1972, Effect of urban development on floods in the Piedmont
province of North Carolina: U.S. Geological Survey Open-File Report.
NORTH DAKOTA
Crosby, O.A., 1975, Magnitude and frequency of floods in small drainage basins of
North Dakota: U.S. Geological Survey Water-Resources Investigations 19-75
(PB-248 480/AS).
OHIO
Roth, D.K., 1985, Estimation of flood peaks from channel characteristics in Ohio:
U.S. Geological Survey Water-Resources Investigations Report 85-4175.
Sherwood, J.M., 1986, Estimating peak discharges, flood volumes, and hydrograph
stages of small urban streams in Ohio: U.S. Geological Survey Water-Resources
Investigations Report 86-4197.
Webber, E.E., and Bartlett, W.P., Jr., 1977, Floods in Ohio magnitude and
frequency: State of Ohio, Department of Natural Resources, Division of Water,
Bulletin 45.
Webber, E.E., and Roberts, J.W., 1981, Floodflow characteristics related to
channel geometry in Ohio: U.S. Geological Survey Open-File Report 81-1105.
A4-6
-------
Appendix 4 - continued
State Hydrology Reports
OKLAHOMA
Sauer, V.B., 1974, An approach to estimating flood frequency for urban areas in
Oklahoma: U.S. Geological Survey Water-Resources Investigations 23-74
(PB-235 307/AS).
Tortorelli, R.L., and Bergman, D.L., 1984, Techniques for estimating flood peak
discharge for unregulated streams and streams regulated by small floodwater
retarding structures in Oklahoma: U.S. Geological Survey Water-Resources
Investigations 84-4358.
OREGON
Harris, D.D., and Hubbard, L.E., 1982, Magnitude and frequency of floods in
eastern Oregon: U.S. Geological Survey Water-Resources Investigations 82-4078.
Harris, D.D., Hubbard, L.E., and Hubbard, L.L., 1979, Magnitude and frequency of
floods in western Oregon: U.S. Geological Survey Open-File Report 79-553.
Laenen, Antonius, 1980, Storm runoff as related to urbanization in the Portland,
Oregon-Vancouver, Washington, area: U.S. Geological Survey Water-Resources
Investigations Open-File Report 80-689.
PENNSYLVANIA
Bailey, J.F., Thomas, W.O., Jr., Wetzel, K.L., and Ross, T.J., 1987, Estimation
of flood-frequency characteristics and the effects of urbanization for streams in
the Philadelphia, Pennsylvania, area: U.S. Geological Survey Water-Resources
Investigations 87-4194.
Flippo, H.N., Jr., 1977, Floods in Pennsylvania: A manual for estimation of their
magnitude and frequency: Pennsylvania Department of Environmental Resources
Bulletin No. 13.
PUERTO RICO
Lopez, M.A., Colon-Dieppa, E., and Cobb, E.D., 1978, Floods in Puerto Rico:
magnitude and frequency: U.S. Geological Survey Water-Resources Investigations
78-141 (PB-300 855/AS).
RHODE ISLAND
Johnson, C.G., and Laraway, G.A., 1976, Flood magnitude and frequency of small
Rhode Island streams--Preliminary estimating relations: U.S. Geological Survey
Open-File Report.
SOUTH CAROLINA
Whetstone, B.H., 1982, Floods in South Carolina--Techniques for estimating
magnitude and frequency of floods with compilation of flood data: U.S. Geological
Survey Water-Resources Investigations 82-1.
A4-7
-------
Appendix 4 - continued
State Hydrology Reports
SOUTH DAKOTA
Becker, L.D., 1974, A method for estimating the magnitude and frequency of floods
in South Dakota: U.S. Geological Survey Water-Resources Investigations 35-74
(PB-239 831/AS).
Becker, L.D., 1980, Techniques for estimating flood peaks, volumes, and
hydrographs on small streams in South Dakota: U.S. Geological Survey Water-
Resources Investigations 80-80 (PB-81 136 145).
TENNESSEE
Neely, B.L., Jr., 1984, Flood frequency and storm runoff of urban areas of
Memphis and Shelby County, Tennessee: U.S. Geological Survey Water-Resources
Investigations 84-4110.
Randolph, W.J., and Gamble, C.R., 1976, A technique for estimating magnitude and
frequency of floods in Tennessee: Tennessee Department of Transportation.
Robbins, C.H., 1984, Synthesized flood frequency of small urban streams in
Tennessee: U.S. Geological Survey Water-Resources Investigations 84-4182.
Wibben, H.C., 1976, Effects of urbanization on flood characteristics in
Nashville-Davidson County, Tennessee: U.S. Geological Survey Water-Resources
Investigations 76-121 (PB-266 654/AS).
TEXAS
Land, L.F., Schroeder, E.E., and Hampton, B.B., 1982, Techniques for estimating
the magnitude and frequency of floods in the Dallas-Fort Worth Metropolitan Area,
Texas: U.S. Geological Survey Water-Resources Investigations 82-18.
Liscum, F., and Massey, B.C., 1980, Techniques for estimating the magnitude and
frequency of floods in the Houston, Texas metropolitan area: U.S. Geological
Survey Water-Resources Investigations 80-17 (ADA-089 495).
Schroeder, E.E., and Massey, B.C., 1977, Techniques for estimating the magnitude
and frequency of floods in Texas: U.S. Geological Survey Water-Resources
Investigations Open-File Report 77-110.
Veenhuis, J.E., and Garrett, D.G., 1986, The effects of urbanization on floods in
the Austin metropolitan area, Texas: U.S. Geological Survey Water-Resources
Investigations 86-4069.
UTAH
Fields, F.K., 1974, Estimating streamflow characteristics for streams in Utah
using selected channel-geometry parameters: U.S. Geological Survey Water-
Resources Investigations 34-74 (PB-241 541/AS).
Thomas, B.E., and Lindskov, K.L., 1983, Methods for estimating peak discharges
and flood boundaries of streams in Utah: U.S. Geological Survey Water-Resources
Investigations 83-4129.
A4-S
-------
Appendix 4 - continued
State Hydrology Reports
VERMONT
Johnson, C.G., and Tasker, G.D., 1974, Flood magnitude and frequency of Vermont
Streams: U.S. Geological Survey Open-File Report 74-130.
VIRGIN ISLANDS
None listed
VIRGINIA
Anderson, D.G., 1970, Effects of urban development on floods in Northern
Virginia: U.S. Geological Survey Water-Supply Paper 2001-C.
Miller, E.M., 1978, Technique for estimating the magnitude and frequency of
floods in Virginia: U.S. Geological Survey Water-Resources Investigations Open-
File Report 78-5.
WASHINGTON
Cummans, J.E., Collins, M.R., and Nassar, E.G., 1974, Magnitude and frequency of
floods in Washington: U.S. Geological Survey Open-File Report 74-336.
Haushild, W.L., 1978, Estimation of floods of various frequencies for the small
ephemeral streams in eastern Washington: U.S. Geological Survey Water-Resources
Investigations 79-81.
WEST VIRGINIA
Runner, G.S., 1980, Technique for estimating the magnitude and frequency of
floods in West Virginia: U.S. Geological Survey Open-File Report 80-1218.
WISCONSIN
Conger, D.H., 1980, Techniques for estimating magnitude and frequency of floods
for Wisconsin streams: U.S. Geological Survey Water-Resources Investigations
Open-File Report 80-1214.
Conger, D.H., 1986, Estimating magnitude and frequency of floods for ungaged
urban streams in Wisconsin: U.S. Geological Survey Water-Resources Investigations
Report 86-4005.
WYOMING
Craig, G.S., Jr., and Rankl, J.G., 1977, Analysis of runoff from small drainage
basins in Wyoming: U.S. Geological Survey Water-Supply Paper 2056.
Lowham, H.W., 1976, Techniques for estimating flow characteristics of Wyoming
streams: U.S. Geological Survey Water-Resources Investigations 76-112
(PB-264 224/AS).
A4-9
-------
Appendix 5
Manning's "n" Values
The value of "n" may be computed by
n = (no + ni + n2 + n3 + n^mj
where: n0 =
n4
m5
basic "n" value for a straight, uniform, smooth channel
value added to correct for the effect of surface irregularities
value added for variation in the shape and size of the channel
cross section
value added for obstructions
value added for vegetation and flow conditions
correction factor for meandering of the channel
Proper values of n0 to n4 and m5 may be selected from the following table according to the
given conditions:
Channel Conditions
Material
involved
Degree of
irregularity
Variations of
channel cross
section
Relative
effect of
obstructions
Vegetation
Degree of
meandering
Earth
Rock cut
Fine gravel
Coarse gravel
Smooth
Minor
Moderate
Severe
Gradual
Alternating occasionally
Alternating frequently
Negligible
Minor
Appreciable
Severe
Low
Medium
High
Very High
Minor
Appreciable
Severe
Values
n0
ni
n2
n3
n4
ms
0.020
0.025
0.024
0.028
0.000
0.005
0.010
0.020
0.000
0.005
0.010-0.015
0.000
0.010-0.015
0.020-0.030
0.040-0.060
0.005-0.010
0.010-0.025
0.025-0.050
0.050-0.100
1.000
1.150
1.300
1.
REFERENCE
Chow, Ven Te, Ph.D.: "Open-Channel Hydraulics," McGraw-Hill Book Company,
New York, 1959, pp. 106-114.
The computed "n" values should be compared with the typical "n" values from the following
pages, or with those in the U.S. Geological Survey Report (Reference 2) or the Federal
Highway Administration Report (Reference 3).
A5-1
-------
Appendix 5 - continued
Manning's "n" Values
Type of channel and description
A. Closed Conduits Flowing Partly Full
A-l. Metal
a. Brass, smooth
b. Steel
1 . Lockbar and welded
2. Riveted and spiral
c. Cast iron
1. Coated
2. Uncoated
d. Wrought iron
1. Black
2. Galvanized
e. Corrugated metal
1 . Subdrain
2. Storm drain
A-2. Nonmetal
a. Lucite
b. Glass
c. Cement
1. Neat, surface
2. Mortar
d. Concrete
1 . Culvert, straight and free of
debris
2. Culvert with bends, connections,
and some debris
3. Finished
4. Sewer with manholes, inlet, etc.,
straight
5. Unfinished, steel form
6. Unfinished, smooth wood form
7. Unfinished, rough wood form
e. Wood
1. Stave
2. Laminated, treated
f. Clay
1 . Common drainage tile
2. Vitrified sewer
3 . Vitrified sewer with manholes,
inlet, etc.
4. Vitrified subdrain with open joint
g. Brickwork
1. Glazed
2. Lined with cement mortar
h. Sanitary sewers coated with sewage
slimes, with bends and connections
i. Paved invert, sewer, smooth bottom
j. Rubble masonry, cemented
Minimum
0.009
0.010
0.013
0.010
0.011
0.012
0.013
0.017
0.021
0.008
0.009
0.010
0.011
0.010
0.011
0.011
0.013
0.012
0.012
0.015
0.010
0.015
0.011
0.011
0.013
0.014
0.011
0.012
0.012
0.016
0.018
Normal
0.010
0.012
0.016
0.013
0.014
0.014
0.016
0.019
0.024
0.009
0.010
0.011
0.013
0.011
0.013
0.012
0.015
0.013
0.014
0.017
0.012
0.017
0.013
0.014
0.015
0.016
0.013
0.015
0.013
0.019
0.025
Maximum
0.013
0.014
0.017
0.014
0.016
0.015
0.017
0.021
0.030
0.010
0.013
0.013
0.015
0.013
0.014
0.014
0.017
0.014
0.016
0.020
0.014
0.020
0.017
0.017
0.017
0.018
0.015
0.017
0.016
0.020
0.030
A5-2
-------
Appendix 5 - continued
Manning's "n" Values
Type of channel and description
B. Lined or Built-up Channels
B-l. Metal
a. Smooth steel surface
1. Unpainted
2. Painted
b. Corrugated
B-2. Nonmetal
a. Cement
1. Neat, surface
2. Mortar
b. Wood
1 . Planed, untreated
2. Planed, creosoted
3. Unplaned
4. Plank with battens
5 . Lined with roofing paper
c. Concrete
1 . Trowel finish
2. Float finish
3 . Finished, with gravel on bottom
4. Unfinished
5. Gunite, good section
6. Gunite, wavy section
7. On good excavated rock
8. On irregular excavated rock
d. Concrete bottom float finished with
sides of
1 . Dressed stone in mortar
2. Random stone in mortar
3 . Cement rubble masonry, plastered
4. Cement rubble masonry
5. Dry rubble or riprap
e. Gravel bottom with sides of
1 . Formed concrete
2. Random stone in mortar
3 . Dry rubble or riprap
f. Brick
1. Glazed
2. In cement mortar
g. Masonry
1 . Cemented rubble
2. Dry rubble
h. Dressed ashlar
i. Asphalt
1. Smooth
2. Rough
j. Vegetal lining
Minimum
0.011
0.012
0.021
0.010
0.011
0.010
0.011
0.011
0.012
0.010
0.011
0.013
0.015
0.014
0.016
0.018
0.017
0.022
0.015
0.017
0.016
0.020
0.020
0.017
0.020
0.023
0.011
0.012
0.017
0.023
0.013
0.013
0.016
0.030
Normal
0.012
0.013
0.025
0.011
0.013
0.012
0.012
0.013
0.015
0.014
0.013
0.015
0.017
0.017
0.019
0.022
0.020
0.027
0.017
0.020
0.020
0.025
0.030
0.020
0.023
0.033
0.013
0.015
0.025
0.032
0.015
0.013
0.016
Maximum
0.014
0.017
0.030
0.013
0.015
0.014
0.015
0.015
0.018
0.017
0.015
0.016
0.020
0.020
0.023
0.025
0.020
0.024
0.024
0.030
0.035
0.025
0.026
0.036
0.015
0.018
0.030
0.035
0.017
0.500
A5-3
-------
Appendix 5 - continued
Manning's "n" Values
Type of channel and description
C. Excavated or Dredged
a. Earth, straight and uniform
1 . Clean, recently completed
2. Clean, after weathering
3 . Gravel, uniform section, clean
4. With short grass, few weeds
b. Earth, winding and sluggish
1 . No vegetation
2. Grass, some weeds
3 . Dense weeds or aquatic plants in
deep channels
4. Earth bottom and rubble sides
5. Stony bottom and weedy banks
6. Cobble bottom and clean sides
c. Dragline-excavated or dredged
1 . No vegetation
2. Light brush on banks
d. Rock cuts
1 . Smooth and uniform
2. Jagged and irregular
e. Channels not maintained, weeds and
brush uncut
1 . Dense weeds, high as flow depth
2. Clean bottom, brush on sides
3 . Same, highest stage of flow
4. Dense brush, high stage
D. Natural Streams
D-l. Minor streams (top width at flood stage
<100 ft)
a. Streams on plain
1 . Clean, straight, full stage, no rifts
or deep pods
2. Same as above, but more stones
and weeds
3 . Clean, winding, some pools and
shoals
4. Same as above, but some weeds
and stones
5. Same as above, lower stages,
more ineffective slopes and
sections
6. Same as 4, but more stones
7. Sluggish reaches, weedy, deep
pools
8. Very weedy reaches, deep pools,
or floodways with heavy stand of
timber and underbrush
Minimum
0.016
0.018
0.022
0.022
0.023
0.025
0.030
0.028
0.025
0.030
0.025
0.035
0.025
0.035
0.050
0.040
0.045
0.080
0.025
0.030
0.033
0.035
0.040
0.045
0.050
0.075
Normal
0.018
0.022
0.025
0.027
0.025
0.030
0.035
0.030
0.035
0.040
0.028
0.050
0.035
0.040
0.080
0.050
0.070
0.100
0.030
0.035
0.040
0.045
0.048
0.050
0.070
0.100
Maximum
0.020
0.025
0.030
0.033
0.030
0.033
0.040
0.035
0.040
0.050
0.033
0.060
0.040
0.050
0.120
0.080
0.110
0.140
0.033
0.040
0.045
0.050
0.055
0.060
0.080
0.150
A5-4
-------
Appendix 5 - continued
Manning's "n" Values
Type of channel and description
b. Mountain streams, no vegetation in
channel, banks usually steep, trees
and brush along banks submerged at
high stages
1. Bottom: gravels, cobbles, and
few boulders
2. Bottom: cobbles with large
boulders
D-2. Floodplains
a. Pasture, no brush
1 . Short grass
2. High grass
b. Cultivated areas
1 . No crop
2. Mature row crops
3 . Mature field crops
c. Brush
1 . Scattered brush, heavy weeds
2. Light brush and trees, in winter
3 . Light brush and trees, in summer
4. Medium to dense brush, in winter
5. Medium to dense brush, in
summer
d. Trees
1 . Dense willows, summer, straight
2. Cleared land with tree stumps, no
sprouts
3 . Same as above, but with heavy
growth of sprouts
4. Heavy stand of timber, a few
down trees, little undergrowth,
flood stage below branches
5. Same as above, but with flood
stage below branches
D-3. Major streams (top width at flood stage
>100 ft). The n value is less than that
for minor streams of similar description,
because banks offer less effective
resistance.
a. Regular section with no boulders or
brush
b. Irregular and rough section
Minimum
0.030
0.040
0.025
0.030
0.020
0.025
0.030
0.035
0.035
0.040
0.045
0.070
0.110
0.030
0.050
0.080
0.100
0.025
0.035
Normal
0.040
0.050
0.030
0.035
0.030
0.035
0.040
0.050
0.050
0.060
0.070
0.100
0.150
0.040
0.060
0.100
0.120
Maximum
0.050
0.070
0.035
0.050
0.040
0.045
0.050
0.070
0.060
0.080
0.110
0.160
0.200
0.050
0.080
0.120
0.160
0.060
0.100
REFERENCES
1. Chow, Ven Te, Ph.D.: "Open-Channel Hydraulics," McGraw-Hill Book Company, New York, 1959, pp.
106-114.
2. Geological Survey, Roughness Characteristics of Natural Channels. Water-Supply Paper 1849,
Washington, D.C., 1967.
3. Department of Transportation, Federal Highway Administration, Guide for Selecting Manning's
Roughness Coefficients for Natural Channels and Floodplains. Report No. FHWA-TS-84-204, McLean,
Virginia, April 1984.
A5-5
-------
Appendix 6
QUICK-2 Computer Program Manual
A6-1
-------
QUICK-2
Computer Program
COMPUTATION OF WATER SURFACE
ELEVATIONS IN OPEN
CHANNELS
VERSION 1.0
JANUARY 1995
-------
QUICK-2
Computation of Water Surface Elevations in Open Channels
User's Guide
Federal Emergency Management Agency
1995
-------
TABLE OF CONTENTS
Chapter 1: INTRODUCTION 1-1
Chapter 2: OVERVIEW 2-1
Chapter 3: GETTING STARTED 3-1
Chapter 4: TUTORIALS 4-1
Normal Depth 4-2
Changing Variables 4-6
Step-Backwater 4-9
Running HEC-2 with QUICK-2 Files 4-18
Rerunning Using Saved Cross-Section Files 4-19
Channel Capacity 4-21
Rating Curve Plot 4-22
PLOT-2 4-23
Profile Plot 4-23
Cross-Section Plot 4-24
Chapter 5: FORMULAS 5-1
Critical Depth 5-2
Channel Capacity 5-4
Normal Depth 5-5
Step-Backwater 5-6
Appendix 1: DEFINITION OF VARIABLES A- 1
-------
QUICK-2 User's GuideIntroduction
Chapter 1: Introduction
QUICK-2 is a user friendly program that assists in the computation of flood
Water Surface Elevations (WSEs) in open channels of all types. It is much
easier to use than the United States Army Corps of Engineers (USAGE) HEC-2
program. However, a QUICK-2 step-backwater file can also be used, as is,
with the HEC-2 program, which is also included in the QUICK-2 package of
programs. Therefore a HEC-2 output file can be generated with a QUICK-2 input
data file, without ever leaving the QUICK-2 environment; and, without having
to know how to set-up and run the HEC-2 program. This version of QUICK-2
(Version 1.0) however, does not perform hydraulic calculations through
bridges or culverts.
QUICK-2 was primarily developed to accompany the FEMA technical guidance
manual titled, "MANAGING FLOODPLAIN DEVELOPMENT IN ZONE A AREAS - A GUIDE FOR
OBTAINING AND DEVELOPING BASE FLOOD ELEVATIONS." That manual is intended to
assist local community officials who are responsible for administering and
enforcing the floodplain management requirements of the National Flood
Insurance Program (NFIP). The purpose of that manual is to provide guidance
for obtaining and developing base flood (100-year) elevations (BFEs) where
Special Flood Hazard Areas (SFHAs) on a community's Flood Hazard Boundary Map
(FHBM) or Flood Insurance Rate Map (FIRM) have been identified and designated
as Zone A.
QUICK-2 will also be useful to community engineers, architect/engineer firms,
developers, builders and others at the local level who may be required to
develop BFEs for use in Special Flood Hazard Areas.
This manual includes four other chapters: Overview, Getting Started,
Tutorials and Formulas. The Formulas section describes the "complex"
equations and methodologies used in the development of the program. An
Appendix is also included that contains a list of Definitions of the
variables shown on the screen and on the printouts.
To get started as quickly as possible in using QUICK-2 we
recommend that the user read the Overview and Getting
Started chapters; and then work through the Tutorials.
MINIMUM SYSTEM REQUIREMENTS
Random Access Memory (RAM) - 512K
Hard disk storage - 800K
Monitor - Color or Monotone
Printer (prints to LPT1) - Dot-matrix to LaserJet
Disk Operating System (DOS) - Version 3.0 or higher
1-1
-------
QUICK-2 User's Guide Overview
Chapter 2: Overview
FOUR OPTIONS
This user friendly program computes:
• Critical Depth,
• Cross Section Capacity (Rating Curves),
• .Normal Depth, and
• Step-Backwater Analysis (similar to the USAGE HEC-2 program)
CRITICAL DEPTH: This option should be used to determine a Base Flood
Elevation (BFE) if a previous calculation using the Normal Depth option
computed a depth that was determined to be SUPERCRITICAL. Super Critical
depths are generally not accepted for use as BFEs.
CHANNEL CAPACITY: This option is used to determine a rating curve for a cross
section. The program computes a discharge based on the entered depth.
Repeating with other depths produces a rating curve. A BFE may be determined
by interpolation with the correct discharge.
NORMAL DEPTH: This is the usual option to use in determining BFEs. The user
should watch the "Flow Type" message to make sure that the calculation is
CRITICAL or SUBCRITICAL. Use Option 1 if SUPERCRITICAL.
STEP-BACKWATER: This option should be used to calculate BFEs if more than one
cross-section is warranted to cover the extent of the property. Generally if
the property parallels more than 500 feet of a flooding source this option
should be used.
HANDLES "REGULAR" AND "IRREGULAR" SHAPED CROSS SECTIONS
The REGULAR shape cross-sections include:
• V-shape,
• Trapezoidal,
• Rectangular, and
• Circular
2-1
-------
QUICK-2 User's Guide Overview
For IRREGULAR cross-sections:
- up to 40 points can be input to describe the ground points
- Ground points are easily modified using the Insert or Delete Keys
- Encroachments or other changes in the floodplain are easily modeled
- An unlimited number of cross sections may be modeled
In addition, ground points and other input variables for the irregular shape
cross-sections can be saved to a file, for later use.
SINGLE SCREEN DATA INPUT, COMPUTATION AND OUTPUT
One of the most user-friendly aspects of this program that sets it apart from
many other computational programs is that all of the data input, the
computation, and the printing or plotting, is performed from the same screen.
You will not get lost in a maze of menus.
GRAPHICS
• Cross-Section Plots,
• Water Surface Elevation Profiles, and
• Rating Curve Plots
Cross section plots and water surface elevation profiles from QUICK-2's step-
backwater analysis can be viewed on the screen using the USAGE PLOT-2 program
that comes with the QUICK-2 package of programs. The channel capacity option
of QUICK-2 can be used to generate rating curve plots of individual cross
sections that can be viewed on screen and printed.
AUTOMATIC ERROR CHECKING
This software is designed to virtually eliminate the need for user's manuals.
The program incorporates error-checking routines and warning messages to
alert the user to incorrect input data or potentially incorrect output data.
The program prompts the user for the required input data so that there is no
need to worry about which columns to put data in; whether or not it needs to
be left-justified, or right- justified, etc.
2-2
-------
QUICK-2 User's Guide Overview
SPECIAL FEATURES OF QUICK-2
»» Critical Depth, Channel Capacity, and Normal Depth Options ««
EASY VIEW: All of the input data is viewed on the same screen (and changes
can be made) before starting the computations
EASY CHANGE: After an initial calculation, the following parameters can be
changed, and the above options can be re-calculated in seconds:
Discharge Channel Slope Manning's N
Base width or Diameter Channel Side Slope Ground Points
Channel Stations
AUTO-SAVE: For irregular channels the program automatically stores all the
input variables to a file designated as "TEMP.XSC", which is stored in the
C:\QUICK2\DATA Directory.
RATING CURVES: A special feature of the Channel Capacity Option for irregular
channels is the Rating Curve Print Option. A rating curve plot can be
automatically generated with 20 computations of water surface elevation
versus discharge. The maximum elevation of the rating curve will be just
lower than the channel depth specified by the user. The rating curve can be
viewed on the screen and/or printed.
»» Step-Backwater Option ««
EASY VIEW: All of the input data is viewed on the same screen (and changes
can be made) before starting the computations
PRECISE: Balances the energy equation to within .01 foot.
COMPUTES CRITICAL DEPTH AUTOMATICALLY: After up to 40 energy balance trials
(without a balance) the program automatically computes critical depth.
OUTPUT OPTIONS: Detailed and Summary printouts are available
AUTO-SAVE: The program automatically saves the first cross-section into a
file designated as TO.XSC, and subsequent cross-sections are saved adding the
Channel distance (XLCH) to the previous cross-section's file name.
Therefore, if we run 3 cross-sections that are 200 feet apart their
filenames will be: TO.XSC, T200.XSC, and T400.XSC. These files are
automatically stored in a directory named C:\QUICK2\DATA.
HEC-2 RUNS WITH QUICK-2 FILES: The backwater option also automatically saves
all of the cross-sections into a HEC-2 compatible file called HEC2.DAT, which
is stored in the C:\QUICK2 Directory. The QUICK-2 program is linked with the
USACE HEC-2 program such that any backwater computation that is run using
QUICK-2 can also be run using the HEC-2 program within the QUICK-2
environment. The user does not need to have any previous experience in
running the HEC-2 model.
2-3
-------
QUICK-2 User's Guide Overview
AUTOMATIC ERROR CHECKS AND WARNING MESSAGES
ERROR CHECKS
Error checks prevent the user from continuing by re-prompting the user for
correct input data. The following are error checks performed automatically by
the program:
- Ground Point (GR) stations should be increasing
- Stations of the left and right bank should match a GR point
WARNING MESSAGES
Warning messages instruct the user that the program has had to modify the
input data in order to complete a calculation, or that the completed
calculation may not be valid. The following are warning messages performed by
the program:
Extended Cross Section
The computed water surface elevation is higher than one or both ends of the
cross-section, and the program automatically extended the end(s) of the
cross-section vertically to complete the computation.
Divided Flow
There is a ground point(s) within the cross-section which is higher than the
computed water surface elevation which is dividing the flow within the cross-
section.
No Energy Balance ... Computing Critical Depth
The program attempted up to 40 trial computations and could not arrive at an
energy balance; and therefore, critical depth is assumed to occur at this
cross-section.
Computing Critical Depth ... Critical Depth Assumed
Either the initial Starting Water Surface Elevation or an energy balance
between two cross sections occurred at an elevation for which the froude
number or the index froude number was equal to or greater than 1. Thus, the
computed water surface elevation is suspected of being below the critical
depth. Therefore the critical depth is computed and compared to the previous
calculated water surface elevation. In this case the critical depth elevation
was higher, and thus Critical Depth is Assumed.
Computing Critical Depth ... Critical Depth Not Assumed
Same as above except, the critical depth is computed and compared to the
previous calculated water surface elevation; and, in this case the critical
depth elevation was lower, and thus Critical Depth is Not Assumed.
2-4
-------
QUICK-2 User's Guide
Getting Started
Chapter 3: Getting Started
This section provides you with
convenient installation and run
procedures that will enable you
to run the program from the hard
disk drive or the floppy disk drive.
HARD DISK INSTALLATION AND RUN PROCEDURE
To install and run QUICK-2 simply place the floppy disk in either your "A"
disk drive or your "B" disk drive.
For "A" Drive users: Type A:\AQ2 and Press
For "B" Drive users: Type B:\BQ2 and Press
Follow the screen message to start the program. That's it!
The program resides in a C:\QUICK2 directory. To run the program in the
future, just change to that directory and type Q2 and press .
FLOPPY DISK INSTALLATION AND RUN PROCEDURE
To install and run QUICK-2 from the floppy disk drive simply place the floppy
disk in either your "A" disk drive or your "B" disk drive.
For "A" Drive users: Type A:\FAQ2 and Press
For "B" Drive users: Type B:\FBQ2 and Press
Follow the screen message to start the program. That's it!
To run the program in the future, just place the disk in your floppy drive,
change to that directory and type Q2 and press . Although the
program will run from the floppy disk drive it will run much faster if
installed and run on the hard disk drive.
REMINDER:
Entering and editing data, as well as moving around within the input screens
is performed using the Function keys, the Backspace Key and the Enter Key. DO
NOT USE THE CURSOR CONTROL KEYS (ARROW KEYS) FOR ENTERING, DELETING, OR
EDITING DATA.
3-1
-------
Chapter 4: TUTORIALS
Normal Depth
Step-Backwater
Channel Capacity
PLOT - 2
{TIME REQUIRED TO COMPLETE ALL THE TUTORIALS is ABOUT ONE HOUR}
4-1
-------
QUICK-2 User's Guide
Normal Depth Tutorial
NORMAL DEPTH {Tutorial Time: 5 to 10 minutes}
After pressing Q2 and to start the program you will come to
Main Menu screen of QUICK-2 as shown below.
the
QUICK - 2
MAIN MENU
Critical Depth
Channel Capacity
Normal Depth
Step -Backwater
QUIT
11 .
Press
1
2
3
4
1
MI
| - Help ||
1. Press 3 and then press to start the Normal Depth Option.
Next you will see the Shape of Cross Section screen:
SHAPE OF CROSS SECTION
V - Ditch
Rectangular Channel
Trapezoidal Channel
Circular Channel
Irregular Channel
| - Main Menu ||
Let's try the Trapezoidal Channel option.
2. Press T and then press to perform a Normal Depth calculation
for a trapezoidal channel.
4-2
-------
QUICK-2 User's Guide
Normal Depth Tutorial
The next screen you will see is the Input / Output screen
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT VARIABLES
L Side Slope (H:V)
Bottom Width (ft)
Discharge (cfs)
Slope (ft/ft)
R Side Slope (H:V)
Manning's n
Depth (ft)
OUTPUT VARIABLES
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
Wet Perimeter (ft)
Hyd Radius
Froude #
Flow Type
Enter Left Side Slope and Press :1
<- Back Tab Main Menu
The program is currently prompting you to enter the Left Side Slope (in
terms of the Number of Horizontal feet (H) to every 1 foot Vertical (H :
1). Let's say our left side slope is 3 to 1 (3:1).
3. Enter 3 and then Press .
The next screen you will see is the Input / Output screen with a new
prompt:
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT VARIABLES
L Side Slope (H:V) 3.0:1 R Side Slope (H:V)
Bottom Width (ft)
Discharge (cfs)
Slope (ft/ft)
Manning's n
Depth (ft)
OUTPUT VARIABLES
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
Wet Perimeter (ft)
Hyd Radius
Froude #
Flow Type
Enter Right Side Slope and Press :1
<- Back Tab Main Menu
Notice that the
3 has been entered
to the right of
"L Side Slope (H:V)"
The program is currently prompting you to enter the Right Side Slope (in
terms of the Number of Horizontal feet (H) to every 1 foot Vertical (H :
1). Let's say our right side slope is 2 to 1 (2:1).
4. Enter 2 and then Press .
4-3
-------
QUICK-2 User's Guide
Normal Depth Tutorial
1 NORMAL DEPTH
1 TRAPEZOIDAL CHANNEL
jj INPUT
1 L Side Slope (H:V) 3.0:1
1 Bottom Width (ft)
1 Discharge (cfs)
1 Slope (ft/ft)
||
1 OUTPUT
1 Area (sq ft)
1 Velocity (ft/s)
1 Top Width (ft)
jj
I <- Back Tab
VARIABLES
R Side Slope (H:V) 2.0:1
Manning ' s n
Depth (ft)
VARIABLES
Wet Perimeter (ft)
Hyd Radius
Froude #
Flow Type
Main Menu
The program will continue to prompt you for input data.
Let's say our channel is 10 feet wide, with a Manning's n value of 0.035,
the discharge is 300 cfs, and the channel slope is .005 ft/ft.
SCREEN PROMPT - "Enter Bottom Width and Press "
5. Enter 10 and then Press .
SCREEN PROMPT - "Enter Manning's n and Press "
6. Enter .035 and then Press .
SCREEN PROMPT - "Enter Discharge and Press "
7. Enter 300 and then Press .
SCREEN PROMPT - "Enter Slope and Press "
8. Enter .005 and then Press .
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT VARIABLES
II ^H
II
L Side Slope (H:V) 3.0:1
Bottom Width (ft) 10.0
Discharge (cfs) 300
Slope (ft/ft) 0.0050
R Side Slope (H:V) 2.0:1
Manning's n 0.0350
Depth (ft) 0.00
OUTPUT VARIABLES
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
Wet Perimeter (ft)
Hyd Radius
Froude #
Flow Type
Begin Calculations
<- Back Tab
Main Menu
After all the data
is input your screen
should look like this
4-4
-------
QUICK-2 User's Guide
Normal Depth Tutorial
To begin the calculation simply ...
9. Press .
After a split second the screen should look like this:
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT VARIABLES
| L Side Slope (H:V) 3.0:1 R Side Slope
| Bottom Width (ft) 10.0 Manning's n
| Discharge (cfs) 300 Depth (ft)
| Slope (ft/ft) 0.0050
(H:V) 2.0:1 |
0.0350 |
3.27 |
1
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
OUTPUT VARIABLES
59.6 Wet Perimeter (ft) 27.7
5.0 Hyd Radius 2.2
26.4 Froude # 0.6
•
•
•
1
Flow Type SUBCRITICAL
Begin Calculations
Print
<- Back Tab
Main Menu
1
Notice that the Depth is no longer 0.00, but equals 3.27 feet, which is the
Normal Depth for this particular Trapezoidal cross-section. If 300 cfs
represents the 100-year discharge, then the 100-year flood depth would
equal 3.27 feet. All of the output variables have also been computed and
listed.
10. To print the output simply Press the Function key.
The printed output is shown below.
QUICK - 2
NORMAL DEPTH
Trapezoidal Channel
INPUT VARIABLES
n = 0.035
Depth
3.0 \
v
\
Base Width = 10.0
Slope = 0.0050
/ 2.0
4-5
OUTPUT VARIABLES
Depth (ft) 3.27
Discharge (cfs) 300.0
Velocity (ft/s) 5.04
Top Width (ft) 26.4
Froude No. 0.59
Flow Type: SUBCRITICAL
-------
QUICK-2 User's Guide
Normal Depth Tutorial
CHANGING THE VARIABLES
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT VARIABLES
L Side Slope (H:V) 3.0:1
Bottom Width (ft) 10.0
Discharge (cfs) 300
Slope (ft/ft) 0.0050
R Side Slope
Manning's n
Depth (ft)
(H:V) 2.0:1
0.0350
3.27
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
OUTPUT VARIABLES
59.6 Wet Perimeter (ft) 27.7
5.0 Hyd Radius 2.2
26.4 Froude # 0.6
I
1
Flow Type SUBCRITICAL
Begin Calculations
Print
<- Back Tab
Main Menu
1
Let's say we want to run this calculation again but with a discharge of 500
cfs instead of 300 cfs.
1. Press the Function Key
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT
L Side Slope (H:V) 3.0:1
Bottom Width (ft) 10.0
Discharge (cfs) 300
Slope (ft/ft) 0.0050
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
Enter Slope and
<- Back Tab
OUTPUT
59.6
5.0
26.4
Press
VARIABLES
R Side Slope (H:V)
Manning ' s n
Depth (ft)
VARIABLES
2.0:1
0.0350
3.27
Wet Perimeter (ft) 27.7
Hyd Radius 2.2
Froude # 0.6
Flow Type SUBCRITICAL
Main Menu
The above screen is what you should be looking at. The key will move
the prompt backwards through all the variables. Note that since we want to
change the Discharge (from 300 to 500) , we will need to Press again to
come to the Enter Discharge prompt. Follow the steps as shown on the
following page to rerun this calculation with a new discharge.
4-6
-------
QUICK-2 User's Guide
Normal Depth Tutorial
SCREEN PROMPT - "Enter Slope and Press "
2. Press .
SCREEN PROMPT - "Enter Discharge and Press "
3. Enter 500 and then Press .
SCREEN PROMPT - "Enter Slope and Press "
4. Press .
After all of the data is input your screen should look like this
NORMAL DEPTH
TRAPEZOIDAL CHANNEL
INPUT VARIABLES
L Side Slope (H:V) 3.0:1
Bottom Width (ft) 10.0
Discharge (cfs) 500
Slope (ft/ft) 0.0050
R Side Slope (H:V) 2.0:1
Manning's n 0.0350
Depth (ft) 3.27
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
OUTPUT VARIABLES
59.6 Wet Perimeter (ft) 27.7
5.0 Hyd Radius 2.2
26.4 Froude # 0.6
Flow Type SUBCRITICAL
Begin Calculations
<- Back Tab
Main Menu
4-7
-------
QUICK-2 User's Guide
Normal Depth Tutorial
5. Press to begin the calculation.
After a split second the screen should look like this
NORMAL DEPTH ||B
| TRAPEZOIDAL CHANNEL ||B
1 «•
| INPUT VARIABLES ||B
•
L Side Slope (H:V)
Bottom Width (ft)
Discharge (cfs)
Slope (ft/ft)
3.0:1
10.0
500
0.0050
OUTPUT
Area (sq ft)
Velocity (ft/s)
Top Width (ft)
Begin Calculations
| Print
j <- Back Tab
86.7
5.8
31.1
R Side Slope (H:V) 2.0:1
Manning's n 0.0350
Depth (ft) 4.22
VARIABLES
Wet Perimeter (ft) 32.8
Hyd Radius 2 . 6
Froude # 0.6
Flow Type SUBCRITICAL
III
Main Menu
•
•
•
•
•
•
•
•
•
•
•
Let's return to the Main Menu.
Just Press the Function Key
QUICK - 2
MAIN MENU
Critical Depth
Channel Capacity
Normal Depth
Step-Backwater
-II
I
QUIT
^ If you want to continue and to perform the Step-Backwater
Tutorial, then turn to the next page.
If you want to exit out of the program for now, Press
4-8
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
STEP-BACKWATER {Tutorial Time: 20 to 25 minutes}
Let's say that we have a piece of property located in an unnumbered Zone A,
and we need to determine if our property is really in or out of the
floodplain. We will be referring to Figure 1 on the next page which
represents a plan view of our proposed floodplain study (step-backwater
analysis) . We have field surveyed 3 cross-sections to use in the step-
backwater analysis. The next page lists all of the data from the field
surveyed cross-sections.
If you have continued from the previous Normal Depth Tutorial you should
see the screen below. If you are just starting the program, you will see
the screen below after pressing Q2 and .
QUICK - 2
MAIN MENU
Critical Depth
Channel Capacity
Normal Depth
Step-Backwater
QUIT
1. Press 4 and then press to start the Step-Backwater Option.
Next you will see the Starting Water Surface Elevation Method screen:
Starting Water Surface Elevation Method
Input
NORMAL DEPTH (Slope-Area)
Enter the Slope in Ft/Ft
(for ex. .0025)
KNOWN WATER SURFACE ELEVATION
Enter the known WS Elevation (for ex. 656.78)
Enter a Slope or an Elevation:
Let's say that we do not have any previous information about flood
elevations for our sample stream. Thus we need to start the step-backwater
analysis assuming that the flow in our first cross-section is at Normal
Depth. (This assumes that the channel slope downstream of our first cross-
section will approximate the slope of the energy grade at the first cross-
section of our study.) Let's assume that our calculated downstream channel
slope is .0029 ft/ft.
2. Type .0029 and then press .
4-9
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
CROSS SECTION INFORMATION
Cross-Section 1
GROUND POINTS
Station Elevation
362 505.0
425 499.1
509 498.0
512 496.9
602 496.9
605 498.2
732 500.1
1020 504.7
CHANNEL BANK STATIONS
Left 509 Right 605
MANNING ' S N VALUES
Left .065
Channel .040
Right .060
CHANNEL REACH LENGTHS
Left
Channel
Right
LOSS COEFFICIENTS
Cont Expan
100 -YEAR DISCHARGE
3000
Cross-Section 2
Station Elevation
0 510.0
150 504.8
233 502.2
236 500.9
331 500.9
334 501.8
402 505.5
591 510.1
Left 233 Right 334
Left .055
Channel .040
Right .060
Left 450
Channel 450
Right 450
Cont 0 . 1 Expan 0 . 3
3000
Cross-Section 3
Station Elevation
0 515.0
433 510.1
600 506.3
614 504.9
701 504.8
725 506.5
866 511.1
1240 514.6
Left 600 Right 725
Left .065
Channel .040
Right .060
Left 490
Channel 490
Right 490
Cont 0 . 1 Expan 0 . 3
3000
FIGURE 1
X SEC. 450
X SEC. 940
4-10
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
The next screen you will see is the Input / Output screen as shown:
XSEC ID:
STAT
CHANNEL
MANNING
CHANNEL
LOSS
WS ELEV
EG ELEV
0 » STEP -
ELEV STAT ELEV
BANK STATIONS: Left
S N VALUES: Left
REACH LENGTHS: Left
COEFFICIENTS: Contractn
Depth
Flow Regime
BACKWATER «
STAT ELEV
Channel
Channel
Expansn
Top Wid
ChanVel
GROUND POINTS
STAT ELEV
Right
Right
Right
:Dschrg
Kratio
Froudtt
||F2}<-Back Tab F5}List Files F6}Retrieve File F7}Main Menu F10}Ed/Ex GrPt| |
F3}lnsert GrPt F4}Delete GrPt Fl }HELP
Before we go on let's read about how data is to be input for this screen.
3. Press to access the Help screen.
4. When you are finished reading the Help screen just Press .
If you refer to the previous page, you will see a tabulation of the Ground
Points for the first field surveyed cross-section listed by Station and
Elevation. You will also see the Channel Bank Stations, Manning's N values,
and Discharge.
5. Following the method explained in the Help Screen, enter the Ground
Points one at a time, by their respective Station and Elevation. Be sure to
Press after you have typed in each correct number.
Once you have entered all of the Ground Points correctly ...
6. Press to Exit from entering Ground Point data
NOTE: The Key will EXIT you from the top of the screen, or it will
RETURN you to the top of the screen if you need to go back to EDIT the
Ground Points.
4-11
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
Your screen should now look like this:
XSEC ID:
STAT
362
602
CHANNEL
MANNING '
CHANNEL
LOSS
WS ELEV
EG ELEV
0
» STEP - BACKWATER «
ELEV STAT ELEV STAT ELEV
505.0 425
496.9 605
BANK STATIONS
S N VALUES
REACH LENGTHS
COEFFICIENTS
499.1 509 498.0
498.2 732 500.1
: Left
: Left Channel
: Left Channel
: Contractn Expansn
Depth Top Wid
Flow Regime ChanVel
1 F2}<- Back Tab F5}PRINT F6}SAVE F7}Main Menu F8}New XSEC
1 Enter LEFT Channel
Bank Station and Press
GROUND POINTS
STAT
512
1020
Right
Right
Right
Dschrg
Kratio
Froudtt
F10}Ed/Ex
ELEV
496.9
504.7
II
GrPt 1
1
The program is currently prompting you to enter the Left Channel Bank
Station. Using the information contained on the previous page, we know that
our Left Channel Bank Station is 509. Therefore ...
7. Enter 509 and then Press . (Notice that the 509 has been
entered to the right of "CHANNEL BANK STATIONS: Left".)
Next you will see the Input / Output screen with a new prompt:
SCREEN PROMPT - "Enter RIGHT Channel Bank Station and Press "
Using the information for Cross-section 1, simply follow the screen prompts
to input the required data, as follows:
SCREEN PROMPT - "Enter RIGHT Channel Bank Station and Press "
8. Type 605 and then Press .
SCREEN PROMPT - "Enter LEFT Manning's n Value and Press "
9. Type .065 and then Press .
SCREEN PROMPT - "Enter CHANNEL Manning's n Value and Press "
10. Type .040 and then Press .
SCREEN PROMPT - "Enter RIGHT Manning's n Value and Press "
11. Type .060 and then Press .
SCREEN PROMPT - "Enter Discharge and Press "
12. Type 3000 and then Press .
4-12
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
Your screen should now look like this:
XSEC ID: 0
STAT ELEV STAT
362 505.0 425
602 496.9 605
CHANNEL BANK STATIONS:
MANNING ' S N VALUES :
CHANNEL REACH LENGTHS:
» STEP - BACKWATER «
ELEV STAT ELEV
499.1 509 498.0
498.2 732 500.1
Left 509.0
Left 0.0650 Channel 0.0400
Left Channel
LOSS COEFFICIENTS: Contractn Expansn
WS ELEV Depth Top Wid
EG ELEV Flow Regime ChanVel
F2}<- Back Tab F5} PRINT
TO BEGIN CALCULATIONS
F6}SAVE F7}Main Menu F8}New XSEC
Press
GROUND POINTS
STAT ELEV
512 496.9
1020 504.7
Right 605.0 |
Right 0.0600 |
Right I
Dschrg 3000 |
Kratio ||
Froudtt ||
II
F10}Ed/Ex GrPt|
!!
The program is now ready to begin the calculations since all of the
required data has been entered for the 1st cross-section of our step-
backwater analysis. Note that even at this point, if any of the data on the
screen has been typed in incorrectly, the user can simply press the
key to toggle backwards through all of the input data, even back to the
Ground Points. Remember that you can instantly go back to the Ground Points
by pressing , also.
13. Press to Begin the Calculations.
Your screen should now look like this:
XSEC ID: 0 » STEP - BACKWATER « GROUND POINTS
STAT ELEV STAT ELEV STAT ELEV STAT ELEV
362 505.0 425 499.1 509 498.0 512 496.9
602 496.9 605 498.2 732 500.1 1020 504.7
CHANNEL BANK STATIONS: Left 509.0
MANNING'S N VALUES: Left 0.0650
CHANNEL REACH LENGTHS: Left
LOSS COEFFICIENTS: Contractn
WS ELEV 501.03 Depth 4.13
EG ELEV 501.32 Flow Regime M-l
F2}<- Back Tab F5}PRINT F6}SAVE F7}Main
1
Right 605.0
Channel 0.0400 Right 0.0600
Channel Right
Expansn :Dschrg 3000
Top Wid 385 Kratio 1.00
ChanVel 5.10 Froudtt 0.50
Menu F8}New XSEC F10}Ed/Ex GrPt|
1
As you can see from the screen, the (100-year) Water Surface Elevation (WS
ELEV) has been computed (501.03), with other variables.
4-13
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
Before we move on to enter the data for the next cross-section let's obtain
a printout of this first calculation.
Press .
The screen prompt will be ...
PRINT: Summary or Detailed? Press S or D
and
Let's obtain a Detailed Printout .
Press D and then Press .
Therefore
Assuming your printer is turned on, the detailed printout will look like
this:
H Cross Section: 0
I XLOB: 0 XLCH: 0 XROB: 0 CC: 0 CE: 0
I NLOB: .065 STCHL: 509 NCHL: .04 STCHR: 605 NROB: .06
i STAT
1 362.00
1 602.00
1 CWSEL
1 Chan Vel
1 Depth
1 Discharge
1 501.03
I 5.10
1 4.13
1 3000
ELEV STAT ELEV
505.00 425
496.90 605
EG
HV
HL
OL
501.317
0.29
0.00
0.00
.00 499.10
.00 498.20
ELMIN
KRATIO
Top Width
Froude #
496.90
1.00
385
0.50
STAT
509.00
732.00
QLOB
ALOB
STAT-L
CH- Slope
493
228
404.4
0.0000
If any of the above variables are unfamiliar,
provided in Appendix 1 .
ELEV STAT ELEV
498.00 512
500.10 1020
QCH
ACH
ST-MIDCH
EG-Slope
2003
392
557.0
0.0029
a description
.00 496.90
.00 504.70
QROB
AROB
STAT-R
FlowRegim
505
265
789.9
of each is
If you want to save the cross-section data to a different name and/or
directory, before pressing , you can Press , (F6}SAVE) , to
perform this.
Now we need to enter the data for the 2nd cross-section. Since we are
entering a new cross-section (New XSEC), we need to ...
_ Press .
Before the Screen changes you will notice that at the bottom of the screen
a message will briefly appear ...
SAVING TEMPORARY FILE C:\QUICK2\DATA\TO.XSC
This alerts you that your cross-section data has been saved to a
called TO.XSC, which is located in your C:\QUICK2\DATA directory.
4-14
file
-------
QUICK-2 User's Guide
Step-Backwater Tutorial
Your screen should be blank again as shown below:
XSEC ID: 0 » STEP - BACKWATER « GROUND POINTS
STAT ELEV STAT ELEV STAT ELEV STAT ELEV
CHANNEL BANK STATIONS: Left
MANNING'S N VALUES: Left
CHANNEL REACH LENGTHS: Left
LOSS COEFFICIENTS: Contractn
WS ELEV Depth
EG ELEV Flow Regime
|F2}<-Back Tab F5}List Files F6}Retrieve
1 F3} Insert GrPt F4}Delete GrPt
Right
Channel Right
Channel Right
Expansn :Dschrg
Top Wid Kratio
ChanVel Froudtt
File F7}Main Menu F10}Ed/Ex GrPt |
Fl }HELP 1
_ Following the method used before, for the 1st cross-section, enter the
Ground Points one at a time, by their respective Station and Elevation for
the 2nd cross-section using the data provided. Be sure to Press
after you have typed in each correct number.
_ Once you have entered all of the Ground Points correctly, remember to
Press to Exit from entering Ground Point data .
_ Follow the on screen prompts to enter all of the other data.
Remember that if any of the data on the screen has been typed in
incorrectly, the user can simply press the key to toggle backwards
through all of the input data, even back to the Ground Points. (You can
also Press to go back to the Ground Points immediately for editing).
After entering all the data your screen should now look like this:
XSEC ID:
STAT
0
331
1 CHANNEL
1 MANNING
1 CHANNEL
1 LOSS
I WS ELEV
EG ELEV
450 » STEP -
ELEV STAT ELEV
510.0 150 504.8
500.9 334 501.8
BANK STATIONS: Left
S N VALUES: Left
REACH LENGTHS: Left
COEFFICIENTS: Contractn
Depth
Flow Regime
BACKWATER «
STAT ELEV
233 502.2
402 505.5
233.0
0.0550 Channel 0.0400
450 Channel 450
0 . 1 Expansn 0 . 3
Top Wid
ChanVel
GROUND
STAT
236
591
Right
Right 0
Right
:Dschrg
Kratio
Froudtt
POINTS
ELEV
500.9
510.1
334.0 1
.0600 1
450 1
3000 1
II F2}<- Back Tab F5}PRINT F6}SAVE F7}Main Menu F8}New XSEC F10}Ed/Ex GrPt| |
TO BEGIN CALCULATIONS Press
4-15
-------
QUICK-2 User's Guide Step-Backwater Tutorial
The program is now ready to begin the calculations since all of the
required data has been entered for the 2nd cross-section of our step-
backwater analysis.
_ Press to Begin the Calculations.
Once the calculation is finished you may ...
_ Press to obtain a printout
Press to save the data to another name and/or directory
Finally, to finish our analysis we need to enter in the data for the 3rd
cross-section.
_ Press
Before the Screen changes you will notice that at the bottom of the screen
a message will briefly appear ...
SAVING TEMPORARY FILE C:QUICK2\DATA\T450.XSC
This alerts you that your 2nd cross-section data has been saved to a file
called T450.XSC, which is located in your C:\QUICK2\DATA directory. Notice
that the 450, represents the channel distance between the 1st and 2nd
cross-sections.
_ Following the method used before for the other cross-sections, enter the
Ground Points one at a time, by their respective Station and Elevation for
the 3rd cross-section using the data provided. Be sure to Press
after you have typed in each correct number.
_ Once you have entered all of the Ground Points correctly, remember to
Press to Exit from entering Ground Point data .
_ Follow the on screen prompts to enter all of the other data.
After entering all the data for the 3rd cross-section ...
_ Press to Begin the Calculations.
Once the calculation is finished you may ...
_ Press to obtain a printout
_ Press to save the data to another name and/or directory
TO EXIT OUT OF THIS SCREEN NOW THAT OUR ANALYSIS IS COMPLETED ...
_ Press
~ 4-16
-------
QUICK-2 User's Guide Step-Backwater Tutorial
You will see a screen prompt at the bottom ...
SUMMARY PRINTOUT: Press , otherwise Press
To print a summary of the output for all 3 cross-sections then ...
_ Press , otherwise just Press
The on screen Summary or the printed summary will look like this:
SEGNO
0
450
940
Q
3000
3000
3000
.0
.0
.0
XLCH
0
450
490
CWSEL
501
503
508
.03
.96
.54
FR#
0.50
1.06
0.71
ELMIN
496
500
504
.90
.90
.80
AVG.VEL.
3.39
7.54
4.95
AREA
885
398
606
.0
.1
.5
TOPWID
385.5
196.9
286.2
If we carefully compare the Computed Water Surface Elevations
(CWSELs) at each cross-section, to the topographic contours on
Figure 1, we will see that the property is clearly higher than
the CWSEL at every cross-section. Therefore this analysis with
more detailed cross-section data has proven that the property
has been inadvertently included in an unnumbered Zone A Special
Flood Hazard Area.
Turn to the next page to continue
4-17
-------
QUICK-2 User's Guide Step-Backwater Tutorial
RUNNING HEC-2 USING QUICK-2 FILES {Tutorial Time: 5 minutes}
_ You will be prompted one more time to Press . The next prompt will
ask you a question concerning running the HEC-2 or PLOT-2 programs.
Press Y and to rerun w/HEC-2 or PLOT-2: If NO Press
_ For purposes of this tutorial let's answer "Y" , (and Press ) to
run the HEC-2 program. The next screen that will appear will include the
following:
To Run Type
QUICK-2 Q2
HEC-2 H2
AUTOHEC-2 AH2
PLOT-2 P2
VIEW/PRINT LIST
_ Type AH2 and Press .
Once the HEC-2 run is complete it will return you to the above-mentioned
screen.
NOTE: Typing AH2 runs the HEC-2 program automatically using the QUICK-2
generated HEC2.DAT, HEC-2 data file.
If you are using a HEC-2 data file other than HEC2.DAT, then Type H2 and
Press . Follow the directions on the screen for naming the Input,
Output and Tape95 files; pressing after each filename is typed in.
_ Type LIST and Press , and then enter your output filename,
(Default is HEC2.0UT), to view the results. Note that you move up, down and
across the screen using the ,, the cursor keys, etc.
_ To Print the data that appears on the screen simply Press P.
_ To Exit from the screen simply Press X or the Escape key.
If you would like to complete the next tutorial example, then
Type Q2 and Press ; and, turn to the next page.
4-18
-------
QUICK-2 User's Guide Step-Backwater Tutorial
RERUNNING USING SAVED CROSS-SECTION FILES
{Tutorial Time: 5 minutes}
Let's say that in the analysis that was performed in the previous tutorial,
we want to change the discharge from 3000 to 5000, and run the step-
backwater option again with the same cross-sections. This is quite easily
done. Just follow the steps as shown below.
1. At the Main Menu Screen Type 4 and Press
2. At the Starting Water Surface Elevation Method Screen
Type .0029 and Press
3. At the Input/Output Screen Press to retrieve a saved cross-
section file
Assuming your 1st cross-section file is stored as
C:\QUICK2\DATA\TO.XSC
Type C and Press when prompted for the directory
Type QUICK2\DATA & Press when prompted for the subdirectory
Type TO and Press when prompted for the filename
4. Press to toggle back to the "Enter Discharge" prompt
5. Type 5000 and Press to enter the new discharge
6. Press to Begin the Calculations
7. Press to input another cross-section
Press to retrieve a saved cross-section file
Assuming your 2nd cross-section file is stored as
C:\QUICK2\DATA\T450.XSC
Type C and Press when prompted for the directory
Type QUICK2\DATA & Press when prompted for the subdirectory
Type T450 and Press when prompted for the filename
8. Press to toggle back to the "Enter Discharge" prompt
9. Type 5000 and Press to enter the new discharge
10. Press to Begin the Calculations
4-19
-------
QUICK-2 User's Guide Step-Backwater Tutorial
11. Press to input another cross-section
Press to retrieve a saved cross-section file
Assuming your 3rd cross-section file is stored as
C:\QUICK2\DATA\T940.XSC
Type C and Press when prompted for the directory
Type QUICK2\DATA & Press when prompted for the subdirectory
Type T940 and Press when prompted for the filename
12 . Press to toggle back to the "Enter Discharge" prompt
13 . Type 5000 and Press to enter the new discharge
14 . Press to Begin the Calculations
_ Press to Exit out of the screen
_ Press to obtain a summary printout
_ Press twice to get back to the main menu
_ Press to leave the program
Q.E.D.
4-20
-------
QUICK-2 User's Guide Channel Capacity Tutorial
CHANNEL CAPACITY OPTION WITH THE RATING CURVE PLOT
{Tutorial Time: 5 to 10 minutes}
Let's say that we need to determine a Base Flood Elevation (BFE) for the
property shown in Figure 1. We do not want to exempt the entire property
from the flood plain, only a structure which is located in the middle of
the property. Therefore, we can use one cross-section (the 2nd cross-
section (T450.XSC) from our previous tutorial and shown on Figure 1), to
compute a BFE.
Let's assume that we know the discharge is between 3000 cfs and 4000 cfs
based on our best estimates.
Let's assume our structure does not have a basement; the lowest adjacent
grade (LAG) to the house is at elevation 510 NGVD; and the first floor
elevation (FFE) is 510.5 NGVD.
Let's determine the maximum carrying capacity of the floodplain using a
depth equal to the lowest adjacent grade (510.0) minus the minimum stream
elevation (500.9). For purposes of this example we'll use a depth of 9 feet
(510-501) .
To perform a channel capacity calculation we also need to know the
downstream slope, which in this case is easy to compute from the
information on page 15. Slope = 500.9 - 496.9 / 450 = .0089.
The graphic below sums up our situation so far:
l< FFE = 510.5
\ %—'< LAG = 510.0
\ X/ WSE = ? / t
\ / I
\ Q=3000 - 4000__ / 9.1'
\" / I
\ / i Stream Invert = 500.9
Slope = .0089
Follow the steps as shown on the next page to compute the rating curve
4-21
-------
QUICK-2 User's Guide Channel Capacity Tutorial
1. At the Main Menu Screen Type 2 and Press
2. At the Shape of Cross Section Screen
Type I and Press
3. At the Input/Output Screen Press
We are using the 2nd cross-section file stored as
C:\QUICK2\DATA\T450.XSC
Type C and Press when prompted for the directory
Type QUICK2\DATA & Press when prompted for the subdirectory
Type T450 and Press when prompted for the filename
4. Type .0089 and Press to enter the slope
5. Type 9 and Press to enter the depth
6. Press to Begin the Calculations
7. Press to Plot to screen .... Press to Print
Looking at the rating curve plot we can see that for a discharge range of
between 3000cfs - 4000cfs the BFE ranges from about 504.3 to 504.8. Since
our lowest adjacent grade and first floor elevation are at or above 510, it
is clear that this structure is above the BFE.
FFE = 510.5
\ %—'< LAG = 510.0
\ WSE = 504.3-504.8 / t
\ I ~/ |
\ P=3000 - 4000__ / 9.1'
\" / I
\ / i Stream Invert = 500.9
Slope = .0089
8. Press to continue
9. Press to go back to the Main Menu
10. Press to Exit the program
4-22
-------
QUICK-2 User's Guide PLOT-2 Tutorial
PLOT-2
In general PLOT-2 will only work on QUICK-2 files that have been converted into HEC-2
format using QUICK-2's Step-Backwater option.
Profile plots from PLOT-2 will work only if the QUICK-2 generated data file (HEC2.DAT) is
also run using the HEC-2 program (see Running HEC-2 Using QUICK-2 Files, page 4-18), since
a HEC2.T95 file needs to be generated by the HEC-2 program for use by PLOT-2.
PLOT-2 Cross-section plots can be generated using the QUICK-2 generated data file
(HEC2.DAT) even if it is not run with HEC-2. However, the Cross-section plot will not show
the computed water surface elevation (CWSEL) unless the QUICK-2 HEC2.DAT file is run with
HEC-2, since the CWSEL is found on the HEC2.T95 file.
Note that the user can compute a normal depth elevation for only one cross-section and
have that cross-section plotted by choosing the Step-Backwater option and the Normal Depth
starting water surface elevation method. Once the computation is finished, the user simply
exits (Presses ), and the QUICK-2 program automatically creates the HEC2.DAT file for
that one cross-section, which can be used by the PLOT-2 program.
Let's say that we want to view the water surface elevation profile and the
cross-section plots from our previous tutorial on the Step-Backwater
option.
_ From the QUICK-2 Title screen Press P2
_ You are now into the PLOT-2 program, Press to continue
PROFILE PLOT {Tutorial Time: 5 to 10 minutes}
1. Let's view the profile first. Press 1 from the PLOT-2 main menu
selection
2. Cursor to the HEC2 Tape95 file name entry and Type ?
This will list all of the data files in the QUICK-2 directory. T95 files
are designated with the 3 letter extension .T95 . Therefore cursor over to
highlight that file (HEC2.T95) and Press .
3. Move up to highlight the Plot profiles entry and Press .
4. Your profile is now plotted. Pressing moves you back to the
Profile plots main menu screen. You can explore the different Profile and
Plotting options and replot the profile if you wish.
5. When you are finished plotting, highlight the Return to main menu
message and Press
4-23
-------
QUICK-2 User's Guide PLOT-2 Tutorial
CROSS-SECTION PLOT {Tutorial Time: 5 to 10 minutes}
1. From the PLOT-2 main menu Press 2 from the menu selection.
2. Cursor down to the HEC2 input file name entry and Type ?
This will list all of the data files in the QUICK-2 directory. Input files
are designated with the 3 letter extension .DAT . Therefore cursor over to
highlight that file (HEC2.DAT) and Press . If we want to view a
different data file than that of the profile we previously viewed, we would
have to specify a different file here before proceeding.
3. Cursor down to the HEC2 Tape95 file name entry
Note that we do not have to re-enter this file since we have already
entered it previously. If we want to view a different Tape95 file than that
of the profile we previously viewed, we would have to Type ?, and then
specify a different file here before proceeding.
4. Move up to highlight the Plot cross sections entry and Press .
5. You now have the option of printing all or selected cross sections from
your data file. Press Y for plotting all, or N for plotting selected cross
sections.
Your first cross-section is now plotted. Pressing moves you back to
the Cross-section plots main menu screen or plots additional cross-sections
depending on how many cross-sect ion plots you have. You can explore the
different Cross-section and Plotting options if you wish.
6. Highlight the Return to main menu message and Press
Pressing 4 at the PLOT-2 main menu exits you from PLOT-2 and back to the
QUICK-2 title screen.
Note: To use PLOT-2 and to access data files that are in another directory
(i.e., they are not in the C:\QUICK2 directory), just change to that data
directory (i.e., CD\dirname) and access PLOT-2 by typing C:\QUICK2\PLOT2
(or A:\PLOT2 if using the program from the floppy drive) from that data
directory.
4-24
-------
Chapter 5: FORMULAS
Critical Depth
Channel Capacity
Normal Depth
Step-Backwater
5-1
-------
QUICK-2 User's Guide Critical Depth Formulas
1. CRITICAL DEPTH
In every cross-section for a given discharge there exists a critical depth, where the
energy grade (depth of water plus velocity head - V2/2G) is at a minimum. Increasing the
discharge above this given discharge will force the flow into the super-critical regime.
Discharges below the given discharge will remain in the sub-critical regime.
Super-critical depths will be lower than the critical depth, and sub-critical depths will
be above the critical depth. Super-critical flow is characterized by small water depths
with large velocity heads; while, sub-critical flow is characterized by large water depths
with small velocity heads. A rule of thumb used to determine critical depth is that when
the Velocity Head equals 1/2 the hydraulic depth (Area/Topwidth) critical flow is
probable.
A formula which can be used to approximate critical depth (Dc) is
given below.
Qc2 / g = A3 / T
Where Qc is the discharge (in cfs) based on critical depth, g is the gravitational
constant (32.15 feet/second squared), A is the cross-section area (in square feet), and T
is the top width of the water surface (in feet). Note: for rectangular channels the above
equation can be reduced so that
Dc = (Qc/5.67 T) -6".
The more exact way to compute critical depth (minimum specific energy) is to find a
specific depth of water within a cross-section for a given discharge which produces the
lowest energy grade. The following represents the process that the Critical Depth option
of QUICK-2 goes through to calculate critical depth.
After the cross-section information (ground points, channel stations, etc.) has been input
the program starts computing the water surface elevation (WSE) and corresponding energy
grade elevation (EG) at a depth of 0.1 foot above the lowest elevation in the
cross-section. It continues to calculate WSE and EG at intervals of 0.5 foot. As the depth
of water in the cross-section increases the EG will decrease. At one point the EG will
begin to increase. This means that between the last 0.5 foot interval there exists a
minimum energy grade. Once this has occurred the program decreases the WSE in intervals of
.02 foot. As the depth of water decreases in the cross-section the EG will also decrease
as it approaches the minimum specific energy. At one point the EG will begin to increase
again.
This means that between the last .02 foot interval critical depth exists. At this point
the screen will display the actual critical water surface elevation (along with other
variables) by assuming that the next to the last iteration was the critical depth.
The calculations performed by the program for a given cross-section are listed on the next
page. The calculations include the iterations that the program goes through to arrive at
critical depth.
5-2
-------
QUICK-2 User's Guide
Critical Depth Formulas
ELMIN =92.5
WSE =92.6
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
Qc= 6.026845E-02 Q= 260
Qc =
Qc =
Qc =
Qc =
Oc =
5.314738
24.18726
61.71717
121.451
204.4488
Oc= 311.15
Q=
Q=
Q=
Q=
Q=
Q=
260
260
260
260
260
260
EG=466080.8
EG=452.6406
EG=125.4259
EG=101.21
EG=96.99928
EG=96.14474
EG=96.14897
EG Decreasing
+ EG Increasing
Therefore Minimum Specific Energy is between WSE's of 95.1 and 95.6. Note also that the
Discharge (Q = 260) is also within the computed Critical Discharge (Qc) range of 204 -
311.
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
WSE =
95
95
95
95
95
95
95
95
95
95
95
95
95
95
.58
.56
.54
.52
.50
.48
.46
.44
.42
.40
.38
.36
.34
.32
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
Qc=
QC =
306.5006
301.8821
297.2962
292.7447
288.2256
283.737
279.2825
274.8583
270.4678
266.1075
261.7806
257.4854
253.22
248.986
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
260
260
260
260
260
260
260
260
260
260
260
260
260
260
EG=96.14179
EG=96.13499
EG=96.12863
EG=96.12271
EG=96.11724
EG=96.11225
EG=96.10776
EG=96.1038
EG=96.10038
EG=96.09752
EG=96.09525
EG=96.09361
EG=96.09262
EG=96.09199
- EG Decreasing
minimum
WSE= 95.30
Qc= 244.76
Q= 260
EG=96.09271 + EG Increasing
We assume that
Critical Depth = 95.32', Minimum Specific Energy = 96.09199'
The Froude number would be, Q / Qc, or 260 / 248.986 = 1.04.
It is not unusual for the Froude number to not equal exactly 1.0, since the calculation of
critical discharge using the formula Qc2 / g = A3 / T, does not always yield a WSE that
is exactly at the True minimum specific energy.
You should notice from the above tabulation, that as you approach critical depth (minimum
specific energy), for very small changes in EG there are large jumps in the water surface
elevation. The EG is only changing by .001' to .003' while the WSE changes by .02'. A
0.01' difference in EG can cause a 0.10' change in WSE.
5-3
-------
QUICK-2 User's Guide Channel Capacity Formulas
2. CHANNEL CAPACITY
In this option, a Normal Depth elevation (see 3. NORMAL DEPTH) is input and the program
computes the corresponding discharge. (In the Normal Depth option, the discharge is input
and the program computes a normal depth elevation) . The Manning's equation is used as the
formula for determining the (normal) discharge.
Q = 1.486 A (R-667) S5 / N
Where Q is the discharge (in cfs) , A is the cross-section area (in square feet) , R is the
hydraulic radius (in feet), S is the energy slope (in feet/feet), and N is the Manning's
roughness value.
After the cross-section information (ground points, channel stations, streambed slope,
normal depth elevation(s), etc.) has been input, the program simply solves for the area
(A) and hydraulic radius (R) below the normal depth elevation (specified by the user) and
computes the (normal) discharge directly using the Manning's equation. This is not an
iterative process. The screen will display the (normal) discharge (which represents the
channel capacity) along with other variables.
5-4
-------
QUICK-2 User's Guide
Normal Depth Formulas
3. NORMAL DEPTH
The standard formula for determining normal depth in a cross-section is the Manning's
formula. Water is flowing at normal depth when the energy grade and the hydraulic grade
(water surface) slopes are the same as the stream bed slope. Normal depth profiles occur,
in general, when the flow is uniform, steady, one-dimensional, and is not affected by
downstream obstructions or flow changes. The standard Manning's equation is:
Q = 1.486 A (R-667) S5 /N
Where Q is the discharge (in cfs) , A is the cross-section area (in square feet) , R is the
hydraulic radius (in feet), S is the energy slope (in feet/feet), and N is the Manning's
roughness value.
The exact method for computing normal depth for a given discharge at a particular
cross-section, is to assume that S is equal to the downstream streambed slope and to solve
iteratively for the depth (this obviously assumes N is known). The following represents
the process that the Normal Depth option of QUICK-2 goes through to calculate normal
depth.
After the cross-section information (ground points, channel stations, discharge, streambed
slope, etc.) has been input, the program starts computing discharge using the Manning's
equation at an initial depth of 0.1 foot above the lowest point in the cross-section, and
from that point in 0.5 foot intervals. At some point, the computed discharge will exceed
the given target discharge. The program then uses a converging technique to compute a
discharge (with a corresponding normal depth) that is within 1% of the given discharge. At
this point the screen will display the actual normal depth water surface elevation (along
with other variables).
The calculations performed by the program for a given cross-section are listed below. The
calculations include the iterations that the program goes through to arrive at normal
depth.
ELMIN= 92.5
WSE =92.6
WSE= 93.1
WSE= 93.6
WSE= 94.1
WSE= 94.6
WSE= 95.1
WSE= 95.6
WSE =
WSE =
96.1
95.65697
Q= 260
Q=
Q=
Q=
Q=
Q=
Q=
Q=
Q=
260
260
260
260
260
260
260
260
Computed Q= .023579 below target Q
Computed Q=
Computed Q=
Computed Q=
Computed Q=
Computed Q=
Computed Q=
2.803083
14.11313
38.33264
80.01045
146.9773
245.9516
below
below
below
below
below
below
Computed Q=
Computed Q=
369.2461
258.7531
above target Q
within 1%
We assume that
Normal depth = 95.66 for a Discharge (Q) of 260 cfs
5-5
-------
QUICK-2 User's Guide Step-Backwater Formulas
4. STEP-BACKWATER
The Energy Equation which represents one-dimensional, uniform, and steady flow in open
channels is shown below.
(1) WSEa + HVa = WSEU + HVU + HL + OL
Where WSEa is the water surface elevation at the downstream cross-section, HVa is the
velocity head at the downstream cross-section, WSEu is the water surface elevation at the
upstream cross-section, HVu is the velocity head at the upstream cross-section, HL is the
friction loss between the two cross-sections, and OL is the eddy (contraction or
expansion) loss between the two cross-sections.
Velocity Head, HV, is calculated as follows:
HV = (a) V2 / 2g
Where (a) is alpha the velocity coefficient, V is velocity (Q/A), and g is the
gravitational constant. Alpha (a) is calculated as follows:
(A2) |K13 Kc3 Kr3
(a) = I--- + --- + ---
(K3) JA12 Ac2 Ar2
Where A and K are the total area and conveyance below the water surface, respectively; and
Kl, Kc, Kr and Al, Ac, Ar, are the conveyance and area in the left overbank, channel, and
right overbank, respectively.
Friction Loss, HL, is calculated as follows:
HL = Lw ( Qa + Qu )2 / ( Ka + Ku )2
Where Lw is the discharge weighted reach length between cross- sections, Qa is the
discharge at the downstream cross-section, Qu is the discharge at the upstream cross
section, Ka is the conveyance at the downstream cross-section, and K^ is the conveyance
at the upstream cross-section. This is derived from the Average Conveyance Friction slope
equation.
The Discharge Weighted Reach Length, Lw, is calculated as follows:
Lw = {(LI * Ql) + (Lc * Qc) + (Lr * Qr)} / Qa
Where Qa is the average total discharge between cross-sections; and, LI, Lc, Lr, and Ql,
Qc, Qr, represent the reach length and average discharge between cross-sections for the
left overbank, channel, and right overbank, respectively.
Eddy Loss, OL, is calculated as follows:
OL = (Ce or Cc) * ABS HVa - HVU
Where Ce is the expansion coefficient, Cc is the contraction coefficient, HVa is the
velocity head at the downstream cross-section, and HVu is the velocity head at the
upstream cross-section. When HVu is greater than HVa Ce is utilized. When HVu is greater
than or equal to HVa Cc is utilized.
5-6
-------
QUICK-2 User's Guide
Step-Backwater Formulas
After the cross-section information for the first cross-section has been input, either a
known water surface elevation is input to start the calculations or the water surface
elevation could have been determined by the Normal or Critical Depth options or by another
source or method. The program then computes all pertinent variables for the first
cross-section that will be needed for an energy balance with the next upstream
cross-section. After this the user must put in the appropriate information for the next
cross-section (i.e., ground points, channel stations, reach lengths, contraction and
expansion coefficients, etc.) . Once this is done the program performs a series of trial
iterations to make sure that the Energy Equation (1) listed previously will balance to
within .014 foot. The sequence of trial elevations is listed below.
1ST TRIAL:
Uses the depth of water (DP) of the previous cross-section added to the lowest elevation
(ELMIN) within the current cross-section. If DP + ELMIN is less than the previous WSE
(i.e., adverse slope condition) then the program uses the previous WSE for the 1st trial
at the current cross-section.
2ND TRIAL:
Uses the average of the computed WSE and the WSE assumed in Trial number 1.
3RD TRIAL AND ON ...:
Uses a formula designed to help converge quickly to balance the energy equation as shown
below:
Trial WSE = WSE- (WSE+HV-DG-HL-OL) / (1- ( (Q/QC)2) + ( (1. 5*HL) / (A/W) ) )
Where WSE, HV, HL, OL, QC, A, and W are the latest computations of water surface
elevation, velocity head, friction loss, eddy loss, critical discharge, total area, and
total wetted perimeter, respectively; and, DG is the computed energy grade elevation from
the previous cross-section; and, Q is the discharge at the current cross-section.
For most energy balances between cross-sections that are not at or near critical flow, the
program will balance the energy equation within 5 trials.
The calculations performed by the program for an energy balance between two cross-sections
are listed below. The calculations include the iterations that the program goes through to
arrive at the energy balance.
WSE
Assumed
98.75
98.53744
98.32476
WSE
Calculated
98.32489
98.32472
98.32513
Difference
+.4251099
+.2127228
-.00037
Trial
1
2
3
We assume that the correct WSE = 98.32
Note: Energy balance in this case was accurate to .00037 foot.
5-7
-------
QUICK-2 User's Guide Definition of Variables
Appendix 1: Definition of Variables
ACH - Area within the specified channel below the water surface elevation
ALOB - Area within the specified left overbank below the water surface elevation
AROB - Area within the specified right overbank below the water surface elevation
ALPHA - Velocity head coefficient
AREA or Area - Total area within the cross-section below the water surface elevation
AVG.VEL or Velocity - Average Velocity within the entire cross-section
Base Width - Channel bottom width of a trapezoidal or rectangular cross-section
Bottom Width - Channel bottom width of a trapezoidal or rectangular cross-section
CC - Contraction Coefficient
CE - Expansion Coefficient
CH-SLOPE - Slope of the streambed, Channel Slope
CHAN-VEL or ChanVel - Velocity within the main channel of cross-section
Critical Slope - Slope of the Energy Grade line at Critical Flow
CWSEL - Computed water surface elevation within a cross-section
Depth - Max depth of water in the cross-sect as measured below the water surface elevation
Diameter - Width or Height of a circular pipe
Discharge - The rate of the flow of a volume of water within a cross-section, usually
expressed in cubic feet per second (cfs)
EG or EG ELEV - Energy grade elevation, expressed as, WSE + HV
EG-Slope - Energy grade slope
ELEV - Elev of a ground pt of a cross-sect, as ref to some datum (i.e., NGVD, NAVD, etc.)
ELMIN - Lowest elevation in a cross section
Flow Regime - Type of water surface profile (Supercritical regimes are not computed)
Ml: EG-Slope <= Ch-Slope and FR# < .8 M2: EG-Slope > Ch-Slope and FR# < .8
Cl: EG-Slope <= Ch-Slope and FR# >= .8 C3: EG-Slope > Ch-Slope and FR# >= .8
Flow Type - either, Supercritical, Critical or Subcritical
Froudett, Froude No., Froudtt or FR# - Froude number, used to determine the flow type
(i.e., sub- (FR# < 1), critical (FR# = 1) or super-critical (FR# > 1) flow)
HL - Friction loss between cross sections
HV - Velocity head
Hyd Radius or Hyd R - Hydraulic Radius: equal to (Area / Wet Perimeter)
A-l
-------
QUICK-2 User's Guide Definition of Variables
KRATIO - Ratio of upstream total conveyance to downstream total conveyance
L Side Slope - Ratio of the slope of the left side of a channel in terms of Horizontal
distance in feet to 1 foot Vertical.
Manning's n - Coefficient used to account for the friction caused by earthen, vegetative,
and/or man-made surfaces within a floodplain cross-section.
Max Discharge - The maximum flow possible within a circular pipe, (usually occurring at
.94 * Diameter).
NCHL, NLOB, NROB - Manning's "N" value for the specified channel, left overbank, and right
overbank, respectively.
OL - Expansion/contraction loss
Q - Total discharge in the cross-section
QC - Critical discharge within entire cross-section for a specific water surface elevation
QCH - Discharge within the specified channel of a cross-section
QIC - Critical discharge within the entire cross-section for a specific water surface
elevation, assuming that critical flow is limited to the channel, even if flow is
occurring in the overbanks
QLOB, QROB - Discharge within the specified left overbank, and right overbank,
respectively, of a cross-section
R Side Slope - Ratio of the slope of the right side of a channel in terms of Horizontal
distance in feet to 1 foot Vertical.
SEGNO - Cross section number or identifier
Slope or EG-Slope - Energy grade slope
STAT-L, STAT-R - Station, within a cross-section, of the left edge, and right edge,
respectively, of the water surface
STAT - Station of a ground point of a cross-section
STCHL, STCHR, ST-MIDCH - Station of the left bank, right bank, and mid-point,
respectively, of a cross-section
Top Width or Top Wid - Top width of the water surface within a cross-section
Velocity - Average Velocity within the entire cross-section
Wet Perimeter or Wet Per - actual width of ground within a cross-section below the water
surface elevation.
WS ELEV or CWSEL - Water surface elevation within a cross-section
XLCH, XLOB, XROB - Distance between cross-sections as measured along the channel, left
overbank, and right overbank, respectively.
A-2
-------
NOTES
-------
Appendix 7
Hydraulic Computer Manuals
HEC-2
U.S. Army Corps of Engineers, Hydrologic Engineering Center (HEC), "Water
Surface Profiles, HEC-2, User's Manual," Davis, California, 1991.
HEC-RAS
U.S. Army Corps of Engineers, Hydraulic Engineering Center (HEC), "HEC-RAS,
River Analysis System, User's Manual - Draft," BETA 2 Test Version, Davis,
California, February 1995.
PSUPRO
Federal Emergency Management Agency, "PSUPRO Encroachment Analysis User's
Manual", Washington, D.C., 1989.
SFD
Federal Emergency Management Agency, "Simplified Floodway Determination
Computer Program User's Manual", Washington, D.C., 1989.
WSPRO
U.S. Geological Survey, "Water Surface PROfiles, WSPRO, User's Manual, Reston,
Virginia, 1990.
WSP2
U.S. Department of Agriculture, Natural Resources Conservation Service, "WSP2
Computer Program User's Manual", Technical Release No. 61, Washington, D.C.,
1976.
A7-1
-------
Appendix 8
Normal Depth Hand Calculation
A8-1
-------
Appendix 8 - continued
Normal Depth Hand Calculation
A8-2
-------
Appendix 8 - continued
Normal Depth Hand Calculation
A8-3
-------
Appendix 9
Weir Flow Hand Calculations
A9-1
-------
Appendix 9 - continued
Weir Flow Hand Calculations
A9-2
-------
Appendix 9 - continued
Weir Flow Hand Calculations
A9-3
-------
Appendix 9 - continued
Weir Flow Hand Calculations
A9-4
-------
Appendix 10
Worksheet
Base Flood Elevations in Zone A Areas
Community Name:
Community ID# :
State:
Panel #:
FIRM Date:
Project Identifier
This request is for: Existing
Single Lot [ ]
Other
[ ] Proposed [ ] <5 acres [ ] >5 acres [ ]
Multi-Lot [ ] <50 lots [ ] >50 lots [ ]
-APPROACH USED TO DEVELOP THE BASE FLOOD ELEVATION (BFE)
EXISTING DATA
Available
FEMA
Federal
Other
Not Available
Did Not Check
SIMPLIFIED
Contour Interpolation [ ]
DETAILED
Hydraulics Normal Depth [ ] Weir Flow [ ]
Other
Data Extrapolation [ ]
Culvert Flow [ ]
Hydrology Regression Equations [ ]
Discharge-Drainage [ ]
Other
Rational Formula [ ]
TR-55 [ ]
Topography Topographic Map [ ]
Map Scale: 1" = '
Field Survey tied to Datum?
Datum: NGVD 1929 [ ] Other
# Cross-Sections [ ] Length of Stream
RESULTS
BFE or Depth of 100-year Flood
First Floor Elevation or Depth
Lowest Adjacent Grade to Structure
Lowest Grade on entire Property N/A
or Field Survey [ ]
Contour Interval:
YES
NO
N/A
ft.
A10-1
-------
A Unofficial Reproduction of the
WETLANDS RESEARCH PROGRAM'S
Technical Report Y-87-1
CORPS OF ENGINEERS WETLANDS
DELINEATION MANUAL
by
Environmental Laboratory
DEPARTMENT OF THE ARMY
Waterways Experiment Station, Corps of Engineers
PO Box 631, Vicksburg, Mississippi 39180-0631
January 1987
Final Report
Approved for Public Release; Distribution Unlimited
Prepared for DEPARTMENT OF THE ARMY
US Army Corps of Engineers
Washington, DC 20314-1000
Reproduced by Environmental Technical Services Company, Austin, Texas as a courtesy.
This reproduction will not be formatted as the original and errors will occur in this
reproduction. Some items are missing due to our inability to reproduce them adequately.
Date: February 25, 1997
-------
Wetland Delination Manual, 1987
ABSTRACT
This document presents approaches and methods for identifying and delineating wetlands for
purposes of Section 404 of the Clean Water Act. It is designed to assist users in making
wetland determinations using a multiparameter approach. Except where noted in the manual,
this approach requires positive evidence of hydrophytic vegetation, hydric soils, and wetland
hydrology for a determination that an area is a wetland. The multiparameter approach provides
a logical, easily defensible, and technical basis for wetland determinations. Technical
guidelines are presented for wetlands, deepwater aquatic habitats, and nonwetlands (uplands).
Hydrophytic vegetation, hydric soils, and wetland hydrology are also characterized, and
wetland indicators of each parameter are listed.
Methods for applying the multiparameter approach are described. Separate sections are
devoted to preliminary data gathering and analysis, method selection, routine determinations,
comprehensive determinations, atypical situations, and problem areas. Three levels of routine
determinations are described, thereby affording significant flexibility in method selection.
Four appendices provide supporting information. Appendix A is a glossary of technical terms
used in the manual. Appendix B contains data forms for use with the various methods.
Appendix C, developed by a Federal interagency panel, contains a list of all plant species
known to occur in wetlands of the-region. Each species has been assigned an indicator status
that describes its estimated probability of occurring in wetlands. A second list contains plant
species that commonly occur in wetlands of the region. Morphological, physiological, and
reproductive adaptations that enable a plant species to occur in wetlands are also described,
along with a listing of some species having such adaptations. Appendix D describes the
procedure for examining the soil for indicators of hydric soil conditions, and includes a
national list of hydric soils developed by the National Technical Committee for Hydric Soils.
PREFACE
This manual is a product of the Wetlands Research Program (WRP) of the US Army Engineer
Waterways Experiment Station (WES), Vicksburg, Miss. The work was sponsored by the
Office, Chief of Engineers (OCE), US Army. OCE Technical Monitors for the WRP were Drs.
John R. Hall and Robert J. Pierce, and Mr. Phillip C. Pierce.
The manual has been reviewed and concurred in by the Office of the Chief of Engineers and
the Office of the Assistant Secretary of the Army (Civil Works) as a method approved for
voluntary use in the field for a trial period of 1 year.
This manual is not intended to change appreciably the jurisdiction of the Clean Water Act
(CWA) as it is currently implemented. Should any District find that use of this method
appreciably contracts or expands jurisdiction in their District as the District currently interprets
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
CWA authority, the District should immediately discontinue use of this method and furnish a
full report of the circumstances to the Office of the Chief of Engineers.
This manual describes technical guidelines and methods using a multiparameter approach to
identify and delineate wetlands for purposes of Section 404 of the Clean Water Act.
Appendices of supporting technical information are also provided.
The manual is presented in four parts. Part II was prepared by Dr. Robert T. Huffman,
formerly of the Environmental Laboratory (EL), WES, and Dr. Dana R. Sanders, Sr., of the
Wetland and Terrestrial Habitat Group (WTHG), Environmental Resources Division (ERD),
EL. Dr. Huffman prepared the original version of Part II in 1980, entitled "Multiple Parameter
Approach to the Field Identification and Delineation of Wetlands." The original version was
distributed to all Corps field elements, as well as other Federal resource and environmental
regulatory agencies, for review and comments. Dr. Sanders revised the original version in
1982, incorporating review comments. Parts I, III, and IV were prepared by Dr. Sanders, Mr.
William B. Parker (formerly detailed to WES by the US Department of Agriculture (USDA),
Soil Conservation Service (SCS)) and Mr. Stephen W. Forsythe (formerly detailed to WES by
the US Department of the Interior, Fish and Wildlife Service (FWS)). Dr. Sanders also served
as overall technical editor of the manual. The manual was edited by Ms. Jamie W. Leach of the
WES Information Products Division.
The authors acknowledge technical assistance provided by: Mr. Russell F. Theriot, Mr. Ellis J.
Clairain, Jr., and Mr. Charles J. Newling, all of WTHG, ERD; Mr. Phillip Jones, former SCS
detail to WES; Mr. Porter B. Reed, FWS, National Wetland Inventory, St. Petersburg, Fla.; Dr.
Dan K. Evans, Marshall University, Huntington, W. Va.; and the USDA-SCS. The authors also
express gratitude to Corps personnel who assisted in developing the regional lists of species
that commonly occur in wetlands, including Mr. Richard Macomber, Bureau of Rivers and
Harbors; Ms. Kathy Mulder, Kansas City District; Mr. Michael Gilbert, Omaha District; Ms.
Vicki Goodnight, Southwestern Division; Dr. Fred Weinmann, Seattle District; and Mr.
Michael Lee, Pacific Ocean Division. Special thanks are offered to the CE personnel who
reviewed and commented on the draft manual, and to those who participated in a workshop
that consolidated the field comments.
The work was monitored at WES under the direct supervision of Dr. Hanley K. Smith, Chief,
WTHG, and under the general supervision of Dr. Conrad J. Kirby, Jr., Chief, ERD. Dr. Smith,
Dr. Sanders, and Mr. Theriot were Managers of the WRP. Dr. John Harrison was Chief, EL.
Director of WES during the preparation of this report was COL Allen F. Grum, USA. During
publication, COL Dwayne G. Lee, CE, was Commander and Director. Technical Director was
Dr. Robert W. Whalin.
This report should be cited as follows:
Environmental Laboratory. 1987. "Corps of Engineers Wetlands
Delineation Manual," Technical Report Y-87-1, US Army Engineer
Waterways Experiment Station, Vicksburg, Miss.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
CONTENTS
Page
ABSTRACT 2
PREFACE 2
PART I: INTRODUCTION 5
Background 5
Purpose and Objectives 5
Scope 5
Organization 7
Use 8
PART II: TECHNICAL GUIDELINES 12
Wetlands 12
Deepwater Aquatic Habitats 13
Nonwetlands 13
PART III: CHARACTERISTICS AND INDICATORS OF HYDROPHYTIC
VEGETATION, HYDRIC SOILS, AND WETLAND HYDROLOGY 14
Hydrophytic Vegetation 14
Hydric Soils 19
Wetland Hydrology 24
PART IV: METHODS 29
Section A. Introduction 29
Section B. Preliminary Data Gathering and Synthesis 29
Section C. Selection of Method 38
Section D. Routine Determinations 38
Section E. Comprehensive Determinations 51
Section F. Atypical Situations 60
Section G. Problem Areas 68
REFERENCES 72
BIBLIOGRAPHY 74
APPENDIX A: GLOSSARY 77
APPENDIX C: VEGETATION 91
APPENDIX D: HYDRIC SOILS 98
Modifications and Clarifications to the 1987 Wetland Delineation Manual 101
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
CORPS OF ENGINEERS WETLANDS DELINEATION MANUAL
PART I: INTRODUCTION
Background
1. Recognizing the potential for continued or accelerated degradation of the Nation's waters,
the US Congress enacted the Clean Water Act (hereafter referred to as the Act), formerly
known as the Federal Water Pollution Control Act (33 U.S.C. 1344). The objective of the Act
is to maintain and restore the chemical, physical, and biological integrity of the waters of the
United States. Section 404 of the Act authorizes the Secretary of the Army, acting through the
Chief of Engineers, to issue permits for the discharge of dredged or fill material into the waters
of the United States, including wetlands.
Purpose and Objectives
Purpose
2. The purpose of this manual is to provide users with guidelines and methods to determine
whether an area is a wetland for purposes of Section 404 of the Act.
Objectives
3. Specific objectives of the manual are to:
a. Present technical guidelines for identifying wetlands and distinguishing them from aquatic
habitats and other nonwetlands. Definitions of terms used in this manual are presented in the
Glossary Appendix A.
b. Provide methods for applying the technical guidelines.
c. Provide supporting information useful in applying the technical guidelines.
4. This manual is limited in scope to wetlands that are a subset of "waters of the United
States" and thus subject to Section 404. The term "waters of the United States" has broad
meaning and incorporates both deep-water aquatic habitats and special aquatic sites, including
wetlands (Federal Register 1982), as follows:
a. The territorial seas with respect to the discharge of fill material.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
b, Coastal and inland waters, lakes, rivers, and streams that are navigable waters of the United
States, including their adjacent wetlands.
c. Tributaries to navigable waters of the United States, including adjacent wetlands.
d. Interstate waters and their tributaries, including adjacent wetlands.
e. All others waters of the United States not identified above, such as isolated wetlands and
lakes, intermittent streams, prairie potholes, and other waters that are not a part of a tributary
system to interstate waters or navigable waters of the United States, the degradation or
destruction of which could affect interstate commerce.
Determination that a water body or wetland is subject to interstate commerce and therefore is a
"water of the United States" shall be made independently of procedures described in this
manual.
Special aquatic sites
5. The Environmental Protection Agency (EPA) identifies six categories of special aquatic
sites in their Section 404 b.(l) guidelines (Federal Register 1980), including:
a. Sanctuaries and refuges.
b. Wetlands.
c. Mudflats.
d. Vegetated shallows.
e. Coral reefs.
f. Riffle and pool complexes.
Although all of these special aquatic sites are subject to provisions of the Clean Water Act, this
manual considers only wetlands. By definition (see paragraph 26a), wetlands are vegetated.
Thus, unvegetated special aquatic sites (e.g. mudflats lacking macrophytic vegetation) are not
covered in this manual.
Relationship to wetland classification systems
6. The technical guideline for wetlands does not constitute a classification system. It only
provides a basis for determining whether a given area is a wetland for purposes of Section 404,
without attempting to classify it by wetland type.
7. Consideration should be given to the relationship between the technical guideline for
wetlands and the classification system developed for the Fish and Wildlife Service (FWS), US
Department of the Interior, by Cowardin et al. (1979). The FWS classification system was
developed as a basis for identifying, classifying, and mapping wetlands, other special aquatic
sites, and deepwater aquatic habitats. Using this classification system, the National Wetland
Inventory (NWI) is mapping the wetlands, other special aquatic sites, and deepwater aquatic
habitats of the United States, and is also developing both a list of plant species that occur in
wetlands and an associated plant database. These products should contribute significantly to
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
application of the technical guideline for wetlands. The technical guideline for wetlands as
presented in the manual includes most, but not all, wetlands identified in the FWS system. The
difference is due to two principal factors:
a. The FWS system includes all categories of special aquatic sites identified in the EPA
Section 404 b.(I) guidelines. All other special aquatic sites are clearly within the purview of
Section 404; thus, special methods for their delineation are unnecessary.
b. The FWS system requires that a positive indicator of wetlands be present for any one of the
three parameters, while the technical guideline for wetlands requires that a positive wetland
indicator be present for each parameter (vegetation, soils, and hydrology), except in limited
instances identified in the manual.
Organization
8. This manual consists of four parts and four appendices. PART I presents the background,
purpose and objectives, scope, organization, and use of the manual.
9. PART II focuses on the technical guideline for wetlands, and stresses the need for
considering all three parameters (vegetation, soils, and hydrology) when making wetland
determinations. Since wetlands occur in an intermediate position along the hydrologic
gradient, comparative technical guidelines are also presented for deepwater aquatic sites and
nonwetlands.
10. PART III contains general information on hydrophytic vegetation, hydric soils, and
wetland hydrology. Positive wetland indicators of each parameter are included.
11. PART IV, which presents methods for applying the technical guideline for wetlands, is
arranged in a format that leads to a logical determination of whether a given area is a wetlands
Section A contains general information related to application of methods. Section B outlines
preliminary data-gathering efforts. Section C discusses two approaches (routine and
comprehensive) for making wetland determinations and presents criteria for deciding the
correct approach to use. Sections D and E describe detailed procedures for making routine and
comprehensive determinations, respectively. The basic procedures are described in a series of
steps that lead to a wetland determination.
12. The manual also describes (PART IV, Section F) methods for delineating wetlands in
which the vegetation, soils, and/or hydrology have been altered by recent human activities or
natural events, as discussed below:
a. The definition of wetlands (paragraph 26a) contains the phrase plunder normal
circumstances, which was included because there are instances in which the vegetation in a
wetland has been inadvertently or purposely removed or altered as a result of recent natural
events or human activities. Other examples of human alterations that may affect wetlands are
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
draining, ditching, levees, deposition of fill, irrigation, and impoundments. When such
activities occur, an area may fail to meet the diagnostic criteria for a wetlands Likewise,
positive hydric soil indicators may be absent in some recently created wetlands. In such cases,
an alternative method must be employed in making wetland determinations.
b. Natural events may also result in sufficient modification of an area that indicators of one or
more wetland parameters are absent. For example, changes in river course may significantly
alter hydrology, or beaver dams may create new wetland areas that lack hydric soil conditions.
Catastrophic events (e.g. fires, avalanches, mudslides, and volcanic activities) may also alter or
destroy wetland indicators on a site.
Such atypical situations occur throughout the United States, and all of these cannot be
identified in this manual.
13. Certain wetland types, under the extremes of normal circumstances, may not always meet
all the wetland criteria defined in the manual. Examples include prairie potholes during
drought years and seasonal wetlands that may lack hydrophytic vegetation during the dry
season. Such areas are discussed in PART IV, Section G, and guidance is provided for making
wetland determinations in these areas. However, such wetland areas may warrant additional
research to refine methods for their delineation.
14. Appendix A is a glossary of technical terms used in the manual. Definitions of some
terms were taken from other technical sources, but most terms are defined according to the
manner in which they are used in the
manual.
15. Data forms for methods presented in PART IV are included in Appendix B. Examples of
completed data forms are also provided.
16. Supporting information is presented in Appendices C and D. Appendix C contains lists of
plant species that occur in wetlands. Section 1 consists of regional lists developed by a Federal
interagency panel. Section 2 consists of shorter lists of plant species that commonly occur in
wetlands of each region. Section 3 describes morphological, physiological, and reproductive
adaptations associated with hydrophytic species, as well as a list of some species exhibiting
such adaptations. Appendix D discusses procedures for examining soils for hydric soil
indicators, and also contains a list of hydric soils of the United States.
Use
17. Although this manual was prepared primarily for use by Corps of Engineers (CE) field
inspectors, it should be useful to anyone who makes wetland determinations for purposes of
Section 404 of the Clean Water Act. The user is directed through a series of steps that involve
gathering of information and decisionmaking, ultimately leading to a wetland determination.
A general flow diagram of activities leading to a determination is presented in Figure 1.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
However, not all activities identified in Figure 1 will be required for each wetland
determination. For example, if a decision is made to use a routine determination procedure,
comprehensive determination procedures will not be employed.
PRELIMINARY DATA
GATHERING AND SYNTHESIS
PART IV, SECTION B
I
SELECT METHOD
PART IV,SECTION C
ROUTINE
DETERMINATION
PART IV,SECTION D
COMPREHENSIVE
DETERMINATION
PART IV, SECTION E
JURISDICTIONAL
DETERMINATION
Figure 1. General schematic diagram of activities leading
to a wetland/nonwetland determination
Premise for use of the manual
18. Three key provisions of the CE/EPA definition of wetlands (see paragraph 26a) include:
a. Inundated or saturated soil conditions resulting from permanent or periodic inundation by
ground water or surface water.
b. A prevalence of vegetation typically adapted for life in saturated soil conditions
(hydrophytic vegetation).
c. The presence of "normal circumstances."
25 February 1997
Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 10
19. Explicit in the definition is the consideration of three environmental parameters:
hydrology, soil, and vegetation. Positive wetland indicators of all three parameters are
normally present in wetlands. Although vegetation is often the most readily observed
parameter, sole reliance on vegetation or either of the other parameters as the determinant of
wetlands can sometimes be misleading. Many plant species can grow successfully in both
wetlands and nonwetlands, and hydrophytic vegetation and hydric soils may persist for
decades following alteration of hydrology that will render an area a nonwetland. The presence
of hydric soils and wetland hydrology indicators in addition to vegetation indicators will
provide a logical, easily defensible, and technical basis for the presence of wetlands. The
combined use of indicators for all three parameters will enhance the technical accuracy,
consistency, and credibility of wetland determinations. Therefore, all three parameters were
used in developing the technical guideline for wetlands and all approaches for applying the
technical guideline embody the multiparameter concept.
Approaches
20. The approach used for wetland delineations will vary, based primarily on t;he complexity
of the area in question. Two basic approaches described in the manual are (a) routine and (b)
comprehensive.
21. Routine approach. The routine approach normally will be used in the vast majority of
determinations. The routine approach requires minimal level of effort, using primarily
qualitative procedures. This approach can be further subdivided into three levels of required
effort, depending on the complexity of the area and the amount and quality of preliminary data
available. The following levels of effort may be used for routine determinations:
a. Level 1 - Onsite inspection unnecessary. (PART IV, Section D, Subsection 1).
b. Level 2 - Onsite inspection necessary. (PART IV, Section D, Subsection 2).
c. Level 3 - Combination of Levels 1 and 2. (PART IV, Section D, Subsection 3).
22. Comprehensive approach. The comprehensive approach requires application of
quantitative procedures for making wetland determinations. It should seldom be necessary,
and its use should be restricted to situations in which the wetland is very complex and/or is the
subject of likely or pending litigation. Application of the comprehensive approach (PART IV,
Section E) requires a greater level of expertise than application of the routine approach, and
only experienced field personnel with sufficient training should use this approach.
Flexibility
23. Procedures described for both routine and comprehensive wetland determinations have
been tested and found to be reliable. However, site-specific conditions may require
modification of field procedures. For example, slope configuration in a complex area may
necessitate modification of the baseline and transect positions. Since specific characteristics
(e.g. plant density) of a given plant community may necessitate the use of alternate methods
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 11
for determining the dominant species, the user has the flexibility to employ sampling
procedures other than those described. However, the basic approach for making wetland
determinations should not be altered (i.e. the determination should be based on the dominant
plant species, soil characteristics, and hydrologic characteristics of the area in question). The
user should document reasons for using a different characterization procedure than described
in the manual. CA UTION: Application of methods described in the manual or the modified
sampling procedures requires that the user be familiar with wetlands of the area and use his
training, experience, and good judgment in making wetland determinations.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 12
PART II: TECHNICAL GUIDELINES
24. The interaction of hydrology, vegetation, and soil results in the development of
characteristics unique to wetlands. Therefore, the following technical guideline for wetlands is
based on these three parameters, and diagnostic environmental characteristics used in applying
the technical guideline are represented by various indicators of these parameters.
25. Because wetlands may be bordered by both wetter areas (aquatic habitats) and by drier
areas (nonwetlands), guidelines are presented for wetlands, deepwater aquatic habitats, and
nonwetlands. However, procedures for applying the technical guidelines for deepwater aquatic
habitats and nonwetlands are not included in the manual.
Wetlands
26. The following definition, diagnostic environmental characteristics, and technical approach
comprise a guideline for the identification and delineation of wetlands:
a. Definition. The CE (Federal Register 1982) and the EPA (Federal Register 1980) jointly
define wetlands as: Those areas that are inundated or saturated by surface or ground water at a
frequency and duration sufficient to support, and that under normal circumstances do support,
a prevalence of vegetation typically adapted for life in saturated soil conditions. Wetlands
generally include swamps, marshes, bogs, and similar areas.
b. Diagnostic environmental characteristics. Wetlands have the following general diagnostic
environmental characteristics:
(1) Vegetation. The prevalent vegetation consists of macrophytes that are typically adapted to
areas having hydrologic and soil conditions described in a above. Hydrophytic species, due to
morphological, physiological, and/or reproductive adaptations), have the ability to grow,
effectively compete, reproduce, and/or persist in anaerobic soil conditions. Footnote: Species
(e.g. Acer rubrum) having broad ecological tolerances occur in both wetlands and
nonwetlands. Indicators of vegetation associated with wetlands are listed in paragraph 35.
(2) Soil. Soils are present and have been classified as hydric, or they possess characteristics
that are associated with reducing soil conditions. Indicators of soils developed under reducing
conditions are listed in paragraphs 44 and 45.
(3) Hydrology. The area is inundated either permanently or periodically at mean water depths
<6.6 ft, or the soil is saturated to the surface at some time during the growing season of the
prevalent vegetation. The period of inundation or soil saturation varies according to the
hydrologic/soil moisture regime and occurs in both tidal and nontidal situations. Indicators of
hydrologic conditions that occur in wetlands are listed in paragraph 49.
c. Technical approach for the identification and delineation of wetlands. Except in certain
situations defined in this manual, evidence of a minimum of one positive wetland indicator
from each parameter (hydrology, soil, and vegetation) must be found in order to make a
positive wetland determination.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 13
Deepwater Aquatic Habitats
27. The following definition, diagnostic environmental characteristics, and technical approach
comprise a guideline for deepwater aquatic habitats:
a. Definition. Deepwater aquatic habitats are areas that are permanently inundated at mean
annual water depths >6.6 ft or permanently inundated areas <6.6 ft in depth that do not support
rooted-emergent or woody plant species. Areas <6.6ft mean annual depth that support only
submergent aquatic plants are vegetated shallows, not wetlands.
b. Diagnostic environmental characteristics. Deepwater aquatic habitats have the following
diagnostic environmental characteristics:
(1) Vegetation. No rooted-emergent or woody plant species are present in-these permanently
inundated areas.
(2) Soil. The substrate technically is not defined as a soil if the mean water depth is >6.6 ft or
if it will not support rooted emergent or woody plants.
(3) Hydrology. The area is permanently inundated at mean water depths >6.6 ft.
c. Technical approach for the identification and delineation of deepwater aquatic habitats.
When any one of the diagnostic characteristics identified in b above is present, the area is a
deepwater aquatic habitat.
Nonwetlands
28. The following definition, diagnostic environmental characteristics, and technical approach
comprise a guideline for the identification and delineation of nonwetlands:
a. Definition. Nonwetlands include uplands and lowland areas that are neither deepwater
aquatic habitats, wetlands, nor other special aquatic sites. They are seldom or never inundated,
or if frequently inundated, they have saturated soils for only brief periods during the growing
season, and, if vegetated, they normally support a prevalence of vegetation typically adapted
for life only in aerobic soil conditions.
b. Diagnostic environmental characteristics. Nonwetlands have the following general
diagnostic environmental characteristics:
(1) Vegetation. The prevalent vegetation consists of plant species that are typically adapted
for life only in aerobic soils. These mesophytic and/or xerophytic macrophytes cannot persist
in predominantly anaerobic soil conditions. Some species, due to their broad ecological
tolerances, occur in both wetlands and nonwetlands (e.g. Acer rubrum).
(2) Soil. Soils, when present, are not classified as hydric, and possess characteristics
associated with aerobic conditions.
(3) Hydrology. Although the soil may be inundated or saturated by surface water or ground
water periodically during the growing season of the prevalent vegetation, the average annual
duration of inundation or soil saturation does not preclude the occurrence of plant species
typically adapted for life in aerobic soil conditions.
(L. Technical approach for the identification and delineation of nonwetlands. When any one of
the diagnostic characteristics identified in b above is present, the area is a nonwetland.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 14
PART III: CHARACTERISTICS AND INDICATORS OF
HYDROPHYTIC VEGETATION, HYDRIC SOILS, AND WETLAND
HYDROLOGY
Hydrophytic Vegetation
Definition
29. Hydrophytic vegetation. Hydrophytic vegetation is defined herein as the sum total of
macrophytic plant life that occurs in areas where the frequency and duration of inundation or
soil saturation produce permanently or periodically saturated soils of sufficient duration to
exert a controlling influence on the plant species present. The vegetation occurring in a
wetland may consist of more than one plant community (species association). The plant
community concept is followed throughout the manual. Emphasis is placed on the assemblage
of plant species that exert a controlling influence on the character of the plant community,
rather than on indicator species. Thus, the presence of scattered individuals of an upland plant
species in a community dominated by hydrophytic species is not a sufficient basis for
concluding that the area is an upland community. Likewise, the presence of a few individuals
of a hydrophytic species in a community dominated by upland species is not a sufficient basis
for concluding that the area has hydrophytic vegetation.
CAUTION: In determining whether an area is "vegetated" for the purpose of Section 404
jurisdiction, users must consider the density of vegetation at the site being evaluated. While it
is not possible to develope a numerical method to determine how many plants or how much
biomass is needed to establish an area as being vegetated or unvegetated, it is intended that
the predominant condition of the site be used to make that characterization. This concept
applies to areas grading from wetland to upland, and from wetland to other waters. This
limitation would not necessarily apply to areas which have been disturbed by man or recent
natural events.
30. Prevalence of vegetation. The definition of wetlands (paragraph 26a) includes the phrase
"prevalence of vegetation." Prevalence, as applied to vegetation, is an imprecise, seldom-used
ecological term. As used in the wetlands definition, prevalence refers to the plant community
or communities that occur in an area at some point in time. Prevalent vegetation is
characterized by the dominant species comprising the plant community or communities.
Dominant plant species are those that contribute more to the character of a plant community
than other species present, as estimated or measured in terms of some ecological parameter or
parameters. The two most commonly used estimates of dominance are basal area (trees) and
percent areal cover (herbs). Hydrophytic vegetation is prevalent in an area when the dominant
species comprising the plant community or communities are typically adapted for life in
saturated soil conditions.
31. Typically adapted. The term "typically adapted" refers to a species being normally or
commonly suited to a given set of environmental conditions, due to some morphological,
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 15
physiological, or reproductive adaptation (Appendix C, Section 3). As used in the CE wetlands
definition, the governing environmental conditions for hydrophytic vegetation are saturated soils
resulting from periodic inundation or saturation by surface or ground water. These periodic events
must occur for sufficient duration to result in anaerobic soil conditions. When the dominant
species in a plant community are typically adapted for life in anaerobic soil conditions, hydrophytic
vegetation is present. Species listed in Appendix C, Section 1 or 2. that have an indicator status of
OBL, FACW, or FAC (Table 1) are considered to be typically adapted for life in anaerobic soil
conditions (see paragraph 35a). Species having a FAC- indicator status are not considered to be
typically adapted for life in anaerobic soil conditions.
Influencing factors
32. Many factors (e.g. light, temperature, soil texture and permeability, man-induced
disturbance, etc.) influence the character of hydrophytic vegetation. However, hydrologic
factors exert an overriding influence on species that can occur in wetlands. Plants lacking
morphological, physiological, and/or reproductive adaptations cannot grow, effectively
compete, reproduce, and/or persist in areas that are subject to prolonged inundation or
saturated soil conditions.
Geographic diversity
33. Many hydrophytic vegetation types occur in the United States due to the diversity of
interactions among various factors that influence the distribution of hydrophytic species.
General climate and flora contribute greatly to regional variations in hydrophytic vegetation.
Consequently, the same associations of hydrophytic species occurring in the southeastern
United States are not found in the Pacific Northwest. In addition, local environmental
conditions (e.g. local climate, hydrologic regimes, soil series, salinity, etc.) may result in broad
variations in hydrophytic associations within a given region. For example, a coastal saltwater
marsh will consist of different species than an inland freshwater marsh in the same region. An
overview of hydrophytic vegetation occurring in each region of the Nation has been published
by the CE in a series of eight preliminary wetland guides (Table 2), and a group of wetland and
estuarine ecological profiles (Table 3) has been published by FWS.
Classification
34. Numerous efforts have been made to classify hydrophytic vegetation. Most systems are
based on general characteristics of the dominant species occurring in each vegetation type.
These range from the use of general physiognomic categories (e.g. overstory, subcanopy,
ground cover, vines) to specific vegetation types (e.g. forest type numbers as developed by the
Society of American Foresters). In other cases, vegetational characteristics are combined with
hydrologic features to produce more elaborate systems. The most recent example of such a
system was developed for the FWS by Cowardin et al. (1979).
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
16
Table 1
Plant Indicator Status Categories*
Indicator Category
OBLIGATE WETLAND
PLANTS
Indicator
Symbol
OBL
FACULTATIVE WETLAND
PLANTS
FACW
FACULTATIVE UPLAND
PLANTS
FACU
OBLIGATE UPLAND
PLANTS
OBL
Definition
Plants that occur almost always (estimated
probability >99%) in wetlands under natural
conditions, but which may also occur rarely
(estimated probability <1%) in nonwetlands.
Examples: Spartina alterniflora, Taxodium
distichum.
Plants that occur usually (estimated probability
>67% to 99%) inwetlands, but may also occur
(estimated probability 1% to 33% in nonwet-
lands). Examples: Fraxinuspennsylvanica,
Cornus stolonifera.
Plants that occur sometimes (estimated prob-
ability 1% to <33%) in wetlands, but occur
more often (estimated probability >67% to
99%) in nonwetlands. Examples: Quercus
rubra, Potentilla arguta.
Plants that occur rarely (estimated probability
<1%) in wetlands, but occur almost always
(estimated probability >99%) in nonwetlands
under natural conditions. Examples: Pinus
echinata, Bromus mollis.
* Categories were originally developed and defined by the USFWS National Wetlands Inventory
and subsequently modified by the National Plant List Panel. The three facultative categories are
subdivided by (+) and (-) modifiers (See Appendix C, Section 1).
25 February 1997
Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 17
Indicators of hydrophytic vegetation
35. Several indicators may be used to determine whether hydrophytic vegetation is present on
a site. However, the presence of a single individual of a hydrophytic species does not mean
that hydrophytic vegetation is present. The strongest case for the presence of hydrophytic
vegetation can be made when several indicators, such as those in the following list, are present.
However, any one of the following is indicative that hydrophytic vegetation is present:
Indicators are listed in order of decreasing reliability. Although all are valid indicators, some
are stronger than others. When a decision is based on an indicator appearing in the lower
portion of the list, re-evaluate the parameter to ensure that the proper decision was reached.
a. More than 50 percent of the dominant species are OBL. FACW. or FAC-** (Table 1) on
lists of plant species that occur in wetlands. A national interagency panel has prepared a
National List of Plant Species that occur in wetlands. This list categorizes species according to
their affinity for occurrence in wetlands. Regional subset lists of the national list, including
only species having an indicator status of OBL, FACW, or FAC, are presented in Appendix C,
Section 1. The CE has also developed regional lists of plant species that commonly occur in
wetlands (Appendix C, Section 2). Either list may be used. Note: A District that, on a
subregional basis, questions the indicator status of FAC species may use the following option:
When FAC species occur as dominants along with other dominants that are not FAC (either
wetter or drier than FAC), the FAC species can be considered as neutral, and the vegetation
decision can be based on the number of dominant species wetter than FAC as compared to the
number of dominant species drier than FAC. When a tie occurs or all dominant species are
FAC, the nondominant species must be considered. The area has hydrophytic vegetation when
more than 50 percent of all considered species are wetter than FAC. When either all
considered species are FAC or the number of species wetter than FAC equals the number of
species drier than FAC, the wetland determination will be based on the soil and hydrology
parameters. Districts adopting this option should provide documented support to the Corps
representative on the regional plant list panel, so that a change in indicator status of FAC
species of concern can be pursued. Corps representatives on the regional and national plant
list panels will continually strive to ensure that plant species are properly designated on both a
regional and subregional basis.
**FAC+ species are considered to be wetter (i.e., have a greater estimated probability of
occurring in wetlands) than FAC species, while FAC- species are considered to be drier (i.e.,
have a lesser estimated probability of occurring in wetlands) than FAC species.
b. Other indicators. Although there are several other indicators of hydrophytic vegetation, it
will seldom be necessary to use them. However, they may provide additional useful
information to strengthen a case for the presence of hydrophytic vegetation. Additional
training and/or experience may be required to employ these indicators.
(1) Visual observation of plant species growing in areas of prolonged inundation and/or soil
saturation. This indicator can only be applied by experienced personnel who have
accumulated information through several years of field experience and written documentation
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 18
(field notes) that certain species commonly occur in areas of prolonged (>10 percent)
inundation and/or soil saturation during the growing season. Species such as Taxodiun
distichum, Typha latifolia, and Spartina alterniflora normally occur in such areas. Thus,
occurrence of species commonly observed in other wetland areas provides a strong indication
that hydrophytic vegetation is present. CA UTION: The presence of standing water or
saturated soil on a site is insufficient evidence that the species present are able to tozerate long
periods of inundation. The user must rezate the observed species to other simizar situations
and determine whether they are normally found in wet areas, taking into consideration the
season and immediatezy preceding weather conditions.
(2) Morphological adaptations. Some hydrophytic species have easily recognized physical
characteristics that indicate their ability to occur in wetlands. A given species may exhibit
several of these characteristics, but not all hydrophytic species have evident morphological
adaptations. A list of such morphological adaptations and a partial list of plant species with
known morphological adaptations for occurrence in wetlands are provided in Appendix C,
Section 3.
(3) Technical literature. The technical literature may provide a strong indication that plant
species comprising the prevalent vegetation are commonly found in areas where soils are
periodically saturated for long periods. Sources of available literature include:
(a) Taxonomic references. Such references usually contain at least a general description of
the habitat in which a species occurs. A habitat description such as, "Occurs in water of
streams and lakes and in alluvial floodplains subject to periodic flooding," supports a
conclusion that the species typically occurs in wetlands. Examples of some useful taxonomic
references are provided in Table 4.
(b) Botanical journals. Some botanical journals contain studies that define species occurrence
in various hydrologic regimes. Examples of such journals include: Ecology. Ecological
Monographs. American Journal of Botany. Journal of American Forestry, and Wetlands: The
Journal of the Society of Wetland Scientists.
(c) Technical reports. Governmental agencies periodically publish reports (e.g. literature
reviews) that contain information on plant species occurrence in relation to hydrologic
regimes. Examples of such publications include the CE preliminary regional wetland guides
(Table 2) published by the US Army Engineer Waterways Experiment Station (WES) and the
wetland community and estuarine profiles of various habitat types (Table 3) published by the
FWS.
(d) Technical workshops, conferences, and symposia. Publications resulting from periodic
scientific meetings contain valuable information that can be used to support a decision
regarding the presence of hydrophytic vegetation. These usually address specific regions or
wetland types. For example, distribution of bottomland hardwood forest species in relation to
hydrologic regimes was examined at a workshop on bottomland hardwood forest wetlands of
the Southeastern United States (Clark and Benforado 1981).
(e) Wetland plant database. The NWI is producing a Plant Database that contains habitat
information on approximately 5,200 plant species that occur at some estimated probability in
wetlands, as compiled from the technical literature. When completed, this computerized
database will be available to all governmental agencies.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 19
(4) Physiological adaptations. Physiological adaptations include any features of the metabolic
processes of plants that make them particularly fitted for life in saturated soil conditions.
NOTE: It is impossible to detect the presence of physiological adaptations in plant species
during onsite visits. Physiological adaptations known for hydrophytic species and species
known to exhibit these adaptations are listed and discussed in Appendix C, Section 3.
(5) Reproductive adaptations. Some plant species have reproductive features that enable them
to become established and grow in saturated soil conditions. Reproductive adaptations known
for hydrophytic species are presented in Appendix C, Section 3.
Hvdric Soils
Definition
36. A hydric soil is a soil that is saturated, flooded, or ponded long enough during the growing
season to develop anaerobic conditions that favor the growth and regeneration of hydrophytic
vegetation (US Department of Agriculture (USDA) Soil Conservation Service (SCS) 1985, as
amended by the National Technical Committee for Hydric Soils (NTCHS) in December 1986).
Criteria for hydric soils
37. Based on the above definition, the NTCHS developed the following criteria for hydric
soils:
a. All Histosols except Folists; Soil nomenclature follows USDA-SCS (1975).
b. Soils in Aquic suborders, Aquic subgroups, Albolls suborder, Salorthids great group, or
Pell great groups of Vertisols that are:
(1) Somewhat poorly drained and have a water table less than 0.5 ft
from the surface for a significant period (usually a week or more) during the growing season,
or
(2) Poorly drained or very poorly drained and have either:
(a) A water table at less than 1.0 ft from the surface for a significant period (usually a week or
more) during the growing season if permeability is equal to or greater than 6.0 in/hr in all
layers within 20 inches; or
(b) A water table at less than 1.5 ft from the surface for a significant period (usually a week or
more) during the growing season if permeability is less than 6.0 in/hr in any layer within 20
inches; or
c. Soils that are ponded for long or very long duration during the growing season; or
d. Soils that are frequently flooded for long duration or very long duration during the growing
season."
A hydric soil may be either drained or undrained, and a drained hydric soil may not continue to
support hydrophytic vegetation. Therefore, not all areas having hydric soils will qualify as
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 20
wetlands. Only when a hydric soil supports hydrophytic vegetation and the area has indicators
of wetland hydrology may the soil be referred to as a "wetland" soil.
38. A drained hydric soil is one in which sufficient ground or surface water has been removed
by artificial means such that the area will no longer support hydrophyte vegetation. Onsite
evidence of drained soils includes:
a. Presence of ditches or canals of sufficient depth to lower the water table below the major
portion of the root zone of the prevalent vegetation.
b. Presence of dikes, levees, or similar structures that obstruct normal inundation of an area.
c. Presence of a tile system to promote subsurface drainage.
d. Diversion of upland surface runoff from an area.
Although it is important to record such evidence of drainage of an area, a hydric soil that has
been drained or partially drained still allows the soil parameter to be met. However, the area
will not qualify as a wetland if the degree of drainage has been sufficient to preclude the
presence of either hydrophytic vegetation or a hydrologic regime that occurs in wetlands.
NOTE: the mere presence of drainage structures in an area is not sufficient basis .for
concluding that a hydric soil has been drained; such areas may continue to have wetland
hydrology.
General information
39. Soils consist of unconsolidated, natural material that supports, or is capable of supporting,
plant life. The upper limit is air and the lower limit is either bedrock or the limit of biological
activity. Some soils have very little organic matter (mineral soils), while others are composed
primarily of organic matter (Histosols). The relative proportions of particles (sand, silt, clay,
and organic matter) in a soil are influenced by many interacting environmental factors. As
normally defined, a soil must support plant life. The concept is expanded to include substrates
that could support plant life. For various reasons, plants may be absent from areas that have
well-defined soils.
40. A soil profile (Figure 2) consists of various soil layers described from the surface
downward. Most soils have two or more identifiable horizons. A soil horizon is a layer
oriented approximately parallel to the soil surface, and usually is differentiated from
contiguous horizons by characteristics that can be seen or measured in the field (e.g., color,
structure, texture, etc.). Most mineral soils have A-, B-, and C-horizons, and many have
surficial organic layers (0-horizon). The A-horizon, the surface soil or topsoil, is a zone in
which organic matter is usually being added to the mineral, soil. It is also the zone from which
both mineral and organic matter are being moved slowly downward. The next major horizon is
the B-horizon, often referred to as the subsoil. The B-horizon is the zone of maximum
accumulation of materials. It is usually characterized by higher clay content and/or more
pronounced soil structure development and lower organic matter than the A-horizon. The next
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 21
major horizon is usually the C-horizon, which consists of unconsolidated parent material that
has not been sufficiently weathered to exhibit characteristics of the B-horizon. Clay content
and degree of soil structure development in the C-horizon are usually less than in the B-
horizon. The lowest major horizon, the R-horizon, consists of consolidated bedrock. In many
situations, this horizon occurs at such depths that it has no significant influence on soil
characteristics.
Influencing factors
41. Although all soil-forming factors (climate, parent material, relief, organisms, and time)
affect the characteristics of a hydric soil, the overriding influence is the hydrologic regime.
The unique characteristics of hydric soils result from the influence of periodic or permanent
inundation or soil saturation for sufficient duration to effect anaerobic conditions. Prolonged
anaerobic soil conditions lead to a reducing environment, thereby lowering the soil redox
potential. This results in chemical reduction of some soil components (e.g. iron and
manganese oxides), which leads to development of soil colors and other physical
characteristics that usually are indicative of hydric soils.
Classification
42. Hydric soils occur in several categories of the current soil classification system, which is
published in Soil Taxonomy (USDA-SCS 1975). This classification system is based on
physical and chemical properties of soils that can be seen, felt, or measured. Lower taxonomic
categories of the system (e.g. soil series and soil phases) remain relatively unchanged from
earlier classification systems.
43. Hydric soils may be classified into two broad categories: organic and mineral. Organic
soils (Histosols) develop under conditions of nearly continuous saturation and/or inundation.
All organic soils are hydric soils except Folists, which are freely drained soils occurring on dry
slopes where excess litter accumulates over bedrock. Organic hydric soils are commonly
known as peats and mucks. All other hydric soils are mineral soils. Mineral soils have a wide
range of textures (sandy to clayey) and colors (red to gray). Mineral hydric soils are those
periodically saturated for sufficient duration to produce chemical and physical soil properties
associated with a reducing environment. They are usually gray and/or mottled immediately
below the surface horizon (see paragraph 44d), or they have thick, dark-colored surface layers
overlying gray or mottled subsurface horizons.
Wetland indicators (nonsandy soils)
44. Several indicators are available for determining whether a given soil meets the definition
and criteria for hydric soils. Any one of the following indicates that hydric soils are present:
Indicators are listed in order of decreasing reliability. Although all are valid indicators, some
are stronger indicators than others. When a decision is based on an indicator appearing in the
lower portion of the list, re-evaluate the parameter to ensure that the proper decision was
reached.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 22
a. Organic soils (Histosols). A soil, is an organic soil when:
(1) more than 50 percent (by volume) of the upper 32 inches of soil is composed of organic
soil material; A detailed definition of organic soil material is available in USDA-SCS (1975),
or
(2) organic soil material of any thickness rests on bedrock. Organic soils (Figure 3) are
saturated for long periods and are commonly called peats or mucks.
b. Histic epipedons. A histic epipedon is an 8- to 16-inch layer at or near the surface of a
mineral hydric soil that is saturated with water for 30 consecutive days or more in most years
and contains a minimum of 20 percent organic matter when no clay is present or a minimum of
30 percent organic matter when clay content is 60 percent or greater. Soils with histic
epipedons are inundated or saturated for sufficient periods to greatly retard aerobic
decomposition of the organic surface, and are considered to be hydric soils.
c. Sulfidic material. When mineral soils emit an odor of rotten eggs, hydrogen sulfide is
present. Such odors are only detected in waterlogged soils that are permanently saturated and
have sulfidic material within a few centimetres of the soil surface. Sulfides are produced only
in a reducing environment.
d. Aquic or peraquic moisture regime. An aquic moisture regime is a reducing one: i.e.. it is
virtually free of dissolved oxygen because the soil is saturated by ground water or by water of
the capillary fringe (USDA-SCS 1975). Because dissolved oxygen is removed from ground
water by respiration of microorganisms, roots, and soil fauna, it is also implicit that the soil
temperature is above biologic zero (5° C) at some time while the soil is saturated. Soils with
peraquic moisture regimes are characterized by the presence of ground water always at or near
the soil surface. Examples include soils of tidal marshes and soils of closed, landlocked
depressions that are fed by permanent streams.
e. Reducing soil conditions. Soils saturated for long or very long duration will usually exhibit
reducing conditions. Under such conditions, ions of iron are transformed from a ferric valence
state to a ferrous valence state. This condition can often be detected in the field by a ferrous
iron test. A simple calorimetric field test kit has been developed for this purpose. When a soil
extract changes to a pink color upon addition of a-a-dipyridil, ferrous iron is present, which
indicates a reducing soil environment. NOTE: This test cannot be used in mineral hydric soils
having low iron content, organic soils, and soils that have been desaturated for significant
periods of the growing season.
f. Soil colors. The colors of various soil components are often the most diagnostic indicator of
hydric soils. Colors of these components are strongly influenced by the frequency and duration
of soil saturation, which leads to reducing soil conditions. Mineral hydric soils will be either
gleyed or will have bright mottles and/or low matrix chroma. These are discussed
below:
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 23
(1) Gleyed soils (gray colors). Gleyed soils develop when anaerobic soil conditions result in
pronounced chemical reduction of iron, manganese, and other elements, thereby producing
gray soil colors. Anaerobic conditions that occur in waterlogged soils result in the
predominance of reduction processes, and such soils are greatly reduced. Iron is one of the
most abundant elements in soils. Under anaerobic conditions, iron in converted from the
oxidized (ferric) state to the reduced (ferrous) state, which results in the bluish, greenish, or
grayish colors associated with the gleying effect (Figure 4). Gleying immediately below the A-
horizon or 10 inches (whichever is shallower) is an indication of a markedly reduced soil, and
gleyed soils are hydric soils. Gleyed soil conditions can be determined by using the gley page
of the Munsell Color Book (Munsell Color 1975).
(2) Soils with bright mottles and/or low matrix chroma. Mineral hydric soils that are saturated
for substantial periods of the growing season (but not long enough to produce gleyed soils) will
either have bright mottles and a low matrix chroma or will lack mottles but have a low matrix
chroma (see Appendix D, Section 1, for a definition and discussion of "chroma" and other
components of soil color). Mottled means "marked with spots of contrasting color." Soils that
have brightlv colored mottles and a low matrix chroma are indicative of a fluctuating water
table. The soil matrix is the portion (usually more than 50 percent) of a given soil layer that
has the predominant color (Figure 5). Mineral hydric soils usually have one of the following
color features in the horizon immediately below the A-horizon or 10 inches (whichever is
shallower):
(a) Matrix chroma of 2 or less* in mottled soils.
(b) Matrix chroma of 1 or less* in unmottled soils.
*FOOTNOTE: Colors should be determined in soils that have been moistened; otherwise,
state that colors are for dry soils.
NOTE: The matrix chroma of some dark (black) mineral hydric soils will not conform to the
criteria described in (a) and (b) above; in such soils, gray mottles occurring at 10 inches or
less are indicative of hydric conditions.
CAUTION: Soils with significant coloration due to the nature of the parent material (e.g. red
soils of the Red River Valley) may not exhibit the above characteristics, in such cases, this
indicator cannot be used.
g. Soil appearing on hydric soils list. Using the criteria for hydric soils (paragraph 37), the
NTCHS has developed a list of hydric soils. Listed soils have reducing conditions for a
significant portion of the growing season in a major portion of the root zone and are frequently
saturated within 12 inches of the soil surface. The NTCHS list of hydric soils is presented in
Appendix D, Section 2. CAUTION: Be sure that the profile description of the mapping unit
conforms to that of the sampled soil.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 24
h. Iron and manganese concretions. During the oxidation-reduction process, iron and
manganese in suspension are sometimes segregated as oxides into concretions or soft masses
(Figure 6). These accumulations are usually black or dark brown. Concretions >2 mm in
diameter occurring within 7.5 cm of the surface are evidence that the soil is saturated for long
periods near the surface.
Wetland indicators (sandy soils)
45. Not all indicators listed in paragraph 44 can be applied to sandy soils. In particular, soil
color should not be used as an indicator in most sandy soils. However, three additional soil
features may be used as indicators of sandy hydric soils, including:
a. High organic matter content in the surface horizon. Organic matter tends to accumulate
above or in the surface horizon of sandy soils that are inundated or saturated to the surface for
a significant portion of the growing season. Prolonged inundation or saturation creates
anaerobic conditions that greatly reduce oxidation of organic matter.
b. Streaking of subsurface horizons by organic matter. Organic matter is moved downward
through sand as the water table fluctuates. This often occurs more rapidly and to a greater
degree in some vertical sections of a sandy soil containing high content of organic matter than
in others. Thus, the sandy soil appears vertically streaked with darker areas. When soil from a
darker area is rubbed between the fingers, the organic matter stains the fingers.
c. Organic pans. As organic matter is moved downward through ,;andy soils, it tends to
accumulate at the point representing the most commonly occurring depth to the water table.
This organic matter tends to become slightly cemented with aluminum, forming a thin layer of
hardened soil (spodic horizon). These horizons often occur at depths of 12 to 30 inches below
the mineral surface. Wet spodic soils usually have thick dark surface horizons that are high in
organic matter with dull, gray horizons above the spodic horizon.
CAUTION: In recently deposited sandy material (e.g. accreting sandbars), it may be
impossible to find any of these indicators, in such cases., consider this as a natural atypical
situation.
Wetland Hydrology
Definition
46. The term "wetland hydrology" encompasses all hydrologic characteristics of areas that are
periodically inundated or have soils saturated to the surface at some time during the growing
season. Areas with evident characteristics of wetland hydrology are those where the presence
of water has an overriding influence on characteristics of vegetation and soils due to anaerobic
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 25
and reducing conditions, respectively. Such characteristics are usually present in areas that are
inundated or have soils that are saturated to the surface for sufficient duration to develop
hydrl.c soils and support vegetation typically adapted for life in periodically anaerobic soil
conditions. Hydrolpgy is often the least exact of the parameters, and indicators of wetland
hydrology are sometimes difficult to find in the field. However, it is essential to establish that a
wetland area is periodically inundated or has saturated soils during the growing season.
Influencing factors
47. Numerous factors (e.g., precipitation, stratigraphy, topography,!' soil permeability, and
plant cover) influence the wetness of an area. Regardless, the characteristic common to all
wetlands is the presence of an abundant supply of water. The water source may be runoff from
direct precipitation, headwater or backwater flooding, tidal influence, ground water, or some
combination of these sources. The frequency and duration of inundation or soil saturation
varies from nearly permanently inundated or saturated to irregularly inundated or saturated.
Topographic position, stratigraphy, and soil permeability influence both the frequency and
duration of inundation and soil saturation. Areas of lower elevation in a floodplain or marsh
have more frequent periods of inundation and/or greater duration than most areas at higher
elevations. Floodplain configuration may significantly affect duration of inundation. When
the floodplain configuration is conducive to rapid runoff, the influence of frequent periods of
inundation on vegetation and soils maylbe reduced. Soil permeability also influences duration
of inundation and soil saturation. For example, clayey soils absorb water more slowly than
sandy or loamy soils, and therefore have slower permeability and remain saturated much
longer. Type and amount of plant cover affect both degree of inundation and duration of
saturated soil conditions. Excess water drains more slowly in areas of abundant plant cover,
thereby increasing frequency and duration of inundation and/or soil saturation. On the other
hand, transpiration rates are higher in areas of abundant plant cover, which may reduce the
duration of soil saturation.
Classification
48. Although the interactive effects of all hydrologic factors produce a continuum of wetland
hydrologic regimes, efforts have been made to classify wetland hydrologic regimes into
functional categories. These efforts have focused on the use of frequency, timing, and duration
of inundation or soil saturation as a basis for classification. A classification system developed
for nontidal areas is presented in Table 5. This classification system was slightly modified from
the system developed by the Workshop on Bottomland Hardwood Forest Wetlands of the
Southeastern United States (Clark and Benforado 1981). Recent research indicates that
duration of inundation and/or soil saturation during the growing season is more influential on
the plant community than frequency of inundation/saturation during the growing season
(Theriot, in press). Thus, frequency of inundation and soil saturation are not included in Table
5. The WES has developed a computer program that can be used to transform stream gage data
to mean sea level elevations representing the upper limit of each hydrologic zone shown in
Table 5.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 26
Wetland indicators
49. Indicators of wetland hydrology may include, but are not necessarily limited to: drainage
patterns, drift lines, sediment deposition, watermarks, stream gage data and flood predictions,
historic records, visual observation of saturated soils, and visual observation of inundation.
Any of these indicators may be evidence of wetland hydrologic characteristics. Methods for
determining hydrologic indicators can be categorized according to the type of indicator.
Recorded data include stream gage data, lake gage data, tidal gage data, flood predictions, and
historical records. Use of these data is commonly limited to areas adjacent to streams or other
similar areas. Recorded data usually provide both short- and long-term information about
frequency and duration of inundation, but contain little or no information about soil saturation,
which must be gained from soil surveys or other similar sources. The remaining indicators
require field observations. Field indicators are evidence of present or past hydrologic events
(e.g. location and height of flooding). Indicators for recorded data and field observations
include: (Indicators are listed in order of decreasing reliability. Although all are valid
indicators, some are stronger indicators than others. When a decision is based on an indicator
appearing in the lower portion of the list, re-evaluate the parameter to ensure that the proper
decision was reached.)
cL Recorded data. Stream gage data, lake gage data, tidal gage data, flood predictions, and
historical data may be available from the following sources:
(1) CE District Offices. Most CE Districts maintain stream, lake, and tidal gage records for
major water bodies in their area. In addition, CE planning and design documents often contain
valuable hydrologic information. For example, a General Design Memorandum (GDM)
usually describes flooding frequencies and durations for a project area. Furthermore, the extent
of flooding within a project area is sometimes indicated in the GDM according to elevation
(height) of certain flood frequencies (1-, 2-, 5-, 10-year, etc.).
(2) US Geological Survey (USGS). Stream and tidal gage data are available from the USGS
offices throughout the Nation, and the latter are also available from the National Oceanic and
Atmospheric Administration. CE Districts often have such records.
(3) State, county, and local agencies. These agencies often have responsibility for flood
control/relief and flood insurance.
(4) Soil Conservation Service Small Watershed Projects. Planning documents from this
agency are often helpful, and can be obtained from the SCS district office in the county.
(5) Planning documents of developers.
b. Field data. The following field hydrologic indicators can be assessed quickly, and although
some of them are not necessarily indicative of hydrologic events that occur only during the
growing season, they do provide evidence that inundation and/or soil saturation has occurred:
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 27
(1) Visual observation of inundation. The most obvious and revealing hydrologic indicator
may be simply observing the areal extent of inundation. However, because seasonal conditions
and recent weather conditions can contribute to surface water being present on a nonwetland
site, both should be considered when applying this indicator.
(2) Visual observation of soil saturation. Examination of this indicator requires digging a soil
pit (Appendix D, Section 1) to a depth of 16 inches and observing the level at which water
stands in the hole after sufficient time has been allowed for water to drain into the hole. The
required time will varv depending on soil texture. In some cases, the upper level at which
water is flowing into the pit can be observed by examining the wall of the hole. This level
represents the depth to the water table. The depth to saturated soils will always be nearer the
surface due to the capillary fringe. For soil saturation to impact vegetation, it must occur
within a major portion of the root zone (usually within 12 inches of the surface) of the
prevalent vegetation. The major portion of the root zone is that portion of the soil profile in
which more than one half of the plant roots occur. CAUTION: In some heavy clay soils, water
may not rapidly accumulate in the hole even when the soil is saturated. If water is observed at
the bottom of the hole but has notfiLLed to the 12-inch depth, examine the sides of the hole
and determine the shallowest depth at which water is entering the hole. When applying this
indicator, both the season of the year and preceding weather conditions must be considered.
(3) Watermarks. Watermarks are most common on woody vegetation. They occur as stains
on bark (Figure 7) or other fixed objects (e.g. bridge pillars, buildings, fences, etc.). When
several watermarks are present, the highest reflects the maximum extent of recent inundation.
(4) Drift lines. This indicator is most likely to be found adjacent to streams or other sources of
water flow in wetlands, but also often occurs in tidal marshes. Evidence consists of deposition
of debris in a line on the surface (Figure 8) or debris entangled in aboveground vegetation or
other fixed objects. Debris usually consists of remnants of vegetation (branches, stems, and
leaves), sediment, litter, and other waterborne materials deposited parallel to the direction of
water flow. Drift lines provide an indication of the minimum portion of the area inundated
during a flooding event; the maximum level of inundation is generally at a higher elevation
than that indicated by a drift line.
(5) Sediment deposits. Plants and other vertical objects often have thin layers, coatings, or
depositions of mineral or organic matter on them after inundation (Figure 9). This evidence
may remain for a considerable period before it is removed by precipitation or subsequent
inundation. Sediment deposition on vegetation and other objects provides an. indication of the
minimum inundation level. When sediments are primarily organic (e.g. fine organic material,
algae), the detritus may become encrusted on or slightly above the soil surface after dewatering
occurs (Figure 10).
(6) Drainage patterns within wetlands. This indicator, which occurs primarily in wetlands
adjacent to streams, consists of surface evidence of drainage flow into or through an area
(Figure 11). In some wetlands, this evidence may exist as a drainage pattern eroded into the
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 28
soil, vegetative matter (debris) piled against thick vegetation or woody stems oriented
perpendicular to the direction of water flow, or the absence of leaf litter (Figure 8). Scouring is
often evident around roots of persistent vegetation. Debris may be deposited in or along the
drainage pattern (Figure 12). CAUTION: Drainage patterns also occur in upland areas after
periods of considerable precipitation; therefore, topographic position must also be considered
when applying this indicator.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 29
PART IV: METHODS
Section A. Introduction
50. PART IV contains sections on preliminary data gathering, method selection, routine
determination procedures, comprehensive determination procedures, methods for
determinations in atypical situations, and guidance for wetland determinations in natural
situations where the three-parameter approach may not always apply.
51. Significant flexibility has been incorporated into PART IV. The user is presented in
Section B with various potential sources of information that may be helpful in making a
determination, but not all identified sources of information may be applicable to a given
situation. The user is not required to obtain information from all identified sources. Flexibility
is also provided in method selection (Section C). Three levels of routine determinations are
available, depending on the complexity of the required determination and the quantity and
quality of existing information. Application of methods presented in both Section D (routine
determinations) and Section E (comprehensive determinations) may be tailored to meet site-
specific requirements, especially with respect to sampling design.
52. Methods presented in Sections D and E vary with respect to the required level of technical
knowledge and experience of the user. Application of the qualitative methods presented in
Section D (routine determinations) requires considerably less technical knowledge and
experience than does application of the quantitative methods presented in Section E
(comprehensive determinations). The user must at least be able to identify the dominant,plant
species in the project area when making a routine determination (Section D), and should have
some basic knowledge of hydric soils when employing routine methods that require soils
examination. Comprehensive determinations require a basic understanding of sampling
principles and the ability to identify all commonly occurring plant species in a project area, as
well as a good understanding of indicators of hydric soils and wetland hydrology. The
comprehensive method should only be employed by experienced field inspectors.
Section B. Preliminary Data Gathering and Synthesis
53. This section discusses potential sources of information that may be helpful in making a
wetland determination. When the routine approach is used, it may often be possible to make a
wetland determination based on available vegetation, soils, and hydrology data for the area.
However, this section deals only with identifying potential information sources, extracting
pertinent data, and synthesizing the data for use in making a determination. Based on the
quantity and quality of available information and the approach selected for use (Section C), the
user is referred to either Section D or Section E for the actual determination. Completion of
Section B is not required, but is recommended because the available information may reduce or
eliminate the need for field effort and decrease the time and cost of making a determination.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 30
However, there are instances in small project areas in which the time required to obtain the
information may be prohibitive. In such cases PROCEED to paragraph 55, complete STEPS 1
through 3, and PROCEED to Section D or E.
Data sources
54. Obtain the following information, when available and applicable:
a. USGS quadrangle maps. USGS quadrangle maps are available at different scales. When
possible, obtain maps at a scale of 1:24,000; otherwise, use maps at a scale of 1:62,500. Such
maps are available from USGS in Reston, Va., and Menlo Park, Calif., but they may already be
available in the CE District Office. These maps provide several types of information:
(1) Assistance in locating field sites. Towns, minor roads, bridges, streams, and other
landmark features (e.g. buildings, cemeteries, water bodies, etc.) not commonly found on road
maps are shown on these maps.
(2) Topographic details, including contour lines (usually at 5- or 10-ft contour intervals).
(3) General delineation of wet areas (swamps and marshes). The actual wet area may be
greater than that shown on the map because USGS generally maps these areas based on the
driest season of the year.
(4) Latitude, longitude, townships, ranges, and sections. These provide legal descriptions of
the area.
(5) Directions, including both true and magnetic north.
(6) Drainage patterns.
(7) General land uses, such as cleared (agriculture or pasture), forested, or urban.
CA UTION: Obtain the most recent USGS maps. Older maps may show features that no
longer exist and will not show new features that have developed since the map was
constructed. Also, USGS is currently changing the mapping scale from 1:24,000 to 1:25,000.
IL National Wetlands Inventory products.
(1) Wetland maps. The standard NWI maps are at a scale of 1:24,000 or, where USGS base
maps at this scale are not available, they are at 1:62,500 (1:63,350 in Alaska). Smaller scale
maps ranging from 1:100,000 to 1:500,000 are also available for certain areas. Wetlands on
NWI maps are classified in accordance with Cowardin et al. (1979). CAUTION: Since not all
delineated areas on NWI maps are wetlands under Department of Army jurisdiction, NWI
maps should not be used as the sole basis for determining whether wetland vegetation is
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 31
present. NWI "User Notes" are available that correlate the classification system with local
wetland community types. An important feature of this classification system is the water
regime modifier, which describes the flooding or soil saturation characteristics. Wetlands
classified as having a temporarily flooded or intermittently flooded water regime should be
viewed with particular caution since this designation is indicative of plant communities that are
transitional between wetland and nonwetland. These are among the most difficult plant
communities to map accurately from aerial photography. For wetlands "wetter" than
temporarily flooded and intermittently flooded, the probability of a designated map unit on
recent NWI maps being a wetland (according to Cowardin et al. 1979) at the time of the
photography is in excess of 90 percent. CAUTION: Due to the scale of aerial photography
used and other factors, all NWI map boundaries are approximate. The optimum use of NWI
maps is to plan field review (i.e. how wet, big, or diverse is the area?) and to assist during field
review, particularly by showing the approximate areal extent of the wetland and its association
with other communities. NWI maps are available either as a composite with, or an overlay for,
USGS base maps and may be obtained from the NWI Central Office in St. Petersburg, Fla., the
Wetland Coordinator at each FWS regional office, or the USGS.
(2) Plant database. This database of approximately 5,200 plant species that occur in wetlands
provides information (e.g., ranges, habitat, etc.) about each plant species from the technical
literature. The database served as a focal point for development of a national list of plants that
occur in wetlands (Appendix C, Section 1).
c. Soil surveys. Soil surveys are prepared by the SCS for political units (county, parish, etc.)
in a state. Soil surveys contain several types of information:
(1) General information (e.g. climate, settlement, natural resources, farming, geology, general
vegetation types).
(2) Soil maps for general and detailed planning purposes. These maps are usually generated
from fairly recent aerial photography. CAUTION: The smallest mapping unit is 3 acres, and a
given soil series as mapped may contain small inclusions of other series.
(3) Uses and management of soils. Any wetness characteristics of soils will be mentioned
here.
(4) Soil properties. Soil and water features are provided that may be very helpful for wetland
investigations. Frequency, duration, and timing of inundation (when present) are described for
each soil type. Water table characteristics that provide valuable information about soil
saturation are also described. Soil permeability coefficients may also be available.
(5) Soil classification. Soil series and phases are usually provided. Published soil surveys will
not always be available for the area. If not, contact the county SCS office and determine
whether the soils have been mapped.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 32
d. Stream and tidal gage data. These documents provide records of tidal and stream flow
events. They are available from either the USGS or CE District office.
e. Environmental impact assessments (EIAsX environmental impact statements (EISsX general
design memora da (GDMX and other similar publications. These documents may be available
from Federal agencies for an area that includes the project area. They may contain some
indication of the location and characteristics of wetlands consistent with the required criteria
(vegetation, soils, and hydrology), and often contain flood frequency and duration data.
f. Documents and maps from State, county, or local governments. Regional maps that
characterize certain areas (e.g., potholes, coastal areas, or basins) may be helpful because they
indicate the type and character of wetlands.
g. Remote sensing. Remote sensing is one of the most useful information sources available for
wetland identification and delineation. Recent aerial photography, particularly color infrared,
provides a detailed view of an area; thus, recent land use and other features (e.g. general type
and areal extent of plant communities and degree of inundation of the area when the
photography was taken) can be determined. The multiagency cooperative National High
Altitude Aerial Photography Program (HAP) has l:59,000-scale color infrared photography for
approximately 85 percent (December 1985) of the coterminous United States from 1980 to
1985. This photography has excellent resolution and can be ordered enlarged to 1:24,000 scale
from USGS. Satellite images provide similar information as aerial photography, although the
much smaller scale makes observation of detail more difficult without sophisticated equipment
and extensive training. Satellite images provide more recent coverage than aerial photography
(usually at 18-day intervals). Individual satellite images are more expensive than aerial
photography, but are not as expensive as having an area flown and photographed at low
altitudes. However, better resolution imagery is now available with remote sensing equipment
mounted on fixed-wing aircraft.
h. Local individuals and experts. Individuals having personal knowledge of an area may
sometimes provide a reliable and readily available source of information about the area,
particularly information on the wetness of the area.
i. USGS land use and land cover maps. Maps created by USGS using remotely sensed data
and a geographical information system provide a systematic and comprehensive collection and
analysis of land use and land cover on a national basis. Maps at a scale of 1:250,000 are
available as overlays that show land use and land cover according to nine basic levels. One
level is wetlands (as determined by the FWS), which is further subdivided into forested and
nonforested areas. Five other sets of maps show political units, hydrologic units, census
subdivisions of counties, Federal land ownership, and State land ownership. These maps can
be obtained from any USGS mapping center.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 33
I, Applicant's survey plans and engineering designs. In many cases, the permit applicant will
already have had the area surveyed (often at 1-ft contours or less) and will also have
engineering designs for the proposed activity.
Data synthesis
55. When employing Section B procedures, use the above sources of information to complete
the following steps:
• STEP 1 - Identify the Project Area on a Map. Obtain a USGS quadrangle map (1:24,000) or
other appropriate map, and locate the area identified in the permit application. PROCEED TO
STEP 2.
* STEP 2 - Prepare a Base Map. Mark the project area boundaries on the map. Either use the
selected map as the base map or trace the area on a mylar overlay, including prominent
landscape features (e.g., roads, buildings, drainage patterns, etc.). If possible, obtain diazo
copies of the resulting base map. PROCEED TO STEP 3.
• STEP 3 - Determine Size of the Project Area. Measure the area boundaries and calculate the
size of the area. PROCEED TO STEP 4 OR TO SECTION D OR E IF SECTION B IS NOT
USED.
• STEP 4 - Summarize Available Information on Vegetation. Examine available sources that
contain information about the area vegetation. Consider the following:
a. USGS quadrangle maps. Is the area shown as a marsh or swamp? CAUTION: Do not use
this as the sole basis for determining that hydrophytic vegetation is present.
b. NWI overlays or maps. Do the overlays or maps indicate that hydrophytic vegetation
occurs in the area? If so, identify the vegetation type(s).
c. EIAs, EISs, or GDMs that include the project area. Extract any vegetation data that pertain
to the area.
d. Federal, State, or local government documents that contain information about the area
vegetation. Extract appropriate data.
e. Recent (within last 5 years) aerial photography of the area. Can the area plant community
type(s) be determined from the photography? Extract appropriate data.
f. Individuals or experts having knowledge of the area vegetation. Contact them and obtain
any appropriate information. CAUTION: Ensure that the individual providing the information
has firsthand knowledge of the area.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 34
g. Any published scientific studies of the area plant communities. Extract any appropriate
data.
h. Previous wetland determinations made for the area. Extract any pertinent vegetation data.
When the above have been considered, PROCEED TO STEP 5.
• STEP 5 - Determine Whether the Vegetation in the Project Area Is Adequately Characterized.
Examine the summarized data (STEP 4) and determine whether the area plant communities are
adequately characterized. For routine determinations, the plant community type(s) and the
dominant species in each vegetation layer of each community type must be known. Dominant
species are those that have the largest relative basal area (overstory) (This term is used because
species having the largest individuals may not be dominant when only a few are present. To
use relative basal area, consider both the size and number of individuals of a species and
subjectively compare with other species present), height (woody understory), number of stems
(woody vines), or greatest areal cover (herbaceous understory). For comprehensive
determinations, each plant community type present in the project area area must have been
quantitatively described within the past 5 years using accepted sampling and analytical
procedures, and boundaries between community types must be known. Record information on
DATA FORM 1. A separate DATA FORM 1 must be used for each plant community type. In
either case, PROCEED TO Section F if there is evidence of recent significant vegetation
alteration due to human activities or natural events. Otherwise, PROCEED TO STEP 6.
* STEP 6. - Summarize Available Information on Area Soils. Examine available information
and describe the area soils. Consider the following:
a. County soil surveys. Determine the soil series present and extract characteristics for each.
CA UTION: Soil mapping units sometimes include more than one soil series.
b. Unpublished county soil maps. Contact the local SCS office and determine whether soil
maps are available for the area. Determine the soil series of the area, and obtain any available
information about possible hydric soil indicators (paragraph 44 or 45) for each soil series.
CL Published EIAs, EISs, or GDMs that include soils information. Extract any pertinent
information.
d. Federal, State, and/or local government documents that contain descriptions of the area
soils. Summarize these data.
e. Published scientific studies that include area soils data. Summarize these data.
f. Previous wetland determinations for the area. Extract any pertinent soils data.
When the above have been considered, PROCEED TO STEP 7.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 35
• STEP 7 - Determine Whether Soils of the Project Area Have Been Adequately Characterized.
Examine the summarized soils data and determine whether the soils have been adequately
characterized. For routine determinations, the soil series must be known. For comprehensive
determinations, both the soil series and the boundary of each soil series must be known.
Record information on DATA FORM 1. In either case, if there is evidence of recent significant
soils alteration due to human activities or natural events, PROCEED TO Section F. Otherwise,
PROCEED TO STEP 8.
• STEP 8 - Summarize Available Hydrology Data. Examine available information and
describe the area hydrology. Consider the following:
a. USGS quadrangle maps. Is there a significant, well-defined drainage through the area? Is
the area within a maior floodplain or tidal area? What range of elevations occur in the area,
especially in relation to the elevation of the nearest perennial watercourse?
b. NWI overlays or maps. Is the area shown as a wetland or deepwater aquatic habitat? What
is the water regime modifier?
c. El As, EISs, or GDMs that describe the project area. Extract any pertinent hydrologic data.
d. Floodplain management maps. These maps may be used to extrapolate elevations that can
be expected to be inundated on a 1-, 2-, 3-year, etc., basis. Compare the elevations of these
features with the elevation range of the project area to determine the frequency of inundation.
e. Federal, State, and local government documents (e.g. CE floodplain management maps and
profiles) that contain hydrologic data. Summarize these data.
f. Recent (within past 5 years) aerial photography that shows the area to be inundated. Record
the date of the photographic mission.
1. Newspaper accounts of flooding events that indicate periodic inundation of the area.
h. SCS County Soil Surveys that indicate the frequency and duration of inundation and soil
saturation for area soils. CA UTION: Data provided only represent average conditions for a
particular soil series in its natural undrained state, and cannot be used as a positive
hydrologic indicator in areas that have significantly altered hydrology.
\. Tidal or stream gage data for a nearby water body that apparently influences the area.
Obtain the gage data and complete (1) below if the routine approach is used, or (2) below if
the comprehensive approach is used (OMIT IF GAGING STATION DATA ARE
UNAVAILABLE):
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 36
(1) Routine approach. Determine the highest water level elevation reached during the growing
season for each of the most recent 10 years of gage data. Rank these elevations in descending
order and select the fifth highest elevation. Combine this elevation with the mean sea level
elevation of the gaging station to produce a mean sea level elevation for the highest water level
reached every other year. Stream gage data are often presented as flow rates in cubic feet per
second. In these cases, ask the CE District's Hydrology Branch to convert flow rates to
corresponding mean sea level elevations and adjust gage data to the site. Compare the
resulting elevations reached biennially with the project area elevations. If the water level
elevation exceeds the area elevation, the area is inundated during the growing season on
average at least biennially.
(2) Comprehensive approach. Complete the following:
(a) Decide whether hydrologic data reflect the apparent hydrology. Data available from the
gaging station may or may not accurately reflect the area hydrology. Answer the following
questions:
• Does the water level of the area appear to fluctuate in a manner that differs from that of the
water body on which the gaging station is located? (In ponded situations, the water level of the
area is usually higher than the water level at the gaging station.)
• Are less than 10 years of daily readings available for the gaging station?
• Do other water sources that would not be reflected by readings at the gaging station appear to
significantly affect the area? For example, do major tributaries enter the stream or tidal area
between the area and gaging station?
If the answer to any of the above questions is YES, the area hydrology cannot be determined
from the gaging station data. If the answer to all of the above questions is NO, PROCEED TO
(b).
(b) Analyze hydrologic data. Subject the hydrologic data to appropriate analytical procedures.
Either use duration curves or a computer program developed by WES (available from the
Environmental Laboratory upon request) for determining the mean sea level elevation
representing the upper limits of wetland hydrology. In the latter case, when the site elevation is
lower than the mean sea level elevation representing a 5-percent duration of inundation and
saturation during the growing season, the area has a hydrologic regime that may occur in
wetlands. NOTE: Duration curves do not reflect the period of soil saturation following
dewatering.
When all of the above have been considered, PROCEED TO STEP 9.
* STEP 9 - Determine Whether Hydrology Is Adequately Characterized. Examine the
summarized data and determine whether the hydrology of the project area is adequately
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 37
characterized. For routine determinations, there must be documented evidence of frequent
inundation or soil saturation during the growing season. For comprehensive determinations,
there must be documented quantitative evidence of frequent inundation or soil saturation
during the growing season, based on at least 10 years of stream or tidal gage data. Record
information on DATA FORM 1. In either case, if there is evidence of recent significant
hydrologic alteration due to human activities or natural events, PROCEED TO Section F.
Otherwise, PROCEED TO Section C.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 38
Section C. Selection of Method
56. All wetland delineation methods described in this manual can be grouped into two general
types: routine and comprehensive. Routine determinations (Section D) involve simple, rapidly
applied methods that result in sufficient qualitative data for making a determination.
Comprehensive methods (Section E) usually require significant time and effort to obtain the
needed quantitative data. The primary factor influencing method selection will usually be the
complexity of the required determination. However, comprehensive methods may sometimes
be selected for use in relatively simple determinations when rigorous documentation is
required.
57. Three levels of routine wetland determinations are described below. Complexity of the
project area and the quality and quantity of available information will influence the level
selected for use.
a. Level 1 - Onsite Inspection Unnecessary. This level may be employed when the
information already obtained (Section B) is sufficient for making a determination for the entire
project area (see Section D., Subsection 1).
b. Level 2 - Onsite Inspection Necessary. This level must be employed when there is
insufficient information already available to characterize the vegetation, soils, and hydrologyof
the entire project area (see Section D, Subsection 2).
c. Level 3 - Combination of Levels 1 and 2. This level should be used when there is sufficient
information already available to characterize the vegetation, soils, and hydrology of a portion,
but not all, of the project area. Methods described for Level 1 may be applied to portions of
the area for which adequate information already exists, and onsite methods (Level 2) must be
applied to the remainder of the area (see Section D, Subsection 3).
58. After considering all available information, select a tentative method (see above) for use,
and PROCEED TO EITHER Section D or E, as appropriate. NOTE: Sometimes it may be
necessary to change to another method described in the manuaz, depending on the quality of
available information and/or recent changes in the prospect area.
Section D. Routine Determinations
59. This section describes general procedures for making routine wetland determinations. It is
assumed that the user has already completed all applicable steps in Section B, (If it has been
determined that it is more expedient to conduct an onsite inspection than to search for
available information, complete STEPS 1 through 3 of Section B. and PROCEED TO
Subsection 2.) and a routine method has been tentatively selected for use (Section C).
Subsections 1-3 describe steps to be followed when making a routine determination using one
of the three levels described in Section C. Each subsection contains a flowchart that defines the
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 39
relationship of steps to be used for that level of routine determinations. The selected method
must be considered tentative because the user may be required to change methods during the
determination.
Subsection 1 - Onsite Inspection Unnecessary
60. This subsection describes procedures for making wetland determinations when sufficient
information is already available (Section B) on which to base the determination. A flowchart
of required steps to be completed is presented in Figure 13, and each step is described below.
Equipment and materials
61. No special equipment is needed for applying this method. The following materials will be
needed:
a. Map of project area (Section B, STEP 2).
b. Copies of DATA FORM 1 (Appendix B).
c. Appendices C and D to this manual.
Procedure
62. Complete the following steps, as necessary:
• STEP 1 - Determine Whether Available Data Are Sufficient for Entire Project Area.
Examine the summarized data (Section B, STEPS 5, 7, and 9) and determine whether the
vegetation, soils, and hydrology of the entire project area are adequately characterized. If so,
PROCEED TO STEP 2. If all three parameters are adequately characterized for a portion, but
not all, of the project area, PROCEED TO Subsection 3. If the vegetation, soils, and hydrology
are not adequately characterized for any portion of the area, PROCEED TO Subsection 2. &
STEP 2 - Determine Whether Hydrophytic Vegetation Is Present. Examine the vegetation data
and list on DATA FORM 1 the dominant plant species found in each vegetation layer of each
community type. NOTE: A separate DATA FORM 1 will be required for each community
type. Record the indicator status for each dominant species (Appendix C, Section 1 or 2).
When more than 50 percent of the dominant species in a plant community have an indicator
status of OBL, FACW, and/or FAC, hydrophytic vegetation is present (For the FAC-neutral
option, see paragraph 35a). If one or more plant communities comprise of hydrophytic
vegetation, PROCEED TO STEP 3. If none of the plant communities comprise hydrophytic
vegetation, none of the area is a wetlands Complete the vegetation section for each DATA
FORM 1.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
40
STEP 1 - DETERMINE WHETHER
AVAILABLE DATA ARE SUFFICIENT
FOR ENTIRE PROJECT AREA
1
YES
I
/
\
ONE OR MORE PARAMETERS MUST
BE CHARACTERIZED OVER ENTIRE
PROJECT AREA
ALL PARAMETERS ADEQUATELY
CHARACTERIZED IN PART, BUT NOT
ALL OF AREA
PROCEED TO
SUBSECTION 2
PROCEED TO
SUBSECTION 3
STEP 2 - DETERMINE WHETHER
HYDROPHYTIC VEGETATION IS
PRESENT
STEP 3 - DETERMINE WHETHER
WETLAND HYDROLOGY IS PRESENT
STEP 4 - DETERMINE WHETHER
SOILS PARAMETER MUST BE
CONSIDERED
STEP 5 - DETERMINE WHETHER
HYDRIC SOILS ARE PRESENT
i
STEP 6-WETLAND DETERMINATION
STEP 7 - DETERMINE WETLAND
BOUNDARY (IF NECESSARY)
Figure 13. Flowchart of steps involved in making
a wetland determination when an
onsite inspection is unnecessary
25 February 1997
Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 41
* STEP 3 - Determine Whether Wetland Hydrology Is Present. When one of the following
conditions applies (STEP 2), it is only necessary to confirm that there has been no recent
hydrologic alteration of the area:
a. The entire project area is occupied by a plant community or communities in which all
dominant species are OBL (Appendix C, Section 1 or 2).
b. The project area contains two or more plant communities, all of which are dominated by
OBL and/or FACW species, and the wetland-nonwetland boundary is abrupt (e.g. a Spartina
alternifzora marsh bordered by a road embankment). * There must be documented evidence of
periodic inundation or saturated soils when the project area:
a. Has plant communities dominated by one or more FAC species;
b. Has vegetation dominated by FACW species but no adjacent community dominated by OBL
species;
c. Has a gradual, nondistinct boundary between wetlands and nonwetlands; and/or
d. Is known to have or is suspected of having significantly altered hydrology.
If either a or b applies, look for recorded evidence of recently constructed dikes, levees,
impoundments, and drainage systems, or recent avalanches, mudslides, beaver dams, etc., that
have significantly altered the area hydrology. If any significant hydrologic alteration is found,
determine whether the area is still periodically inundated or has saturated soils for sufficient
duration to support the documented vegetation (a or b above). When a or b applies and there is
no evidence of recent hydrologic alteration, or when a or b do not applv and there is
documented evidence that the area is periodically inundated or has saturated soils, wetland
hydrology is present. Otherwise, wetland hydrology does not occur on the area. Complete the
hydrology section of DATA FORM 1 and PROCEED TO STEP 4.
• STEP 4 - Determine Whether the Soils Parameter Must Be Considered. When either a or b
of STEP 3 applies and there is either no evidence of recent hydrologic alteration of the project
area or if wetland hydrology presently occurs on the area, hydric soils can be assumed to be
present. If so, PROCEED TO STEP 6. Otherwise PROCEED TO STEP 5. o STEP 5^
Determine Whether Hydric Soils Are Present. Examine the soils data (Section B, STEP 7) and
record the soil series or soil phase on DATA FORM 1 for each community type. Determine
whether the soil is listed as a hydric soil (Appendix D, Section 2). If all community types have
hydric soils, the entire project area has hydric soils. (CAUTION: If the soil series description
makes reference to inclusions of other soil types, data must be field verified). Any portion of
the area that lacks hydric soils is a nonwetland. Complete the soils section of each DATA
FORM 1 and PROCEED TO STEP 6.
• STEP 6 - Wetland Determination. Examine the DATA FORM 1 for each community type.
Any portion of the project area is a wetland that has:
a. Hydrophytic vegetation that conforms to one of the conditions identified in STEP 3a or 3b
and has either no evidence of altered hydrology or confirmed wetland hydrology.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 42
b. Hydrophytic vegetation that does not conform to STEP 3a or 3b. has hydric soils, and has
confirmed wetland hydrology.
If STEP 6a or 6b_ applies to the entire project area, the entire area is a wetlands Complete a
DATA FORM 1 for all plant community types. Portions of the area not qualifying as a
wetland based on an office determination might or might not be wetlands. If the data used for
the determination are considered to be highly reliable, portions of the area not qualifying as
wetlands may properly be considered nonwetlands. PROCEED TO STEP 7. If the available
data are incomplete or questionable, an onsite inspection (Subsection 2) will be required.
• STEP 7 - Determine Wetland Boundary. Mark on the base map all community types
determined to be wetlands with a W and those determined to be nonwetlands with an N.
Combine all wetland community types into a single mapping unit. The boundary of these
community types is the interface between wetlands and nonwetlands.
Subsection 2 - Onsite Inspection Necessary
63. This subsection describes procedures for routine determinations in which the available
information (Section B) is insufficient for one or more parameters. If only one or two
parameters must be characterized, apply the appropriate steps and return to Subsection 1 and
complete the determination. A flowchart of steps required for using this method is presented
in Figure 14, and each step is described below.
Equipment and materials
64. The following equipment and materials will be needed:
a. Base map (Section B, STEP 2).
b. Copies of DATA FORM 1 (one for each community type and additional copies for
boundary determinations).
c. Appendices C and D.
d. Compass.
e^ Soil auger or spade (soils only).
f. Tape (300 ft).
g. Munsell Color Charts (Munsell Color 1975) (soils only).
Procedure
65. Complete the following steps, as necessary:
• STEP 1 - Locate the Project Area. Determine the spatial boundaries of the project area using
information from a USGS quadrangle map or other appropriate map, aerial photography,
and/or the project survey plan (when available). PROCEED TO STEP 2.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 43
• STEP 2 - Determine Whether an Atypical Situation Exists. Examine the area and determine
whether there is evidence of sufficient natural or human-induced alteration to significantly
alter the area vegetation, soils, and/or hydrology. NOTE: Include possible of/site modifications
that may affect the area hydrology. If not, PROCEED TO STEP 3.
If one or more parameters have been significantly altered by an activity that would normally
require a permit, PROCEED TO Section F and determine whether there is sufficient evidence
that hydrophytic vegetation, hydric soils, and/or wetland hydrology were present prior to this
alteration. Then, return to this subsection and characterize parameters not significantly
influenced by human activities. PROCEED TO STEP 3.
• STEP 3 - Determine the Field Characterization Approach to be Used. Considering the size
and complexity of the area, determine the field characterization approach to be used. When
the area is equal to or less than 5 acres in size (Section B,-STEP 3) and the area is thought to
be relatively homogeneous with respect to vegetation, soils, and/or hydrologic regime,
PROCEED TO STEP 4. When the area is greater than 5 acres in size (Section B, STEP 3) or
appears to be highly diverse with respect to vegetation, PROCEED TO STEP 18.
Areas Equal to or Less Than 5 Acres in Size
• STEP 4 - Identify the Plant Community Type(s). Traverse the area and determine the
number and locations of plant community types. Sketch the location of each on the base map
(Section B, STEP 2), and give each community type a name. PROCEED TO STEP 5.
• STEP 5 - Determine UThether Normal Environmental Conditions Are Present. Determine
whether normal environmental conditions are present by considering the following:
a. Is the area presently lacking hydrophytic vegetation or hydrologic indicators due to annual
or seasonal fluctuations in precipitation or ground-water levels?
b. Are hydrophytic vegetation indicators lacking due to seasonal fluctuations in temperature?
If the answer to either of these questions is thought to be YES, PROCEED TO Section G. If
the answer to both questions is NO, PROCEED TO STEP 6.
• STEP 6 - Select Representative Observation Points. Select a representative observation
point in each community type. A representative observation point is one in which the apparent
characteristics (determine visually) best represent characteristics of the entire community.
Mark on the base map the approximate location of the observation point. PROCEED TO
STEP 7.
• STEP 7 - Characterize Each Plant Community Type. Visually determine the dominant plant
species in each vegetation layer of each community type and record them on DATA FORM 1
(use a separate DATA FORM 1 for each community type). Dominant species are those having
the greatest relative basal area (woody overstory) (This term is used because species having
the larsest individuals may not be dominant when only a few are present. To determine
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 44
relative basal area, consider both the size and number of individuals of a species and subjectly
compare with other species present.) ,greatest height (woody understory), greatest percentage
of areal cover (herbaceous understory), and/or greatest number of stems (woody vines).
PROCEED TO STEP 8.
• STEP 8 - Record Indicator Status of Dominant Species. Record on DATA FORM 1 the
indicator status (Appendix C, Section 1 or 2) of each dominant species in each community
type. PROCEED TO STEP 9.
• STEP 9 - Determine Whether Hydrophytic Vegetation Is Present. Examine each DATA
FORM 1. When more than 50 percent of the dominant species in a community type have an
indicator status (STEP 8) of OBL, FACW, and/or FAC (for the FAC-neutral option, see
paragraph 35a.), hydrophytic vegetation is present. Complete the vegetation section of each
DATA FORM 1. Portions of the area failing this test are not wetlands. PROCEED TO STEP
10.
• STEP 10 - Apply Wetland Hydrologic Indicators. Examine the portion of the area occupied
by each plant community type for positive indicators of wetland hydrology (PART III,
paragraph 49). Record findings on the appropriate DATA FORM 1. PROCEED TO STEP 11.
• STEP 11 - Determine Whether Wetland Hydrology Is Present. Examine the hydrologic
information on DATA FORM 1 for each plant community type. Any portion of the area
having a positive wetland hydrology indicator has wetland hydrology. If positive wetland
hydrology indicators are present in all community types, the entire area has wetland hydrology.
If no plant community type has a wetland hydrology indicator, none of the area has wetland
hydrology. Complete the hydrology portion of each DATA FORM 1. PROCEED TO STEP
12.
This term is used because species having the largest individuals may not be dominant when
only a few are present. To determine relative basal area, consider both the size and number of
individuals of a species and subjectively compare with other species present. For the FAC-
neutral option, see paragraph 3 5a.
• STEP 12 - Determine Whether Soils Must Be Characterized. Examine the vegetation section of
each DATA FORM 1. Hydric soils are assumed to be present in any plant community type in
which:
a. All dominant species have an indicator status of OBL.
b. All dominant species have an indicator status of OBL or FACW, and the wetland boundary
(when present) is abrupt. The soils parameter must be considered in any plant community in
which:
a. The community is dominated by one or more FAC species.
b. No community type dominated by OBL species is present.
c. The boundary between wetlands and nonwetlands is gradual or nondistinct.
d. The area is known to or is suspected of having significantly altered hydrology.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 45
When either a or b occurs and wetland hydrology is present, check the hydric soils blank as
positive on DATA FORM 1 and PROCEED TO STEP 16. If neither a nor b applies,
PROCEED TO STEP 13.
• STEP 13 - Dig a Soil Pit. Using a soil auger or spade, dig a soil pit at the representative
location in each community type. The procedure for digging a soil pit is described in
Appendix D, Section 1. When completed, approximately 16 inches of the soil profile will be
available for examination. PROCEED TO STEP 14.
• STEP 14 - Apply Hydric Soil Indicators. Examine the soil at each location and compare its
characteristics immediately below the A-horizon or 10 inches (whichever is shallower) with
the hydric soil indicators described in PART III, paragraphs 44 and/or 45. Record findings on
the appropriate DATA FORM 1's. PROCEED TO STEP 15.
• STEP 15 - Determine Whether Hydric Soils Are Present. Examine each DATA FORM 1
and determine whether a positive hydric soil indicator was found. If so, the area at that
location has hydric soil. If soils at all sampling locations have positive hydric soil indicators,
the entire area has hydric soils. If soils at all sampling locations lack positive hydric soil
indicators, none of the area is a wetlands Complete the soil section of each DATA FORM 1.
PROCEED TO STEP 16.
• STEP 16 - Make Wetland Determination. Examine DATA FORM 1. If the entire area presently or
normally has wetland indicators of all three parameters (STEPS 9, 11, and 15), the entire area is a
wetlands If the entire area presently or normallv lacks wetland indicators of one or more parameters,
the entire area is a nonwetland. If only a portion of the area presently or normally has wetland
indicators for all three parameters, PROCEED TO STEP 17.
• STEP 17 - Determine Wetland-Nonwetland Boundary. Mark each plant community type on the
base map with a W if wetland or an N if nonwetland. Combine all wetland plant communities into
one mapping unit and all nonwetland plant communities into another mapping unit. The wetland-
nonwetland boundary will be represented by the interface of these two mapping units.
Areas Greater Than 5 Acres in Size
• STEP 18 - Establish a Baseline. Select one project boundary as a baseline. The baseline should
parallel the major watercourse through the area or should be perpendicular to the hydrologic gradient
(Figure 15). Determine the approximate baseline length. PROCEED TO STEP 19.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
46
• STEP 19 - Determine the Required Number and Position of Transects. Use the following to
determine the required number and position of transacts (specific site conditions may necessitate
changes in intervals):
Number of
Baseline length, miles Required Transects
<0.25
>0.25-0.50
>0. 50-0. 75
>0. 75-1. 00
>1. 00-2. 00
>2. 00-4. 00
>4.00
3
3
3
3
3-5
5-8
8 or more*
*Transect intervals should not exceed 0.5 mile.
BASELINE
SEGMENT
B
C
D
STARTING POINT
OF BASELINE
A
\
STREAM
TRANSECT 1
Figure 15. General Orientation of baseline and transects (dotted lines) in a hypothetical project area.
Alpha characters represent different plant communities. All transects start at the midpoint of a
basline segment except the first, which may be repositioned to include community type A.
Divide the baseline length by the number of required transacts. Establish one transect in each
resulting baseline increment. Use the midpoint of each baseline increment as a transect
starting point. For example, if the baseline is 1,200 ft in length, three transacts would be
25 February 1997
Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 47
established—one at 200 ft, one at 600 ft, and one at 1,000 ft from the baseline starting point.
CAUTION: All plant community types must be included. This may necessitate relocation of
one or more transect lines. PROCEED TO STEP 20.
• STEP 20 - Sample Observation Points Along the First Transect. Beginning at the starting
point of the first transect, extend the transect at a 90-deg angle to the baseline. Use the
following procedure as appropriate to simultaneously characterize the parameters at each
observation point. Combine field-collected data with information already available and make
a wetland determination at each observation point. A DATA FORM 1 must be completed for
each observation point.
cL Determine whether normal environmental conditions are present. Determine whether
normal environmental conditions are present by considering the following:
(1) Is the area presently lacking hydrophytic vegetation and/or hydrologic indicators due to
annual or seasonal fluctuations in precipitation or ground-water levels?
(2) Are hydrophytic vegetation indicators lacking due to seasonal fluctuations in temperature?
If the answer to either of these questions is thought to be YES, PROCEED TO Section G. If
the answer to both questions is NO, PROCEED TO STEP 20b.
b. Establish an observation point in the first plant community type encountered. Select a
representative location along the transect in the first plant community type encountered. When
the first plant community type is large and covers a significant distance along the transect,
select an area that is no closer than 300 ft to a perceptible change in plant community type.
PROCEED TO STEP 20c.
c. Characterize parameters. Characterize the parameters at the observation point by
completing (1), (2), and (3) below:
(1) Vegetation. Record on DATA FORM 1 the dominant plant species in each vegetation
layer occurring in the immediate vicinity of the observation point. Use a 5-ft radius for herbs
and saplings/shrubs, and a 30-ft radius for trees and woody vines (when present). Subjectively
determine the dominant species by estimating those having the largest relative basal area
(woody overstory) This term is used because species having the largest individuals may not be
dominant when only a few are present. To use relative basal area, consider both the size and
number of individuals of a species and subjectively compare with other species present.,
greatest height (woody understory), greatest percentage of areal cover (herbaceous
understory), and/or greatest number of stems (woody vines). NOTE: Plot size may be
estimated, and plot size may also be varied when site conditions warrant. Record on DATA
FORM 1 any dominant species observed to have morphological adaptations (Appendix C,
Section 3) for occurrence in wetlands, and determine and record dominant species that have
known physiological adaptations for occurrence in wetlands (Appendix C, Section 3). Record
on DATA FORM 1 the indicator status (Appendix C, Section 1 or 2) of each dominant species.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 48
Hydrophytic vegetation is present at the observation point when more than 50 percent of the
dominant species have an indicator status of OBL, FACW, and/or FAC (For the FAC-neutral
option, see paragraph 35a); when two or more dominant species have observed
morphological or known physiological adaptations for occurrence in wetlands; or when other
indicators of hydrophytic vegetation (PART III, paragraph 35) are present. Complete the
vegetation section of DATA FORM 1. PROCEED TO (2).
(2) Soils. In some cases, it is not necessary to characterize the soils. Examine the vegetation
of DATA FORM 1. Hydric soils can be assumed to be present when:
(a) All dominant plant species have an indicator status of OBL.
(b) All dominant plant species have an indicator status of OBL and/or FACW (at least one
dominant species must be OBL). Soils must be characterized when any dominant species has
an indicator status of FAC.
When either (a) or (b) applies, check the hydric soils blank as positive and PROCEED TO (3).
If neither (a) nor (b) applies but the vegetation qualifies as hydrophytic, dig a soil pit at the
observation point using the procedure described in Appendix D, Section 1. Examine the soil
immediately below the A-horizon or 10-inches (whichever is shallower) and compare its
characteristics (Appendix D, Section 1) with the hydric soil indicators described in PART III,
paragraphs 44 and/or 45. Record findings on DATA FORM 1. If a positive hydric soil
indicator is present, the soil at the observation point is a hydric soil. If no positive hydric soil
indicator is found, the area at the observation point does not have hvdric soils and the area at
the observation point is not a wetland. Complete the soils section of DATA FORM 1 for the
observation point. PROCEED TO (3) if hydrophytic vegetation (1) and hydric soils (2) are
present. Otherwise, PROCEED TO STEP 20d.
(3) Hydrology. Examine the observation point for indicators of wetland hydrology (PART III,
paragraph 49), and record observations on DATA FORM 1. Consider the indicators in the
same sequence as presented in PART III, paragraph 49. If a positive wetland hydrology
indicator is present, the area at the observation point has wetland hydrology. If no positive
wetland hydrologic indicator is present, the area at the observation point is not a wetland.
Complete the hydrology section of DATA FORM 1 for the observation point. PROCEED TO
STEP 20d.
d. Wetland determination. Examine DATA FORM 1 for the observation point. Determine
whether wetland indicators of all three parameters are or would normally be present during a
significant portion of the growing season. If so, the area at the observation point is a wetland.
If no evidence can be found that the area at the observation point normally has wetland
indicators for all three parameters, the area is a nonwetland. PROCEED TO STEP 20e.
e. Sample other observation points along the first transect. Continue along the first transect
until a different community type is encountered. Establish a representative observation point
within this community type and repeat STEP 20c - 20d. If the areas at both observation points
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 49
are either wetfands -or nonwetlands, continue along the transect and repeat STEP 20c - 20d for
the next community type encountered. Repeat for all other community types along the first
transect. If the area at one observation point is wetlands and the next observation point is
nonwetlands (or vice versa), PROCEED TO STEP 20f.
f. Determine wetland-nonwetland boundary. Proceed along the transect from the wetland
observation point toward the nonwetland observation point. Look for subtle changes in the
plant community (e.g. the first appearance of upland species, disappearance of apparent
hydrology indicators, or slight changes in topography). When such features are noted,
establish an observation point and repeat the procedures described in STEP 20c - 20d. NOTE:
A new DATA FORM 1 must be completed for this observation point, and all three parameters
must be characterized by field observation. If the area at this observation point is a wetlands
proceed along the transect toward the nonwetland observation point until upland indicators are
more apparent. Repeat the procedures described in STEP 20c - 20d. If the area at this
observation point is a nonwetland, move Halfway back along the transect toward the last
documented wetland observation point and repeat the procedure described in STEP 20c - 20d.
Continue this procedure until the wetland-nonwetland boundary is found. It is not necessary to
complete a DATA FORM 1 for all intermediate points, but a DATA FORM 1 should be
completed for the wetland-nonwetland boundary. Mark the position of the wetland boundary
on the base map, and continue along the first transect until all community types have been
sampled and all wetland boundaries located. CAUTION: In areas where wetlands are
interspersed among nonwetlands (or vice versa), several boundary determinations will be
required. When all necessary wetland determinations have been completed for the first
transect, PROCEED TO STEP 21.
• STEP 21 - Sample Other Transects. Repeat procedures described in STEP 21 for all other
transacts. When completed, a wetland determination will have been made for one observation
point in each community type along each transect, and all wetland-nonwetland boundaries
along each transect will have been determined. PROCEED TO STEP 22.
• STEP 22 - Synthesize Data. Examine all completed copies of DATA FORM 1, and mark
each plant community type on the base map. Identify each plant community type as either a
wetland (W) or nonwetland (N). If all plant community types are identified as wetlands, the
entire area is wetlands. If all plant community types are identified as nonwetlands, the entire
area is nonwetlands. If both wetlands and nonwetlands are present, identify observation points
that represent wetland boundaries on the base map. Connect these points on the map by
generally following contour lines to separate wetlands from nonwetlands. Walk the contour
line between transacts to confirm the wetland boundary. Should anomalies be encountered, it
will be necessary to establish short transacts in these areas, apply the procedures described in
STEP 20f, and make any necessary adjustments on the base map.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 50
Subsection 3 - Combination of Levels 1 and 2
66. In some cases, especially for large projects, adequate information may already be available
(Section B) to enable a wetland determination for a portion of the project area, while an onsite
visit will be required for the remainder of the area. Since procedures for each situation have
already been described in Subsections 1 and 2, they will not be repeated. Apply the following
steps:
• STEP 1 - Make Wetland Determination for Portions of the Project Area That Are Already
Adequately Characterized. Apply procedures described in Subsection 1. When completed, a
DATA FORM 1 will have been completed for each community type, and a map will have been
prepared identifying each community type as wetland or nonwetland and showing any wetland
boundary occurring in this portion of the project area. PROCEED TO STEP 2.
• STEP 2 - Make Wetland Determination for Portions of the Project Area That Require an
Onsite Visit. Apply procedures described in Subsection 2. When completed, a DATA FORM
1 will have been completed for each plant community type or for a number of observation
points (including wetland boundary determinations). A map of the wetland (if present) will
also be available. PROCEED TO STEP 3.
• STEP 3 - Synthesize Data. Using the maps resulting from STEPS 1 and 2, prepare a
summary map that shows the wetlands of the entire project area. CAUTION: Wetland
boundaries for the two maps will not always match exactly. When this occurs, an additional
site visit will be required to refine the wetland boundaries. Since the degree of resolution of
wetland boundaries will be greater when determined onsite, it may be necessary to employ
procedures described in Subsection 2 in the vicinity of the boundaries determined from
Subsection 1 to refine these boundaries.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 51
Section E. Comprehensive Determinations
67. This section describes procedures for making comprehensive wetland determinations.
Unlike procedures for making routine determinations (Section D), application of procedures
described in this section will result in maximum information for use in making determinations,
and the information usually will be quantitatively expressed. Comprehensive determinations
should only be used when the project area is very complex and/or when the determination
requires rigorous documentation. This type of determination may be required in areas of any
size, but will be especially useful in large areas. There may be instances in which only one
parameter (vegetation, soil, or hydrology) is disputed. In such cases, only procedures
described in this section that pertain to the disputed parameter need be completed. It is
assumed that the user has already completed all applicable steps in Section B. NOTE:
Depending on site characteristics, it may be necessary to alter the sampling design and/or
data collection procedures.
68. This section is divided into five basic types of activities. The first consists of preliminary
field activities that must be completed prior to making a determination (STEPS 1-5). The
second outlines procedures for determining the number and locations of required
determinations (STEPS 6-8). The third describes the basic procedure for making a
comprehensive wetland determination at any given point (STEPS 9-17). The fourth describes
a procedure for determining wetland boundaries (STEP 18). The fifth describes a procedure
for synthesizing the collected data to determine the extent of wetlands in the area (STEPS 20-
21). A flowchart showing the relationship of various steps required for making a
comprehensive determination is presented in Figure 16.
Equipment and material
69. Equipment and materials needed for making a comprehensive determination include:
a. Base map (Section B, STEP 2).
b. Copies of DATA FORMS 1 and 2.
c. Appendices C and D.
d. Compass.
e. Tape (300 ft).
f. Soil auger or spade.
g. Munsell Color Charts (Munsell Color 1975).
h. Quadrat (3.28 ft by 3.28 ft).
i. Diameter or basal area tape (for woody overstory).
Field procedures
70. Complete the following steps:
• STEP I - Identify the Project Area. Using information from the USGS quadrangle or other
appropriate map (Section B), locate and measure the spatial boundaries of the project area.
Determine the compass heading of each boundary and record on the base map (Section B,
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 52
STEP 2). The applicant's survey plan may be helpful in locating the project boundaries.
PROCEED TO STEP 2.
• STEP 2 - Determine Whether an Atypical Situation Exists. Examine the area and determine
whether there is sufficient natural or human-induced alteration to significantly change the area
vegetation, soils, and/or hydrology. If not, PROCEED TO STEP 3. If one or more parameters
have been recently altered significantly.- PROCEED TO Section F and determine whether
there is sufficient evidence that hydrophytic vegetation, hydric soils, and/or wetland hydrology
were present on the area prior to alteration. Then return to this section and characterize
parameters not significantly influenced by human activities. PROCEED TO STEP 3. o STEP
3 - Determine Homogeneity of Vegetation. While completing STEP 2, determine the number
of plant community types present. Mark the approximate location of each community type on
the base map. The number and locations of required wetland determinations will be strongly
influenced by both the size of the area and the number and distribution of plant community
types; the larger the area and greater the number of plant community types, the greater the
number of required wetland determinations. It is imperative that all plant community types
occurring in all portions of the area be included in the investigation. PROCEED TO STEP 4.
• STEP 4 - Determine the Type and Number of Layers in Each Plant Community. Examine
each identified plant community type and determine the type(s) and number of layers in each
community. Potential layers include trees (woody overstory), saplings/shrubs (woody
understory), herbs (herbaceous understory), and/or woody vines. PROCEED TO STEP 5. o
• STEP 5 - Determine Whether Normal Environmental Conditions Are Present. Determine
whether normal environmental conditions are present
at the observation point by considering the following:
a. Is the area at the observation point presently lacking hydrophytic vegetation and/or
hydrologic indicators due to annual or seasonal fluctuations in precipitation or groundwater
levels?
b. Are hydrophytic vegetation indicators lacking due to seasonal fluctuations in temperature?
If the answer to either of these questions is thought to be YES,
PROCEED TO Section G. If the answer to both questions is NO, PROCEEDTO STEP 6.
• STEP 6 - Establish a Baseline. Select one project boundary area as a baseline. The baseline
should extend parallel to any major watercourse and/or perpendicular to a topographic gradient
(see Figure 17). Determine the baseline length and record on the base map both the baseline
length and its compass heading. PROCEED TO STEP 7.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 53
• STEP 7. Establish Transect Locations. Divide the baseline into a number of equal segments
(Figure 17). Use the following as a guide to determine the appropriate number of baseline
segments:
Length of
Baseline Length, ft Number of Segments Baseline Segment ft
>50 - 500
>500 - 1,000
>1,000- 5,000
>5,000 - 10,000
>10,000*
3
3
5
7
variable
18-
167
200
700
-2,
167
-333
- 1,000
- 1,400
000
*If the baseline exceeds 5 miles, baseline segments should be 0.5 mile in length.
Use a random numbers table or a calculator with a random numbers generation feature to
determine the position of a transect starting point within each baseline segment. For example,
when the baseline is 4,000 ft, the number of baseline segments will be five, and the baseline
segment length will be 4,000/5 = 800 ft. Locate the first transect within the first 800 ft of the
baseline. If the random numbers table yields 264 as the distance from the baseline starting
point, measure 264 ft from the baseline starting point and establish the starting point of the first
transect. If the second random number selected is 530, the starting point of the second
transect will be located at a distance of 1,330 ft (800 + 530 ft) from the baseline starting point.
CA UTION: Make sure that each plant community type is included in at
least one transect. If not, modify the sampling design accordingly. When the starting point
locations for all required transacts have been determined, PROCEED TO STEP 8.
• STEP 8 - Determine the Number of Required Observation Points Along Transects. The
number of required observation points along each transect will be largely dependent on
transect length. Establish observation points along each transect using the following as a
guide:
Transect Number of Interval Between
Length, ft Observation Points Observation Points, ft
<1,000 2-10 100
1,000-<5,000 10 100-500
5,000-<10,000 10 500-1,000
>10,000 >10 1,000
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 54
Establish the first observation point at a distance of 50 ft from the baseline (Figure 17). When
obvious nonwetlands occupy a long portion of the transect from the baseline starting point,
establish the first observation point in the obvious nonwetland at a distance of approximately
300 ft from the point that the obvious nonwetland begins to intergrade into a potential wetland
community type. Additional observation points must also be established to determine the
wetland boundary between successive regular observation points when one of the points is a
wetland and the other is a nonwetland. CA UTION: In large areas having a mosaic of plant
community types, several wetland boundaries may occur along the same transect. PROCEED
TO STEP 9 and apply the comprehensive wetland determination procedure at each required
observation point. Use the described procedure to simultaneously characterize the vegetation,
soil, and hydrology at each required observation point along each transect, and use the
resulting characterization to make a wetland determination at each point. NOTE: All required
wetland boundary determinations should be made while proceeding along a transect.
• STEP 9 - Characterize the Vegetation at the First Observation Point Along the First Transect.
There is no single best procedure for characterizing vegetation. Methods described in STEP 9
afford standardization of the procedure. However, plot size and descriptors for determining
dominance may vary. Record on DATA FORM 2 the vegetation occurring at the first
observation point along the first transect by completing the following (as appropriate):
a. Trees. Identify each tree occurring within a 3 0-ft radius (A larger sampling plot may be
necessary when trees are large and widely spaced.) of the observation point, measure its basal
area (square inches) or diameter at breast height (DBH) using a basal area tape or diameter
tape, respectively, and record. NOTE: If DBH is measured, convert values to basal area by
applying the formula A = pi r2. This must be done on an individual basis. A tree is any
nonclimbing, woody plant that has a DBH of >3.0 in., regardless of height.
b. Saplings/shrubs. Identify each sapling/shrub occurring within a 10-ft radius of the
observation point, estimate its height, and record the midpoint of its class range using the
following height classes (height is used as an indication of dominance; taller individuals exert
a greater influence on the plant community):
Height Height Class Midpoint of
Class Range, ft Range, ft
113 2
235 4
357 6
479 8
5 9 11 10
6 >11 12
A sapling/shrub is any woody plant having a height >3.2 ft but a stem diameter of <3.0 in.,
exclusive of woody vines.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 55
(L, Herbs. Place a 3.28- by 3.28-ft quadrat with one corner touching the observation point and
one edge adjacent to the transect line. As an alternative, a 1.64-ft-radius plot with the center of
the plot representing the observation point position may be used. Identify each plant species
with foliage extending into the quadrat and estimate its percent cover by applying the
following cover classes:
Cover Class Midpoint of
Class Range. % Class Range. %
1 0-5 2.5
2 >5-25 15.0
3 >25 -50 37.5
4 >50 -75 62.5
5 >75 -95 85.0
6 >95 - 100 97.5
Include all nonwoody plants and woody plants <3.2 ft in height.
NOTE: Total percent cover for all species will often exceed 100 percent.
d. Woody vines (lianas). Identify species of woody vines climbing each tree and
sapling/shrub sampled in STEPS 9a and 9b above, and record the number of stems of each.
Since many woody vines branch profusely, count or estimate the number of stems at the
ground surface. Include only individuals rooted in the 10-ft radius plot. Do not include
individuals <3.2 ft in height. PROCEED TO STEP 10.
• STEP 10 - Analyze Field Vegetation Data. Examine the vegetation data (STEP 9) and
determine the dominant species in each vegetation layer (The same species may occur as a
dominant in more than one vegetation layer) by completing the following:
a. Trees. Obtain the total basal area (square inches) for each tree species identified in STEP
9a by summing the basal area of all individuals of a species founi in the sample plot. Rank the
species in descending order of dominance based on total basal area. Complete DATA FORM
2 for the tree layer.
b. Saplings/shrubs. Obtain the total height for each sapling/ shrub species identified in STEP
9b_. Total height, which is an estimate of dominance, is obtained by summing the midpoints of
height classes for all individuals of a species found in the sample plot. Rank the species in
descending order of dominance based on sums of midpoints of height class ranges. Complete
DATA FORM 2 for the sapling/shrub layer.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 56
c. Herbs. Obtain the total cover for each herbaceous and woody seedling species identified in
STEP 9c. Total cover is obtained by using the midpoints of the cover class range assigned to
each species (only one estimate of cover is made for a species in a given plot). Rank herbs and
woody seedlings in descending order of dominance based on percent cover. Complete DATA
FORM 2 for the herbaceous layer.
d. Woody vines (lianas). Obtain the total number of individuals of each species of woody
vine identified in STEP 9d. Rank the species in descending order of dominance based on
number of stems. Complete DATA FORM 2 for the woody vine layer. PROCEED TO STEP
11.
• STEP 11 - Characterize Soil. If a soil survey is available (Section B), the soil type may
already be known. Have a soil scientist confirm that the soil type is correct, and determine
whether the soil series is a hydric soil (Appendix D, Section 2). CAUTION: Mapping units on
soil surveys sometimes have inclusions of soil series or phases not shown on the soil survey
map, If a hydric soil type is confirmed, record on DATA FORM I and PROCEED TO STEP
12. If not, dig a soil pit using a soil auger or spade (See Appendix D, Section 1) and look for
indicators of hydric soils immediately below the A-horizon or 10 inches (whichever is
shallower) (PART III, paragraphs 44 and/or 45). Record findings on DATA FORM 1.
PROCEED TO STEP 12.
• STEP 12 - Characterize Hydrology. Examine the observation point for indicators of wetland
hydrology (PART III, paragraph 49), and record observations on DATA FORM 1. Consider
indicators in the same sequence as listed in paragraph 49. PROCEED TO STEP 13.
• STEP 13 - Determine Whether Hydrophytic Vegetation Is Present. Record the three
dominant species from each vegetation layer (five species if only one or two layers are present)
on DATA FORM 1.* Record all dominant species when less than three are present in a
vegetation layer. Determine whether these species occur in wetlands by considering the
following:
a. More than 50 percent of the dominant plant species are OBL. FACW. and/or FAC** on
lists of plant species that occur in wetlands. For the FAC-neutral option, see paragraph 35a.
Record the indicator status of all dominant species (Appendix C, Section I or 2) on DATA
FORM 1. Hydrophytic vegetation is present when the majority of the dominant species have
an indicator status of OBL, FACW, or FAC. CA UTION: Not necessarily all plant communities
composed of only FAC species are hydrophytic commnities. They are hydrophytic
communities only when positive indicators of hydric soils and wetland hydrology are also
found. If this indicator is satisfied, complete the vegetation portion of DATA FORM 1 and
PROCEED TO STEP 14. If not, consider other indicators of hydrophytic vegetation.
b. Presence of adaptations for occurrence in wetlands. Do any of ihe-species listed on DATA
FORM I have observed morphological or known physiological adaptations (Appendix C,
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 57
Section 3) for occurrence in wetlands? If so, record species having such adaptations on DATA
FORM 1. When two or more dominant species have observed morphological adaptations or
known physiological adaptations for occurrence in wetlands, hydrophytic vegetation is present.
If so, complete the vegetation portion of DATA FORM I and PROCEED TO STEP 14. If not,
consider other indicators of hydrophytic vegetation.
c. Other indicators of hydrophytic vegetation. Consider other indicators (see PART III,
paragraph 35) that the species listed on DATA FORM I are commonly found in wetlands. If
so, complete the vegetation portion of DATA FORM I bv recording sources of supporting
information, and PROCEED TO STEP 14. If no indicator of hydrophytic vegetation is
present, the area at the observation point is not a wetlands In such cases, it is unnecessary to
consider soil and hydrology at that observation point. PROCEED TO STEP 17.
• STEP 14 - Determine Whether Hydric Soils Are Present. Examine DATA FORM 1 and
determine whether any indicator of hydric soils is present. If so, complete the soils portion of
DATA FORM I and PROCEED TO STEP 15. If not, the area at the observation point is not a
wetlands. PROCEED TO STEP 17.
• STEP 15 - Determine Whether Wetland Hydrology Is Present. Examine DATA FORM 1
and determine whether any indicator of wetland hydrology is present. Complete the hydrology
portion of DATA FORM I and PROCEED TO STEP 16.
• STEP 16 - Make Wetland Determination. When the area at the observation point presently
or normally has wetland indicators of all three parameters, it is a wetlands When the area at the
observation point presently or normally lacks wetland indicators of one or more parameters, it
is a nonwetland. PROCEED TO STEP 17.
• STEP 17 - Make Wetland Determination at Second Observation Point. Locate the second
observation point along the first transect and make a wetland determination by repeating
procedures described in STEPS 9-16. When the area at the second observation point is the
same as the area at the first observation point (i.e. both wetlands or both nonwetlands),
PROCEED TO STEP 19. When the areas at the two observation points are different (i.e. one
wetlands, the other nonwetlands), PROCEED TO STEP 18.
• STEP 18 - Determine the Wetland Boundary Between Observation Points. Determine the
position of the wetland boundary by applying the following procedure:
a. Look for a change in vegetation or topography. NOTE: The changes may sometimes be
very subtle. If a change is noted, establish an observation point and repeat STEPS 9-16.
Complete a DATA FORM 1. If the area at this point is a wetlands proceed toward the
nonwetland observation point until a more obvious change in vegetation or topography is
noted and repeat the procedure. If there is no obvious change, establish the next observation
point approximately halfway between the last observation point and the nonwetland
observation point and repeat STEPS 9-16.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 58
b. Make as many additional wetland determinations as necessary to find the wetland
boundary. NOTE: The completed DATA FORM 1's for the original two observation points
often will provide a clue as to the parameters that change between the two points.
c. When the wetland boundary is found, mark the boundary location on the base map and
indicate on the DATA FORM 1 that this represents a wetland boundary. Record the distance
of the boundary from one of the two regular observation points. Since the regular observation
points represent known distances from the baseline, it will be possible to accurately pinpoint
the boundary location on the base map. PROCEED TO STEP 19.
• STEP 19 - Make Wetland Determinations at All Other Required Observation Points Along
All Transects. Continue to locate and sample all required observation points along all
transacts. NOTE: The procedure described in Step 18 must be applied at every position where
a wetzand boundary occurs between successive observation points. Complete a DATA FORM
1 for each observation point and PROCEED TO STEP 20.
• STEP 20 - Synthesize Data to Determine the Portion of the Area Containing Wetlands.
Examine all completed copies of DATA FORM 1 (STEP 19), and mark on a copy of the base
map the locations of all-observation points that are wetlands with a W and all observation
points that are nonwetlands with an N. Also, mark all wetland boundaries occurring along
transacts with an X. If all the observation points are wetlands, the entire area is wetlands. If all
observation points are nonwetlands, none of the area is wetlands. If some wetlands and some
nonwetlands are present, connect the wetland boundaries (X) by following contour lines
between transacts. CAUTION: -If the determination is considered to be highly controversial, it
may be necessary to be more precise in determining the wetland boundary between transacts.
This is also true for very large areas where the distance between transacts is greater. If this is
necessary, PROCEED TO STEP 21.
• STEP 21 - Determine Wetland Boundary Between Transects. Two procedures may be used
to determine the wetland boundary between transacts, both of which involve surveying:
a. Survey contour from wetland boundary along transacts. The first method involves
surveying the elevation of the wetland boundaries along transacts and then extending the
survey to determine the same contour between transacts. This procedure will be adequate in
areas where there is no significant elevational change between transacts. However, if a
significant elevational change occurs between transacts, either the surveyor must adjust
elevational readings to accommodate such changes or the second method must be used.
NOTE: The surveyed wetland boundary must be examined to ensure that no anomalies exist.
If these occur, additional wetland determinations will be required in the portion of the area
where the anomalies occur, and the wetland boundary must be adjusted accordingly.
b. Additional wetland determinations between transacts. This procedure consists of traversing
the area between transacts and making additional wetland determinations to locate the wetland
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 59
boundary at sufficiently close intervals (not necessarily standard intervals) so that the area can
be surveyed. Place surveyor flags at each wetland boundary location. Enlist a surveyor to
survey the points between transacts. From the resulting survey data, produce a map that
separates wetlands from nonwetlands.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 60
Section F. Atypical Situations
71. Methods described in this section should be used only when a determination has already
been made in Section D or E that positive indicators of hydrophytic vegetation, hydric soils,
and/or wetland hydrology could not be found due to effects of recent human activities or
natural events. This section is applicable to delineations made in the following types of
situations:
a. Unauthorized activities. Unauthorized discharges requiring Tnf-orcement actions may
result in removal or covering of indicators of one or more wetland parameters. Examples
include, but are not limited to: (1) alteration or removal of vegetation; (2) placement of
dredged or fill material over hydric soils; and/or (3) construction of levees, drainage systems,
or dams that significantly alter the area hydrology. NOTE: This section should not be used for
activities that have been previously authorize 3-or those that are exempted from CE regulation.
For example, this section is not applicable to areas that have been drained under CE
authorization or that did not require CE authorization. Some of these areas may still be
wetlands, but procedures described in Section D or E must be used in these cases.
b. Natural events. Naturally occurring events may result in either creation or alteration of
wetlands. For example, recent beaver dams may impound water, thereby resulting in a shift of
hydrology and vegetation to wetlands. However, hydric soil indicators may not have
developed due to insufficient time having passed to allow their development. Fire, avalanches,
volcanic activity, and changing river courses are other examples. NOTE: It is necessary to
determine whether alterations to an area have resulted in changes that are now the "normal
circumstances. " The relative permanence of the change and whether the area is now
functioning as a wetland must be considered.
c. Man-induced wetlands. Procedures described in Subsection 4 are for use in delineating
wetlands that have been purposely or incidentally created by human activities, but in which
wetland indicators of one or more parameters are absent. For example, road construction may
have resulted in impoundment of water in an area that previously was nonwetland, thereby
effecting hydrophytic vegetation and wetland hydrology in the area. However, the area may
lack hydric soil indicators. NOTE: Subsection D is not intended to bring into CE jurisdiction
those manmade wetlands that are exempted under CE regulations or policy. It is also
important to consider whether the man-induced changes are now the "normal circumstances"
for the area. Both the relative permanence of the change and the functioning of the area as a
wetland are implied.
72. When any of the three types of situations described in paragraph 71 occurs, application of
methods described in Sections D and/or E will lead to the conclusion that the area is not a
wetland because positive wetland indicators for at least one of the three parameters will be
absent. Therefore, apply procedures described in one of the following subsections (as
appropriate) to determine whether positive indicators of hydrophytic vegetation, hydric soils,
and/or wetland hydrology existed prior to alteration of the area. Once these procedures have
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 61
been employed, RETURN TO Section D or E to make a wetland determination. PROCEED
TO the appropriate subsection.
Subsection 1 - Vegetation
73. Employ the following steps to determine whether hydrophytic vegetation previously
occurred:
• STEP 1 - Describe the Type of Alteration. Examine the area and describe the type of
alteration that occurred. Look for evidence of selective harvesting, clear cutting, bulldozing,
recent conversion to agriculture, or other activities (e.g., burning, discing, or presence of
buildings, dams, levees, roads, parking lots, etc.). Determine the approximate date* when the
alteration occurred. Record observations on DATA FORM 3, and PROCEED TO STEP 2.
• STEP 2 - Describe Effects on Vegetation. Record on DATA FORM 3 a general description
of how the activities (STEP 1) have affected the plant communities. Consider the following:
a. Has all or a portion of the area been cleared of vegetation?
b. Has only one layer of the plant community (e.g. trees) been removed?
c. Has selective harvesting resulted in removal of some species?
d. Has all vegetation been covered by fill, dredged material, or structures?
e. Have increased water levels resulted in the death of some individuals?
It is especially important to determine whether the alteration occurred prior to implementation
of Section 404. PROCEED TO STEP 3.
• STEP 3 - Determine the Type of Vegetation That Previously Occurred. Obtain all possible
evidence of the type of plant communities that occurred in the area prior to alteration.
Potential sources of such evidence include:
a. Aerial photography. Recent (within 5 years) aerial photography
an often be used to document the type of previous vegetation. The general type of plant
communities formerly present can usually be determined, and species identification is
sometimes possible.
b. Onsite inspection. Many types of activities result in only partial removal of the previous
plant communities, and remaining species may be indicative of hydrophytic vegetation. In
other cases, plant fragments (e.g. stumps, roots) may be used to reconstruct the plant
community types that occurred prior to site alteration. Sometimes, this can be determined by
examining piles of debris resulting from land-clearing operations or excavation to uncover
identifiable remains of the previous plant community.
c. Previous site inspections. Documented evidence from previous inspections of the area may
describe the previous plant communities, particularly in cases where the area was altered after
a permit application was denied.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 62
d. Adjacent vegetation. Circumstantial evidence of the type of plant communities that
previously occurred may sometimes be obtained by examining the vegetation in adjacent
areas. If adjacent areas have the same topographic position, soils, and hydrology as the altered
area, the plant community types on the altered area were probably similar to those of the
adjacent areas.
e. SCS records. Most SCS soil surveys include a description of the plant community types
associated with each soil type. If the soil type on the altered area can be determined, it may be
possible to generally determine the type of plant communities that previously occurred.
f. Permit applicant. In some cases, the permit applicant may provide important information
about the type of plant communities that occurred prior to alteration.
g. Public. Individuals familiar with the area may provide a good general description of the
previously occurring plant communities.
h. NWI wetland maps. The NWI has developed wetland type maps for many areas. These
may be useful in determining the type of plant communities that occurred prior to alteration.
To develop the strongest possible record, all of the above sources should be considered. If the
plant community types that occurred prior to alteration can be determined, record them on
DATA FORM 3 and also record the basis used for the determination. PROCEED TO STEP 4.
If it is impossible to determine the plant community types that occurred on the area prior to
alteration, a determination cannot be made using all three parameters. In such cases, the
determination must be based on the other two parameters. PROCEED TO Subsection 2 or 3 if
one of the other parameters has been altered, or return to the appropriate Subsection of Section
D or to Section E, as appropriate.
• STEP 4 - Determine Whether Plant Community Types Constitute Hydrophytic Vegetation.
Develop a list of species that previously occurred on the site (DATA FORM 3). Subject the
species list to applicable indicators of hydrophytic vegetation (PART III, paragraph 35). If
none of the indicators are met, the plant communities that previously occurred did not
constitute hydrophytic vegetation. If hydrophytic vegetation was present and no other
parameter was in question, record appropriate data on the vegetation portion of DATA FORM
3, and return to either the appropriate subsection of Section D or to Section E. If either of the
other parameters was also in question, PROCEED TO Subsection 2 or 3.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 63
Subsection 2 - Soils
74. Employ the following steps to determine whether hydric soils previously occurred:
• STEP 1 - Describe the Type of Alteration. Examine the area and describe the type of
alteration that occurred. Look for evidence of:
a. Deposition of dredged or fill material or natural sedimentation. In many cases the presence
of fill material will be obvious. If so, it will be necessary to dig a hole to reach the original soil
(sometimes several feet deep). Fill material will usually be a different color or texture than the
original soil (except when fill material has been obtained from like areas onsite). Look for
decomposing vegetation between soil layers and the presence of buried organic or hydric soil
layers. In accreting or recently formed sandbars in riverine situations, the soils may support
hydrophytic vegetation but lack hydric soil characteristics.
b. Presence of nonwoody debris at the surface. This can only be applied in areas where the
original soils do not contain rocks.
Nonwoody debris includes items such as rocks, bricks, and concrete fragments.
c. Subsurface plowing. Has the area recently been plowed below he A-horizon or to depths of
greater than 10 in.?
d. Removal of surface layers. Has the surface soil layer been removed by scraping or natural
landslides? Look for bare soil surfaces with exposed plant roots or scrape scars on the surface.
e. Presence of man-made structures. Are buildings, dams, levees, roads, or parking lots
present?
Determine the approximate date (It is especially important to determine whether the alteration
occurred prior to implementation of Section 404.} when the alteration occurred. This may
require checking aerial photography, examining building permits, etc. Record on DATA
FORM 3, and PROCEED TO STEP 2.
• Step 2 - Describe Effects on Soils. Record on DATA FORM 3 a general description of how
identified activities in STEP 1 have affected the soils. Consider the following:
a. Has the soil been buried? If so, record the depth of fill material and determine whether the
original soil is intact.
b. Has the soil the original been mixed at a depth below the A-horizon or greater than 10
inches? If so, it will be necessary to examine soil at a depth immediately below the plowed
zone. Record supporting evidence.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 64
c. Has the soil been sufficiently altered to change the soil phase? Describe these changes.
PROCEED TO STEP 3.
• STEP 3 - Characterize Soils That Previously Occurred. Obtain all possible evidence that
may be used to characterize soils that previously occurred on the area. Consider the following
potential sources of information:
a. Soil surveys. In many cases, recent soil surveys will be available. If so, determine the soil
series that were mapped for the area, and compare these soil series with the list of hydric soils
(Appendix D, Section 2). If all soil series are listed as hydric soils, the entire area had hydric
soils prior to alteration.
b. Characterization of buried soils. When fill material has been placed over the original soil
without physically disturbing the soil, examine and characterize the buried soils. To
accomplish this, dig a hole through the fill material until the original soil is encountered.
Determine the point at which the original soil material begins. Remove 12 inches of the
original soil from the hole and look for indicators of hydric soils (PART III, paragraphs 44
and/or 45) immediately below the A-horizon or 10 inches (whichever is shallower). Record on
DATA FORM 3 the color of the soil matrix, presence of an organic layer, presence of mottles
or gleying, and/or presence of manganese concretions. If the original soil is motthe chroma of
the soil matrix is 2 or less, (The matrix chroma must be 1 or less if no mottles are present (see
paragraph 44). The soil must be moist when colors are determined.), a hydric formerly
present on the site. If any of these indicafound, the original soil was a hydric soil. (NOTE: fill
material is a thick layer, it might be necessary to use backhoe or posthole digger to excavate
the soil pit.) If USGS quadrangle maps indicate distinct variation in area topography, this
procedure must be applied in each portion of the area that originally had a different surface
elevation. Record findings on DATA FORM 3.
c. Characterization of plowed soils. Determine the depth to which the soil has been disturbed
by plowing. Look for hydric soil characteristics (PART III, paragraphs 44 and/or 45)
immediately below this depth. Record findings on DATA FORM 3.
d. Removal of surface layers. Dig a hole (Appendix D, Section 1) and determine whether the
entire surface layer (A-horizon) has been removed. If so, examine the soil immediately below
the top of the subsurface layer (B-horizon) for hydric soil characteristics. As an alternative,
examine an undisturbed soil of the same soil series occurring in the same topographic position
in an immediately adjacent area that has not been altered. Look for hydric soil indicators
immediately below the A-horizon or 10 inches (whichever is shallower), and record findings
on DATA FORM 3.
If sufficient data on soils that existed prior to alteration can be obtained to determine whether a
hydric soil was present, PROCEED TO STEP 4. If not, a determination cannot be made using
soils. Use the other parameters (Subsections 1 and 3) for the determination.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 65
• STEP 4 - Determine Whether Hydric Soils Were Formerly Present. Examine the available
data and determine whether indicators of hydric soils (PART III, paragraphs 44 and/or 45)
were formerly present. If no indicators of hydric soils were found, the original soils were not
hydric soils. If indicators of hydric soils were found, record the appropriate indicators on
DATA FORM 3 and PROCEED TO Subsection 3 if the hydrology of the area has been
significantly altered or return either to the appropriate subsection of Section D or to Section E
and characterize the area hydrology.
Subsection 3 - Hydrology
75. Apply the following steps to determine whether wetland hydrology previously occurred:
• STEP 1 - Describe the Type of Alteration. Examine the area and describe the type of
alteration that occurred. Look for evidence of:
a. Dams. Has recent construction of a dam or some natural event (e.g. beaver activity or
landslide) caused the area to become increasingly wetter or drier? NOTE: This activity could
have occurred a considerable distance away from the site in question.
b. Levees, dikes, and similar structures. Have levees or dikes recently been constructed that
prevent the area from becoming periodically inundated by overbank flooding?
c. Ditching. Have ditches been constructed recently that cause the area to drain more rapidly
following inundation?
d. Filling of channels or depressions (land-leveling). Have natural channels or depressions
been recently filled?
e. Diversion of water. Has an upstream drainage pattern been altered that results in water
being diverted from the area?
f. Ground-water extraction. Has prolonged and intensive pumping of ground water for
irrigation or other purposes significantly lowered the water table and/or altered drainage
patterns?
g. Channelization. Have feeder streams recently been channelized sufficiently to alter the
frequency and/or duration of inundation?
Determine the approximate date* It is especially important to determine whether the alteration
occurred prior to implementation of Section 404. when the alteration occurred. Record
observations on DATA FORM 3 and PROCEED TO STEP 2.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 66
• STEP 2 - Describe Effects of Alteration on Area Hydrology. Record on DATA FORM 3 a
general description of how the observed alteration (STEP 1) has affected the area. Consider
the following:
a. Is the area more frequently or less frequently inundated than prior to alteration? To what
degree and why?
b. Is the duration of inundation and soil saturation different than prior to alteration? How
much different and why? PROCEED TO STEP 3.
• STEP 3 - Characterize the Hydrology That Previously Existed in the Area. Obtain all
possible evidence that may be used to characterize the hydrology that previously occurred.
Potential sources of information include:
a. Stream or tidal gage data. If a stream or tidal gaging station is located near the area, it may
be possible to calculate elevations representing the upper limit of wetlands hydrology based on
duration of inundation. Consult hydrologists from the local CE District Office for assistance.
The resulting mean sea level elevation will represent the upper limit of inundation for the area
in the absence of any alteration. If fill material has not been placed on the area, survey this
elevation from the nearest USGS benchmark. Record elevations representing zone boundaries
on DATA FORM 3. If fill material has been placed on the area, compare the calculated
elevation with elevations shown on a USGS quadrangle or any other survey map that predated
site alteration.
b. Field hydrologic indicators. Certain field indicators of wetland hydrology (PART III,
paragraph 49) may still be present. Look for watermarks on trees or other structures, drift
lines, and debris deposits. Record these on DATA FORM 3. If adjacent undisturbed areas are
in the same topographic position and are similarly influenced by the same sources of
inundation, look for wetland indicators in these areas.
c. Aerial photography. Examine any available aerial photography and determine whether the
area was inundated at the time of the photographic mission. Consider the time of the year that
the aerial photography was taken and use only photography taken during the growing season
and prior to site alteration.
d. Historical records. Examine any available historical records for evidence that the area has
been periodically inundated. Obtain copies of any such information and record findings on
DAT A FORM 3.
e. Floodplain Management Maps. Determine the previous frequency of inundation of the area
from Floodplain Management Maps (if available). Record flood frequency on DATA FORM
O
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 67
f. Public or local government officials. Contact individuals who might have knowledge that
the area was periodically inundated.
If sufficient data on hydrology that existed prior to site alteration can be obtained to determine
whether wetland hydrology was previously present, PROCEED TO STEP 4. If not, a
determination involving hydrology cannot be made. Use other parameters (Subsections 1 and
2) for the wetland determination. Return to either the appropriate subsection of Section D or to
Section E and complete the necessary data forms. PROCEED TO STEP 4 if the previous
hydrology can be characterized. *
• STEP 4 - Determine Whether Wetland Hydrology Previously Occurred. Examine the
available data and determine whether indicators of wetland hydrology (PART III, paragraph
49) were present prior to site alteration. If no indicators of wetland hydrology were found, the
original hydrology of the area was not wetland hydrology. If indicators of wetland hydrology
were found, record the appropriate indicators on DATA FORM 3 and return either to the
appropriate subsection of Section D or to Section E and complete the wetland determination.
Subsection 4 - Man-Induced Wetlands
76. A man-induced wetland is an area that has developed at least some characteristics of
naturally occurring wetlands due to either intentional or incidental human activities. Examples
of man-induced wetlands include irrigated wetlands, wetlands resulting from impoundment
(e.g. reservoir shorelines), wetlands resulting from filling of formerly deepwater habitats,
dredged material disposal areas, and wetlands resulting from stream channel realignment.
Some man-induced wetlands may be subject to Section 404. In virtually all cases, man-
induced wetlands involve a significant change in the hydrologic regime, which may either
increase or decrease the wetness of the area. Although wetland indicators of all three
parameters (i.e. vegetation, soils, and hydrology) may be found in some man-induced
wetlands, indicators of hydric soils are usually absent. Hydric soils require long periods
(hundreds of years) for development of wetness characteristics, and most man-induced
wetlands have not been in existence for a sufficient period to allow development of hydric soil
characteristics. Therefore, application of the multiparameter approach in making wetland
determinations in man-induced wetlands must be based on the presence of hydrophytic
vegetation and wetland hydrology. (Uplands that support hydrophytic vegetation due to
agricultural irrigation and that have an obvious hydrologic connection to other "waters of the
United States" should not be delineated as wetlands under this subsection). There must also
be documented evidence that the wetland resulted from human activities. Employ the
following steps to determine whether an area consists of wetlands resulting from human
activities:
• STEP I - Determine Whether the Area Represents a Potential Man-Induced Wetland.
Consider the following questions:
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 68
a. Has a recent man-induced change in hydrology occurred that caused the area to become
significantly wetter?
b. Has a major man-induced change in hydrology that occurred in the past caused a former
deepwater aquatic habitat to become significantly drier?
c. Has man-induced stream channel realignment significantly altered the area hydrology?
d. Has the area been subjected to long-term irrigation practices? If the answer to any of the
above questions is YES, document the approximate time during which the change in
hydrology occurred, and PROCEED TO STEP 2. If the answer to all of the questions is NO,
procedures described in Section D or E must be used.
• STEP 2 - Determine Whether a Permit Will be Needed if the Area is Found to be a Wetland.
Consider the current CE regulations and policy regarding man-induced wetlands. If the type of
activity resulting in the area being a potential man-induced wetland is exempted by regulation
or policy, no further action is needed. If not exempt, PROCEED TO STEP 3.
• STEP 3 - Characterize the Area Vegetation. Soils, and Hydrology. Apply procedures
described in Section D (routine determinations) or Section E (comprehensive determinations)
to the area. Complete the appropriate data forms and PROCEED TO STEP 4.
• STEP 4 - Wetland Determination. Based on information resulting from STEP 3, determine
whether the area is a wetland. When wetland indicators of all three parameters are found, the
area is a wetland. When indicators of hydrophytic vegetation and wetland hydrology are
found and there is documented evidence that the change in hydrology occurred so recently that
soils could not have developed hydric characteristics, the area is a wetland. In such cases, it is
assumed that the soils are functioning as hydric soils. CAUTION: if hydrophytic vegetation is
being-maintained only because of man-induced wetland hydrology that would no longer exist
if the activity (e.g. irrigation) were to be terminated., the area should not be considered a
wetland.
Section G - Problem Areas
77. There are certain wetland types and/or conditions that may make application of indicators
of one or more parameters difficult, at least at certain times of the year. These are not
considered to be atypical situations. Instead, they are wetland types in which wetland
indicators of one or more parameters may be periodically lacking due to normal seasonal or
annual variations in environmental conditions that result from causes other than human
activities or catastrophic natural events.
Types of problem areas
78. Representative examples of potential problem areas, types of variations that occur, and
their effects on wetland indicators are presented in the following subparagraphs. Similar
situations may sometimes occur in other wetland types. Note: This section is not intended to
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 69
bring nonwetland areas having wetzand indicators of two, but not all three, parameters into
Section 404 jurisdiction.
a. Wetlands on drumlins. Slope wetlands occur in glaciated areas Tn-which thin soils cover
relatively impermeable glacial till or in which layers of glacial till have different hydraulic
conditions that produce a broad zone of ground-water seepage. Such areas are seldom, if ever,
flooded, but downslope groundwater movement keeps the soils saturated for a sufficient
portion of the growing season to produce anaerobic and reducing soil conditions. This fosters
development of hydric soil characteristics and selects for hydrophytic vegetation. Indicators of
wetland hydrology may be lacking during the drier portion of the growing season.
b. Seasonal wetlands. In many regions (especially in western states), depression areas occur
that have wetland indicators of all three parameters during the wetter portion of the growing
season, but normally lack wetland indicators of hydrology and/or vegetation during the drier
portion of the growing season. Obligate hydrophytes and facultative wetland plant species
(Appendix C, Section 1 or 2) normally are dominant during the wetter portion of the growing
season, while upland species (annuals) may be dominant during the drier portion of the
growing season. These areas may be inundated during the wetter portion of the growing
season, but wetland hydrology indicators may be totally lacking during the drier portion of the
growing season. It is important to establish that an area truly is a water body. Water in a
depression normally must be sufficiently persistent to exhibit an ordinary high-water mark or
the presence of wetland characteristics before it can be considered as a water body potentially
subject to Clean Water Act jurisdiction. The determination that an area exhibits wetland
characteristics for a sufficient portion of the growing season to qualify as a wetland under the
Clean Water Act must be made on a case-by-case basis. Such determinations should consider
the respective length of time that the area exhibits upland and wetland characteristics, and the
manner in which the area fits into the overall ecological system as a wetland. Evidence
concerning the persistence of an area's wetness can be obtained from its history, vegetation,
soil, drainage characteristics, uses to which it has been subjected, and weather or hydrologic
records.
c. Prairie potholes. Prairie potholes normally occur as shallow depressions in glaciated
portions of the north-central United States. Many are landlocked, while others have a drainage
outlet to streams or other potholes. Most have standing water for much of the growing season
in years of normal or above normal precipitation, but are neither inundated nor have saturated
soils during most of the growing season in years of below normal precipitation. During dry
years, potholes often become incorporated into farming plans, and are either planted to row
crops (e.g. soybeans) or are mowed as part of a haying operation. When this occurs, wetland
indicators of one or more parameters may be lacking. For example, tillage would eliminate
any onsite hydrologic indicator, and would make detection of soil and vegetation indicators
much more difficult.
d. Vegetated flats. In both coastal and interior areas throughout the Nation, vegetated flats are
often dominated by annual species that are categorized as OBL. Application of procedures
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 70
described in Sections D and E during the growing season will clearly result in a positive
wetland determination. However, these areas will appear to be unvegetated mudflats when
examined during the nongrowing season, and the area would not qualify at that time as a
wetland due to an apparent lack of vegetation.
Wetland determinations in problem areas
79. Procedures for making wetland determinations in problem areas are presented below.
Application of these procedures is appropriate only when a decision has been made in Section
D or E that wetland indicators of one or more parameters were lacking, probably due to normal
seasonal or annual variations in environmental conditions. Specific procedures to be used will
vary according to the nature of the area, site conditions, and parameters) affected by the
variations in environmental conditions. A determination must be based on the best evidence
available to the field inspector, including:
a. Available information (Section B).
b. Field data resulting from an onsite inspection.
c. Basic knowledge of the ecology of the particular community
type(s) and environmental conditions associated with the community type.
NOTE: The procedures described below should only be applied to parameters not adequately
characterized in Section D or E. Complete the following steps:
• STEP 1 - Identify the Parameter(s) to be Considered. Examine the DATA FORM 1 (Section
D or E) and identify the parameter(s) that must be given additional consideration. PROCEED
TO STEP 2.
• STEP 2 - Determine the Reason for Further Consideration. Determine the reason why the
parameters) identified in STEP 1 should be given further consideration. This will require a
consideration and documentation of:
a. Environmental condition(s) that have impacted the parameters).
b. Impacts of the identified environmental condition(s) on the parameter(s) in question.
Record findings in the comments section of DATA FORM 1. PROCEED TO STEP 3.
• STEP 3 - Document Available Information for Parameter(s) in Question. Examine the
available information and consider personal ecological knowledge of the range of normal
environmental conditions of the area. Local experts (e.g. university personnel) may provide
additional information. Record information on DATA FORM 1. PROCEED TO STEP 4.
• STEP 4 - Determine Whether Wetland Indicators are Normally Present During a Portion of
the Growing Season. Examine the information resulting from STEP 3 and determine whether
wetland indicators are normally present during part of the growing season. If so, record on
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 71
DATA FORM 1 the indicators normally present and return to Section D or Section E and make
a wetland determination. If no information can be found that wetland indicators of all three
parameters are normally present during part of the growing season, the determination must be
made using procedures described in Section D or Section E.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 72
REFERENCES
Clark, J. R., and Benforado, J., eds. 1981. Wetlands of Bottomland Hardwood Forests.
Proceedings of a Workshop on Bottomland Hardwood Forest Wetlands of the Southeastern
United States. Elsevier Scientific Publishing Company, New York.
Correll, D. S., and Correll, H. B. 1972.. Aquatic and Wetland Plants of the Southwestern
United States. Publ. No. 16030 DNL 01/72, Environmental Protection Agency, Washington,
D.C.
Cowardin, L. M., Carter, V., Golet, F. C., and LaRoe, E. T. 1979. "Classification of Wetlands
and Deepwater Habitats of the United States," FWS/OBS79/31, US Fish and Wildlife Service,
Office of Biological Services, Washington, D.C.
Cronquist, A., Holmgren, A. H., Holmgren, N. H., and Reveal, J. L. 1972. Intermountain Flora
- Vascular Plants of the Intermountain West. USA. Vols I and II, Hafner Publishing Company,
New York.
Davis, R. J. 1952. Flora of Idaho. William C. Brown Company, Dubuque, Iowa.
Federal Register. 1980. "40 CFR Part 230: Section 404(b)(l) Guidelines for Specification of
Disposal Sites for Dredged or Fill Material," Vol 45, No. 249, pp 85352-85353, US
Government Printing Office, Washington, D.C.
Federal Register. 1982. "Title 33: Navigation and Navigable Waters; Chapter II, Regulatory
Programs of the Corps of Engineers," Vol 47, No. 138, p 31810, US Government Printing
Office, Washington, D.C.
Fernald, M. L. 1950. Gray's Manual of Botany. 8th ed., American Book Company, New York.
Gleason, H. A., and Cronquist, A. 1963. Manual of Vascular Plants of Northeastern United
States and Adjacent Canada. Van Nostrand, Princeton, N. J.
Godfrey, R. K., and Wooten, J. W. 1979. Aquatic and Wetland Plants of the Southeastern
United States. Vols I and II, University of Georgia Press, Athens, Ga.
Harrington, H. D. 1979. Manual of the Plants of Colorado. 2nd ed., Sage Books, Denver,
Colo.
Hitchcock, A. S. 1950. Manual of Grasses of the United States. US Department of Agriculture
Miscellaneous Publication No. 200, US Government Printing Office, Washington, D.C.
Hitchcock, C. L., and Cronquist, A. 1973. Flora of the Pacific Northwest. University of
Washington Press, Seattle, Wash.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 73
Kearney, T. H., and Peebles, R. H. 1960. Arizona Flora. 2nd ed., University of California
Press, Berkeley, Calif.
Long, R. W., and Lakela, 0. 1976. A Flora of Tropical Florida. Banyan Books, Miami, Fla.
Munsell Color. 1975. Munsell Soil Color Charts. Kollmorgen Corporation, Baltimore, Md.
Munz, P. A., and Keck, D. D. 1959. A California Flora. University of California Press,
Berkeley, Calif.
Radford, A. E., Ahles, H. E., and Bell, C. R. 1968. Manual of the Vascular Flora of the
Carolinas. The University of North Carolina Press, Chapel Hill, N. C.
Small, J. K. 1933. Manual of the Southeastern Flora. The University of North Carolina Press,
Chapel Hill, N. C.
Steyermark, J. A. 1963. Flora of Missouri. The Iowa State University Press, Ames, Iowa.
Theriot, R. F. In Review. "Flood Tolerance Indices of Plant Species of Southeastern
Bottomland Forests," Technical Report, US Army Engineer Waterways Experiment Station,
Vicksburg, Miss.
US Department of Agriculture - Soil Conservation Service. 1975. Soil Taxonomy. Agriculture
Handbook No. 436, US Government Printing Office, Washington, D.C.
US Department of Agriculture - Soil Conservation Service. 1983. "List of Soils with Actual
or High Potential for Hydric Conditions," National Bulletin No. 430-3-10, Washington, D.C.
US Department of Agriculture - Soil Conservation Service. 1985. "Hydric Soils of the United
States," USDA-SCS National Bulletin No. 430-5-91 Washington, D.C.
US Department of the Interior. 1970. National Atlas of the United States. US Geological
Survey, US Government Printing Office, Washington, D.C., pp 110-111.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 74
BIBLIOGRAPHY
Copeland, B. I, Hodson, R. G., and Riggs, S. R. 1984. "The Ecology of the Pamlico River,
North Carolina: An Estuarine Profile," FWS/OBS-82/06, US Fish and Wildlife Service,
Washington, D.C.
Foster, M. S., and Schiel, D. R. 1985. "The Ecology of Giant Kelp Forests in California: A
Community Profile," FWS/OB8-85(7.2), US Fish and Wildlife Service, Washington, D.C.
Gosselink, J. G. 1984. "The Ecology of Delta Marshes of Coastal Louisiana: A Community
Profile," FWS/OBS-84/09, US Fish and Wildlife Service, Washington, D.C.
Hobbie, J. E. 1984. "The Ecology of Tundra Ponds of the Arctic Coastal Plain: A Community
Profile," FWS/OBS-83/25, US Fish and Wildlife Service, Washington, D.C.
Huffman, R. T., and Tucker, G. E. 1984. "Preliminary Guide to the Onsite Identification and
Delineation of the Wetlands of Alaska," Technical Report Y-78-9, US Army Engineer
Waterways Experiment Station, Vicksburg, Miss.
Huffman, R. T., Tucker, G. E, Wooten, J. W., Kilmas, C. V., Freel, M. W., Forsythe, S. W.,
and Wilson, J. S. 1982. "Preliminary Guide to the Onsite Identification and Delineation of the
Wetlands of the South Atlantic United States," Technical Report Y-78-7, US Army Engineer
Waterways Experiment Station, Vicksburg, Miss.
. 1982. "Preliminary Guide to the Onsite Identification and Delineation of the
Wetlands of the North Atlantic United States," Technical Report Y-78-8, US Army Engineer
Waterways Experiment Station, Vicksburg, Miss.
Japp, W. C. 1984. "The Ecology of the South Florida Coral Reefs: A Community Profile,"
FWS/OBS-82/08, US Fish and Wildlife Service, Washington, D.C.
Josselyn, M. 1983. "The Ecology of San Francisco Bay Tidal Marshes: A Community
Profile," FWS/OBS-83/23, US Fish and Wildlife Service, Washington, D.C.
Livingston, R. J. 1984. "The Ecology of the Apalachicola Bay System: An Estuarine Profile,"
FWS/OBS-82/05, US Fish and Wildlife Service, Washington, D.C.
Nixon, S. W. 1982. "The Ecology of New England High Salt Marshes: A Community
Profile," FWS/OBS-81/55, US Fish and Wildlife Service, Washington, D.C.
Odum, W. E., Mclvor, C. C., and Smith, T. J., III. 1982. "The Ecology of the Mangroves of
South Florida: A Community Profile," FWS/OB S-81/24, US Fish and Wildlife Service,
Washington, D.C.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 75
Odum, W. E., Smith, T. I, III, Hoover, J. K., and Mclvor, C. C. 1984. "The Ecology of
Tidal Freshwater Marshes of the United States East Coast: A Community Profile," FWS/OBS-
83/17, US Fish and Wildlife Service, Washington, D.C.
Peterson, C. H., and Peterson, N. M. 1979. "The Ecology of Intertidal Flats of North Carolina:
A Community Profile," FWS/OBS-79/39, US Fish and Wildlife Service, Washington, D.C.
Phillips, R. C. 1984. "The Ecology of Eelgrass Meadows in the Pacific Northwest: A
Community Profile," FWS/OBS-84/24, US Fish and Wildlife Service, Washington, D.C.
Seliskar, D. M., and Gallagher, J. L. 1983. "The Ecology of Tidal Marshes of the Pacific
Northwest Coast: A Community Profile," FWS/OBS-82/32, US Fish and Wildlife Service,
Washington, D.C.
Sharitz, R. R., and Gibbons, J. W. 1982. "The Ecology of Southeastern Shrub Bogs (Pocosins)
and Carolina Bays: A Community Profile," FWS/OBS-82/04, US Fish and Wildlife Service,
Washington, D.C.
Thayer, G. W., Kenworthy, W. J., and Fonseca, M. S. 1984. "The Ecology of Eelgrass
Meadows of the Atlantic Coast: A Community Profile," FWS/OBS-84/02, US Fish and
Wildlife Service, Washington, D.C.
US Army Corps of Engineers. 1978. "Preliminary Guide to Wetlands of Peninsular Florida,"
Technical Report Y-78-2, US Army Engineer Waterways Experiment Station, Vicksburg,
Miss.
. 1978. "Preliminary Guide to Wetlands of Puerto Rico," Technical Report Y-78-3,
US Army Engineer Waterways Experiment Station, Vicksburg, Miss.
. 1978. "Preliminary Guide to Wetlands of the West Coast States," Technical Report
Y-78-4, US Army Engineer Waterways Experiment Station, Vicksburg, Miss.
. 1978. "Preliminary Guide to Wetlands of the Gulf Coastal Plain," Technical
Report Y-78-5, US Army Engineer Waterways Experiment Station, Vicksburg, Miss.
. 1982. "Preliminary Guide to the Onsite Identification and Delineation of the
Wetlands of the Interior United States," Technical Report Y-78-6, US Army Engineer
Waterways Experiment Station, Vicksburg, Miss.
Witlatch, R. B. 1982. "The Ecology of New England Tidal Flats: A Community Profile,"
FWS/OBS-81/01, US Fish and Wildlife Service, Washington, D.C.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 76
Wharton, C. H., Kitchens, W. M., and Sipe, T. W. 1982. "The Ecology of Bottomland
Hardwood Swamps of the Southeast: A Community Profile," FWS/OBS-81/37, US Fish and
Wildlife Service, Washington, D.C.
Zedler, J. B. 1984. "The Ecology of Southern California Coastal Salt Marshes: A Community
Profile," FWS/OBS-81/54, US Fish and Wildlife Service, Washington, D.C.
Zieman, J. C. 1982. "The Ecology of the Seagrasses of South Florida: A Community Profile,"
FWS/OBS-82/25, US Fish and Wildlife Service, Washington, D.C.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 77
APPENDIX A: GLOSSARY
Active water table - A condition in which the zone of soil saturation fluctuates, resulting in
periodic anaerobic soil conditions. Soils with an active water table often contain bright mottles
and matrix chromas of 2 or less.
Adaptation - A modification of a species that makes it more fit for existence under the
conditions of its environment. These modifications are the result of genetic selection
processes.
Adventitious roots Roots found on plant stems in positions where they normally do not occur.
Aerenchymous tissue A type of plant tissue in which cells are unusually large and arranged in
a manner that results in air spaces in the plant organ. Such tissues are often referred to as
spongy and usually provide increased buoyancy.
Aerobic - A situation in which molecular oxygen is a part of the environment.
Anaerobic - A situation in which molecular oxygen is absent (or effectively so) from the
environment.
Aquatic roots - Roots that develop on stems above the normal position occupied by roots in
response to prolonged inundation.
Aguic moisture regime - A mostly reducing soil moisture regime nearly free of dissolved
oxygen due to saturation by ground water or its capillary fringe and occurring at periods when
the soil temperature at 19.7 in. is greater than 50 C.
Arched roots - Roots produced on plant stems in a position above the normal position of roots,
which serve to brace the plant during and following periods of prolonged inundation.
Areal cover - A measure of dominance that defines the degree to which aboveground portions
of plants (not limited to those rooted in a sample plot) cover the ground surface. It is possible
for the total areal cover in a community to exceed 100 percent because (a) most plant
communities consist of two or more vegetative strata; (b) areal cover is estimated by vegetative
layer; and (c) foliage within a single layer may overlap.
Atypical situation - As used herein, this term refers to areas in which one or more parameters
(vegetation, soil, and/or hydrology) have been sufficiently altered by recent human activities or
natural events to preclude the presence of wetland indicators of the parameter.
Backwater flooding - Situations in which the source of inundation is overbank flooding from a
nearby stream.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 78
Basal area - The cross-sectional area of a tree trunk measured in square inches, square
centimetres, etc. Basal area is normally measured at 4.5 ft above the ground level and is used
as a measure of dominance. The most easily used tool for measuring basal area is a tape
marked in square inches. When plotless methods are used, an angle gauge or prism will
provide a means for rapidly determining basal area. This term is also applicable to the
crosssectional area of a clumped herbaceous plant, measured at 1.0 in. above the soil surface.
Bench mark - A fixed, more or less permanent reference point or object, the elevation of which
is known. The US Geological Survey (USGS) installs brass caps in bridge abutments or
otherwise permanently sets bench marks at convenient locations nationwide. The elevations
on these marks are referenced to the National Geodetic Vertical Datum (NGVD), also
commonly known as mean sea level (MSL). Locations of these bench marks on USGS
quadrangle maps are shown as small triangles. However, the marks are sometimes destroyed
by construction or vandalism. The existence of any bench mark should be field verified before
planning work that relies on a particular reference point. The USGS and/or local state
surveyor's office can provide information on the existence, exact location, and exact elevation
of bench marks.
Biennial - An event that occurs at 2-year intervals.
Buried soil - A once-exposed soil now covered by an alluvial, loessal, or other deposit
(including man-made).
Canopy layer - The uppermost layer of vegetation in a plant community. In forested areas,
mature trees comprise the canopy layer, while the tallest herbaceous species constitute the
canopy layer in a marsh.
Capillary fringe - A zone immediately above the water table (zero gauge pressure) in which
water is drawn upward from the water table by capillary action.
Chemical reduction - Any process by which one compound or ion acts as an electron donor.
In such cases, the valence state of the electron donor is decreased.
Chroma - The relative purity or saturation of a color; intensity of distinctive hue as related to
grayness; one of the three variables of color.
Comprehensive wetland determination - A type of wetland determination that is based on the
strongest possible evidence, requiring the collection of quantitative data.
Concretion - A local concentration of chemical compounds (e.g. calcium carbonate, iron
oxide) in the form of a grain or nodule of varving size, shape, hardness, and color.
Concretions of significance in hydric soils are usually iron and/or manganese oxides occurring
at or near the soil surface, which develop under conditions of prolonged soil saturation.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 79
Contour - An imaginary line of constant elevation on the ground surface. The corresponding
line on a map is called a "contour line."
Criteria - Standards, rules, or tests on which a judgment or decision may be based.
Deepwater aquatic habitat - Any open water area that has a mean annual water depth >6.6 ft,
lacks soil, and/or is either unvegetated or supports only floating or submersed macrophytes.
Density - The number of individuals of a species per unit area.
Detritus - Minute fragments of plant parts found on the soil surface. When fused together by
algae or soil particles, this is an indicator that surface water was recently present.
Diameter at breast height (DBH) - The width of a plant stem as measured at 4.5 ft above the
ground surface.
Dike - A bank (usually earthen) constructed to control or confine water.
Dominance - As used herein, a descriptor of vegetation that is related to the standing crop of a
species in an area, usually measured by height, areal cover, or basal area (for trees).
Dominant species - As used herein, a plant species that exerts a controlling influence on or
defines the character of a community.
Drained - A condition in which ground or surface water has been reduced or eliminated from
an area by artificial means.
Drift line An accumulation of debris along a contour (parallel to the water flow) that represents
the height of an inundation event.
Duration (inundation/soil saturation) - The length of time during which water stands at or
above the soil surface (inundation), or during which the soil is saturated. As used herein,
duration refers to a period during the growing season.
Ecological tolerance - The range of environmental conditions in which a plant species can
grow.
Emergent plant - A rooted herbaceous plant species that has parts extending above a water
surface.
Field capacity - The percentage of water remaining in a soil after it has been saturated and after
free drainage is negligible.
Fill material - Any material placed in an area to increase surface elevation.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 80
Flooded - A condition in which the soil surface is temporarily covered with flowing water from
any source, such as streams overflowing their banks, runoff from adjacent or surrounding
slopes, inflow from high tides, or any combination of sources.
Flora - A list of all plant species that occur in an area.
Frequency (inundation or soil saturation) - The periodicity of coverage of an area by surface
water or soil saturation. It is usually expressed as the number of years (e.g. 50 years) the soil is
inundated or saturated at least once each year during part of the growing season per 100 years
or as a 1-, 2-, 5-year, etc., inundation frequency.
Frequency (vegetation) - The distribution of individuals of a species in an area. It is
quantitatively expressed as
Number of samples containing species A X 100
Total number of samples
More than one species may have a frequency of 100 percent within the same area.
Frequently flooded - A flooding class in which flooding is likely to occur often under normal
weather conditions (more than 50-percent chance of flooding in any year or more than 50
times in 100 years).
Gleyed - A soil condition resulting from prolonged soil saturation, which is manifested by the
presence of bluish or greenish colors through the soil mass or in mottles (spots or streaks)
among other colors. Gleying occurs under reducing soil conditions resulting from soil
saturation, by which iron is reduced predominantly to the ferrous state.
Ground water - That portion of the water below the ground surface that is under greater
pressure than atmospheric pressure.
Growing season - The portion of the year when soil temperatures at 19.7 inches below the soil
surface are higher than biologic zero (5' C) (US Department of Agriculture - Soil Conservation
Service 1985). For ease of determination this period can be approximated by the number of
frost-free days (US Department of the interior 1970).
Habitat - The environment occupied by individuals of a particular species, population, or
community.
Headwater flooding - A situation in which an area becomes inundated directly by surface
runoff from upland areas.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 81
Herb - A nonwoody individual of a macrophytic species. In this manual, seedlings of woody
plants (including vines) that are less than 3.2 ft in height are considered to be herbs.
Herbaceous layer - Any vegetative stratum of a plant community that is composed
predominantly of herbs.
Histic epipedon - An 8- to 16-in. soil layer at or near the surface that is saturated for 30
consecutive days or more during the growing season in most years and contains a minimum of
20 percent organic matter when no clay is present or a minimum of 30 percent organic matter
when 60 percent or greater clay is present.
Histosols - An order in soil taxonomy composed of organic soils that have organic soil
materials in more than half of the upper 80 cm or that are of any thickness if directly overlying
bedrock.
Homogeneous vegetation - A situation in which the same plant species association occurs
throughout an area.
Hue - A characteristic of color that denotes a color in relation to red, yellow, blue, etc; one of
the three variables of color. Each color chart in the Munsell Color Book (Munsell Color 1975)
consists of a specific hue.
Hydric soil - A soil that is saturated, flooded, or ponded long enough during the growing
season to develop anaerobic conditions that favor the growth and regeneration of hydrophytic
vegetation (US Department of Agriculture-Soil Conservation Service 1985). Hydric soils that
occur in areas having positive indicators of hydrophytic vegetation and wetland hydrology are
wetland soils.
Hydric soil condition - A situation in which characteristics exist that are associated with soil
development under reducing conditions.
Hydrologic regime - The sum total of water that occurs in an area on average during a given
period.
Hydrologic zone - An area that is inundated or has saturated soils within a specified range of
frequency and duration of inundation and soil saturation.
Hydrology - The science dealing with the properties, distribution, and circulation of water.
Hydrophyte - Any macrophyte that grows in water or on a substrate that is at least periodically
deficient in oxygen as a result of excessive water content; plants typically found in wet
habitats.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 82
Hydrophytic vegetation - The sum total of macrophytic plant life growing in water or on a
substrate that is at least periodically deficient in oxygen as a result of excessive water content.
When hydrophytic vegetation comprises a community where indicators of hydric soils and
wetland hydrology also occur, the area has wetland vegetation.
Hypertrophied lenticels - An exaggerated (oversized) pore on the surface of stems of woody
plants through which gases are exchanged between the plant and the atmosphere. The
enlarged lenticels serve as a mechanism for increasing oxygen to plant roots during periods of
inundation and/or saturated soils.
Importance value - A quantitative term describing the relative influence of a plant species in a
plant community, obtained by summing any combination of relative frequency, relative
density, and relative dominance.
Indicator - As used in this manual, an event, entity, or condition that typically characterizes a
prescribed environment or situation; indicators determine or aid in determining whether or not
certain stated circumstances exist.
Indicator status - One of the categories (e.g. OBL) that describes the estimated probability of a
plant species occurring in wetlands.
Intercellular air space - A cavity between cells in plant tissues, resulting from variations in cell
shape and configuration. Aerenchymous tissue (a morphological adaptation found in many
hydrophytes) often has large intercellular air spaces.
Inundation - A condition in which water from any source temporarily or permanently covers a
land surface.
Levee - A natural or man-made feature of the landscape that restricts movement of water into
or through an area.
Liana - As used in this manual, a layer of vegetation in forested plant communities that
consists of woody vines. The term may also be applied to a given species.
Limit of biological activity - With reference to soils, the zone below which conditions preclude
normal growth of soil organisms. This term often is used to refer to the temperature (5' C) in a
soil below which metabolic processes of soil microorganisms, plant roots, and animals are
negligible.
Long duration (flooding) - A flooding class in which the period of inundation for a single
event ranges from 7 days to 1 month.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 83
Macrophyte Any plant species that can be readily observed without the aid of optical
magnification. This includes all vascular plant species and mosses (e.g., Sphagnum spp.), as
well as large algae (e.g. Chora spp., kelp).
Macrophytic - A term referring to a plant species that is a macrophyte.
Major portion of the root zone. The portion of the soil profile in which more than 50 percent
of plant roots occur. In wetlands, this usually constitutes the upper 12 in. of the profile.
Man-induced wetland - Any area that develops wetland characteristics due to some activity
(e.g., irrigation) of man.
Mapping unit - As used in this manual, some common characteristic of soil, vegetation, and/or
hydrology that can be shown at the scale of mapping for the defined purpose and objectives of
a survey.
Mean sea level - A datum, or "plane of zero elevation," established by averaging all stages of
oceanic tides over a 19-year tidal cycle or "epoch." This plane is corrected for curvature of the
earth and is the standard reference for elevations on the earth's surface. The correct term for
mean sea level is the National Geodetic Vertical Datum (NGVD).
Mesophytic - Any plant species growing where soil moisture and aeration conditions lie
between extremes. These species are typically found in habitats with average moisture
conditions, neither very dry nor very wet.
Metabolic processes - The complex of internal chemical reactions associated with life-
sustaining functions of an organism.
Method - A particular procedure or set of procedures to be followed.
Mineral soil - A soil consisting predominantly of, and having its properties determined
predominantly by, mineral matter usually containing less than 20-percent organic matter.
Morphological adaptation - A feature of structure and form that aids in fitting a species to its
particular environment (e.g. buttressed base, adventitious roots, aerenchymous tissue).
Mottles - Spots or blotches of different color or shades of color interspersed within the
dominant color in a soil layer, usually resulting from the presence of periodic reducing soil
conditions.
Muck - Highly decomposed organic material in which the original plant parts are not
recognizable.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 84
Multitrunk - A situation in which a single individual of a woody plant species has several
stems.
Nonhydric soil - A soil that has developed under predominantly aerobic soil conditions. These
soils normally support mesophytic or xerophytic species.
Nonwetland - Any area that has sufficiently dry conditions that indicators of hydrophytic
vegetation, hydric soils, and/or wetland hydrology are lacking. As used in this manual, any
area that is neither a wetland, a deepwater aquatic habitat, nor other special aquatic site.
Organic pan - A layer usually occurring at 12 to 30 inches below the soil surface in coarse-
textured soils, in which organic matter and aluminum (with or without iron) accumulate at the
point where the top of the water table most often occurs. Cementing of the organic matter
slightly reduces permeability of this layer.
Organic soil - A soil is classified as an organic soil when it is: (1) saturated for prolonged
periods (unless artificially drained) and has more than 30-percent organic matter if the mineral
fraction is more than 50-percent clay, or more than 20-percent organic matter if the mineral
fraction has no clay; or (2) never saturated with water for more than a few days and having
more than 34-percent organic matter.
Overbank flooding - Any situation in which inundation occurs as a result of the water level of
a stream rising above bank level.
Oxidation-reduction process - A complex of biochemical reactions in soil that influences the
valence state of component elements and their ions. Prolonged soil saturation during the
growing season elicits anaerobic conditions that shift the overall process to a reducing
condition.
Oxygen pathway - The sequence of cells, intercellular spaces, tissues, and organs, through
which molecular oxygen is transported in plants. Plant species having pathways for oxygen
transport to the root system are often adapted for life in saturated soils.
Parameter - A characteristic component of a unit that can be defined. Vegetation, soil, and
hydrology are three parameters that may be used to define wetlands.
Parent material - The unconsolidated and more or less weathered mineral or organic matter
from which a soil profile develops.
Ped - A unit of soil structure (e.g. aggregate, crumb, prism, block, or granule) formed by
natural processes.
Peraquic moisture regime - A soil condition in which a reducing environment always occurs
due to the presence of ground water at or near the soil surface.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 85
Periodically - Used herein to define detectable regular or irregular saturated soil conditions or
inundation, resulting from ponding of ground water, precipitation, overland flow, stream
flooding, or tidal influences that occur(s) with hours, days, weeks, months, or even years
between events.
Permeability- A soil characteristic that enables water or air to move through the profile,
measured as the number of inches per hour that water moves downward through the saturated
soil. The rate at which water moves through the least permeable layer governs soil
permeability.
Physiognomy - A term used to describe a plant community based on the growth habit (e.g.,
trees, herbs, lianas) of the dominant species.
Physiological adaptation - A feature of the basic physical and chemical activities that occurs in
cells and tissues of a species, which results in it being better fitted to its environment (e.g.
ability to absorb nutrients under low oxygen tensions).
Plant community - All of the plant populations occurring in a shared habitat or environment.
Plant cover - See areal cover.
Pneumatophore - Modified roots that may function as a respiratory organ in species subjected
to frequent inundation or soil saturation (e.g., cypress knees).
Ponded - A condition in which water stands in a closed depression. Water may be removed
only by percolation, evaporation, and/or transpiration.
Poorly drained - Soils that commonly are wet at or near the surface during a sufficient part of
the year that field crops cannot be grown under natural conditions. Poorly drained conditions
are caused by a saturated zone, a layer with low hydraulic conductivity, seepage, or a
combination of these conditions.
Population - A group of individuals of the same species that occurs in a given area.
Positive wetland indicator - Any evidence of the presence of hydrophytic vegetation, hydric
soil, and/or wetland hydrology in an area.
Prevalent vegetation - The plant community or communities that occur in an area during a
given period. The prevalent vegetation is characterized by the dominant macrophytic species
that comprise the plant community.
Quantitative - A precise measurement or determination expressed numerically.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 86
Range - As used herein, the geographical area in which a plant species is known to occur.
Redox potential - A measure of the tendency of a system to donate or accept electrons, which
is governed by the nature and proportions of the oxidizing and reducing substances contained
in the system.
Reducing environment - An environment conducive to the removal of oxygen and chemical
reduction of ions in the soils.
Relative density - A quantitative descriptor, expressed as a percent, of the relative number of
individuals of a species in an area; it is calculated by
Number of individuals of species A X 100
Total number of individuals of all species
Relative dominance - A quantitative descriptor, expressed as a percent, of the relative size or
cover of individuals of a species in an area; it is calculated by
Amount* of species A X 100
Total amount of all species
*The amount of a species may be based on percent areal cover, basal area, or height.
Relative frequency - A quantitative descriptor, expressed as a percent, of the relative
distribution of individuals of a species in an area; it is calculated by
Frequency of species A X 100
Total frequency of all species
Relief - The change in elevation of a land surface between two points; collectively, the
configuration of the earth's surface, including such features as hills and valleys.
Reproductive adaptation - A feature of the reproductive mechanism of a species that results in
it being better fitted to its environment (e.g. ability for seed germination under water).
Respiration - The sum total of metabolic processes associated with conversion of stored
(chemical) energy into kinetic (physical) energy for use by an organism.
Rhizosphere - The zone of soil in which interactions between living plant roots and
microorganisms occur.
Root zone - The portion of a soil profile in which plant roots occur.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 87
Routine wetland determination - A type of wetland determination in which office dat and/or
relatively simple, rapidly applied onsite methods are employed to determine whether or not an
area is a wetland. Most wetland determinations are of this type, which usually does not require
collection of quantitative data.
Sample plot - An area of land used for measuring or observing existing conditions.
Sapling/shrub - A layer of vegetation composed of woody plants <3.0 in. in diameter at breast
height but greater than 3.2 ft in height, exclusive of woody vines.
Saturated soil conditions - A condition in which all easily drained voids (pores) between soil
particles in the root zone are temporarily or permanently filled with water to the soil surface at
pressures greater than atmospheric.
Soil - Unconsolidated mineral and organic material that supports, or is capable of supporting,
plants, and which has recognizable properties due to the integrated effect of climate and living
matter acting upon parent material, as conditioned by relief over time.
Soil horizon - A layer of soil or soil material approximately parallel to the land surface and
differing from adjacent genetically related layers in physical, chemical, and biological
properties or characteristics (e.g. color, structure, texture, etc.).
Soil matrix - The portion of a given soil having the dominant color. In most cases, the matrix
will be the portion of the soil having more than 50 percent of the same color.
Soil permeability - The ease with which gases, liquids, or plant roots penetrate or pass through
a layer of soil.
Soil phase - A subdivision of a soil series having features (e.g. slope, surface texture, and
stoniness) that affect the use and management of the soil, but which do not vary sufficiently to
differentiate it as a separate series. These are usually the basic mapping units on detailed soil
maps produced by the Soil Conservation Service.
Soil pore - An area within soil occupied by either air or water, resulting from the arrangement
of individual soil particles or peds.
Soil profile - A vertical section of a soil through all its horizons and extending into the parent
material.
Soil series - A group of soils having horizons similar in differentiating characteristics and
arrangement in the soil profile, except for texture of the surface horizon.
Soil structure - The combination or arrangement of primary soil particles into secondary
particles, units, or peds.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 88
Soil surface - The upper limits of the soil profile. For mineral soils, this is the upper limit of
the highest (Al) mineral horizon. For organic soils, it is the upper limit of undecomposed.
dead organic matter.
Soil texture - The relative proportions of the various sizes of particles in a soil.
Somewhat poorly drained - Soils that are wet near enough to the surface or long enough that
planting or harvesting operations or crop growth is markedly restricted unless artificial
drainage is provided. Somewhat poorly drained soils commonly have a layer with low
hydraulic conductivity, wet conditions high in the profile, additions of water through seepage,
or a combination of these conditions.
Stilted roots - Aerial roots arising from stems (e.g., trunk and branches), presumably providing
plant support (e.g., Rhizophora mangle).
Stooling - A form of asexual reproduction in which new shoots are produced at the base of
senescing stems, often resulting in a multitrunk growth habit.
Stratigraphy - Features of geology dealing with the origin, composition, distribution, and
succession of geologic strata (layers).
Substrate - The base or substance on which an attached species is growing.
Surface water - Water present above the substrate or soil surface.
Tidal - A situation in which the water level periodically fluctuates due to the action of lunar
and solar forces upon the rotating earth.
Topography - The configuration of a surface, including its relief and the position of its natural
and man-made features.
Transect - As used herein, a line on the ground along which observations are made at some
interval.
Transition zone - The area in which a change from wetlands to nonwetlands occurs. The
transition zone may be narrow or broad.
Transpiration - The process in plants by which water vapor is released into the gaseous
environment, primarily through stomata.
Tree - A woody plant >3.0 in. in diameter at breast height, regardless of height (exclusive of
woody vines).
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 89
Typical - That which normally, usually, or commonly occurs.
Typically adapted - A term that refers to a species being normally or commonly suited to a
given set of environmental conditions, due to some feature of its morphology, physiology, or
reproduction.
Unconsolidated parent material - Material from which a soil develops, usually formed by
weathering of rock or placement in an area by natural forces (e.g. water, wind, or gravity).
Under normal circumstances - As used in the definition of wetlands, this term refers to
situations in which the vegetation has not been substantially altered by man's activities.
Uniform vegetation - As used herein, a situation in which the same group of dominant species
generally occurs throughout a given area.
Upland - As used herein, any area that does not qualify as a wetland because the associated
hydrologic regime is not sufficiently wet to elicit development of vegetation, soils, and/or
hydrologic characteristics associated with wetlands. Such areas occurring within floodplains
are more appropriately termed nonwetlands.
Value (soil color) - The relative lightness or intensity of color, approximately a function of the
square root of the total amount of light reflected from a surface; one of the three variables of
color.
Vegetation - The sum total of macrophytes that occupy a given area.
Vegetation layer - A subunit of a plant community in which all component species exhibit the
same growth form (e.g., trees, saplings/shrubs, herbs).
Very long duration (flooding) - A duration class in which the length of a single inundation
event is greater than I month.
Very poorly drained - Soils that are wet to the surface most of the time. These soils are wet
enough to prevent the growth of important crops (except rice) unless artificially drained.
Watermark - A line on a tree or other upright structure that represents the maximum static
water level reached during an inundation event.
Water table - The upper surface of ground water or that level below which the soil is saturated
with water. It is at least 6 in. thick and persists in the soil for more than a few weeks.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 90
Wetlands - Those areas that are inundated or saturated by surface or ground water at a
frequency and duration sufficient to support, and that under normal circumstances do support,
a prevalence of vegetation typically adapted for life in saturated soil conditions. Wetlands
generally include swamps, marshes, bogs, and similar areas.
Wetland boundary - The point on the ground at which a shift from wetlands to nonwetlands or
aquatic habitats occurs. These boundaries usually follow contours.
Wetland determination - The process or procedure by which an area is ad udged a wetland or
nonwetland.
Wetland hydrology - The sum total of wetness characteristics in areas that are inundated or
have saturated soils for a sufficient duration to support hydrophytic vegetation.
Wetland plant association - Any grouping of plant species that recurs wherever certain wetland
conditions occur.
Wetland soil - A soil that has characteristics developed in a reducing atmosphere, which exists
when periods of prolonged soil saturation result in anaerobic conditions. Hydric soils that are
sufficiently wet to support hydrophytic vegetation are wetland soils.
Wetland vegetation - The sum total of macrophytic plant life that occurs in areas where the
frequency and duration of inundation or soil saturation produce permanently or periodically
saturated soils of sufficient duration to exert a controlling influence on the plant species
present. As used herein, hydrophytic vegetation occurring in areas that also have hydric soils
and wetland hydrology may be properly referred to as wetland vegetation.
Woody vine - See liana.
Xerophytic - A plant species that is typically adapted for life in conditions -where a lack of
water is a limiting factor for growth and/or reproduction. These species are capable of growth
in extremely dry conditions as a result of morphological, physiological, and/or reproductive
adaptations.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 91
APPENDIX C: VEGETATION
1. This appendix contains three sections. Section 1 is a subset of the regional list of plants
that occur in wetlands, but includes only those species having an indicator status of OBL,
F ACW, or FAC. Section 2 is a list of plants that commonly occur in wetlands of a given
region. Since many geographic areas of Section 404 responsibility include portions of two or
more plant list regions, users will often need more than one regional list; thus, Sections 1 and 2
will be published separately from the remainder of the manual. Users will be furnished all
appropriate regional lists.
2. Section 3, which is presented herein, describes morphological, physiological, and
reproductive adaptations that can be observed or are known to occur in plant species that are
typically adapted for life in anaerobic soil conditions.
Section 3 - Morphological, Physiological, and Reproductive
Adaptations of Plant Species for Occurrence in Areas
Having Anaerobic Soil Conditions
Morphological adaptations
3. Many plant species have morphological adaptations for occurrence in wetlands. These
structural modifications most often provide the plant with increased buoyancy or support. In
some cases (e.g. adventitious roots), the adaptation may facilitate the uptake of nutrients and/or
gases (particularly oxygen). However., not all species occurring in areas having anaerobic soil
conditions exhibit morphological adaptations for such conditions. The following is a list of
morphological adaptations that a species occurring in areas having anaerobic soil conditions
may possess (a partial list of species with such adaptations is presented in Table Cl):
a. Buttressed tree trunks. Tree species (e.g. Taxodium distichum) may develop enlarged
trunks (Figure Cl) in response to frequent inundation. This adaptation is a strong indicator of
hydrophytic vegetation in nontropical forested areas.
b. Pneumatophores. These modified roots may serve as respiratory organs in species
subjected to frequent inundation or soil saturation. Cypress knees (Figure C2) are a classic
example, but other species (e.g., Nyssa aquatics, Rhizophora mangze) may also develop
pneumatophores.
c. Adventitious roots. Sometimes referred to as "water roots," Adventitious roots occur
on plant stems in positions where roots normally are not found. Small fibrous roots protruding
from the base of trees (e.g. Salix nigra) or roots on stems of herbaceous plants and tree
seedlings in positions immediately above the soil surface (e.g. Ludwigia spp.) occur in
response to inundation or soil saturation (Figure C3). These usually develop during periods of
sufficiently prolonged soil saturation to destroy most of the root system. CAUTION: Not all
adventitious roots develop as a result of inundation or soil saturation. For example, aerial
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 92
roots on woofty vines are not normally produced as a response to inundation or soil
saturation.
d. Shallow root systems. When soils are inundated or saturated for long periods during
the growing season, anaerobic conditions develop in the zone of root growth. Most species
with deep root systems cannot survive in such conditions. Most species capable of growth
during periods when soils are oxygenated only near the surface have shallow root systems. In
forested wetlands, windthrown trees (Figure C4) are often indicative of shallow root systems.
e. Inflated leaves, stems, or roots. Many hydrophytic species, particularly herbs (e.g.
Limnobium spongia, Ludwigia spp.), have or develop spongy (aerenchymous) tissues in
leaves, stems, and/or roots that provide buoyancy or support and serve as a reservoir or
passageway for oxygen needed for metabolic processes. An example of inflated leaves is
shown in Figure C5.
f. Polymorphic leaves. Some herbaceous species produce different types of leaves,
depending on the water level at the time of leaf formation. For example, Alisma spp. produce
strap-shaped leaves when totally submerged, but produce broader, floating leaves when plants
are emergent. CA UTION: Many upland species also produce polymorphic leaves.
g. Floating leaves. Some species (e.g. Nymphaea spp.) produce leaves that are uniquely
adapted for floating on a water surface (Figure C6). These leaves have stomata primarily on
the upper surface and a thick waxy cuticle that restricts water penetration. The presence of
species with floating leaves is strongly indicative of hydrophytic vegetation.
h. Floating stems. A number of species (e.g., Alternantheraphiloxeroides) produce
matted stems that have large internal air spaces when occurring in inundated areas. Such
species root in shallow water and grow across the water surface into deeper areas. Species
with floating stems often produce adventitious roots at leaf nodes.
i. Hypertrophied lenticels. Some plant species (e.g. GZeditsia aquatica) produce enlarged
lenticels on the stem in response to prolonged inundation or soil saturation. These are thought
to increase oxygen uptake through the stem during such periods.
j.. Multitrunks or stooling. Some woody hydrophytes characteristical.ly produce several
trunks of different ages (Figure C7) or produce new stems arising from the base of a senescing
individual (e.g. Forestiera acuminata, Nyssa ogechee) in response to inundation.
k. Oxygen pathway to roots. Some species (e.g. Spartina alterniflora) have a specialized
cellular arrangement that facilitates diffusion of gaseous oxygen from leaves and stems to the
root system.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 93
Physiological adaptations
4. Most, if not all, hydrophytic species are thought to possess physiological adaptations
for occurrence in areas that have prolonged periods of anaerobic soil conditions. However,
relatively few species have actually been proven to possess such adaptations, primarily due to
the limited research that has been conducted. Nevertheless, several types of physiological
adaptations known to occur in hydrophytic species are discussed below, and a list of species
having one or more of these adaptations is presented in Table C2. NOTE: Since it is
impossible to detect these adaptations in the field, use of this indicator will be limited to
observing the species in the field and checking the list in Table C2 to determine whether the
species is known to have a physiological adaptation for occurrence in areas having anaerobic
soil conditions):
a. Accumulation of malate. Malate, a nontoxic metabolite, accumuYat-es in roots of
many hydrophytic species (e.g. Glyceria maxima, Nyssa sylvatica var. biflord). Nonwetland
species concentrate ethanol, a toxic by-product of anaerobic respiration, when growing in
anaerobic soil conditions. Under such conditions, many hydrophytic species produce high
concentrations of malate and unchanged concentrations of ethanol, thereby avoiding
accumulation of toxic materials. Thus, species having the ability to concentrate malate instead
of ethanol in the root system under anaerobic soil conditions are adapted for life in such
conditions, while species that concentrate ethanol are poorly adapted for life in anaerobic soil
conditions.
b. Increased levels of nitrate reductase. Nitrate reductase is an enzyme involved in
conversion of nitrate nitrogen to nitrite nitrogen, an intermediate step in ammonium
production. Ammonium ions can accept electrons as a replacement for gaseous oxygen in
some species, thereby allowing continued functioning of metabolic processes under low soil
oxygen conditions. Species that produce high levels of nitrate reductase (e.g. Larix Zaricina)
are adapted for life in anaerobic soil conditions.
c. Slight increases in metabolic rates. Anaerobic soil conditions effect short-term
increases in metabolic rates in most species. However, the rate of metabolism often increases
only slightly in wetland species, while metabolic rates increase significantly in nonwetland
species. Species exhibiting only slight increases in metabolic rates (e.g. Larix laricina,
Senecio vulgaris) are adapted for life in anaerobic soil conditions.
d. Rhizosphere oxidation. Some hydrophytic species (e.g. Nyssa aquatica, Myrica gale)
are capable of transferring gaseous oxygen from the root system into soil pores immediately
surrounding the roots. This adaptation prevents root deterioration and maintains the rates of
water and nutrient absorption under anaerobic soil conditions.
e. Ability for root growth in low oxygen tensions. Some species (e.g. Typha angustifotia,
Juncus effusus) have the ability to maintain root growth under soil oxygen concentrations as
low as 0.5 percent. Although prolonged (>1 year) exposure to soil oxygen concentrations
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 94
lower than 0.5 percent generally results in the death of most individuals, this adaptation
enables some species to survive extended periods of anaerobic soil conditions.
f. Absence of alcohol dehydrogenase (ADH) activity. ADH is an enzyme associated with
increased ethanol production. When the enzyme is not functioning, ethanol production does
not increase significantly. Some hydrophytic species (e.g. Potentilla anserina, Polygonum
amphibium) show only slight increases in ADH activity under anaerobic soil conditions.
Therefore, ethanol production occurs at a slower rate in species that have low concentrations of
ADH.
Reproductive adaptations
5. Some plant species have reproductive features that enable them to become established
and grow in saturated soil conditions. The following have been identified in the technical
literature as reproductive adaptations that occur in hydrophytic species:
a. Prolonged seed viability. Some plant species produce seeds that may remain viable for
20 years or more. Exposure of these seeds to atmospheric oxygen usually triggers
germination. Thus, species (e.g., Taxodium distichum) that grow in very wet areas may
produce seeds that germinate only during infrequent periods when the soil is dewatered.
NOTE: Many upland species also have prolonged seed viability, but the trigger mechanism for
germination is not exposure to atmospheric oxygen.
b. Seed germination under low oxygen concentrations. Seeds of some hydrophytic
species germinate when submerged. This enables germination during periods of early-spring
inundation, which may provide resulting seedlings a competitive advantage over species
whose seeds germinate only when exposed to atmospheric oxygen.
c. Flood-tolerant seedlings. Seedlings of some hydrophytic species (e.g. Fraxinus
pennsylvanica) can survive moderate periods of total or partial inundation. Seedlings of these
species have a competitive advantage over seedlings of flood-intolerant species.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
95
Tabled
Partial List of Species With Known Morphological Adaptations for
Occurrence in Wetlands*
Species
Acer negundo
Acer rubrum
Acer saccharinum
Alisma spp.
Alternanthera philoxeroides
Avicennia nitida
Brasenia schreberi
Cladium mariscoides
Cyperus spp. (most species)
Eleocharis spp. (most species)
Forestiera acuminata
F-raxinus pennsylvanica
Gleditsia aquatics
Juncus spp.
Limnobium spongia
Ludwigia spp.
Menyanthes trifoliata
Myrica gale
Nelumbo spp.
Nuphar spp.
Nymphaea spp.
Nyssa aquatics
Nyssa ogechee
Common Name
Box elder
Red maple
Silver maple
Water plantain
Alligatorweed
Black mangrove
Watershield
Twig rush
Flat sedge
Spikerush
Swamp privet
Green ash
Water locust
Rush
Frogbit
Waterprimrose
Buckbean
Sweetgale
Lotus
Cowlily
Waterlily
Water tup el o
Ogechee tupelo
Adaptation
Adventitious roots
Hypertrophied lenticels
Hypertrophied lenticels:
adventitious roots
(juvenile plants)
Polymorphic leaves
Adventitious roots; inflated,
floating stems
Pneumatophores; hypertrophied
lenticels
Inflated, floating leaves
Inflated stems
Inflated stems and leaves
Inflated stems and leaves
Multi-trunk, stooling
Buttressed trunks; adventi-
tious roots
Hypertrophied lenticels
Inflated stems and leaves
Inflated, floating leaves
Adventitious roots; inflated
floating stems
Inflated stems (rhizome)
Hypertrophied lenticels
Floating leaves
Floating leaves
Floating leaves
Buttressed trunks; pneuma-
tophores; adventitious
roots
Buttressed trunks; multi-
trunk; stooling
* Many other species exhibit one or more morphological adaptations for occurrence in
wetlands. However, not all individuals of a species will exhibit these adaptations under field
conditions, and individuals occurring in uplands characteristically may not exhibit them.
25 February 1997
Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 96
Tabled (Continued)
Partial List of Species With Known Morphological Adaptations for
Occurrence in Wetlands*
Species Common Name Adaptation
Nyssa sylvatica Swamp blackgum Buttressed trunks
var. biflora
Platanus occidentalis Sycamore Adventitious roots
Populus deltoides Cottonwood Adventitious roots
Quercus laurifolia Laurel oak Shallow root system
Quercus palustris Pin oak Adventitious roots
Rhizophora mangle Red mangrove Pneumatophores
Sagittaria spp. Arrowhead Polymorphic leaves
Salix spp. Willow Hypertrophied lenticels;
adventitious roots; oxygen
pathway to roots
Scirpus spp. Bulrush Inflated stems and leaves
Spartina alterniflora Smooth Oxygen pathway to roots
cordgrass
Taxodium distichum Bald cypress Buttressed trunks;
pneumatophores
* Many other species exhibit one or more morphological adaptations for occurrence in
wetlands. However, not all individuals of a species will exhibit these adaptations under field
conditions, and individuals occurring in uplands characteristically may not exhibit them.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987
97
Species
Alnus incana
Alnus rubra
Baccharis viminea
Betula pubescens
Carex arenaria
Carexflacca
Carex lasiocarpa
Deschampsia cespitosa
Filipendula ulmaria
Fraxinus pennsyzvanica
Glyceria maxima
Juncus effusus
Larix laricina
Lobelia dortmanna
Lythrum salicaria
Molinia caerulea
Myrica gale
Nuphar lutea
Nyssa aquatica
Nyssa sylvatica var. biflora
Phalaris arundinacea
Phragmites australis
Pinus contorta
Polygonum amphibium
Potentilla anserina
Ranunculus flammula
Salix cinerea
Salix fragilis
Salix lasiolepis
Scirpus maritimus
Senecio vulgaris
Spartina alterniflora
Trifoliun subterraneum
Typha angustifolia
Table C2
Species Exhibiting Physiological Adaptations for
Occurrence in Wetlands
Physiological Adaptation
Increased levels of nitrate reductase; malate accum.
Increased levels of nitrate reductase
Ability for root growth in low oxygen tensions
Oxidizes the rhizosphere; malate accumulation
Malate accumulation
Absence of ADH activity
Malate accumulation
Absence of ADH activity
Absence of ADH activity
Oxidizes the rhizosphere
Malate accumulation; absence of ADH activity
Ability for root growth in low oxygen tensions;
absence of ADH activity
Slight increases in metabolic rates; increased
levels of nitrate reductase
Oxidizes the rhizosphere
Absence of ADH activity
Oxidizes the rhizosphere
Oxidizes the rhizosphere
Organic acid production
Oxidizes the rhizosphere
Oxidizes the rhizosphere; malate accumulation
Absence of ADH activity; ability for root
growth in low oxygen tensions
Malate accumulation
Slight increases in metabolic rates; increased
levels of nitrate reductase
Absence of ADH activity
Absence of ADH activity; ability for root
Malate accumulation; absence of ADH activity
Malate accumulation
Oxidizes the rhizosphere
Ability for root growth in low oxygen tensions
Ability for root growth in low oxygen tensions
Slight increases in metabolic rates
Oxidizes the rhizosphere
Low ADH activity
Ability for root growth in low oxygen tensions
25 February 1997
Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 98
APPENDIX D: HYDRIC SOILS
1. This appendix consists of two sections. Section 1 describes the basic procedure for
digging a soil pit and examining for hydric soil indicators. Section 2 is a list of hydric soils of
the United States.
Section 1 - Procedures for Digging a Soil Pit and Examining
for Hydric Soil Indicators
Digging a soil pit
2. Apply the following procedure: Circumscribe a 1-ft-diameter area, preferably with a tile
spade (sharpshooter). Extend the blade vertically downward, cut all roots to the depth of the
blade, and lift the soil from the hole. This should provide approximately 16 inches of the soil
profile for examination. Note: Observations are usually made immediately below the A-
horizon or 10 inches (whichever is shallower). In many cases, a soil auger or probe can be
used instead of a spade. If so, remove successive cores until 16 inches of the soil profile have
been removed. Place successive cores in the same sequence as removed from the hole. Note:
An auger or probe cannot be effectively used when the soil profile is loose, rocky, or contains
a large volume of water (e.g. peraquic moisture regime).
Examining the soil
3. Examine the soil for hydric soils indicators (paragraphs 44 and/or 45 of main text (for
sandy soils)). Note: It may not be necessary to conduct a classical characterization (e.g.
texture, structure, etc.) of the soil. Consider the hydric soil indicators in the following
sequence (Note: THE SOIL EXAMINATION CAN BE TERMINATED WHEN A POSITIVE
HYDRIC S01-LINDICA TOR IS FOUND) :
Nonsandy soils.
a. Determine whether an organic soil is present (see paragraph 44 of the
main text). If so, the soil is hydric.
b. Determine whether the soil has a histic epipedon (see paragraph 44 of the
main text). Record the thickness of the histic epipedon on DATA FORM 1.
c. Determine whether sulfidic materials are present by smelling the soil.
The presence of a "rotten egg" odor is indicative of hydrogen sulfide, which
forms only under extreme reducing conditions associated with prolonged
inundation/soil saturation.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 99
d. Determine whether the soil has an aquic or peraquic moisture regime
(see paragraph 44 of the main text). If so, the soil is hydric.
e. Conduct a ferrous iron test. A calorimetric field test kit has been
developed for this purpose. A reducing soil environment is present when the
soil extract turns pink upon addition of a-a-dipyridil.
f. Determine the color(s) of the matrix and any mottles that may be present.
Soil color is characterized by three features: hue, value, and chroma. Hue refers
to the soil color in relation to red, yellow, blue, etc. Value refers to the lightness
of the hue. Chroma refers to the strength of the color (or departure from a
neutral of the same lightness). Soil colors are determined by use of a Munsell
Color Book (Munsell Color 1975). Each Munsell Color Book has color charts
of different hues, ranging from 10R to 5Y. Each page of hue has color chips
that show values and chromas. Values are shown in columns down the page
from as low as 0 to as much as 8, and chromas are shown in rows across the
page from as low as 0 to as much as 8. In writing Munsell color notations, the
sequence is always hue, value, and chroma (e.g. 10YR5/2). To determine soil
color, place a small portion of soil (moistened) in the openings behind the color
page and match the soil color to the appropriate color chip. Note: Match the soil
to the nearest color chip. Record on DATA FORM 1 the hue, value, and
chroma of the best matching color chip. CAUTION: Never place soil on the
face or front of the color page because this might smear the color chips.
Mineral hydric soils usually have one of the following color features
immediately below the A-horizon or 10 inches (whichever is shallower):
(1) Gleyed soil.
Determine whether the soil is gleyed. If the matrix color best fits a color chip
found on the gley page of the Munsell soil color charts, the soil is gleyed. This
indicates prolonged soil saturation, and the soil is highly reduced.
(2) Nongleyed soil.
(a) Matrix chroma of 2 or less in mottled soils, (moistened)
(b) Matrix chroma of 1 or less in unmottled soils, (moistened)
(c) Gray mottles within 10 inches of the soil surface in dark (black) mineral
soils (e.g., Mollisols) that do not have characteristics of (a) or (b) above.
Soils having the above color characteristics are normally saturated for
significant duration during the growing season. However, hydric soils with
significant coloration due to the nature of the parent material (e.g. red soils of
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 100
the Red River Valley) may not exhibit chromas within the range indicated
above. In such cases, this indicator cannot be used.
g. Determine whether the mapped soil series or phase is on the national list
of hydric soils (Section 2). CAUTION: It will often be necessary to compare
the profile description of the soil with that of the soil series or phase indicated
on the soil map to verify that the soil was correctly mapped. This is especially
true when the soil survey indicates the presence of inclusions or when the soil is
mapped as an association of two or more soil series.
h. Look for iron and manganese concretions. Look for small (>0.08-inch)
aggregates within 3 inches of the soil surface. These are usually black or dark
brown and reflect prolonged saturation near the soil surface.
Sandy soils.
Look for one of the following indicators in sandy soils:
a. A layer of organic material above the mineral surface or high organic
matter content in the surface horizon (see paragraph 45a of the main text). This
is evidenced by a darker color of the surface layer due to organic matter
interspersed among or adhering to the sand particles. This is not observed in
upland soils due to associated aerobic conditions.
h. Streaking of subsurface horizons (see paragraph 45c of the main text).
Look for dark vertical streaks in subsurface horizons. These streaks represent
organic matter being moved downward in the profile. When soil is rubbed
between the fingers, the organic matter will leave a dark stain on the fingers.
c. Organic pans (see paragraph 45b of the main text). This is evidenced by
a thin layer of hardened soil at a depth of 12 to 30 inches below the mineral
surface.
Section 2. Hydric Soils of the United States
(Note: not included here as the list from 1987 is out of date)
(Contact your County or State NRCS office)
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 101
Modifications and Clarifications
to the 1987 Wetland Delineation Manual
These modifications and clarifications were issued after Congress removed the 1989 Manual
from use in making wetland determinations.
CECW-OR 23 August 1991
MEMORANDUM
SUBJECT: Wetlands Delineation and the 1992 Energy and Water Development Appropriations
Act
1. The 1992 Energy and Water Development Appropriations Act (Act) contains the following
provisions:
a. None of the funds of the Act shall be used to identify or delineate any land as a "water of the
United States" under the Federal Manual for Identifying and Delineating Jurisdictional Wetlands
that was adopted in January 1989 (1989 Manual) or any subsequent manual not adopted in
accordance with the requirements for notice and public comment of the rule-making process of
the Administrative Procedure Act.
b. In addition, regarding Corps of Engineers ongoing enforcement actions and permit
applications involving lands which the Corps or the Environmental Protection Agency (EPA)
has delineated as "waters of the United States" under the 1989 Manual, and which have not yet
been completed on the date of enactment of the Act (i.e., August 17, 1991), the landowner or
permit applicant shall have the option to elect a new delineation under the Corps 1987 Wetlands
Delineation Manual (1987 Manual) (Technical Report Y-87-1, Waterways Experiment Station
(WES), January 1987) or completion of the permit process or enforcement action based on the
1989 Manual delineation, unless the Corps of Engineers determines, after investigation and
consultation with other appropriate parties, including the landowner or permit applicant, that the
delineation would be substantially the same under either the 1987 or the 1989 Manual.
c. None of the funds in the Act shall be used to finalize or implement the proposed regulations to
amend the fee structure for the Corps of Engineers regulatory program which were published in
Federal Register, Vol. 55, No. 197, Thursday, October 11, 1991.
The provisions of the Act make it necessary for the Corps to change some regulatory procedures,
as described below. This guidance has been reviewed and approved by the Office of the
Assistant Secretary of the Army for Civil Works.
2. After August 17, 1991, initial delineations will be made using the Corps 1987 Manual. WES
will provide field of floes copies of the 1987 Manual. Supplementary guidance will be provided
by CECW-OR under separate cover.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 102
3. The following general guidance will apply:
a. For the purposes of this guidance, ongoing permit applications are defined as: formal
individual permit applications (ENG FORM 4345), letters requesting verification of
authorization under regional or nationwide permits, or predischarge notices as required by
nationwide permits, received before August 17, 1991, where no permit has been issued, verified
or denied, and existing jurisdictional delineations where no permit has been requested.
Individual permit decisions are considered final when the District Engineer signs the decision
document. The 20 day predischarge notification clock (nationwide permit 26) will be stopped
and the information considered incomplete until the options are explained and the general
permitted responds, at which time the clock restarts.
b. Only applications involving section 404, where the jurisdictional delineation (Corps or EPA)
involved wetlands identified as waters of the United States using the 1989 Manual are subject to
the options provided by the Act.
4. Landowners/applicants involved in ongoing applications, as defined in 3. a. & b., will be
notified of their options, as follows:
a. The district may investigate and determine, for some or all of the district's area, on a generic
or case-by-case basis, whether delineations using the 1989 Manual are substantially the same or
substantially different than would have been made using the 1987 Manual. In every case the
Corps must notify the landowner/applicant by letter of the Corps initial determination and
provide the landowner/applicant an opportunity for consultation (not to exceed 30 days), before
proceeding with the final determination in accordance with paragraph b. or c. below. In the
interest of timeliness, districts are encouraged to use generic initial determinations and form
letters requesting the landowner's/applicant's consultation.
b. Where, after considering all applicable information (including any obtained during
consultation with the landowner/applicant and other appropriate parties), the Corps determines
that a delineation using the 1989 Manual is substantially the same as would have been made
using the 1987 Manual, each landowner/applicant will be notified by letter of the Corps'
determination. The Corps letter will inform the landowner/applicant that the delineation will be
considered binding and the evaluation of any formal application will proceed based on this
delineation.
c. Where the Corps determines that delineations using the 1989 Manual are substantially
different than would have been made using the 1987 Manual, each landowner/applicant will be
notified by letter of the Corps determination. The letter will provide the landowner/applicant the
following options: (1) proceeding with the ongoing application using the delineation made under
the 1989 Manual; or (2), electing to have the area redelineated using the 1987 Manual. The
Corps letter will also include an indication of the approximate time required by the district to
complete a redelineation of the area using the 1987 Manual and a statement that the
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 103
landowner/applicant must reply in writing concerning their choice of the above options.
Evaluation, including a redelineation if necessary, of any application will proceed upon receipt
of the landowner's/applicant's letter.
5. Where an ongoing permit application, involving wetlands delineations made using the 1989
Manual, does not fit the circumstances described above, districts will compare the old (1989
Manual) delineation with the provisions of the 1987 Manual to determine if there would be a
substantial change in the area delineated. If there would be a substantial change, the
landowner/applicant will be provided the opportunity to have a new delineation made using the
1987 Manual.
6. The final decision concerning whether a substantial difference exists between delineations
based on the use of the 1989 and the 1987 Manuals rests solely with the Corps. The Corps may
consult with other experts in delineating wetlands, such as the Waterways Experiment Station,
EPA, the Fish and Wildlife Services or the Soil Conservation Service, in reaching a decision
Landowners/applicants and other appropriate parties may provide the Corps documentation for
use in tile decision making process, but in all cases the Corps district decision will be final.
7. The Office of Counsel (CECC-K) is providing through Counsel channels guidance concerning
the steps to be taken for ongoing enforcement actions that fall within the provision of the Act. A
copy of that guidance is enclosed for reference and use.
8. The Army will take no further action, at this time, on its proposal to amend regulatory fees, as
published in the Federal Register, Vol. 55, No. 197, Thursday, October 11,1990. No decision
has been made concerning future actions to change the fee structure.
FOR THE DIRECTOR OF CIVIL WORKS
SIGNED
JOHN P. ELMORE Chief
Operations, Construction and Readiness Division
Directorate of Civil Works
CECW-OR 27 August 1991
MEMORANDUM
SUBJECT: Implementation of the 1987 Corps Wetland Delineation Manual
1. The purpose of this memorandum is to provide guidance concerning the implementation of
the 1987 Corps of Engineers Wetlands Delineation Manual (1987 Manual)(Waterways
Experiment Station Technical Report Y-87-1, January 1987). This guidance supersedes the
guidance provided in John Studt's 21 August memorandum to the field. Further, this guidance is
to be used in conjunction with memoranda dated 23 August, 1991, concerning wetlands
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 104
delineation and the 1992 Energy and Water Development Appropriations Act (Act). In
accordance with the Act and the 23 August memoranda, the 1987 Manual is now used to
delineate potentially jurisdictional wetlands in place of the 1989 Federal Manual for Identifying
and Delineating Jurisdictional Wetlands.
2. The guidance in paragraph 3 will be followed in the application of the 1987 Manual.
3. Use of the 1987 Manual is mandatory, however, the Appendices are modified as discussed
below:
a. Appendix A: The definition of "under normal circumstances" provided in this glossary is
modified pursuant to Regulatory Guidance Letter (RGL) #90-7;
b. Appendix B: Use of the data sheets provided is recommended, but is not mandatory;
c. Appendix C: Sections 1 and 2 - These sections are replaced with the May 1988 National List
of Plant Species That Occur in Wetlands and associated regional lists (U.S. Fish and Wildlife
Service, Summary 88(24) and Biological Reports 88(26.1- 26.13). The referenced lists will be
used to determine the wetland indicator status of plant species and any subsequent updates will
be adopted;
d. Appendix D: Section 2 - The most recent Hydric Soils of the United States list developed by
the U.S. Department of Agriculture, Soil Conservation Service (SCS), will be used to determine
if a particular soil has been designated as hydric by the National Technical Committee for
Hydric Soils. The current hydric soils list was published by SCS in December 1987, and any
subsequent updates will be adopted.
4. All other current policy considerations concerning wetlands in general (e.g., RGLs) remain in
effect during interim use of the 1987 Corps Manual.
5. The Waterways Experiment Station will provide each division and district copies of the 1987
Manual. In addition, a copy of the Environmental Effects of Dredging Technical Notes (EEDP-
04-7) dated January 1988, an article which summarizes the methods for delineating wetlands as
presented in the 1987 Manual, will follow under separate cover. The article does not reflect the
guidance contained in this memorandum; however, it does provide a general overall summary of
the 1987 Manual.
JOHN P. ELMORE
Chief, Operations, Construction and Readiness Division
Directorate of Civil Works
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 105
SUBJECT: Questions & Answers on 1987 Manual
1. In response to questions from the field, the Qs & As on the 1987 Corps of Engineers Wetland
Delineation Manual (1987 Manual) have been further clarified (in particular, questions #7 and
#8). We clarified that for saturated-only systems, the saturation must be to the surface for the
appropriate number of days during the growing season. Furthermore, we clarified that the
number of days for inundation or saturation to the surface are consecutive, not cumulative. The
enclosed Qs and As dated 7 October, 1991 supersede those previously distributed under the
cover memorandum of 16 September, 1991.
2.1 want to again emphasize that the 1987 Manual stresses the need to verify that all three
parameters exist prior to identifying and delineating an area as a wetland. Further, the 1987
Manual focuses on hydrology (i.e., inundation and/or saturation to the surface). In situations
where hydrology is questionable, the 1987 Manual requires stronger evidence regarding the
hydrophytic nature of the vegetation. The 1987 Manual also stresses the need to use sound
professional judgment, providing latitude to demonstrate whether an area is a wetland or not
based on a holistic and careful consideration of evidence for all three parameters. As indicated in
the 1987 Manual and the attached Qs and As, careful professional judgment must be used in
situations where indicators of hydrology are not clear and the dominant vegetation is facultative.
JOHNF. STUDT
Chief, Regulatory Branch
Operations, Construction and Readiness Division
Directorate of Civil Works
October 7, 1991
Questions and Answers on 1987 Corps of Engineers Manual
l.Q. What is the definition and practical interpretation of the growing season which should be
used in the application of the 1987 Manual?
A. The 1987 Manual defines the growing season as "the portion of the year when soil
temperatures at 19.7 inches below the soil surface are higher than biological zero (5 degrees C)".
This is the definition found in Soil Taxonomy, and growing season months can be assumed
based on temperature regimes (e.g., mesic: March- October). The 1987 Manual further states this
period can be approximated by the number of frost-free days. The Waterways Experiment
Station (WES) indicates that the county soil surveys, which utilize 32 degrees, provide the
growing season for each county. There is some flexibility in the determination of the growing
season in the 1987 Manual. The growing season, based on air temperature in the county soil
surveys can be approximated as the period of time between the average date of the first killing
frost to average date of the last killing frost, which sometimes does not accurately reflect the
period of time when the soil temperatures are higher than biological zero. The source of the
information may vary, however, the growing season generally is to be determined by the number
of killing frost-free days. In certain parts of the country where plant communities in general have
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 106
become more adapted to regional conditions, local means of determining growing season may be
more appropriate and can be used.
2.Q. Should the determination of hydric soils be based on the presence of an indicator listed in
the 1987 Manual higher on the list than the soils list or on the fact that a soil appears on the
Hydric Soils of the United States, an indicator which is listed as less reliable in the hierarchy of
hydric soil indicators in the 1987 Manual?
A. The order of soil indicators reliability as listed in the 1987 Manual remains valid and will be
used. The reliability of the indicators is based on the fact that field verification of a soil's hydric
characteristics is more accurate than mapping or soils lists. Soils listed on the most recent Hydric
Soils list have been determined by the National Technical Committee for Hydric Soils to meet
the criteria for hydric soils. When in the field, verification that mapped hydric soils actually
exhibit indicators identified in me 1987 Manual for hydric soils is recommended. Although a
soil may appear on the list of hydric soils, inclusions or disturbances may alter this designation
to some degree, so the list alone may not always be reliable. In obvious wetlands, if the soil is on
the list and the area meets the hydrology and vegetation criteria, the area is a wetland. As found
with me 1989 Manual, one cannot rely solely on the fact that a soil is mapped as hydric in
making the wetland delineation, all cases, best professional judgment should be used. The
county lists provide valuable information, but again should not solely be relied on to make a
final determination as to whether hydric soils are present. Verification of the presence of at least
one of the indicators for hydric soils on the list (pages. 30-34) is required in conjunction with the
use of a county soils list. The national soils list to be used has recently been updated by the
NTCHS (June 1991), and this list will be used by the Corps in conjunction with the 1987
Manual.
3.Q. How should the 1987 Manual be applied with respect to the definition of "normal
circumstances"?
A. The definition of "under normal circumstances" in the 1987 Manual states briefly that "this
term refers to situations in which the vegetation has not been substantially altered by man's
activities". As stated in item #3 of the memorandum of 27 August, 1991, the definition of
normal circumstances used in the 1987 Manual has been clarified by Regulatory Guidance Letter
(RGL) 90-7. Although this RGL deals primarily with agricultural activities in wetlands,
paragraphs #3 & #4 discuss normal circumstances with respect to all areas potentially subject to
404. Further guidance on normal circumstances is found in RGL 86-9 regarding construction
sites and irrigated wetlands. The guidance should be followed in preferential sequence of; 1)
RGL 90-7, 2) RGL 86-9, and 3) 1987 Manual.
4.Q. Does the vegetation criteria in the 1987 manual require the use of the facultative(FAC) -
neutral vegetation test (i.e., count the dominant species wetter and drier than FAC, and ignore all
of the FACs in the vegetation determination)?
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 107
A. While the 1987 Manual mentions use of the FAC - neutral test for determining the presence
of wetland vegetation in several places, the first indicator of wetland vegetation criteria is the
presence of more than 50% of the dominant plant species FAC or wetter (not including FAC-
species, which are considered non-wetland indicators under the 1987 manual). The indicator
status of each of the dominant species is determined by consulting the current regional plant list
published by the FWS. The 1987 Manual provides an option in this determination of applying
the FAC - neutral test in cases where the delineator questions the status designation of a
particular plant species on a subregional basis (see page 233. As always, any deviation from
established protocol requires documentation. The FAC - neutral option may also prove useful in
questionable areas or when the determination relies on the vegetation call in an area that is not
otherwise an obvious wetland. Specifically, the 1987 Manual is replete with cautions and
guidance that the Corps regulators must be confident that the area is wetland when the area has a
FAC-dominated plant community. Uncertainty regarding the status of an area as a wetland
where the dominant vegetation is FAC would be a valid reason to use the FAC - neutral option.
Situations exist where use of the FAC - neutral method will not serve to provide any additional
information as to the hydrophytic nature of the plant community (e.g., all species are FAC or
there is an equal number of species wetter and drier than FAC such that they cancel each other
out). In these cases, it is appropriate to consider the + and - modifiers associated with some FAC
species, which indicate the species frequency of occurrence in a wetter or drier environment, in
the overall assessment of the vegetation parameter. Documentation supporting reasons for using
the FAC - neutral option must always be provided and acceptance of delineations, as always,
remains up to the discretion of the District.
5.Q. Can indicators for any of the criteria in the 1989 Manual be used as indicators for
verification of the same or other criteria presented in the 1987 Manual?
A. The indicators of hydrology in the 1987 Manual differ from those of the 1989 Manual, and
are not interchangeable. In particular, the hydrology determination in the 1989 Manual often
relied on evidence of properties from the soil and/or vegetation parameters. Indicators provided
in the 1989 Manual for field verification of a certain criterion that are not presented in the 1987
Manual for application with the same criterion cannot be used except as additional information
in support of the verification. It is unlikely that an area which is a wetland will fail to meet a
criteria utilizing the indicators which are listed in the 1987 Manual.
6.Q. Will the other Federal agencies be utilizing the 1987 Manual in their wetland
determinations as well as the Corps of Engineers?
A. EPA has concurred with the Corps using the 1987 Manual for all actions. Further, we
understand that EPA will likely use the 1987 Manual for EPA's delineations as well. The other
agencies (SCS & FWS) typically do not make delineations for purposes of Section 404.
7.Q. To what depth should one look in the soil to find indicators of hydrology?
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 108
A. In accordance with the 1987 Manual's guidance on reading soil color (D2), after digging a
16" soil pit observations should be made immediately below the A-horizon or within 10" of the
soil surface (whichever is shallower). This guidance pertains to observations of indicators of the
soil criterion. For indicators of saturation to the surface in the hydrology criterion, observations
are made within a major portion of the root zone (usually within 12"), again in the 16" pit.
Visual observation of standing water within 12" of the surface may, under certain circumstances,
be considered a positive indicator of wetland hydrology (i.e., saturation to the surface) as stated
on page 38. When using water table within 12" of the surface as an indicator of hydrology, care
must be used to consider conditions and the soil types (i.e., to ensure that the capillary ability of
the soil texture is considered in regard to the water table depth). Vegetation and soil properties
used in the determination of hydrology in the 1989 Manual, are typically not available for field
verification of this criterion in the 1987 Manual. However, the 1987 Manual allows for some
flexibility with regards to indicators of wetland hydrology, and states that indicators are not
limited to those listed on pages 37-41. Other indicators, such as some type of recorded data (e.g.,
soil surveys which provide specific and strong information about the soil series' hydrology) may
be used to verify a wetland hydrology call in a saturated but not inundated area. Appropriate
documentation to support the call is necessary in all cases.
8.Q. What length of time must wetland hydrology be present for an area to be determined a
wetland under the 1987 Manual?
A. In the hydrology section of Part III, the 1987 Manual discusses the hydrologic zones which
were developed through research at WES to indicate the duration of inundation and/or soil
saturation during the growing season. Wetland hydrology is defined in the 1987 Manual as the
sum total of wetness characteristics in areas that are inundated or have saturated soils for a
sufficient duration to support hydrophytic vegetation. The 1987 Manual discusses hydrology in
terms of a percent of the growing season when an area is wet (page 36). Generally speaking,
areas which are seasonally inundated and/or saturated to the surface for more than 12.5% of the
growing season are wetlands. Areas saturated to the surface between 5% and 12.5% of the
growing season are sometimes wetlands and sometimes uplands. Areas saturated to the surface
for less than 5% of the growing season are non-wetlands. The percent of growing season
translates to a number of days, depending on the length of the growing season in any particular
area (e.g., 12.5% of a 170-day growing season is 21 consecutive days). This system for
classification of hydrologic zones based on stream gauge data transformed to mean sea level
elevations is useful as a guide to time frames of wetness sufficient to create wetlands. The length
of time an area is wet for hydrology is based on consecutive days during the growing season. If
an area is only saturated to the surface for a period of between 5% and 12.5% of the growing
season and no clear indicators of wetland hydrology exist (i.e., recorded or field data; also see
answer #7 above), then the vegetation test should be critically reviewed. Specifically, in such
cases a vegetative community dominated by FAC species would generally indicate that the area
is not a wetland (unless the FAC- neutral test was indicative of wetlands). The actual number of
days an area is inundated and/or saturated to the surface for an area to be called a wetland
varies; the identification of an indicator of recorded or field data is necessary to document that
an area meets the wetland hydrology criterion of the 1987 Manual (i.e., the list of hydrology
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 109
indicators on pages 37-41, which are to be used in the preferential order shown; also see
question #7). The number of days specified in the June 1991 Hydric Soils of the United States
(i.e., usually more than 2 weeks during the growing season) as the criteria for hydric soils
pertains to hydric soils and not the hydrology criterion of the 1987 Manual, which varies with
the growing season as previously discussed.
9.Q. Will delineations made now under the 1987 Corps Manual be subject to redelineation under
the revised 1989 Manual after it is finalized?
A. Wetland determinations made after 17 August, 1991, are made following the guidance
provided in the 1987 Corps Manual and memoranda of 23 and 27 August, 1991 and these
questions and answers. These delineations are subject to and remain valid for the period of time
described in RGL 90-6. As discussed in Issue #4 of the preamble to the proposed revisions to the
1989 Federal Manual for Identifying and Delineating Jurisdictional Wetlands issued 14 August
in the Federal Register, wetland calls made after the issuance date of these revisions but prior to
finalization of the revised manual may be subject to redelineation under the new manual at the
request of the landowner. Final actions will generally not be reopened. Wetland calls made
under the 1989 Manual are already subject to redelineation under the 1987 Manual in
accordance with the guidance issued 23 August. Until such time as the proposed revisions to the
1989 Manual are finalized, it is unclear as to what effect, if any, the equity provision in the
preamble to the proposed revisions will have on the 404 program. Therefore, written
delineations made with the 1987 Manual will explicitly state they are final for a period of three
years as specified in RGL 90-6, subject to any equity provisions that may be adopted as part of
implementation of the final revisions to the 1989 Manual.
10.Q. How does the 1987 Manual compare to the 1989 Manual or its proposed revisions?
A. The various manuals have been compared by WES and the side-by-side comparison is
available for your information.
1 l.Q. Will applicants be subject to delay with use of the 1987 Manual?
A. During the initial transition to use of the 1987 Manual for wetland delineations as of 17
August, some delays are unavoidable. The Corps field offices must adhere to the procedures
provided in the 23 August memorandum, while striving to expedite the review process to the
extent possible. No offices should indicate that they cannot operate due to lack of guidance
during this transition period. HQUSACE recognizes that there will be delays associated with
implementing the Corps 1987 Manual and we will take these delays into account when
reviewing district application performance data (e.g., % of IPs evaluated in 60 days). Districts
should not stop the permit clock, but should indicate where substantial impacts to permit
evaluation performance have resulted from implementation of the 1987 Manual.
Field testing questions:
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 110
l.Q. Which procedure should be followed for the field testing since the guidance provided
initially states both the routine and comprehensive?
A. Since the criteria for determining the presence of hydrophytic vegetation is proposed to be the
prevalence index, the comprehensive method must be used as neither the routine nor
intermediate on-site methods will yield a prevalence index. The point intercept sampling
procedure on pages 40471-40473 yields a PI directly and is stated that it is the procedure to be
used in the determination of the PI on page 40455 of the Federal Register. Routine or
intermediate on-site determinations methods can be used for determination of the limits of hydric
soils (Step 1, page 40471) and wetland hydrology (Step 16, page 40473) necessary to be
considered in conjunction with the vegetation assessment as described on pages. 40472-40473
(Steps 6-15). Documentation of deviations from the prescribed methodology should be noted in
the field testing reports.
2.Q. Must all plants or just the dominants be looked at to determine the prevalence index (data
sheets say dominants while procedure looks at individuals)?
A. All plant species, not just dominant plants, are considered in the prevalence index
determination of hydrophytic vegetation. However, the point intercept sampling procedure in the
comprehensive methods which is used to yield a PI requires the identification of only those
species which are intercepted by an imaginary line extended from the points every 2 feet along
the transect.
3.Q. How is one to compare site determinations made using the 1989 Manual and the proposed
revisions when different methods (e.g., routine vs. comprehensive) were used?
A. Comparing the data from delineations previously performed under the 1989 Manual
(regardless of the method employed at the time) with the data recorded during the field testing of
the proposed revisions of the 1989 Manual, should be done to the level of detail possible. The
resultant delineations from these different manuals and/or procedures are what these tests are
comparing; some discussion of any differences and why they exist (e.g., changes in the criteria,
indicators, and methodologies utilized) should be provided based on the teams or individual's
observations, experience and best professional judgment.
4.Q. Are the regional indicators listed in the proposed revisions to the 1989 Manual the same for
soils and hydrology?
A. Yes, they are proposed for use as an indicator of the soil criterion as well as corroborative
information for use with the secondary indicators of hydrology.
5.Q. Do you always need to prove 21 days saturation?
A. Yes, but identification of one or more of the primary indicators of hydrology is proof that the
area meets the 21 day hydrology criterion.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 111
6.Q. Is there a map that goes with the regional indicators?
A. Yes, the regional indicators of soil saturation continue to be developed and are being
proposed for use with the manual for the four Department of Agriculture, Soil Conservation
Service main regions; Northeast, South, Midwest and West (see enclosed map).
7.Q. Can the soils list be used as corroborative information with a secondary indicator for
verification that the hydrology criterion is met?
A. No, only the information listed can be used. Comments on the appropriateness of the
information and suggestions for others should be provided during the comment period.
8.Q. Can blackened leaves be used as hydrology indicator?
A. No, only those listed in the list of hydrology indicators.
9.Q. Must three transects always be evaluated for the methodology described to determine the
prevalence index?
A. No, in most cases, only 1 transect is required for purposes of these tests. The following
guidance provides the exact circumstances under which options concerning the number of
transects are permitted before proceeding to Step 11 and repeating the two transects: - If the
prevalence index value determined for the first transect is less than 2.5 or more than 3.5, and, in
the professional judgment of the investigators repeated transects are likely to provide the same
result, the investigators need not repeat the transect. In cases where this shortened method is
used, this should be reflected in the results. While the agencies recognize that this change in
protocol does not strictly adhere to the method contained in the revised manual, it is nevertheless
a practical step to reduce the demands of using the comprehensive method.
10.Q. How are the frost-free dates determined under the proposed revisions which the field
testing should utilize?
A. The frost-free dates are to be determined by using the summary of growing season data
developed by NOAA referred to on the first page of the Interagency Protocol. This information
is available for each state and will be provided to the Divisions for distribution to the field in the
near future.
1 l.Q. Must sites visited and data collected prior to the recently released revised data sheets be
redone using the new forms?
A. No, both versions collect essentially the same data. Sections I (Background) and VI (Wetland
Evaluation) of the revised data sheets are identified as sections that will generally apply to the
entire site, and in most cases will not have to be repeated for each individual transect. Besides
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 112
the reorganization of these forms, revisions have been made to collect data on the 1987 Manual,
include questions on the regional indicators, and clarify the functions and values assessments.
12.Q. How should the Corps provide comments on the proposed revisions to the 1989 manual?
A. Each District is to submit comments to their Division office for consolidation into one
response to EPA (in accordance with the Federal Register issued 14 August) prior to the end of
the comment period (currently 15 October); a copy of these comments are to be provided to
CECW-OR. Each Division is encouraged to provide comments to HQUSACE at least one week
prior to the end of the comment period so that key issues and concerns raised may be stressed in
the comments which will come from Headquarters. In addition, the Corps has the opportunity to
comment through the interagency field testing reports, which are due to HQUSACE by 1
November. HQUSACE and WES will continue to participate in interagency efforts to finalize
revisions to the 1989 Manual until such time as the revisions become final. There is
consideration being given as to the need for a time extension to the comment period (and
therefore the field testing). If this becomes final, we will let the Divisions know as soon as
possible.
13.Q. Is the information gathered during the inter agency field testing available for release to the
public and may the public accompany the teams in the field?
A. The interagency field testing of the proposed revisions to the 1989 Federal Manual for
Identifying and Delineating Jurisdictional Wetlands published in the 14 August, 1991 Federal
Register is being performed to evaluate the technical validity, practical utility, and clarity of
understanding of the proposed revisions. The results of the field testing will be taken into
consideration during the final revisions to the 1989 Manual. The information gathered during the
field testing is considered to be predecisional by HQUSACE and inappropriate for release prior
to the finalization of the proposed revisions. Counsel is preparing written guidance on this issue
and it will be distributed to the Divisions as soon as possible.
14.Q. Will delineations made now under the 1987 Corps Manual be subject to redelineation
under the revised 1989 Manual after it is finalized?
A. Wetland determinations made after 17 August, 1991, are made following the guidance
provided in the 1987 Corps Manual and memoranda of 23 and 27 August, 1991. These
delineations are subject to and remain valid for the period of time described in RGL 9-6. As
discussed in Issue #4 of the preamble to the proposed revisions to the 1989 Federal Manual for
Identifying and Delineating Jurisdictional Wetlands issued 14 August in the Federal Register,
wetland calls made after the issuance date of these revisions but prior to finalization of the
revised manual may be subject to redelineation under the new manual at the request of the
landowner. Final actions will generally not be reopened. Wetland calls made under the 1989
Manual are already subject to redelineation under the 1987 Manual in accordance with the
guidance issued 23 August. Until such time as the proposed revisions to the 1989 Manual are
finalized, it is unclear as to what effect, if any, the equity provision in the preamble to the
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 113
proposed revisions will have on the 4U4 program. Therefore, written delineations made with the
1987 Manual will explicitly state they are final for a period of three years as specified in RGL
9-6, subject to any equity provisions that may be adopted as part of implementation of the final
revisions to the 1989 Manual.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 114
CECW-OR20Feb 1992
MEMORANDUM FOR ALL MAJOR
SUBORDINATE COMMANDS,
DISTRICT COMMANDS
SUBJECT: Regional Interpretation of the 1987 Manual
1. The purpose of this memorandum is to provide clarification to the divisions and districts
concerning application of the 1987 Corms Wetlands Delineation Manual (1987 Manual). As you
are aware, we have been using the 1987 Manual since 17 August 1991. All indications are that
the 1987 Manual, when used in conjunction with the guidance provided in the 7 October 1991
Questions and Answers, is working well for the identification and delineation of wetlands. While
the Administration's effort to finalize a revised Manual continues, the Corps of Engineers will
continue to use the 1987 Manual.
2. Local procedures on the implementation of the 1987 Manual must be fully consistent with
both the 1987 Manual and the Questions and Answers issued 7 October 1991. Any efforts to
provide additional guidance regarding the use of the 1987 Manual must be reviewed and
approved by HQUSACE (CECW-OR) prior to regional implementation. The data forms
provided in the 1987 Manual are to be used, however, additional fields may be added to collect
more detailed site specific information when taken from the list of indicators in the 1987
Manual. As pointed out in the 7 October 1991 Questions and Answers, there is flexibility in the
1987 Manual which can be applied on a case-by-case basis only. Local procedures must not add
indicators of any of the three wetland parameters to the data sheets. We recognize that the
indicators of hydrology in the 1987 Manual are sometimes difficult to demonstrate. However,
additional regional indicators must only be used on a case-by-case basis to demonstrate that a
parameter is met. For example, blackened leaves and oxidized rhizospheres could be used
together to support a delineation where the listed indicators in the 1987 Manual are not directly
observed, but where the FAC neutral test is met.
3. All guidance on the use of the 1987 Manual must come from HQUSACE to ensure the a
consistent national approach is taken in the Corps application of the 1987 Manual.
FOR THE COMMANDER:
Signed
ARTHUR E. WILLIAMS
Major General, USA
Director of Civil Works
CECW-OR March 6, 1992
SUBJECT: Clarification and Interpretation of the 1987 Manual
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 115
1. The purpose of this memorandum is to provide additional clarification and guidance
concerning the application of the Corps of Engineers Wetlands Delineation Manual, Technical
Report Y-87-1, January 1987, Final Report (1987 Manual). As discussed in my 20 February
1992 memorandum, procedures for the identification and delineation of wetlands must be fully
consistent with both the 1987 Manual and the Questions and Answers issued 7 October 1991.
The technical and procedural guidance contained in paragraphs 2 thru 6 below has been
prepared by the Waterways Experiment Station (WES) and is provided as further guidance. The
following guidance is considered to be consistent with the 1987 Manual and the 7 October
Questions and Answers. Further, this guidance will be presented in the upcoming Regulatory IV
wetlands delineation training sessions in FY 92. The alternative technical methods of data
gathering discussed below are acceptable as long as the basic decision rules (i.e., criteria and
indicators) established in the 1987 Manual are applied. Also enclosed is a revised data form
which may be used in lieu of the routine data sheet provided with the 1987 Manual, if desired.
As discussed in my 20 February 1992 memorandum to the field, regional approaches and/or
alternative data sheets must be reviewed and approved by HQUSACE (CECW-OR) prior to
regional implementation. Notwithstanding this requirement, we encourage interagency
coordination and cooperation on implementation of the 1987 Manual. Such cooperation can
facilitate the continued success of our use of the 1987 Manual
2. Vegetation:
a. Basic rule: More than 50 percent of dominant species from all strata are OBL, FACW, or FAC
(excluding FAC-) on the appropriate Fish and Wildlife Service regional list of plant species that
occur in wetlands.
b. The 1987 Manual provides that the 3 most dominant species be selected from each stratum
(select 5 from each stratum if only 1-2 strata are present). However, alternative ecologically
based methods for selecting dominant species from each stratum are also acceptable. The
dominance method described in the 1989 interagency manual is an appropriate alternative
method. (1989 Manual, p. 9, para.3.3)
c. The 4 vegetation strata (tree, sapling/shrub, herb, and woody vine) described in the 1987
Manual are appropriate. However, a 5-stratum approach (tree, sapling, shrub, herb, and woody
vine) is an acceptable alternative.
d. The 1987 Manual states on page 79 that hydrophytic vegetation is present if 2 or more
dominant species exhibit morphological adaptations or have known physiological adaptations
for wetlands. This rule should be used only after the basic rule is applied; use caution with
adaptations (e.g., shallow roots) that can develop for reasons other than wetness. Furthermore,
the morphological adaptations must be observed on most individuals of the dominant species.
e. In areas where the available evidence of wetlands hydrology or hydric soil is weak (e.g., no
primary indicators of hydrology), the Facultative Neutral (FAC neutral) option may be used to
help clarify a wetland delineation. Use of the FAC neutral option is explained in paragraph
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 116
3 5 (a), page 23, of the 1987 Manual. Use of the FAC neutral option is at the discretion of the
District. Further, the FAC neutral option cannot be used to exclude areas that meet the "basic
vegetation rule" and the hydrology and hydric soil requirements.
3. Hydrology:
a. Areas which are seasonally inundated and/or saturated to the surface for a consecutive number
of days for more than 12.5 percent of the growing season are wetlands, provided the soil and
vegetation parameters are met. Areas wet between 5 percent and 12.5 percent of the growing
season in most years (see Table 5, page 36 of the 1987 Manual) may or may not be wetlands.
Areas saturated to the surface for less than 5 percent of the growing season are non-wetlands.
Wetland hydrology exists if field indicators are present as described herein and in the enclosed
data sheet.
b. To evaluate hydrologic data (e.g., from stream gages or groundwater wells) growing season
dates are required. Soil temperature regime (i.e., period of the year when soil temperature at 20
inches below the surface is above 5 degrees C) is the primary definition of growing season, but
data are rarely available for individual sites. Broad regions based on soil temperature regime
(e.g., mesic, thermic) are not sufficiently site-specific. For wetland determinations, growing
season can be estimated from climatological data given in most SCS county soil surveys (usually
in Table 2 or 3 of modern soil surveys). Growing season starting and ending dates will generally
be determined based on the "28 degrees F or lower" temperature threshold at a frequency of "5
years in 10". In the south, at the discretion of the district, it may be more appropriate to use the
32 degree F threshold.
c. In groundwater-driven systems, which lack surface indicators of wetland hydrology, it is
acceptable to use local Soil Conservation Service (SCS) soil survey information to evaluate the
hydrology parameter (p. 37 in the Manual) in conjunction with other information, such as the
FAC neutral test. Use caution in areas that may have been recently drained.
d. Oxidized rhizospheres surrounding living roots are acceptable hydrology indicators on a case-
by-case basis and may be useful in groundwater systems. Use caution that rhizospheres are not
relicts of past hydrology. Rhizospheres should also be reasonably abundant and within the upper
12 inches of the soil profile. Oxidized rhizospheres must be supported by other indicators of
hydrology such as the FAC neutral option if hydrology evidence is weak.
4. Soil:
a. The most recent version of National Technical Committee for Hydric Soils hydric soil criteria
will be used. At this writing, criteria published in the June 1991 Hydric Soils of the United States
are current. These criteria specify at least 15 consecutive days of saturation or 7 days of
inundation during the growing season in most years.
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
Wetland Delination Manual, 1987 117
b. Local Lists of Hydric Soil Mapping Units recently developed by SCS and available from
county or State SCS offices give local information about presence of hydric soils on a site.
When available, these local lists take precedence over the national list for hydric soil
determinations.
c. SCS is currently developing regional indicators of significant soil saturation. Until finalized
and adopted, these indicators may not be used for hydrology or hydric soil determinations.
d. The statement (p.31 of the 1987 Manual) that gleyed and low-chroma colors must be observed
"immediately below the A-horizon or 10 inches (whichever is shallower)" is intended as general
guidance. Certain problem soils may differ.
5. Methods:
a. As stated in the 1987 Manual (footnote p. 76), alternative plot sizes and dominance measures
are acceptable.
h. For comprehensive determinations involving a patchy or diverse herb layer, a single, centrally
located 3.28 x 3.28-foot quadrat may not give a representative sample. As an alternative, the
multiple-quadrat procedure presented in the 1989 Manual (p. 42) is recommended.
6. Problem Areas
a. Page 93, paragraph 78 of the 1987 Manual states that similar problem situations may occur in
other wettand types, therefore, problem areas are not limited this list.
b. Problem soil situations mentioned elsewhere in the Manual include soils derived from red
parent materials, some Entisols, Mollisols, and Spodosols.
7. Questions concerning this information should be directed to Ms. Karen A. Kochenbach,
HQUSACE (CECW-OR), at (202) 272-1784, or Mr. James S. Wakeley, WES, at (601) 634-
3702.
Signed by,:
Hugh F. Boyd III
for
ARTHUR E. WILLIAMS
Major General, USA
Director of Civil Works
25 February 1997 Environmental Techncal Services Co. 834 Castle Ridge Rd Austin, Texas 78746
-------
United States
Environmental Protection
Agency
Office of Environmental
Information
Washington, DC 20460
EPA/600/R-96/055
August 2000
f/EPA
Guidance for the Data Quality
Objectives Process
EPA QA/G-4
-------
-------
Foreword
The U.S. Environmental Protection Agency (EPA) has developed the Data Quality
Objectives Process as the Agency's recommended planning process when environmental data are
used to select between two opposing conditions. The Data Quality Objectives Process is used to
develop Data Quality Objectives that clarify study objectives, define the appropriate type of data,
and specify tolerable levels of potential decision errors that will be used as the basis for
establishing the quality and quantity of data needed to support decisions. When this Process is
not immediately applicable (i.e., the objective of the program is estimation, research, or any other
objective that does not select between two distinct conditions), the Agency requires the use of a
systematic planning method for defining performance criteria. This document, Guidance for the
Data Quality Objectives Process (EPA QA/G-4) provides a standard working tool for project
managers and planners to develop Data Quality Objectives for determining the type, quantity, and
quality of data needed to reach defensible decisions.
As required by EPA Manual 5360 (May 2000), this document is valid for a period of up to
five years from the official date of publication. After five years, this document will be reissued
without change, revised, or withdrawn from the EPA Quality System series documentation.
This document is one of the EPA Quality System Series documents which describe EPA
policies and procedures for planning, implementing, and assessing the effectiveness of a quality
system. Questions regarding this document or other EPA Quality System Series documents
should be directed to:
U.S. EPA
Quality Staff (2811R)
1200 Pennsylvania Ave, NW
Washington, DC 20460
Phone: (202)564-6830
Fax: (202)565-2441
e-mail: qualityfg), epa.gov
Copies of EPA Quality System Series documents may be obtained from the Quality Staffer by
downloading them from the Quality Staff Home Page:
www.epa.gov/quality
Final
EPA QA/G-4 i August 2000
-------
Final
EPA QA/G-4 ii August 2000
-------
TABLE OF CONTENTS
Pas
CHAPTER 0.
0.1
0.2
0.3
0.4
0.5
CHAPTER 1.
1.1
1.2
1.3
1.4
CHAPTER 2.
2.1
2.2
2.3
2.4
CHAPTER 3.
3.1
3.2
o o
J.J
3.4
CHAPTER 4.
4.1
4.2
4.3
4.4
CHAPTER 5.
5.1
5.2
5.3
5.4
INTRODUCTION 0-1
EPA Program Requirements for Environmental Data and Technology . 0-1
Systematic Planning and the Data Quality Objectives Process 0-4
Benefits of Using the DQO Process 0-9
Organization of this Document 0-12
Background for the Three Examples 0-13
STEP 1. STATE THE PROBLEM 1-1
Background 1-1
Activities 1-2
Outputs 1-5
Examples 1-5
STEP 2. IDENTIFY THE DECISION 2-1
Background 2-1
Activities 2-2
Outputs 2-4
Examples 2-4
STEP 3. IDENTIFY THE INPUTS TO THE DECISION 3-1
Background 3-1
Activities 3-2
Outputs 3-4
Examples 3-4
STEP 4. DEFINE THE BOUNDARIES OF THE STUDY 4-1
Background 4-1
Activities 4-3
Outputs 4-7
Examples 4-8
STEP 5. DEVELOP A DECISION RULE 5-1
Background 5-1
Activities 5-2
Outputs 5-4
Examples 5-5
EPA QA/G-4
in
Final
August 2000
-------
TABLE OF CONTENTS (continued)
Pa
CHAPTER 6. STEP 6. SPECIFY TOLERABLE LIMITS ON DECISION
ERRORS 6-1
6.1 Background 6-1
6.2 Activities 6-2
6.3 Outputs 6-12
6.4 Examples 6-14
CHAPTER 7. OPTIMIZE THE DESIGN FOR OBTAINING DATA 7-1
7.1 Background 7-1
7.2 Activities 7-2
7.3 Outputs 7-6
7.4 Examples 7-6
CHAPTER 8. BEYOND THE DATA QUALITY OBJECTIVES PROCESS ... 8-1
8.1 Planning 8-1
8.2 Implementation 8-3
8.3 Assessment 8-3
APPENDIX A. DERIVATION OF SAMPLE SIZE FORMULA FOR TESTING
MEAN OF NORMAL DISTRIBUTION VERSUS AN ACTION
LEVEL A-l
APPENDIX B. BIBLIOGRAPHY AND REFERENCES B-l
APPENDIX C. DQO EXAMPLE: USE OF THE MEDIAN C-l
APPENDIX D. GLOSSARY OF TERMS D-l
EPA QA/G-4
IV
Final
August 2000
-------
LIST OF FIGURES
Page
Figure 0-1. The EPA Quality System 0-2
Figure 0-2. The Scientific Method 0-4
Figure 0-3. The Data Quality Objectives Process 0-6
Figure 0-4. Repeated Application of the DQO Process Throughout the Life
Cycle of a Project 0-8
Figure 2-1. An Example of the DQO Process Applied to Multiple Decisions
for a Hazardous Site Investigation 2-3
Figure 4-1. Geographic Boundaries Delineated Using a Map 4-4
Figure 6-1. An Example of How Total Study Error Can Be Broken Down
by Components 6-3
Figure 6-2. An Example of a Decision Performance Curve 6-7
Figure 6-3. An Example of a Decision Performance Goal Diagram (Baseline Condition:
Parameter exceeds the Action Level) 6-9
Figure 6-4. An Example of a Decision Performance Goal Diagram (Baseline Condition:
Parameter is less than the Action Level) 6-13
Figure 6-5. Decision Performance Goal Diagram for Example 1 6-13
Figure 6-6. Decision Performance Goal Diagram for Example 2 6-15
Figure 8-1. The Data Quality Assessment Process 8-4
Figure C-l. Decision Performance Goal Diagram for Dust Lead Loading C-4
Final
EPA QA/G-4 v August 2000
-------
LIST OF TABLES
Page
Table 0-1. Elements of the Systematic Planning Process 0-5
Table 2-1. An Example of a Principal Study Question and Alternative Actions 2-2
Table 5-1. Population Parameters and Their Applicability to a Decision Rule 5-3
Table 6-1. False Acceptance and False Rejection Decisions 6-5
Table 7-1. False Acceptance Decision Error Rates and Alternative
Sampling Frequencies 7-8
Table 8-1. Elements of a Quality Assurance Project Plan 8-2
Table C-l. Number of Samples Required for Determining
if the True Median Dust Lead Loading is Above the Standard C-5
EPA QA/G-4
VI
Final
August 2000
-------
CHAPTER 0
INTRODUCTION
After reading this chapter you should understand the structure and
function of EPA 's Quality System, the kinds of programs that are apart
of this System, and the benefits of using the Data Quality Objectives
Process.
When data are being used to select between two alternative conditions (e.g., compliance
or non-compliance with a standard), the Agency's recommended systematic planning tool is the
Data Quality Objectives (DQO) Process. The DQO Process is a systematic planning process that
is part of the EPA's Quality System.
Who can use this guidance document! This guidance is intended for project managers,
technical staff, regulators, stakeholders, and others who wish to use the DQO Process to plan
data collection efforts and develop an appropriate data collection design to support decision
making.
0.1 EPA Quality System Requirements
EPA Order 5360.1 A2 (EPA 2000a) and the applicable Federal regulations establish a
mandatory Quality System that applies to all EPA organizations and organizations funded by
EPA. Components of the Quality System are presented in Figure 0-1. Organizations must ensure
that data collected for the characterization of environmental processes and conditions are of the
appropriate type and quality for their intended use and that environmental technologies are
designed, constructed, and operated according to defined expectations. Systematic planning is a
key project-level component of the EPA Quality System (see Figure 0-1).
EPA policy is based on the national consensus standard, ANSI/ASQC E4-1994,
Specifications and Guidelines for Environmental Data Collection and Environmental
Technology Programs, developed by the American National Standards Institute and the American
Society for Quality. This document describes the necessary management and technical area
elements for developing and implementing a quality system by using a tiered approach to a quality
system. The standard recommends first documenting each organization-wide quality system in a
Quality Management Plan or Quality Manual (to address requirements of Part A: Management
Systems of the standard), and then documenting the applicability of the quality system to technical
activity-specific efforts in a Quality Assurance Project Plan or similar document (to address the
requirements of Part B: Collection and Evaluation of Environmental Data of the standard).
EPA has adopted this tiered approach for its mandatory Agency-wide Quality System. This
document addresses Part B requirements of the standard for systematic planning for
environmental data operations.
Final
EPA QA/G-4 0 - 1 August 2000
-------
o
LLJ
O
or
Q.
o
o
Q.
Z
g
i
z
o
o
O
_l
O
Q.
Consensus Standards
ANSI/ASQC E4
ISO 9000 Series
Internal EPA Policies
EPA Order 5360.1
EPA Manual 5360
External Policies
Contracts - 48 CFR 46
Assistance Agreements -
40 CFR 30, 31, and 35
EPA Program &
Regional Policy
Quality System
Documentation
(e.g., Quality Management Plan)
Supporting System Elements
(e.g., Procurements,
Computer Hardware/Software)
Training/Communication
(e.g., Training Plan,
Conferences)
Annual Review and Planning
(e.g., QA Annual Report
and Work Plan)
System Assessment
(e.g., Quality System Audit)
Systematic
Planning
(e.g., DQO Process)
Acquire Data
1
Standard
Operating
Procedures
i
Technical
Assessments
PLANNING
i
IMPLEMENTATION
ASSESSMENT
Defensible Products and Decisions
Figure 0-1. EPA Quality System Components and Tools
EPA QA/G-4 0 - 2
Final
August 2000
-------
In accordance with EPA Order 5360.1 A2, the Agency requires that:
• Environmental programs performed for or by the Agency be supported by data of
the type and quality appropriate to their expected use. EPA defines environmental
data as information collected directly from measurements, produced from models,
or compiled from other sources such as databases or literature.
Decisions involving the design, construction, and operation of environmental
technology be supported by appropriate quality assured engineering standards and
practices. Environmental technology includes treatment systems, pollution control
systems and devices, waste remediation, and storage methods.
EPA Order 5360.1 A2 is supported by the EPA Quality Manual for Environmental
Programs (U.S. EPA, 2000b) that defines requirements for implementing EPA's Quality System.
The Order defines the quality requirements and the Manual presents the mandatory "how to" for
implementing these requirements.
EPA's Quality System (presented in Figure 0-1) comprises three levels - Policy,
Organization/Program, and Project:
• Policy - this level addresses Agency-wide quality policies and regulations that both
EPA organizations and external EPA-funded organizations must address;
Organization/Program - this level addresses the management and implementation
component of the individual Quality System; and
• Project - this level addresses the project-specific components that are applied to
individual projects to ensure that the needs of the organization are met.
EPA has developed a Quality System Series of documents that provide guidelines to help
organizations ensure that data collected for the characterization of environmental processes and
conditions are of the appropriate type and quality for their intended use. Documents useful in
planning for data collection include:
• Decision Error Feasibility Trials (DEFT) Software for the Data Quality
Objectives Process (EPA QA/G-4D),
• Guidance for the Data Quality Objectives Process for Hazardous Waste Sites
(EPA QA/G-4HW),
• Guidance on Quality Assurance Project Plans (EPA QA/G-5),
• Guidance for the Preparation of Standard Operating Procedures for Quality-
Related Documents (EPA QA/G-6), and
• Guidance for Data Quality Assessment: Practical Me thods for Data Analysis
(EPA QA/G-9).
Final
EPA QA/G-4 0 - 3 August 2000
-------
0.2 Systematic Planning and the DQO Process
EPA Order 5360.1 A2 requires that all EPA organizations (and organizations with
extramural agreements with EPA) follow a systematic planning process to develop acceptance or
performance criteria for the collection, evaluation, or use of environmental data. A systematic
planning process is the first component in the planning phase of the project tier, while the actual
data collection activities are in the implementation phase of this tier (Figure 0-1).
1. Observe some aspect of the universe.
2. Invent a tentative theory or hypothesis
consistent with what is observed.
3. Use this hypothesis to make predictions.
4. Test these predictions by planned
experiment or the collection of further observations.
Draw the conclusion
that the theory is true.
5. Modify the theory or hypothesis
in light of the results.
Figure 0-2. The Scientific Method
What is systematic planning? Systematic planning is a planning process that is based on the
scientific method and includes concepts such as objectivity of approach and acceptability of
results (Figure 0-2). Systematic planning is based on a common sense, graded approach to ensure
that the level of detail in planning is commensurate with the importance and intended use of the
work and the available resources. This framework promotes communication between all
organizations and individuals involved in an environmental program. Through a systematic
planning process, a team can develop acceptance or performance criteria for the quality of the
data collected and for the quality of the decision. When these data are being used in decision
EPA QA/G-4
0-4
Final
August 2000
-------
making by selecting between two clear alternative conditions (e.g., compliance/non-compliance
with a standard), the Agency's recommended systematic planning tool is called the DQO Process.
Elements of the systematic planning process (from Section 3.3.8 of the EPA Quality Manual) and
relationship to the DQO Process are shown in Table 0-1.
Table 0-1. Elements of the Systematic Planning Process
Elements of Systematic Planning Process
Identifying and involving the project
manager/decision maker, and project personnel
Identifying the project schedule, resources,
milestones, and requirements
Describing the project goal(s) and objective(s)
Identifying the type of data needed
Identifying constraints to data collection
Determining the quality of the data needed
Determining the quantity of the data needed
Describing how, when, and where the data will be
obtained
Specifying quality assurance and quality control
activities to assess the quality performance criteria
Describing methods for data analysis, evaluation,
and assessment against the intended use of the data
and the quality performance criteria
Corresponding Step in the DQO Process
Step 1 . Define the problem
Step 1 . Define the problem
Step 2. Identify the problem
Step 3. Identify information needed for the
decision
Step 4. Define the boundaries of the study
Step 5. Develop a decision rule
Step 6. Specify limits on decision errors
Step 7. Optimize the design for obtaining data
Step 7. Optimize the design for obtaining data
Part B of QA Project Plan
Part D of QA Project Plan; DQA Process
What are acceptance or performance criteria? Acceptance or performance criteria are based on
the ultimate use of the data to be collected and needed quality assurance (QA) and quality control
(QC) practices required to support the decision. In the decision making process, these criteria
allow a user to limit decision errors to a fixed level for determining whether or not an Action
Level (regulatory or risk-based) has been exceeded.
What is the DQO Process? The DQO Process is a seven-step planning approach to develop
sampling designs for data collection activities that support decision making. This process uses
systematic planning and statistical hypothesis testing to differentiate between two or more clearly
defined alternatives. A summary of the seven steps is presented in Figure 0-3.
EPA QA/G-4
0-5
Final
August 2000
-------
Step 1. State the Problem
Define the problem; identify the planning team;
examine budget, schedule.
1
Step 2. Identify the Decision
State decision; identify study question; define
alternative actions.
1
Step 3. Identify the Inputs to the Decision
Identify information needed for the decision (information
sources, basis for Action Level, sampling/analysis method).
1
Step 4. Define the Boundaries of the Study
Specify sample characteristics; define
spatial/temporal limits, units of decision making.
1
Step 5. Develop a Decision Rule
Define statistical parameter (mean, median); specify Action
Level; develop logic for action.
1
Step 6. Specify Tolerable Limits on Decision Errors
Set acceptable limits for decision errors relative to
consequences (health effects, costs).
1
Step 7. Optimize the Design for Obtaining Data
Select resource-effective sampling and analysis plan that
meets the performance criteria.
Figure 0-3. The Data Quality Objectives Process
The DQO Process is iterative and allows the planning team to incorporate new
information and modify outputs from previous steps as inputs for a subsequent step. Although
the principles of systematic planning and the DQO Process are applicable to all scientific studies,
the DQO Process is particularly designed to address problems that require making a decision
between two clear alternatives. The final outcome of the DQO Process is a design for collecting
data (e.g., the number of samples to collect, and when, where, and how to collect samples),
together with limits on the probabilities of making decision errors.
What are DQOs? DQOs are qualitative and quantitative statements, developed using the DQO
Process, that clarify study objectives, define the appropriate type of data, and specify tolerable
EPA QA/G-4
0-6
Final
August 2000
-------
levels of potential decision errors that will be used as the basis for establishing the quality and
quantity of data needed to support decisions. DQOs define the performance criteria that limit the
probabilities of making decision errors by considering the purpose of collecting the data; defining
the appropriate type of data needed; and specifying tolerable probabilities of making decision
errors.
What projects are covered by the DQO Process? The DQO Process may be applied to all
programs involving the collection of environmental data used in decision making. The principles
used in the DQO Process are also applicable to programs with objectives other than decision
making (e.g., estimation and research studies).
Who should be included in the DQO Process? When applying the DQO Process, a planning
team of senior program staff, technical experts, managers, data users (usually with some statistical
expertise), a quality assurance specialist, regulators, and stakeholders are usually involved. It is
important that the key persons participate (or stay informed) throughout the DQO Process so that
each individual understands the problem/decision and objectives of the decision-making process.
Individuals with specific areas of technical expertise may decide to be involved only in the steps of
the DQO Process that require technical input.
When should the DQO Process be used? The DQO Process should be used during the planning
stage of any study that requires data collection, before the data are collected. As the DQO
Process is iterative by nature, steps within the process can be revisited before a final decision is
reached. As shown in Figure 0-4, the planning team may choose to revisit selected parts of the
DQO Process or to investigate the entire process cyclically.
Is the DQO Process only applicable to large studies or studies that require multiple decisions?
The DQO Process applies to any study, regardless of its size. However, the depth and detail of
DQO development will depend on the study objectives. The DQO Process is particularly
applicable to a study in which multiple decisions must be reached because, by using this planning
process, the planning team can clearly separate and delineate data requirements for each
problem/decision. For projects that require multiple decisions or answers to more than one
question, it is likely that the resolution of one decision will lead to the evaluation of subsequent
decisions. In these cases, the DQO Process can be used repeatedly throughout the life cycle of a
project. Often, the decisions that are made early in the project will be preliminary in nature; they
might require only a limited planning and evaluation effort. As the study nears conclusion and the
consequences of making a decision error become more critical, however, the level of effort
needed to resolve a decision generally will become greater. Figure 0-4 illustrates this point.
What are the outputs of the DQO Process? The DQO Process leads to the development of
acceptance or performance criteria based on the ultimate use of the data to be collected and define
the quality required for the decision in terms of acceptance limits on the probabilities of
committing a decision error. Each step of the DQO Process defines criteria that will be used to
Final
EPA QA/G-4 0 - 7 August 2000
-------
STUDY PLANNING
COMPLETED
START
DEVELOPING
DQOs
STUDY PLANNING
COMPLETED
STUDY PLANNING
COMPLETED
APPROACH
INCREASING LEVEL OF EFFORT
Figure 0-4. Repeated Application of the DQO Process throughout the Life Cycle of
a Project
establish the final data collection design. The first five steps of the DQO Process are primarily
focused on identifying qualitative criteria, such as:
the nature of the problem that has initiated the study and a conceptual model of the
environmental hazard to be investigated;
the decisions that need to be made and the order of priority for resolving them;
• the type of data needed (i.e., geographic area, environmental medium, overall
timing of data collection, etc.); and
• a decision rule that defines how the data will be used to choose among alternative
actions.
The sixth step defines quantitative criteria, expressed as limits on the probability or chance
(risk) of making a decision error, that the decision maker can tolerate. The seventh step is used to
develop a data collection design based on the criteria developed in the first six steps. In this step
the planning team considers the final product of the DQO Process, a data collection design that
meets the quantitative and qualitative needs of the study using a specified number of samples that
can be accommodated by the budget available. The outputs of the DQO Process are used to
develop a QA Project Plan and for performing Data Quality Assessment (Chapter 8).
What is a data collection design? A data collection design specifies the number, location,
physical quantity, and type of samples that should be collected to satisfy the DQOs. The sampling
design designates where, when, and under what conditions samples should be collected; what
variables are to be measured; and the QA and QC activities that will ensure that sampling design
EPA QA/G-4
0-8
Final
August 2000
-------
and measurement errors are managed sufficiently to meet the tolerable decision error rates
specified in the DQOs. These QA and QC activities together with details of the data collection
design are documented in the QA Project Plan.
Can existing data be used in the DQO Process to support your decision making? Existing data
can be very useful. For example, pilot studies are often performed to provide a preliminary
assessment of variability. In these cases, the existing data may provide valuable information to
help develop a design for collecting data. It is critical to examine the existing data to ensure that
their quality is acceptable for use, or for integration into a new data set. Some considerations
include:
• determining if the existing data were collected within approximately the same
spatial and temporal boundaries as the new data;
• examining the existing data to determine if this data set includes identical media
and analytes;
• examining the performance of the analytical methods for the existing data
(accuracy, precision, detection limits) and comparing this to the specifications in
Step 3 of the DQO Process for new data to be collected; and
examining the variability among samples in the existing and new data sets.
Combining existing data and new data can be a very complex operation and you should
undertake this with great care. In many cases, statistical expertise is required to evaluate both
data sets before they can be combined with confidence.
Will you always develop statistical/probabilistic sampling designs for data collection if you use
the DQO Process? No. Although statistical methods for developing the data collection design
are strongly encouraged, this guidance recognizes that not every sampling problem can be
resolved with probabilistic sampling designs. However, the DQO Process can and should be used
as a planning tool for studies even if a statistical data collection design ultimately will not be used.
In these cases, the planning team is encouraged to seek expert advice on how to develop a non-
statistical data collection design and how to evaluate the results of the data collection. When
nonprobabilistic, judgmental, or quota sampling methods are used, be sure to consult with an EPA
representative to ensure that program-specific QA requirements are satisfied.
How should you use this guidance? You should use this guidance as a tool to structure the
planning activities for collecting environmental data. It should be used to organize meetings,
focus the collection of background information, and facilitate communication between a team that
includes technical experts, program managers, stakeholders, regulators, and decision makers.
0.3 Benefits of Using the DQO Process
The DQO Process integrates a multidisciplinary team and offers the advantages of using
experience and resources of individuals who have different backgrounds, different kinds of
Final
EPA QA/G-4 0 - 9 August 2000
-------
knowledge, and who can collectively focus on achieving a successful project conclusion. During
the initial planning stages, the planning team can concentrate on developing requirements for
collecting the data and work to reach consensus on the type, quantity, and quality of data needed
to support Agency decisions. This interaction results in a clear understanding of the problem and
the options available for addressing it, the development of acceptance or performance criteria for
decision making, a consensus-based approach to understanding the problem, and data being
collected of appropriate quality. Organizations that have used the DQO Process have observed
that:
• The structure of the DQO Process provides a convenient way to document
activities and decisions and to communicate the data collection design to others.
This documentation facilitates rapid review and approval by regulators and
stakeholders.
• The DQO Process enables data users and relevant technical experts to participate
collectively in data collection planning and to specify their particular needs prior to
data collection. The DQO process fosters communication among all participants,
one of the central tenets of quality management practices, and directs efforts to
achieving consensus between decision makers, stakeholders, and regulators.
• The DQO Process helps to focus studies by encouraging data users to clarify
vague objectives and to limit the number of decisions that will be made. Due to
this clarification, the consequences of decision errors are examined and correct
decisions will be made most frequently when the DQO Process is employed.
• The DQO Process is a planning tool that can save resources by making data
collection operations more resource-effective. Good planning will streamline the
study process and increase the likelihood of efficiently collecting appropriate and
useful data.
• The DQO Process provides a method for defining decision performance
requirements that are appropriate for the intended use of the data. This is done by
considering the consequences of decision errors and then placing tolerable limits
on the chance that the data will mislead the decision maker into committing a
decision error. A statistical sampling design can then be generated to provide the
most efficient method for managing decision errors and satisfying the DQOs.
Upon implementing the DQO Process, your environmental programs may be strengthened by:
• focused data requirements and optimized design for data collection,
• use of clearly developed work plans for collecting data in the field,
• uniformly documented data collection, evaluation, and use,
• clearly developed analysis plans,
Final
EPA QA/G-4 0 - 10 August 2000
-------
• sound, comprehensive quality assurance project plans, and
up-front buy-in by stakeholders to the sampling design and data collection process.
This can lead to:
• rapid review by regulators and other stakeholders,
• defensible results on which to base decisions,
• increased credibility with regulators and stakeholders, and
• a better use of resources.
Where else can the DQO Process be applied? The DQO Process is widely applicable. For
example, the Department of Energy Environmental Management program considers the following
potential applications for the DQO Process (Grumley, 1994):
• Waste management
S Characterizing waste, using process knowledge verified by minimal
sampling/ analysis data to meet acceptance criteria for treatment, storage,
and disposal.
S Designing optimal monitoring networks for ground water and surface
water discharges, and air emissions.
• Environmental restoration
S Focusing regulatory and public concerns associated with remediation.
S Identifying target analytes of concern for remedial activities.
S Determining when remediation has met cleanup levels.
• Facility transition and management
S Performing characterization assessments, using existing information or
collecting new data, to verify facilities for environmental management
acceptance.
S Evaluating alternative end-state conditions and planning facility
deactivation in preparation for eventual decontamination and
decommi ssioning.
S Designing optimized short- and long-term environmental monitoring.
• Decontamination and decommissioning
S Determining the location and levels of facility contamination.
S Determining when decontamination and decommissioning is complete.
Final
EPA QA/G-4 0-11 August 2000
-------
• Technology development
S Determining what constitutes and acceptably demonstrates success in
technology development and evaluation.
0.4 Organization of This Document
This document provides EPA's guidance specific to the design plans for collecting data for
decision-making activities. EPA recognizes that by using systematic planning and the DQO
Process to design environmental data collection efforts, the effectiveness, efficiency, and
defensibility of decisions will be improved. This document presents:
• the advantages of using systematic planning for data collection,
the seven steps of the DQO Process, including activities and outputs for each step,
and
three scenarios that each use a different statistical parameter (mean, median, and
upper percentile) to develop a design for collecting data.
The objective of this guidance document is to describe how a planning team can use the DQO
Process to generate a plan to collect data of appropriate quality and quantity for defensible
decision making. This guidance replaces in its entirety EPA's September 1994 document,
Guidance for the Data Quality Objectives Process (EPA QA/G-4), (U.S. EPA, 1994a), and is
consistent with the Data Quality Objectives Process for Hazardous Waste Site Investigations
(EPA QA/G-4HW) (U.S. EPA, 1999).
This document contains an introductory chapter that is followed by seven chapters that
correspond to the seven steps of the DQO Process. Each chapter is divided into four sections:
1. Background — Provides background information on the DQO Process step,
including the rationale for the activities in that step and the objective(s) of the
chapter.
2. Activities — Describes the activities recommended for completing the DQO
Process step, including how inputs to the step are used.
3. Outputs — Identifies the results that may be achieved by completing the DQO
Process step.
4. Examples — Presents outputs from two different DQO scenarios for
environmental contamination.
Chapter 8 shows how outputs of the DQO Process are used to develop a QA Project Plan.
Appendix A shows the derivation of the formula used to calculate sample size, and
Appendix B gives a Bibliography of referenced books, papers, and publications. Appendix C
Final
EPA QA/G-4 0 - 12 August 2000
-------
shows a complete DQO example using the median as the parameter of interest and Appendix D
contains a glossary of terms used in this document.
0.5 Background for the Two Examples
The following examples have been derived from real-life DQO development efforts to
illustrate the use of mean and percentile in planning for decision making:
Example 1 - Use of the mean to make a decision about waste disposal of material.
Example 2 - Use of the percentile to make a decision relative to a regulatory limit value.
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
Cadmium is primarily used for corrosion protection on metal parts of cars, electrical
appliances, and in some batteries. Cadmium and cadmium salts have been shown to be
toxic to humans through both ingestion and inhalation. Ingestion of concentrations as
low as 0.1 mg/kg/day causes mild to severe irritation of the gastrointestinal tract.
Exposure from chronic (long-term) inhalation can cause increased incidence of
emphysema and chronic bronchitis, as well as kidney damage.
A waste incineration facility located in the Midwest routinely removes fly ash from its
flue gas scrubber system and disposes of it in a municipal landfill. Previously the waste
fly ash was determined not to be hazardous according to RCRA program regulations.
The incinerator, however, recently began accepting and treating a new waste stream.
The representatives of the incineration company are concerned that the waste fly ash in a
new waste stream could contain hazardous levels of cadmium from new waste sources.
They have decided to test the ash to determine whether it should be sent to a hazardous
waste landfill or continue to be sent to the municipal landfill.
As a precursor to the DQO Process, the incineration company has conducted a pilot
study of the fly ash to determine the variability in the concentration of cadmium within
loads of waste fly ash leaving the facility and has determined that each load is fairly
homogeneous. There is considerable variability between loads, however, due to the
nature of the waste stream. The company has decided that testing each container load
before it leaves the facility would be an economical approach to evaluating the potential
hazard. They could then send containers of ash that exceeded the regulated standards to
the higher cost RCRA landfills and continue to send the other containers to the
municipal landfill. This example demonstrates use of the mean as the population
parameter of concern. (The derivation of a sampling design using the mean is provided
in Appendix A).
Final
EPA QA/G-4 0-13 August 2000
-------
Example 2. Making Decisions About Urban Air Quality Compliance
In July 1997, the EPA established new ambient air quality standards for PM25,
paniculate matter smaller than 2.5 microns (40 CFR 50). PM25 is comprised of fine
particles about I/30th the thickness of a human hair that are a complex mixture of acids,
metals, and carbon. Because the health risks of the chemical components ofPM25 are
not fully understood, EPA is implementing PM25 standards and investigating scientific
uncertainties associated with these components.
This example involves monitoring urban air for the presence ofPM25. Representatives
of a primary metropolitan statistical area (PMSA) in the northeast wish to determine
whether their PMSA is in attainment for PM25 according to the National Ambient Air
Quality Standards (NAAQS). If determined to be in nonattainment, control strategies
will be implemented for the PMSA, as defined in its associated State Implementation
Plan (SIP). This example uses an upper percentile as the primary population parameter
of concern as it is specified in the Standards. Additionally, this example highlights DQO
activities and outputs for the case when a data collection design (i.e., number of
samples) already has been determined, but not necessarily in accordance with the DQO
Process.
Final
EPA QA/G-4 0 - 14 August 2000
-------
CHAPTER 1
STEP 1. STATE THE PROBLEM
1.
State
the
Problem
2. Identify the Decision"
3. Identify the Inputs to
4. Define the Boundaries of the
5. Develop a Decision Rule
6. Specify Tolerable Limits on Decision
Errors
7. Optimize the Design for Obtaining Data
1. State the Problem
Identify the planning team members
including decision makers.
Describe the problem; develop a
conceptual model of the
environmental hazard to be
investigated.
Determine resources - budget,
personnel, and schedule.
After reading this chapter you should understand how to assemble an
effective planning team and how to describe the problem and examine
your resources for investigating it.
1.1 Background
The first step in any systematic planning process is to define the problem that has initiated
the study. As environmental problems are often complex combinations of technical, economic,
social, and political issues, it is critical to the success of the process to separate each problem,
define it completely, and express it in an uncomplicated format. A proven effective approach to
solving a problem is to use a planning team composed of experts and stakeholders who have
multidisciplinary backgrounds. A team of individuals with diverse backgrounds offers:
the ability to develop a complete, concise description of complex problems,
multilateral experience and awareness of potential data uses.
and
If there is a potential that the data collected in the current investigation could be used in future
studies (secondary uses of the data), it is important to consult, if possible, potential data users
during the planning process.
EPA QA/G-4
1 -1
Final
August 2000
-------
1.2 Activities
The most important activities in this step are to:
• establish the planning team including the decision makers;
describe the problem and develop a conceptual model of the environmental hazard
to be investigated; and
identify available resources, constraints, and deadlines.
How do you establish the planning team and decision makers? The DQO planning team is
usually composed of the project manager, technical staff, data users (including those with a
statistical background), and stakeholders. It is important to carefully select the planning team and
leaders (or decision makers) because this team will work together through all seven steps of the
planning process. The development of DQOs does not necessarily require a large planning team,
particularly if the problem appears to be straightforward. The size of the planning team is usually
directly proportional to the complexity and importance of the problem. As the DQO Process is
iterative, team members may be added to address areas of expertise not initially considered.
Prior to or during the first meeting of the DQO team, members should identify the
decision makers. The decision maker may be one or more individuals familiar with the problem,
or with a vested interest in it. As the technical project manager is familiar with the problem and
the budgetary/time constraints the team is facing, he or she will usually serve as one of the
decision makers and will actively participate in all steps of DQO development. The decision
makers will have the ultimate authority for making final decisions based on the recommendations
of the planning team. In cases where the decision makers cannot attend DQO planning meetings,
alternate staff members should attend and keep the decision makers informed of important
planning issues.
The technical staff and data users should include individuals who are knowledgeable
about technical issues (such as geographical layout, sampling constraints, analysis, statistics, and
data interpretation). The planning team of multidisciplinary experts may include quality assurance
managers, chemists, modelers, soil scientists, engineers, geologists, health physicists, risk
assessors, field personnel, regulators, and data users with statistical experience.
Stakeholders are individuals or organizations who are directly affected by a decision,
interested in a problem, and want to be involved, offer input, or seek information. Usually
stakeholders will have multiple perspectives about a problem. The involvement of stakeholders
early on in the DQO Process can provide a forum for communication as well as foster trust in the
decision making process. An environmental example is the Common Sense Initiative Council, a
group of stakeholders convened to offer EPA advice and recommendations on a number of topics.
The Common Sense Initiative Council recognizes that involving stakeholders improves
communication and assists in analyzing situations to determine the tools and expertise needed to
address problems and maintain lasting agreements.
Final
EPA QA/G-4 1 - 2 August 2000
-------
The identification of stakeholders is influenced by the issues under consideration, as well
as the ability of stakeholders to articulate their interests. Because EPA is organized into multiple
program areas that are concerned with different environmental media that address different
regulatory areas (e.g., Clean Air Act, Resource Conservation and Recovery Act), stakeholder
involvement activities are not centralized. EPA has developed a web page Introduction to
Stakeholder Involvement (http://www.epa.gov/ooaujeag/stakeholders/people.htm) that identifies
individuals in various EPA program offices who can offer assistance in stakeholder involvement
activities. EPA provides additional information/resources on stakeholder involvement, including:
• EPA Resources for Non-Profit Organizations,
• Children's Health Protection Advisory Committee,
EPA Voluntary Programs, and
• Partners of Wetlands, Oceans, and Watersheds.
Information for stakeholder involvement and consensus building processes for other federal
agencies is also provided at this website. At the state level, information on potential stakeholders
is often available. For example, the State of California has developed a directory of citizen
groups, government agencies, and environmental education programs concerned with California
environmental issues (Harbinger Communications, 1996).
You should identify the roles of team members and group members that have key and secondary
roles, then consider the roles of the planning team members when coordinating meetings. While it
is important for key members (e.g., decision makers and members involved in policy decisions) to
either attend all meetings, or designate a representative to attend meetings that are missed,
technical members (e.g., technical managers, field and laboratory personnel, data users,
statisticians) may decide to be involved only in meetings where technical input is required.
Stakeholders and regulators may elect to attend initial meetings, but miss meetings that address
technical issues (e.g., sampling and analysis). When possible, the use of a facilitator or recorder at
these meetings is encouraged.
How do you describe the problem and the environmental hazard to be investigated? In Step 1,
the planning team describes the conditions or circumstances that are causing the problem and the
reasons for undertaking the study. Typical examples for environmental problems include
conditions that may pose a threat to human health or the environment or circumstances of
potential noncompliance with regulations.
The team may be able to describe the problem as it is currently understood by briefly
summarizing existing information, or they may conduct literature searches and examine past or
ongoing studies. This will ensure that the problem is correctly defined and has not been solved
previously. As you define the problem, you should consider similar studies and document
information about the performance of sampling and analytical methods observed in these studies.
This information may prove to be particularly valuable later in the DQO Process. You should
Final
EPA QA/G-4 1 - 3 August 2000
-------
organize and review all relevant information, indicate the source of the information, and evaluate
its reliability.
The planning team should:
examine the study objectives from a regulatory standpoint as necessary;
• identify individuals or organizations who are involved or have an interest in the
study;
• examine political issues associated with the study;
look at results of similar studies performed previously from the standpoint of:
S study parameters,
S regulatory or other constraints on sampling designs,
S variability and quality of data collected;
consider non-technical issues that may influence the sample design; and
• examine possible future uses of the data to be collected (e.g., the data to be
collected may be eventually linked to an existing database).
It is critical to carefully develop an accurate conceptual model of the environmental
problem in this step of the DQO Process, as this model will serve as the basis for all subsequent
inputs and decisions. Errors in the development of the conceptual model will be perpetuated
throughout the other steps of the DQO Process and are likely to result in developing a sampling
and analysis plan that may not achieve the data required to address the relevant issues.
The conceptual model of the potential environmental hazard developed at the beginning of
the DQO Process is often a diagram that shows:
known or expected locations of contaminants,
• potential sources of contaminants,
media that are contaminated or may become contaminated, and
• exposure scenarios (location of human health or ecological receptors).
If the problem is complex, the team may consider breaking it into more manageable pieces,
which might be addressed by separate studies. Priorities may be assigned to individual segments
of the problem and the relationship between the segments examined.
How do you identify available resources, constraints, and deadlines? You should examine
limitations on resources and time constraints for collecting data. This estimate should include
developing acceptance or performance criteria, preparing the QA Project Plan, collecting and
analyzing samples, and interpreting data. At this time the planning team should also examine
available personnel, and contracts (if applicable) and identify intermediate and final deadlines for
collecting data.
Final
EPA QA/G-4 1 - 4 August 2000
-------
1.3 Outputs
The major outputs of this step are:
• a list of the planning team members and their roles,
identification of decision makers,
• a concise description of the problem and a conceptual model of the environmental
problem to be investigated, and
• a summary of available resources and relevant deadlines for the study including
budget, availability of personnel, and schedule.
1.4 Examples
Given the background of the three examples as outlined in Section 0.5, the following
DQO Step 1 outputs were derived.
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
How were the planning team members selected? The planning team included the
incineration plant manager, a plant engineer, a quality assurance specialist with
statistical experience, and a chemist with sampling experience in the RCRA program.
The plant manager was the decision maker.
How was the problem described and a conceptual model of the potential hazard
developed? The problem was described as determining which container loads of waste
fly ash from a new waste stream needed to be sent to a RCRA landfill as a result of a
change in an incinerator process that possibly increased the levels of cadmium in waste
fly ash. The plant manager wanted to avoid expensive RCRA disposal of waste, if
possible, but also needed to comply with regulations and permits.
The conceptual model described fly ash that was created from industrial waste
incineration and is a potential source of toxic metals that include cadmium. Ash is
transferred to large containers via a conveyer belt. Containers are filled and trucked to
a disposal site. If the waste fly ash is hazardous but disposed in a municipal (sanitary)
landfill, then metals can leach into ground water and create runoff to streams, and other
surface water bodies, which could pose a hazard to human health and ecological
receptors. If such waste is disposed in a RCRA approved landfill, the hazards are
contained.
What were the available resources and relevant deadlines? Although the project was
not constrained by cost, the waste generator (the incineration company) wished to hold
sampling costs below $2,500. The incineration company also requested that the testing
of the waste fly ash in each container be completed within one week.
Final
EPA QA/G-4 1 - 5 August 2000
-------
Example 2. Making Decisions About Urban Air Quality Compliance
How were the planning team members selected? The planning team included senior
program staff, technical experts, senior managers, a QA specialist, and an individual
with statistical experimental design expertise. The most senior program staff member
served as the decision maker.
How was the problem described and a conceptual model of the potential hazard
developed? EPA had setNAAQSfor fine paniculate matter (PM2^) and other air
pollutants (40 CFR 50). The problem was described as determining whether the primary
metropolitan statistical area (PMSA) of concern was in attainment for fine paniculate
matter.
The conceptual model of the potential hazard was considering the concentration of fine
particulates in urban air that were primarily combustion products from point and mobile
sources. The particulates posed potential sources of exposure from inhalation. As a
rule, the PMSA was not concerned with long-term transport because over time
particulates aggregated or became deposited on other materials such that the particles
came within the purview of the Pmlorule. The PMSA developed a Cartesian map
indicating local PM25 point sources, main roadways, and predominant wind patterns to
identify areas of maximum potential exposure.
What were the available resources and relevant deadlines? The monitoring network
was already in place. It consisted of three fixed-site multiple filter gravimetric devices
for measuring daily concentrations (24-hr average) once every 3 days. Thus, about 365
readings were obtained each year.
Looking Ahead to other DQO Steps:
• Careful description of the problem will assist in Step 3, Identify the
Inputs to the Decision, when considering additional use of data (link to
databases, etc.).
• The conceptual model will assist in Step 4, Define the Boundaries of the
Study, when
S establishing spatial boundaries, and
S considering regulatory and practical constraints for sampling.
Final
EPA QA/G-4 1 - 6 August 2000
-------
CHAPTER 2
STEP 2. IDENTIFY THE DECISION
The DQO Process
1. State the Problem
2. Identify the Decision
3. Identify the Inputs to the Decision
4. Define the Boundaries of the StuC
5. Develop a Decision Rule
6. Specify Tolerable Limits on Decision
Errors
7. Optimize the Design for Obtaining Data
2. Identify the Decision
• Identify the principal study question.
• Define alternative actions.
• Develop a decision statement.
• Organize multiple decisions.
After reading this chapter you should know how to identify the principal study
question and how to define options for addressing it (alternative actions).
2.1 Background
This step builds on the output of the previous step where you have:
• identified members of a planning team, including decision makers;
concisely described the problem; and
• developed a conceptual model of the environmental problem to be investigated.
In Step 2 of the DQO Process, you should identify the key question that the study
attempts to address and alternative actions that may be taken, depending on the answer to the key
study question. Then you are able to combine these two elements to develop a decision
statement. The decision statement is critical for defining decision performance criteria later in
Step 6 of the Process.
In cases of multiple or complex problems, you should identify multiple decisions, organize
the decisions sequentially (or logically), and examine the decisions to ensure consistency with the
statement of the problem in Step 1. If the principal study question is not obvious and specific
alternative actions cannot be identified, then the study may fall in the category of exploratory
research, in which case this particular step of the DQO Process may not be needed.
EPA QA/G-4
2- 1
Final
August 2000
-------
2.2 Activities
In this step you should:
• identify the principal study question;
define alternative actions;
• combine the principal study question and alternative actions into a decision
statement and state each decision in terms of whether to take action. In some
cases, this decision statement will be based on regulatory guidelines; and
organize multiple decisions into an order of priority.
How do you identify the principal study question? Based on a review of the problem described
in Step 1, you should identify the principal study question and state it as specifically as possible.
A specific statement of the principal study question focuses the search for information needed to
address the problem. The principal study question identifies key unknown conditions or
unresolved issues that reveal the solution to the problem being investigated. EPA recommends
that initially you should concentrate on only one principal study question and expand to other
issues later. The following are examples of typical principal study questions:
• Does the concentration of contaminants in ground water exceed acceptable levels?
Does the pollutant concentration exceed the National Ambient Air Quality
Standard?
Does a contaminant pose a human health or ecological risk?
• Is the contaminant concentration significantly above background levels (suggesting
a release)?
In each case, the answer to the principal study question will provide the basis for determining the
course of action that should be taken to solve the problem.
What are alternative actions and how should you define them? During this step, the planning
team should identify the possible actions that may be taken to solve the problem, including an
alternative that requires no action. The team should confirm that the alternative actions can
resolve the problem (if it exists) and determine whether the actions satisfy regulations. An
example of a principal study question and alternative actions is given in Table 2-1.
Table 2-1. An Example of a Principal Study Question and Alternative Actions
Principal Study Question
Are there significant levels of lead in floor dust at a
children's residence.
Alternative Actions
Remove the children from the residence.
Initiate a clean-up removal of lead -based
paint.
Take no action.
EPA QA/G-4
2-2
Final
August 2000
-------
How do you develop a decision statement? After examining the alternative actions, you should
combine the principal study question and alternative actions into a decision statement that
expresses a choice among alternative actions. The following template may be helpful in drafting
decision statements:
Determine whether or not [unknown environmental conditions/issues/criteria from
the principal study question] require (or support) [taking alternative actions].
Does the DQO Process address multiple decisions? If several separate decision statements must
be defined to address the problem, you should examine how the decisions relate to one another
and prioritize them in the order of the importance and sequence for resolving them. It may be
helpful to document the prioritizing process proposed to resolve the problem using a diagram or a
flow chart. An example is presented in Figure 2-1.
Figure 2-1. An Example of the DQO Process Applied to
Multiple Decisions for a Hazardous Waste Investigation
EPA QA/G-4
2-3
Final
August 2000
-------
2.3 Outputs
The output for this step is a decision statement that links the principal study question to
possible actions that will solve the problem.
2.4 Examples
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
What was the Decision Statement? The decision statement was determining whether
waste fly ash was hazardous under RCRA regulations.
What were the alternative actions? If the waste was hazardous, disposal in a RCRA
landfill was required. If it was not, the team decided that disposal in a sanitary landfill
was acceptable.
Example 2. Making Decisions About Household Dust for Lead Hazard Assessment
What was the Decision Statement? The decision statement was determining if there
were significant levels of lead in floor dust at the residence.
What were the alternative actions? If yes, the team planned follow-up testing to
determine whether immediately dangerous contamination existed and the location of the
contamination in the property. If no, the team decided there was not a potential lead
hazard, and testing was discontinued.
Example 3. Making Decisions About Urban Air Quality Compliance
What was the decision statement? The decision statement was determining if the PMSA
of concern was in attainment for PM25.
What were the alternative actions? If yes, monitoring was continued. If no, monitoring
was continued and the PM2i control strategies outlined in the State Implementation Plan
(SIP) were implemented.
Looking Ahead to other DQO steps:
The principal study question will be used in constructing the baseline and
alternative conditions in Step 6.
• Alternative actions will form the basis for determining the potential
consequences of committing a decision error as addressed in Step 6.
Final
EPA QA/G-4 2 - 4 August 2000
-------
CHAPTER 3
STEP 3. IDENTIFY THE INPUTS TO THE DECISION
6.
7.
The
1. State the Problem
2. Identify the Decision
13. Identify the Inputs to the Decision
4. Define the Boundaries of the Study
5. Develop a Decision Rule
Specify Tolerable Limits on Decision
Errors
Optimize the Design for Obtaining Data
3. Identify the Inputs
to the Decision
Identify the information needed.
Determine sources for this information.
Determine the basis for determining the
Action Level.
Identify sampling and analysis methods
that can meet the data requirements.
After reading this chapter you should know the kinds of information
that are required to investigate the problem and whether appropriate
sampling and analytical methods are available.
3.1 Background
This step builds on the previous steps where you have:
• identified members of a planning team, including decision makers;
concisely described the problem;
• developed a conceptual model of the environmental problem to be investigated;
and
• identified the decision that needs to be made.
In Step 3 of the DQO Process you should identify the kind of information that is needed to
resolve the decision statement and potential sources of this information (new data or existing
data). This information should include the decision values (e.g., concentration of contaminants)
information about its derivation. You should also determine if the appropriate analytical
methodology exists to measure the environmental characteristics. Once you have determined
what needs to be measured, you may refine the specifications and criteria for these measurements
in later steps of the DQO Process.
EPA QA/G-4
- 1
Final
August 2000
-------
3.2 Activities
In this step you should:
• identify the kinds of information needed;
identify the sources of information;
• determine the basis for setting the Action Level; and
confirm the appropriateness of proposed sampling and analyses methods.
How do you identify the kinds of information that you will need? You may identify
information needs by asking the following questions:
Is information on the physical properties of the media required?
• Is information on the chemical or radiological characteristics of the matrix needed?
Can existing data be used to make the decision?
• Do we need to make new measurements of environmental characteristics?
If you decide that new measurements are needed, you should develop a list of characteristics that
need to be measured to make the decision. For example, if the information can be obtained as an
output from an environmental model (e.g., ground water transport), then the list of characteristics
should include the inputs required for the model.
If the decision can be based on existing data, then the sources of these data should be
examined to the extent possible to ensure that they are acceptable. If you consider integrating
new data with existing data, parameters in the existing database need to be examined so that new
samples can be collected (or analyzed) in a similar way and that the databases for new and existing
data include common parameters. In some cases, statistical expertise is required to evaluate
databases for possible aggregation because data collected for different purposes may not be
compatible. For example, studies that model exposure to environmental contaminants may link
environmental, lexicological, biological, geological, and census data. In these cases, issues such
as physical properties of contaminants, environmental media, ingestion and inhalation rates,
cancer slope factors, plant uptake rates, meteorological conditions, latitude, longitude, location of
population centers and water bodies, and population density are inputs for evaluating exposure to
the contaminant. Meta-data analysis offers the planning team options for using existing databases
in conjunction with newly collected data. Existing data will also be evaluated quantitatively in
Step 7, Optimize the Design for Obtaining Data.
How should you identify the source of the information? You should identify and document the
sources for the information needed to resolve the decision. These sources may include results of
previous data collections, historical records, regulatory guidance, professional judgment, scientific
literature, or new data collections.
Final
EPA QA/G-4 3 - 2 August 2000
-------
How do you determine the basis for the Action Level? The value for action is the threshold
value (chosen in Step 5 of the DQO Process) that provides the criterion for choosing among
alternative actions (e.g., whether to take action or not to take action or whether to choose action
1 versus action 2). Action Levels are concentrations of contaminants that are either based on
regulatory requirements, based on risk assessments, based on performance criteria for analytical
methodology (limitations of technology), or based on a reference standard. In this step, it is
important for you to understand how the Action Level will be derived. In other words, you need
to understand what information will be used to determine the Action Level, such as a promulgated
regulation or a project-specific risk assessment. The actual numerical value of the Action Level
need not be specified until DQO Process Step 5, Develop a Decision Rule, but a potential Action
Level should be established. If the Action Level is based on a regulatory requirement, then the
planning team will know the numerical value of the Action level at this step. However, if the
Action Level is based on a risk assessment or other performance criterion, it may be best to defer
the specification of the numerical value until after the study boundaries have been specified in
DQO Process Step 4.
If the decision will be made relative to background concentrations (rather than a
quantitative limit), then you should determine what constitutes background. Characteristics of the
background need to be consistent with the characteristics of the area to be investigated. The
actual numerical value of the Action Level will be established in Step 5, Develop a Decision Rule.
How should you identify that sampling and analysis methods that can meet the data
requirements? Using the list of environmental characteristics that pertain to the decision, you
should develop a list of sampling and analytical methods that may be appropriate for the problem
being investigated. For example, you should specify sampling considerations (e.g., quantities)
required for detecting analytes at low concentrations and procedures required to collect these
sample quantities. You should also identify analytical methods that have appropriate detection
limits (the minimum concentration that can be measured and reported with a specific confidence
that the analyte concentration is greater than zero). Detection limits are analyte-, matrix- and
instrument-specific. For example, atomic absorption spectroscopy or inductively coupled plasma
emission spectrometry may not be sensitive enough to measure lead levels in water samples;
however, graphite furnace atomic absorption spectroscopy would be capable of making these
measurements.
Great importance should be given to the problem of minimizing bias as this is an important
performance characteristic of sampling and analysis. The decision error rates to be established in
Step 6 of the DQO Process rely on bias being kept to a minimum. Six major causes of bias have
been identified for environmental sampling and analysis (1) non-representative sampling; (2)
instability of samples between sampling and analysis; (3) interferences and matrix effects in
analysis; (4) inability to determine the relevant forms of the parameter being measured; (5)
calibration; and (6) failure to blank-correct. Some of the EPA methods are particularly subject to
bias in calibration. For example, EPA methods for analyses of phenols in water exhibit around
50% bias due to calibration. Methods known to exhibit large biases should be avoided if possible.
Final
EPA QA/G-4 3 - 3 August 2000
-------
Additional considerations include requirements for certification of personnel, and
laboratory accreditation or Performance-Based Measurement Systems (PBMS). Laboratories
analyzing environmental samples should follow standard protocols and procedures or use
performance-based methods. When measurement requires the analysis of chemical, biological, or
radioactive samples, it is advisable to select a laboratory that is accredited to perform the analyses.
Requirements for accreditation include having qualified personnel, appropriate instrumentation,
standard operating procedures, and proficiency in the analysis of samples for specific analytes or
programs. For example, laboratories analyzing lead in paint, dust, and soil samples must be
accredited through the National Lead Laboratory Accreditation Program (NLLAP) to become
"EPA recognized." According to the Department of Housing and Urban Development's
Guidelines (HUD, 1995), "property owners, risk assessors, inspector technicians, and contractors
should ensure that laboratory analyses are performed by an 'EPA-recognized' laboratory;" a
requirement also of EPA and many States.
3.3 Outputs
The outputs from Step 3 are:
• a list of environmental characteristics that will be measured to enable the planning
team to make the decision;
• a list of information sources or methods that indicate how each Action Level will
be derived;
• a list of information that may be applicable to uses of the data in future
investigations [e.g., inputs to models, associated meta-data analysis (e.g., using
latitude, longitude, census data) that may be appropriate to use for combining
existing databases with newly collected data]; and
• confirmation that sampling and analytical methods exist (or can be developed) to
meet the detection limit criteria required for collecting data, given the appropriate
magnitude of the Action Level.
3.4 Examples
It is in this step that numerical quantities start making their appearance in the DQO
Process.
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
Identify the kind of information. To resolve the decision statement, the planning team
decided to measure the cadmium concentration in the leachate resulting from Toxicity
Characteristic Leaching Procedure (TCLP) extraction. Existing pilot study data
provided information about variability, but there was not enough information to resolve
the decision statement.
Final
EPA QA/G-4 3 - 4 August 2000
-------
Identify the source of information. The Action Level was based on RCRA toxicity
regulations for cadmium in TCLP leachate which is specified as 1.0 mg L.
What sampling and analytical methods were appropriate? Cadmium was measured
in the leachate according to the method specified in 40 CFR 261, App. II. The
detection limit was well below the Action Level.
Example 2. Making Decisions About Urban Air Quality Compliance
Identify the kind of information. To resolve the decision statement, the planning
team obtained three years ofPM2,5 concentration measurements from the existing
monitoring network within the PMSA of concern.
Identify the source of information. The 24-hr PM2,5 federal standard of 65 ^g/m3 is
attained when 98 percent of the daily concentrations, measured over three years,
are equal to or less than the standard.
What sampling and analytical methods were appropriate? The existing network
consisted of three IMPROVE samplers, each equipped with a
polytetrafluoroethylene (PTFE) membrane filter to collect aerosols for mass
measurement. Gravimetry (electro-microbalance) was used as the method of
quantitative analysis. The detection limit was well below the standard used for the
Action Level.
Looking Ahead to other DQO Steps:
• The effect of sampling methods (e.g., compositing) may affect the required
detection limit and should be considered relative to analytical measurement
methods. These issues are also considered in Steps 5 and Step 7.
• Criteria for existing data will be examined in Step 7, Optimize the Design for
Collecting Data.
Method detection limit and method quantitation limits identified in this step will
be revisited in Step 7, Optimizing the Design for Collecting Data.
Final
EPAQA/G-4 3-5 August 2000
-------
Final
EPA QA/G-4 3 - 6 August 2000
-------
CHAPTER 4
STEP 4. DEFINE THE BOUNDARIES OF THE STUDY
The DQO Process
1. State the Problem
2. Identify the Decision
3. Identify the Inputs to the Decision
14. Define the Boundaries of the Study
5. Develop a Decision Rule
6. Specify Tolerable Limits on Decision
Errors
7. Optimize the Design for Obtaining Data
4. Define the Boundaries
of the Study
• Define the target population of interest.
• Specify the spatial boundaries that
clarify what the data must represent.
• Determine the time frame for collecting
data and making the decision.
• Determine the practical constraints on
collecting data.
• Determine the smallest subpopulation,
area, volume, or time for which
separate decisions must be made.
After reading this chapter you should understand how to define the
geographic and temporal boundaries of the problem, how to examine any
practical constraints to collecting data, and factors that affect your
selection of the unit for decision making.
4.1 Background
This step builds on the previous steps where you have:
identified members of the planning team, including decision makers;
• concisely described the problem;
developed a conceptual model of the environmental problem to be investigated;
• identified the decision that needs to be made; and
identified sources of information, potential Action Levels, and possible
measurement methods that are appropriate.
In Step 4 of the DQO Process, you should identify the target population of interest and specify
the spatial and temporal features of that population that are pertinent for decision making.
It is difficult to interpret data that have not been drawn from a well-defined target
population. The term "target population" refers to the total collection or universe of objects, or
sampling units, to be studied and from which samples will be drawn. (In this context, the term
EPA QA/G-4
4-1
Final
August 2000
-------
"sample" means the individual member, or unit, of the target population that is selected and
measured or observed, such as a 100-gram scoop of soil, a cubic meter of air, a single fish, or
single radiation measurement.) The term "sampling unit" is used in the more general and
theoretical context when defining how the target population will be broken down into elementary
components or members that can be selected and measured or observed. When the target
population is made up of "natural units," such as people, plants, or fish, then the definition of a
sampling unit is straightforward. However, many environmental studies involve target
populations made up of continuous media, such as air, water, or soil. In this context, the
sampling unit must be defined as some volume or mass to be selected which is often called the
sample support (Myers, 1997). The actual determination of the optimal size of a sampling unit
for environmental data collection efforts can be complicated, and usually will be addressed as a
part of the sampling design in Step 7. Here in Step 4, the planning team should be able to provide
a first approximation of the sampling unit definition when specifying the target population.
Quite often in environmental studies the target population is the set of all possible
environmental samples (e.g., volume of soil, water, or air) that, taken together, constitute the
geographic area of interest. The purpose of this step is to unambiguously define the spatial and
temporal features of each environmental medium within a specific area or time period covered in
the decision. A clear definition of the target population and its characteristics to the decision
maker will make data interpretation more straightforward. The boundaries of the population
include:
• spatial boundaries that define the physical area to be studied and generally where
samples will be collected, and
• temporal boundaries that describe the time frame that the study will represent and
when the samples should be taken.
You should use boundaries to ensure that the data collection design incorporates the time
periods in which the study and decision should be implemented, areas where samples will be
collected, and the time period to which the decision should apply. This should help you collect
data that are representative of the population being studied. Defining boundaries before the data
are collected can also prevent inappropriate combining of data sets in a way that masks useful
information. The conceptual model that you developed in Step 1 of the DQO Process should
provide essential input into defining the spatial boundaries.
Practical constraints that could interfere with sampling should also be identified in this
step. A practical constraint is any hindrance or obstacle (such as fences, property access, water
bodies) that may interfere with collecting a complete data set. These constraints may limit the
spatial and/or temporal boundaries or regions that will be included in the study population and
hence, the inferences (conclusions) that can be made with the study data.
As the final decision depends on data that are aggregated, you should carefully identify the
size of "decision" units within which the data will be combined to make the decision. Factors
Final
EPA QA/G-4 4 - 2 August 2000
-------
such as areas of potential risk, limits of remediation technology, future land uses, and activity
patterns, may impact the size of the decision unit selected.
4.2 Activities
In this step you should:
define the target population,
• determine the spatial and temporal boundaries,
identify practical constraints, and
• define the scale of decision making.
How do you define the target population? It is important for you to clearly define the target
population to be sampled. The target population is usually the set of all environmental samples
about which the decision maker wants to draw conclusions. In a number of cases, defining the
target population for an environmental study requires specifying the medium, such as ground
water, ambient air, surface soil, etc. It may be helpful to "work backwards" and think of how you
would define an individual sampling unit when trying to develop a clear definition of the target
population.
How do you determine the spatial boundaries of the decision statement?
1. Define the geographic area applicable for the decision making.
You should define the entire geographic area where data are to be collected using
distinctive physical features such as volume, length, width, or boundaries. Some examples
of geographic areas are the metropolitan city limits, the soil within the property boundaries
down to a depth of 6 inches, a specific water body, length along a shoreline, or the natural
habitat range of a particular animal species. It is important to state as definitively as
possible the media and geographic area; this statement may include soil depth, water
depth, or distance inside a fence line. You should be careful when designating areas that
are on the periphery of the geographic area because peripheral samples are subject to edge
effects and contamination that is not associated with the spatial boundaries designated for
the decision making. In Figure 4-1 the geographic area of the study has been indicated on
a map in the area with a grid.
2. Divide the population into subsets that have relatively homogeneous
characteristics.
You may consider dividing the target population into subpopulations that are relatively
homogeneous within each area or subunit. When combined with an appropriate sampling
design in Step 7, Optimize the Design for Obtaining Data, this approach can reduce the
Final
EPA QA/G-4 4 - 3 August 2000
-------
Figure 4-1. Geographic Boundaries Delineated Using a Map
number of samples required to meet the tolerable limits on decision errors (Step 6), and,
thus, allow more efficient use of resources. It is often helpful to consider subdividing the
target population in this way at this step because the planning team is focused on their
understanding of how the target population's features and characteristics relate to the
decision. The planning team can use its knowledge of the conceptual model (developed in
Step 1, State the Problem) to consider how the characteristics of interest for the target
population vary or change over space and time. This information will be useful when
completing the subsequent activities in this step, and when considering alternative
sampling designs (such as stratified random sampling) in Step 7, Optimize the Design for
Collecting Data.
How do you determine the temporal boundaries of the decision statement?
1.
Determine when to collect data.
Conditions may vary over the course of a study because of time-related phenomena such
as weather conditions, seasons, operation of equipment under different environmental
conditions, or activity patterns. Examples of these variations include seasonal ground
EPA QA/G-4
4-4
Final
August 2000
-------
water levels, daily or hourly airborne contaminant levels in metropolitan areas, and
fluctuations in pollutant discharges from industrial sources. These variations may impact
the success of your data collection and the interpretation of data results. You should
determine when conditions are most favorable for collecting data and select the most
appropriate time period to collect data. For example, you may consider the measurement
stability of the following:
• measurement of lead in dust on window sills may show higher concentrations
during the summer when windows are raised and paint/dust accumulates on the
window sill;
terrestrial background radiation levels may change due to shielding effects related
to soil dampness;
• measurement of pesticides on surfaces may show greater variations in the summer
because of higher temperatures and volatilization;
instruments may not give accurate measurements when temperatures are colder; or
measurements of airborne paniculate matter may not be accurate if the sampling is
conducted in the wetter winter months rather than the drier summer months.
2. Determine the time frame for decision making.
It may not be possible to collect data over the full time period to which the decision will
apply. This is particularly true for decisions that project future uses, such as
"Brownfields" (an inactive property being put back into productive economic use after the
relevant environmental agencies agree that contaminants once present at the property no
longer pose an unacceptable risk to human health or to the environment). You should
evaluate the population and determine the optimum time frame for collecting data, given
that the medium may change over time, or the time constraints of the study relative to the
decision making. You should specify if you are making a decision on whether the current
medium meets a criterion, or if the medium will meet the criterion for some future time
periods. You should define time frames for the overall population and for any
subpopulation of interest; then address discrepancies that may arise from the short time
frame of data collection relative to the long time periods for implementing decisions. For
example, you may develop a statement for the decision to be based on:
• the condition of contaminant leaching into ground water over a period of a
hundred years, or
the risk conditions of an average resident over their average length of residence,
which is estimated to be 8 years.
Final
EPA QA/G-4 4 - 5 August 2000
-------
What kinds of practical constraints on collecting data should you identify? You should discuss
the proposed data collection activities in light of any practical constraints that are related to the
spatial or temporal boundaries of the study, to the availability of personnel, or to time and
budgetary constraints (identified in Step 1 of the DQO Process). These constraints could include
access to the property, availability and operation of equipment, and environmental conditions
when sampling is not possible (high humidity, freezing temperatures). For example:
• it may not be possible to take surface soil samples beyond the east boundaries of a
property under investigation because permission has not been granted by the
owner of the adjacent property, or
• it may not be possible to collect dust wipe samples (for lead) if certified risk
assessors are not available to supervise the sampling.
How do you define the scale of decision making? The scale of decision making refers to the
way the planning team has delineated decision units and identified the smallest unit of area,
volume, or time where data will be collected, analyzed, aggregated, and interpreted to make a
decision and control decision error. The consequences of making incorrect decisions (Step 6) are
associated with the size, location, and shape of the decision unit. It is important to consider
present and future uses for the decision unit, where the decision unit is located (remote area
versus densely populated area) and requirements for potential remediation. The consequences of
a wrong decision (even if quite small) should be carefully considered. For example, if a decision,
based on the data collected, results in a large land area being cleaned (soil removed to a certified
disposal area) when the true conditions would not warrant a cleanup action, then the decision
maker may have to incur a large cost unnecessarily. The area of land being sampled (decision
unit) should be appropriate to the potential risk of an incorrect decision. When establishing the
scale of decision making, take care that this is not so large that an incorrect decision could result
in either an unacceptable resource expense or unacceptable threat to human health or the
environment.
The question of using one large decision unit versus a number of small decision units is
also an important consideration for the planning team. If there are many decision units and
multiple decisions are made, then the team needs to consider whether they want to limit the
probability of leaving at least one contaminated unit unremediated (rather than just any one unit).
The chance of at least one incorrect decision increases exponentially. This is known as
"comparison-wise" versus "experiment-wise" error rates. If multiple decisions are expected, and
the planning team determines that the overall probability of making at least one decision error
must be controlled, then consultation with a statistician is advisable.
The planning team may establish decision units based on several considerations:
Risk - The scale of decision making based on risk is determined by the potential
exposure an area presents; an individual unit of risk is called an exposure unit. For
Final
EPA QA/G-4 4 - 6 August 2000
-------
example, in a study where the decision statement is, "Determine whether or not the
concentration of lead in soil poses an unacceptable health risk to children and
requires remediation," the geographic area is the top 6 inches of soil within the
property boundaries, and the population is the collection of individual volumes of
soil that could be selected for inclusion in a sample. The scale of decision making
could be the size that corresponds to the area where children derive the majority of
their exposure (such as a play area or an average residential lot size if the future
land use will be residential). Studying the area at this scale will be protective of
children, a sensitive population in risk assessment.
• Technological Considerations - A technological scale for decision making is
defined as the most efficient area or volume that can be remediated with a selected
technology. An example of a remediation unit would be the area of soil that can be
removed by available technology under estimated working conditions if the
decision will be made on the basis of bulldozer-pass-volume.
• Temporal Considerations - A temporal scale of decision making is based on
exposure from constituents in media that change over time. For example, in order
to regulate water quality, it would be useful to set a scale of decision making that
reduces the time between sampling events. Using this scale the planning team
could minimize the potential adverse effects in case the water quality changed
between sampling events.
Financial Scale - The financial scale is based on the actual cost to remediate a
specified decision unit. For example, if a large exposure unit is identified, the costs
of remediation could be prohibitive. In this case, the planning team may want to
develop a different scale to narrow the data collection process and identify the
distinct areas of contamination.
Other - The possibility of "hot spots" (areas of high concentration of a
contaminant) may be apparent to the planning team from the history of the
property. In cases where previous knowledge (or planning team judgment)
includes identification of areas that have a higher potential for contamination, a
scale may be developed to specifically represent these areas.
Further information on sampling designs and associated definitions on methods may be
obtained from Gilbert (1987) and Thompson (1992).
4.3 Outputs
The outputs of this step are:
Final
EPA QA/G-4 4 - 7 August 2000
-------
• detailed descriptions of the characteristics that define the population to be
sampled,
• detailed descriptions of geographic limits (spatial boundaries) that are appropriate
for the data collection and decision making,
• time frame appropriate for collecting data and making the decision,
list of practical constraints that may interfere with the data collection, and
• appropriate scale for decision making.
4.4 Examples
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
What population was sampled? Individual samples of fly ash that comprise the
container were sampled and analyzed. The fly ash was not mixed with any other
constituents except water (used for dust control). Each container of ash filled at least
70% of the waste container. In cases where the container was less than 70% full, the
container was kept on-site until more ash was produced and the container was filled to
capacity.
What were the spatial boundaries? Decisions applied to each container load of fly ash
waste as the actual container made a natural physical boundary.
What was an appropriate time frame for sampling? The decision was to be based on
the current concentration of cadmium in the waste fly ash. Contained in the containers,
the waste did not pose a threat to humans or the environment. Additionally, since the fly
ash was not subject to change, disintegration, or alteration, the decision about the waste
characteristics was not influenced by temporal constraints. To expedite decision
making, however, the planning team placed deadlines on sampling and reporting. The
waste fly ash was tested within 48 hours of being loaded onto waste containers. The
analytical results from each sampling round were completed and reported within 5
working days of sampling. The container was not used again until analysis had been
completed and evaluated.
What were the practical constraints for collecting data? The most important practical
constraint was the ability to take samples from the waste fly ash stored in the containers.
Although the containers had open access, special procedures and methods based on EPA
protocols were implemented so that samples were representative of the entire depth of
the waste fly ash.
What was the scale for decision making? The decision unit was each container of waste
fly ash.
Example 2. Making Decisions about Urban Air Quality Compliance
What population was sampled? The volume from samplers that represented fine
paniculate matter from urban air was sampled and analyzed.
Final
EPA QA/G-4 4 - 8 August 2000
-------
What were the spatial boundaries? The spatial boundary was defined by the region
represented by the PMSA of concern.
What was an appropriate time frame for sampling? The temporal boundaries had two
components. Individual observations (i.e., daily concentrations) were based on 24-hour
averages obtained each day of monitoring. The standard required that a decision be
made, and subsequent action taken, after 3 years of data collection. Monitoring results
were assumed to characterize both the near past (i.e., previous 3 years) and current air
quality, unless substantial upward or downward trends were observed in daily PM2.5
concentrations.
What were the practical constraints for collecting data? Given that the monitoring
network and sampling plan were already established, the only potential practical
constraint was the continual operation of the monitoring network. If a monitor became
defective, the planning team decided to either collect a smaller sample size (number of
samples) over the 3-year period, or to extend the period for collecting data to obtain the
required number of samples.
What was the scale for decision making? The decision unit was the geographic region
represented by the PMSA over the 3-year period of data collection.
Looking ahead to other DQO steps:
The way in which you divide the problem into strata may affect the
number of samples required to meet the tolerable limits for decision
errors specified in Step 6.
The scale of decision making may have an impact on the performance
criteria and the consequences of decision errors in Step 6.
• Outputs from Step 4 may potentially affect the sampling design
developed in Step 7.
EPA QA/G-4
4-9
Final
August 2000
-------
Final
EPA QA/G-4 4 - 10 August 2000
-------
CHAPTER 5
STEP 5. DEVELOP A DECISION RULE
The DQO Process
1. State the Problem
2. Identify the Decision
3. Identify the Inputs to the Decisi
4. Define the Boundaries of t
15. Develop a Decision Rule |
6. Specify Tolerable Limits
Errors
7. Optimize the Design for Obtaining Data
5. Develop a Decision Rule
• Specify an appropriate population
parameter (mean, median, percentile).
• Confirm the Action Level exceeds
measurement detection limits.
• Develop a decision rule
(If... then... statement).
After reading this chapter you should know how to construct a
theoretical "If...then... " decision rule that defines how the decision
maker would choose among alternative actions if the true state of
nature could be known with certainty.
5.1 Background
This step builds on the previous steps where you have:
identified members of the planning team, including decision makers;
concisely described the problem;
developed a conceptual model of the environmental problem to be investigated;
identified the decision that needs to be made;
identify sources of information, potential Action Levels, and possible measurement
methods that are appropriate; and
decided on the spatial/temporal boundaries of the decision.
In Step 5 of the DQO Process, you should imagine that perfect information will be
available for making decisions. Under the assumption that there is no uncertainty in the decision
making process, the planning team integrates the outputs from previous steps with inputs
developed in Step 5 into an unambiguous "If...then..." statement (theoretical decision rule). This
rule describes the conditions under which possible alternative actions would be chosen.
EPA QA/G-4
5- 1
Final
August 2000
-------
You need to conduct the following activities in this step:
• specify the population parameter (e.g., mean, median, percentile, or total amount)
that the DQO planning team considers to be important to make decisions about the
target population;
• choose an Action Level (if not already established) that sets the boundary between
one outcome of the decision process and another outcome;
• select the measurement and analysis methods capable of performing over the
expected rate of values and verify that the Action Level is greater than the
detection limit of the measurement method that will be used; and
• construct the theoretical "If...then..." decision rule by combining the true value of
the selected population parameter and the Action Level (from above) with the
scale of decision making (from Step 4) and the alternative actions (from Step 2).
This decision rule will state the alternative actions that would be taken depending
on the true value of the parameter relative to the Action Level.
Note that the "If... then..." decision rule is a theoretical rule because it is stated in terms of
the true value of the population parameter, even though in reality the true value is never known.
In practice, the decision is made by using an operational decision rule that uses an estimate (based
on the actual data) of the true value of the population parameter. The reason for specifying the
theoretical rule is to focus the attention of the DQO planning team on how they would make
decisions if they had perfect knowledge of the population. This helps clarify what the team really
wants to know to support the decision. In Step 7 of the DQO Process, the planning team will
select the operational decision rule they believe will most efficiently meet the requirements
specified in the first six steps of the DQO process.
5.2 Activities
In this step you should:
• define the population parameter;
determine what action is needed; and
• confirm that the Action Level exceeds minimum detection limits.
What population parameter best characterizes the population of interest? In this step you
should select a population parameter (such as the true mean, median, percentile, or total amount)
that summarizes the critical characteristic or feature of the population that will be compared to the
Action Level to make a decision. In some cases, the parameter that must be used may be
specified in a regulation. In other cases, the DQO planning team will select the parameter based
Final
EPA QA/G-4 5 - 2 August 2000
-------
on specific needs and considerations. A comparison of the different population parameters and
their application to a decision rule is presented in Table 5-1.
Table 5-1. Population Parameters and Their Applicability to a Decision Rule
Parameter
Definition
Example of Use
Mean
Average
Central tendency: Comparison of middle part of population to
Action Level. Appropriate for chemical that could cause cancer
after a long-term chronic exposure. Use of the mean and the total
amount of media (e.g., mass of soil or water) allows a planning
team to estimate the total amount of a contaminant contained in
the soil or water body. The mean is greatly influenced by
extremes in the contaminant distribution, and not very useful if a
large proportion of values are below the detection limit.
Median
Middle observation of
distribution; 50th
percentile; half of data
is above and half is
below
Better estimate of central tendency for a population that is highly
skewed (nonsymmetrical). Also may be preferred if the population
contains many values that are less than the measurement detection
limit. The median is not a good choice if more than 50% of the
population is less than the detection limit because a true median
does not exist in this case. The median is not influenced by the
extremes of the contaminant distribution.
Percentile
Specifies percent of
sample that is below the
given value; e.g., the
80th percentile should be
chosen if you are
interested in the value
that is greater than 80%
of the population.
For cases where only a small portion of the population can be
allowed to exceed the Action Level. Sometimes selected if the
decision rule is being developed for a chemical that can cause
acute health effects. Also useful when a large part of the
population contains values less than the detection limit. Often
requires larger sample sizes than mean or median.
It must be noted, however, that the more complex the parameter chosen, the more
complex will be the decision rule and accompanying data collection design. The most common
parameter used in decision making is the population mean because the mean is frequently used to
model random exposure to environmental contamination. Aside from scientific or policy
considerations, the mathematical and statistical properties of the mean are well understood. You
should consult a statistician if you are uncertain as to the choice of an appropriate parameter.
What Action Level is needed for the decision? In addition to specifying the population
parameter, you will need to specify the Action Level that will be used to choose between courses
of action. For example, the decision maker may take one action if the true value of the parameter
exceeds a specified value (Action Level) and a different action otherwise. There are basically two
kinds of Action Levels - those predetermined and those determined during the DQO Process.
Examples of predetermined Action Levels are fixed standards such as drinking water
standards or technology-based standards. For example, in the area of childhood lead poisoning
EPA QA/G-4
5-3
Final
August 2000
-------
prevention, EPA's Office of Pollution Prevention and Toxics has proposed hazard levels for lead
in residential dust and soil to protect children from significant lead exposures (40 CFR 745).
Also, in the area of air quality control, EPA's Office of Air and Radiation has promulgated
National Ambient Air Quality Standards for priority pollutants such as carbon monoxide, lead,
nitrogen dioxide, ozone, particulate matter (PM10), and sulfur dioxide, as well as other pollutants
that include fine particulate matter (PM2 5) (40 CFR 50).
Examples of investigation-specific Action Levels are background standards or specific
risk-based standards. For the case of investigation-specific Action Levels, one consideration in
selecting the Action Level is its degree of conservatism, i.e., whether the level is a very low value
or a higher value. You will need to decide whether to set the Action Level at a threshold of real
concern, or at a lower (more conservative) value that, if exceeded to some degree, may not
necessarily pose a serious risk. A more conservative Action Level may require a more sensitive
analytical method that has appropriate detection limits.
Does the Action Level exceed measurement detection limits? You will need to determine the
detection limit for each potential measurement method identified in Step 3. If the detection limit
for a measurement method exceeds the Action Level, then a more sensitive method should be
specified or a different approach should be used.
Detection limits are defined specific to an intended purpose. The DQO planning team
should choose the definition that is most appropriate to the "If...then..." decision rule being used.
For example, if the decision rule is used to decide if a contaminant exists at the study site, then the
detection limit should be one that provides for a high probability of positive identification and
presence in the matrix and a low probability of false confirmation. However, if the decision rule is
used to compare a mean to a threshold action level, then the detection limit should be defined in
terms of the reliability of quantitation.
5.3 Outputs
After you have completed the above activities, you can construct the theoretical
"If...then..." decision rule by combining the selected population parameter and Action Level with
the scale of decision making (from Step 4) and the alternative actions (from Step 2). An example
of a theoretical decision rule is:
//"the true mean dioxin concentration in the surface 2 inches of soil
of a decision unit (20 ft by 100 ft) exceeds 1 ppb, then remove a 6
inch layer of soil, //"the true mean is not greater than 1 ppb, then
do nothing.
Final
EPA QA/G-4 5 - 4 August 2000
-------
5.4 Examples
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
What was the decision rule and Action Level? The planning team was interested in the
true mean concentration of cadmium in the TCLP leachate for each container. If the
true mean concentration of cadmium from the fly ash leachate in each container load
was greater than 1.0 mg/L, then the waste was considered hazardous and disposed of at
a RCRA landfill. If the true mean concentration of cadmium from the waste fly ash
leachate was less than 1.0 mg/L, then the waste was considered nonhazardous and
disposed of in a sanitary landfill.
Example 2. Making Decisions About Urban Air Quality Compliance
What was the decision rule and Action Level? The population parameter of interest
that characterizes PM25 air quality was the true long-run proportion of daily
concentrations falling below the 24-hr PM25 federal standard of 65 ^g/m3. If the true
proportion of daily concentrations less than or equal to 65 jug/m3 was greater than or
equal to 0.98, then the local region was considered in attainment for PM25, so
monitoring was continued, but no other action was taken. If the true proportion of daily
concentrations less than or equal to 65 ^g/m3 was less than 0.98, then the local region
was considered in nonattainment for PM25, so monitoring was continued and the PM25
control strategies outlined in the State Implementation Plan were implemented.
Looking ahead to other DQO steps:
Step 6 provides key information that will be used with the outputs of this
step to select the sampling and analysis methods.
• Step 6 addresses the questions of what risk of an incorrect decision can
be tolerated.
Final
EPA QA/G-4 5 - 5 August 2000
-------
Final
EPA QA/G-4 5 - 6 August 2000
-------
CHAPTER 6
STEP 6. SPECIFY TOLERABLE LIMITS ON DECISION ERRORS
The DQO Process
1. State the Problem
2. Identify the Decision
3. Identify the Inputs to the Decision
4. Define the Boundaries of the Study
5. Develop a Decision Rule
6. Specify Tolerable Limits on Decision
Errors
7. Optimize the Design for Obtaining Data
6. Specify Tolerable Limits
on Decision Errors
• Determine the range of the parameter
of interest.
• Choose a null hypothesis.
• Examine consequences of making an
incorrect decision.
• Specify a range of values where
consequences are minor (gray region).
• Assign probability values to points
above and below the Action Level that
reflect tolerable probability for
potential decision errors.
After reading this chapter you should under stand why specifying
tolerable limits on decision errors is required to continue the DQO
Process and the meaning of the concepts and terms used in completing
this task. You should be able to specify tolerable limits on decision
errors for your problem.
6.1 Background
This step builds on the previous steps where you have:
• identified members of the planning team, including decision makers;
concisely described the problem;
• developed a conceptual model of the environmental problem to be investigated;
identified the decision that needs to be made;
• determined the type of information required, the Action Level, and probable
measurement methods;
• decided on the spatial/temporal boundaries of the decision; and
• decided on the theoretical "if... then" decision rule.
EPA QA/G-4
6- 1
Final
August 2000
-------
In Step 6 of the DQO Process you no longer imagine that perfect information on unlimited
data will be available for making decisions as you did in Step 5. You now face the reality that you
will not have perfect information upon which to base your decisions. Instead you will be making
decisions based on a set of sample data subject to various errors which is only part of the much
larger population of interest. Inherent in the use of sampled data for making decisions is the fact
that those decisions can, and occasionally will, be wrong. In this step of the DQO Process,
numerical values will be considered in an attempt to keep the possibility of a decision error to a
minimum.
The purpose of Step 6 is to specify quantitative performance goals for choosing between
the two alternative actions decision rule. These goals are expressed as probabilities of making
errors in your decision at selected true values of the parameter of interest specified in Step 5.
These decision performance goal probabilities are a statement of the amount of uncertainty you
are willing to tolerate in your decisions at a few specific critical true values of the parameter of
interest.
6.2 Activities
You should conduct the following activities in Step 6:
• determine the sources of error in the sample data set;
establish a plausible range of values for the parameter of interest;
• define the two types of potential decision errors and the consequences of making
those errors;
• determine how to manage potential decision errors;
• select the baseline condition of the environment that will be assumed to be true in
the absence of overwhelming evidence to the contrary;
specify a range of possible parameter values where the consequences of a false
acceptance decision error are considered tolerable (gray region); and
assign probability values at several true value points above and below the Action
Level that reflect your tolerable probability for the occurrence of decision errors.
What are sources of error in the sample data set? A decision error occurs when the sample data
set misleads you into making the wrong decision and, therefore, taking the wrong response action.
The possibility of a decision error exists because your decision is based on sample data that are
incomplete and never perfect. Even though the data collection method and analysis method may
be unbiased, the sample data are subject to random and systematic errors at different stages of
acquisition, from field collection to sample analysis. The combination of all these errors is called
"total study error." There can be many contributors to total study error, but there are typically
two main components:
• Sampling design error - This error is influenced by the inherent variability of the
population over space and time, the sample collection design, and the number of
Final
EPA QA/G-4 6 - 2 August 2000
-------
samples. It is usually impractical to measure the entire decision unit, and limited
sampling may miss some features of the natural variation of the measurement of
interest. Sampling design error occurs when the data collection design does not
capture the complete variability within the decision unit to the extent appropriate
for the decision of interest. Sampling design error can lead to random error (i.e.,
variability or imprecision) and systematic error (bias) in estimates of population
parameters.
Measurement error - This error (variability) is influenced by imperfections in the
measurement and analysis system. Random and systematic measurement errors are
introduced in the measurement process during physical sample collection, sample
handling, sample preparation, sample analysis, data reduction, transmission, and
storage.
Total study error directly affects the
probability of making decision errors.
Therefore, it is essential for you to manage
total study error by your choice of sample
design and measurement system. This will
enable you to control the possibility of making
decision errors to acceptable levels. Figure 6-1
shows an example of how total study error
(also known as Total Variability) can be
broken down further into components that will
relate to the data collection process.
How do you establish a plausible range of
values for the parameter of interest? You
should establish a plausible range of values for
the parameter of interest by approximating its
upper and lower bounds based on currently
available information, professional judgment,
or historical data. This helps focus the process
of defining probability limits on decision errors
only on the relevant values of the parameter.
For example, if the parameter of interest is a
mean, the range might be defined using the
lowest and highest concentrations at which the
contaminant is thought to exist at the property.
This range of values is useful when discussing
the Decision Performance Goal Diagram (to be
discussed later).
1— Homogenization
Sampling Frame Selection
Sampling Unit Definition
- Selection Probabilities
Number of Samples
— Support Volume/mass
— Sample Delineation
1—Sample Extraction
Sample Handling
"Preservation
—Packaging
—Labeling
—Transport
—Storage
— Preparation
— Subsampling
— Extraction
— Analytical
Determination
— Data Reduction
Figure 6-1. An Example of How Total Study
Error Can Be Broken Down by Components
EPA QA/G-4
6-3
Final
August 2000
-------
How are decision errors defined? If perfect knowledge of the true value of the parameter of
interest in a decision unit were available to you, you could simply apply the theoretical decision
rule from Step 5 to the known true value, make your decision, and not be concerned with decision
errors. However, in real life you use sample data to make the decision and, consequently, the
chance of a decision error becomes reality.
Due to the uncertainty inherent in decisions based on sample data, it is possible to get
results that will not clearly tell you if the true value is below the Action Level or above the Action
Level. It becomes necessary to label one of these two possibilities as the baseline condition so
that a decision can still be made in these situations. The baseline condition then becomes the de
facto decision outcome when there is insufficient evidence to refute it and the other condition then
becomes the alternative decision. For example, in legal decisions on human behavior, the baseline
condition is "innocent until proven guilty."
In environmental decisions affecting human health and the environment, the baseline
condition is more flexible and may depend on your situation. In certain instances, the baseline
condition for your problem may be prescribed for you in regulations. For example the baseline
condition in RCRA facility monitoring is that the concentrations in ground water are less than or
equal to the background concentrations. If the baseline condition is not specified for you, you
must select it based on careful consideration of the consequences of making decision errors and
taking the wrong actions. This selection may be based on your conceptual model for the decision
unit, i.e., based on prior information, you have good cause to think that the true value for the
decision unit is above the Action Level.
The probabilities of making decision errors with sample data can be quantified through the
use of a statistical decision procedure known as hypothesis testing. When hypothesis testing is
applied to decision making, the sample data are used to choose between a baseline condition of
the environment and an alternative condition. The test can then be used to show either that there
is insufficient evidence to indicate that the baseline condition is false (and therefore you accept the
default that the baseline condition is presumed to be true), or that the baseline condition is
probably false (and therefore the alternative condition is probably true). The burden of proof is
placed on rejecting the baseline condition. This approach is taken because the test-of-hypothesis
structure maintains the baseline condition as being true until overwhelming evidence is presented
to indicate that the baseline condition is not true. It is critical to understand that selection of the
baseline condition is important to the outcome of the decision process. The exact same set of
sample data from a decision unit can lead to different decisions depending on which possibility
was chosen as the baseline condition.
A false rejection decision error1 occurs when the limited amount of sample data lead you
to decide that the baseline condition is probably false when it is really true. In the reverse case, a
In previous editions of Guidance for Data Quality Objectives Process (EPA QA/G-4) (U.S. EPA, 1994a),
false rejection was called "false positive" and false acceptance was called "false negative."
Final
EPA QA/G-4 6 - 4 August 2000
-------
false acceptance decision occurs when the sample data lead you to decide that the baseline
condition is probably true when it is really false. To understand these definitions you may find it
helpful to note that an acceptance decision is to decide the baseline condition is true and a
rejection decision is to decide the alternative condition is true. Hence, a false rejection decision
incorrectly decides that the alternative is true, and a false acceptance decision incorrectly decides
that the baseline is true (see Table 6-1). For example, suppose you strongly believe that the true
value of the parameter of interest exceeds the Action Level (i.e., the baseline condition states that
the true value of the parameter of interest exceeds the Action Level). If your baseline assumption
is actually correct and the sample data, by chance, contain an abnormally large proportion of low
values, you would conclude that the true value of the parameter of interest does not exceed the
Action Level. In reality, the true value did exceed the Action Level; therefore, you would then be
making a false rejection decision error.
Table 6-1. False Acceptance and False Rejection Decisions
Decision Based on Sample
Data
Decide baseline is true
Decide alternative is true
True Condition
Baseline is True
Correct Decision
Decision Error (False Rejection)
Alternative is True
Decision Error (False Acceptance)
Correct Decision
Another example would be a regulatory situation in which an effluent discharge should not
exceed the permitted level. Your baseline condition would be that the true parameter value of the
effluent is less than or equal to the permitted level; your alternative would be that the true
parameter exceeds the permitted level. If the baseline condition was actually correct, but your
sample data happened to have a preponderance of high values, you could conclude the effluent
exceeds the permitted level. This would be a false rejection decision error and is sometimes called
a false positive decision error. The reverse (a false acceptance decision error) is sometimes called
a false negative decision error.
In the statistical language of hypothesis testing, the baseline condition is called the null
hypothesis (H0) and the alternative condition is called the alternative hypothesis (Ha). A false
rejection decision error occurs when the decision maker rejects the null hypothesis when it is
really true; a false acceptance decision error occurs when the decision maker fails to reject the null
hypothesis when it is really false. Statisticians label a false rejection decision error as a Type I
error and the measure of the size of this error (probability) is labeled alpha (a), the hypothesis
test's level of significance. Statisticians label a false acceptance decision error as a Type II error;
the measure of the size of this error (probability) is labeled beta (P). Both alpha and beta are
expressed numerically as probabilities. The statistical power of a test of hypothesis is equal to
How can you manage potential decision errors? Although the possibilities of making decision
errors can never be eliminated totally, you can manage them. To manage the possibilities of
Final
EPA QA/G-4 6 - 5 August 2000
-------
decision errors, your planning team focuses mostly on the largest components of total study error.
If the sampling design error is believed to be relatively large, you can manage the chance of
making a decision error by collecting a larger number of samples or developing a better sampling
design, i.e., a better way of deciding where and when to sample. If the analytical component of
the measurement error is believed to be relatively large, you can manage it by analyzing multiple
individual samples, or by using more precise and accurate analytical methods. In some instances
your planning team will actually be able to address both components of total error.
In some cases, placing a stringent (i.e., very small) limit on the possibility of both types of
decision errors is unnecessary for making a defensible decision. If the consequences of one
decision error are relatively minor, it may be possible for you to make a defensible decision based
on relatively imprecise data or on a small amount of data. For example, in the early phases of a
hazardous site assessment, the consequences of deciding that an area of a site is hazardous, when
in reality it is not, may be relatively minor. In this case, you may make a decision during this stage
of the investigation by using a moderate amount of data, analyzed using a field screening
analytical method, and only using a limited number of confirmatory analyses.
Conversely, if the consequences of decision errors are severe (i.e., human health effects),
you will want to develop a data collection design that exercises more control over sampling
design and measurement error. For example, in a waste discharge investigation, deciding that a
discharge is not hazardous when it truly is hazardous may have serious consequences because the
discharge may pose a risk to human health and to the environment. Therefore, the decision made
during this phase of the investigation may need to be supported by a large amount of data and
analyzed using very precise and accurate analytical methods.
You will need to balance the consequences of decision errors against the cost of limiting
the possibility of these errors. It may be necessary to iterate between Step 6 and Step 7 several
times before this balance between limits on decision errors and costs of data collection design is
achieved. This is not an easy part of the DQO Process. The balancing of the risk of incorrect
decisions with potential consequences should be explored fully by your planning team. Resorting
to arbitrary values such as "false rejection = 0.05, false acceptance = 0.20" is not recommended.
The circumstances of the investigation may allow for a less stringent option, or possibly a more
stringent requirement. In the early stages of DQO development, it is recommended that a very
stringent choice be made and the consequences of that choice be investigated by your planning
team during their activities under Step 7 of the DQO Process.
Decision errors can also occur that are independent of the use of statistical hypothesis
testing. An example could be that the data were manipulated prior to use in decision making by
an outside agent censoring the reported values. This is sometimes found in the collection of
screening data where insufficient training on the importance of adherence to QA protocol and
practice has resulted in data being recorded in an erroneous fashion. If data has been manipulated
prior to use in decision making, the assumed false rejection and false acceptance error rates
become invalid.
Final
EPA QA/G-4 6 - 6 August 2000
-------
How can you represent the quality of a decision process? There is a graphical construct called a
Decision Performance Curve that represents the quality of a decision process. In statistical
hypothesis testing usage, an operating characteristic curve or a power curve serve similar
purposes. Figure 6-2 depicts an example Decision Performance Curve and shows the range of
possible true values of the parameter of interest (including the Action Level) decided in Step 6, on
the x-axis and the range of probabilities (0 to 1) of deciding that the parameter of interest exceeds
the Action Level along the y-axis. Intuitively, the probability of deciding the parameter of interest
exceeds the Action Level is small for low true values and increases as the true value increases. A
full Decision Performance Curve is actually a continuous curve from the lowest true value to the
highest true value. If you had perfect knowledge of the true value of the parameter of interest, a
Decision Performance Curve would have a probability of 0 for any true value less than the Action
Level and jump to a probability of 1 for any true value above the Action Level. Since you are
dealing with sampled data (containing error), the probabilities will more realistically increase
gradually from near 0 for true values far below the Action Level, to near 1 for true values far
above the Action Level. The shape and steepness of this curve is a consequence of the sample
design and number of samples taken.
« a
^ I
» ^
•S M
"« '
•Q ^
M
0.9 _
0.8 _
0.6 _
0.5 _
0.4 -
0.2
0.1
Low •*-
Action Level
True Value of the Parameter —
High
Figure 6-2. An Example of a Decision Performance Curve
The following subsections describe the process of selecting a baseline condition, defining a
gray region, and establishing Decision Performance Goals (DPGs) by stating tolerable decision
error probabilities at a few critical true values of the parameter of interest. The combined
EPA QA/G-4
6-7
Final
August 2000
-------
information from these activities can then be displayed graphically as a Decision Performance
Goal Diagram (DPGD) that approximates the Decision Performance Curve. This DPGD
stipulates your tolerable risks of decision errors and allows you to communicate them to others,
including your sample design team and all stakeholders. The Decision Performance Curve is then
the overlay on the diagram and can be used to assess performance.
How do you select the baseline condition? If your baseline is not established by regulatory
considerations, your planning team should define the baseline condition based on the relative
consequences of the decision errors.
The baseline condition is the one that will be kept until overwhelming evidence (in the
form of data to be collected) is presented to make you reject the baseline condition in favor of the
alternative. You should use your evaluation of the potential consequences of the decision errors
to establish which decision error has the more severe consequences near the Action Level. For
example, you would judge the threat to public health against spending unnecessary resources.
Define the baseline condition and the alternative condition and assign the terms "false
rejection" and "false acceptance" to the appropriate decision error. An alternative name for "false
rejection" is "false positive" or Type I Error (by statisticians principally). The alternative name for
"false acceptance" is "false negative" or Type II Error. A false rejection decision error
corresponds to the more severe decision error, and a false acceptance decision error corresponds
to the less severe decision error.
You should designate the areas above and below the Action Level as the range where the
two types of decision errors may occur. This activity has two steps:
1. Define both types of decision errors and establish the "true state of nature" for
each decision error. The "true state of nature" is the actual condition of the
parameter of interest in the decision unit which is unknown to the decision maker.
You should state both decision errors in terms of the parameter of interest, the
Action Level, and the alternative actions.
2. Specify and evaluate the potential consequences of each decision error. For
example, the consequences of incorrectly deciding that the parameter is below the
Action Level (when in fact it is above the Action Level) include potential threats to
human health and to the environment. Conversely, the consequences of incorrectly
deciding that the value of the parameter of interest is above the Action Level
(when in fact it does not exceed the Action Level) include spending unnecessary
resources for further study.
You should evaluate the potential consequences of decision errors at several points
within the false rejection and false acceptance ranges. For example, the
consequences of a decision error when the true parameter value is only 10% above
Final
EPA QA/G-4 6 - 8 August 2000
-------
the Action Level may be minimal because it may cause only a moderate increase in
the risk to human health. Conversely, the consequences of a decision error when
the true parameter is an order of magnitude above the Action Level may be severe
because it could significantly increase the risk to human health and threaten the
local ecosystem.
How do you specify a range of possible true parameter values where the consequences of a
false acceptance decision error are considered tolerable (gray region)? The gray region is one
component of the quantitative decision performance criteria that is specifically used to limit
impractical and nonfeasible number of samples. The gray region is a range of true parameter
values within the alternative condition near the Action Level where it is "too close to call." This
gray region is where sampled data may correctly reject the baseline condition, but the sampled
data frequently do not provide sufficient evidence to be overwhelming. In essence, the gray
region is an area where it is not considered feasible to control the false acceptance decision error
limits to lower levels because the high costs of sampling and analysis outweigh the potential
consequences of choosing the wrong course of action (see Figure 6-3 for example).
• Alternative
« oc
•5 et
I I
g i*
I-3
si
.0
o
0.9
0.8
0.7 -
0.6
0.5
0.4
0.3
0.2
0.1
0
Gray Region
Relatively Large
Decision Error Rates
are Considered
Tolerable
Tolerable False
Acceptance Decision
Error
0 20 40
Action Level
True Value of the Parameter (Mean Concentration, ppm)
200
Figure 6-3. An Example of a Decision Performance Goal Diagram
(Baseline Condition: Parameter Exceeds the Action Level)
The first boundary of the gray region is the Action Level itself. Your planning team
establishes the other boundary of the gray region by evaluating the consequences of a false
EPA QA/G-4
6-9
Final
August 2000
-------
acceptance decision error over the range of possible parameter values in which this error may
occur. This boundary corresponds to the parameter value at which the consequences of a false
acceptance decision error are significant enough to have to set a low limit on the probability of
this decision error occurring.
For example, suppose the baseline condition is that the true mean level of contaminant
does not exceed 1.0 mg/L and the result of a sample of five observations reveals a sample mean of
1.05 mg/L. Is this sufficient evidence to reject the baseline condition? If the natural variability of
the contaminant was low, then probably this would be enough evidence. If the natural variability
was quite high (i.e., a coefficient of variation of 50%), then the evidence would not be
overwhelming, as a result of this happening quite naturally. On the other hand, if the sample mean
had been 1.50 mg/L, even high variability could not hide the fact that the baseline condition had
been exceeded. The second boundary of the gray region is that value that you decide represents
overwhelming evidence to reject the baseline condition.
In general, the narrower the gray region, the greater the number of samples needed to
meet the criteria because the area of uncertainty has been reduced. The width of the gray region
may be wide during early phases of the study process, but narrowed at later stages to determine if
the parameter of interest is only slightly different than the Action Level.
In statistical hypothesis testing language, the width of the gray region is called the
"minimum detectable difference" and is often expressed as the Greek letter delta (A). This value
is an essential part of many calculations for determining the number of samples that need to be
collected so that you will have your stated confidence in decisions made based on the data
collected.
How do you assign probability values to points above and below the action level that reflect
the tolerable probability for the occurrence of decision errors? A decision error limit is the
probability that a decision error may occur for a specific value of the parameter of interest when
making the decision using sampled data. This probability is an expression of the decision maker's
tolerance for uncertainty but does not imply that a decision error will occur. Instead it is only a
measure of the risk a decision maker is willing to assume of making an incorrect decision.
At a minimum, you should specify a false rejection decision error limit at the Action Level
and a false acceptance decision error limit at the other end of the gray region based on the
consequences of the respective errors. Severe consequences (such as extreme risks to human
health) should have stringent limits (small probabilities), whereas moderate consequences may
have less stringent limits (large probabilities). In general, the tolerable limits for making a
decision error should decrease as the consequences of a decision error become more severe
farther away from the Action Level.
The most stringent limits on decision errors that are typically encountered for
environmental data are 0.01 (1%) for both the false rejection and false acceptance decision errors.
Final
EPA QA/G-4 6 - 10 August 2000
-------
This guidance recommends using 0.01 as the starting point for setting decision error rates. If the
consequences of a decision error are not severe enough to warrant this stringent decision error
limit, this value may be relaxed (a larger probability may be selected). However, if this limit is
relaxed from a value of 0.01 for either the decision error rate at the Action Level or the other
bound of the gray region, your planning team should document the rationale for relaxing the
decision error rate. This rationale may include regulatory guidelines; potential impacts on cost,
human health, and ecological conditions; and sociopolitical consequences.
The value of 0.01 should not be considered a prescriptive value for setting decision error
rates, nor should it be considered EPA policy to encourage the use of any particular decision error
rate. Some programs, for example Superfund, give guidance on alternative values for starting
points. In the Soil Screening Guidance: User's Guide (U.S. EPA, 1996), the starting value for
false rejection is 0.05, and for false acceptance, 0.20. The actual values finally selected by the
planning team will depend on the specific characteristics of the problem being investigated.
Figures 6-3 and 6-4 illustrate some key outputs of Step 6 of the DQO Process for an
example, but with opposite baseline conditions and different project specific-considerations. The
DPGD is a special schematic representation of a Decision Performance Curve. While the
Decision Performance Curve is a continuous curve, the schematic DPGD depicts only a few
critical points on that curve. These few points represent your tolerable error limits, or DPGs, at a
few critical values. Your sampling design team will use this as the criteria for any sampling plan
they design. As the explanation progresses, it may be helpful to keep in mind that the DPGD
represents a set of "what if?" conditions in the following sense. You are answering the question
at several selected true values of the parameter of interest:
If the true value of the parameter of interest were at this level, how
strong of an aversion would I have if the data misled me into
making the wrong decision and taking action?
Figure 6-3 shows the case where a decision maker considers the more severe decision
error to occur above the Action Level and has labeled that as baseline. Figure 6-4 shows the
reverse, the case where the decision maker considers the more severe decision error to occur
below the Action Level.
Consider Figure 6-3 where the baseline condition is that the parameter exceeds the Action
Level (in statistical terms, H0: the parameter equals or exceeds the Action Level and HA: the
parameter is less than the Action Level). The plausible range of values based on professional
judgment was from the Detection Limit (as the Detection Limit was 0.01, it is essentially zero for
purposes of the DPGD) to 200 ppm. The Action Level was 100 ppm (from the permit for this
investigation). A false rejection would be saying the parameter is less than the Action Level,
when, in fact, it is really greater. A false acceptance would be saying the parameter level is above
the Action Level, when, in reality, it is below the Action Level. The gray region is the area where
you consider it is tolerable to make a decision error as it is "too close to call." For example,
Final
EPA QA/G-4 6-11 August 2000
-------
suppose you decided the true parameter level was above the Action Level (100 ppm) when in
reality it was 99 ppm. Although an error has occurred (false acceptance), it is not particularly
severe because the difference of 1 ppm on human health and financial resources is minimal. On
the other hand, suppose you decided the true parameter level was above the Action Level (100
ppm) when in reality it was 80 ppm. Again, an error has occurred (false acceptance), but it is
severe because a difference of 20 ppm is quite considerable. In this particular case the planning
team chose 80 ppm as the edge of their gray region because it represented the case where errors
in decision making have a great impact on resources. The planning team then assigned risk
probabilities to the chance of making decision errors for various true values of the parameter.
They agreed that, if the true value was 80 ppm and they decided (from the data yet to be
collected) that the true value exceeded 100 ppm, they were only willing to accept a 10% risk of
this happening. The team then considered the implications of what adverse effect would occur if
the true value was 60 ppm, but they decided the parameter was greater than 100 ppm. The
analysis showed a huge expenditure of resources, so the planning team elected to take only a 5%
risk of this happening. They did a similar exercise with the tolerable false rejection error rates.
Now consider Figure 6-4 where the baseline condition is that the parameter is less than the
Action Level (in statistical terms, H0: the parameter is less than or equal to the Action Level and
HA: the parameter is greater than the Action Level). Notice how the DPGD looks very similar to
that of Figure 6-3, except that the gray region is on the other side of the Action Level, and false
rejection and false acceptance have now been switched. In statistical terms, this is because a false
rejection is defined as rejecting H0 when H0 is really true, and false acceptance to be accepting H0
when H0 is really false.
Figure 6-4 shows that at the Action Level the decision maker will tolerate a 10% chance
of deciding that the true value is below the Action Level when it is really above the Action Level.
If the true value is 140 ppm, the decision maker will tolerate only a 1% chance of deciding the
true value is below the Action Level when it is really above the Action Level. At the edge of the
gray region, 120 ppm, the decision maker is willing to tolerate a 10% risk of saying it is above the
Action Level when it is really below the Action Level. At 60 ppm, the decision maker is only
willing to tolerate a 5% risk of a decision error. These probabilities represent the risk to the
decision maker of making an incorrect decision for the selected true values.
6.3 Outputs
The outputs from this step are your baseline condition, your gray region, and your set of
tolerable decision error limits at selected true values of the parameter of interest. These selections
are based on a consideration of the consequences of making incorrect decisions. The baseline
condition, the gray region, and your tolerable limits on decision errors are summarized in a
Decision Performance Goal Diagram.
Final
EPA QA/G-4 6 - 12 August 2000
-------
ter Exceeds the
Data
o o
00 ^D h-
, 1 , 1 , 1
It °'7-
£ £ °-6 "
^ a^
•S W) 0.5 -
*- .£
Probability of Deciding th
Action Level Us
o o o o
0 H* M W *.
— 1 1 1 II 1 1 1 1 1
i i i i i i i i i
Tolerable False
Rejection Decision
i
i i i i i i i
<4
Tolerable False
Acceptance Decision
Error Rates
Gray Region
Relatively Large
Decision Error Rates
are Considered
Tolerable
-
;
-
-
20 40 60 80 100 120 140 160 180 200
Action Leve
True Value of the Parameter (Mean Concentration, ppm)
Figure 6-4. An Example of a Decision Performance Goal
Diagram (Baseline Condition: Parameter is less than the
Action Level)
a> ^ Baseline ^
- i
•e
8 °-9 ~
X
lu •§ °-8 ~
la
S-O 0.7 -
fS V
IB. :
** a^
•5 ex 0.5 -
*- c
II 0.4 -
~ | 0.3 -
™
o | °-2 ~
*^- *-
0 H
2 o
C5
i 1 i 1 i 1 i 1 i
Tolerable False
Rejection Decision
Erro/kHjes
/ ..^>
................£.
i i i i i i i i i
^ Alternative ^
I
•4
i 1 i 1 i 1 i
*
^\
^^
Tolerable False
Acceptance Decision
Error Rates
Gray Region
1 Relatively Large
Decision Error Rates
are Considered
Tolerable
i i | i | i |
-
-
1
-
-
-
-
•g 0 20 40 60 80 100 120 140 160 180 200
1 t
Action Level
Figure 6-5. Decision Performance Goal Diagram for
Example 1
EPA QA/G-4
6- 13
Final
August 2000
-------
6.4 Examples
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
How was the baseline condition set? The baseline condition [i.e., the null hypothesis
(HJ] was established as "the waste is hazardous. " The consequences of deciding that
the waste was not hazardous when it truly was hazardous were that the incinerator
company disposed of the waste in a sanitary landfill, possibly endangering human health
and the environment. In this situation, the incinerator company could be held liable for
future damages and environmental cleanup costs. Additionally, the consequences of this
decision error were to compromise the reputation of the incinerator company,
jeopardizing its future profitability. The planning team concluded that this decision
error (false rejection) had the more severe consequences near the Action Level since the
risk of jeopardizing human health outweighed the consequences of having to pay more
for disposal.
How was the gray region specified? The gray region was designated as that area
adjacent to the Action Level where the planning team considered that the consequences
of a false acceptance decision error were minimal. The planning team specified a width
of 0.25 mg/Lfor this gray region based on their preferences to guard against false
acceptance decision errors at a concentration ofO. 75 mg/L (the lower bound of the gray
region).
How were tolerable decision error limits set? RCRA regulations specify a 5% decision
error rate at the Action Level. Below the Action Level, the planning team set the
maximum tolerable probability of making a false acceptance error at 20% when the true
parameter was from 0.25 to 0.75 mg/L and 10% when it was below 0.25 mg/L. These
limits were based on both experience and an economic analysis that showed that these
decision error rates reasonably balanced the cost of sampling versus the consequence of
sending clean ash to the RCRA facility.
Example 2. Making Decisions About Urban Air Quality Compliance
How was the baseline condition set? In most applications of the DQO Process, when,
where, and how many samples to collect is not determined until Step 7. However, given
that the monitoring network and sampling frequency were already established, the DQO
Process in this case was conducted to establish the quality and quantity of data needed
for making attainment decisions and to determine if the present network design achieved
those quality and quantity specifications. As the planning team was most concerned
about protecting public health, the baseline condition in this case was that the 98th
percentile of daily concentrations was above 65 jug/m3 (i.e., less than 98% of daily
concentrations are below 65 ^g/m3). That is, the null hypothesis was set as the state of
nature the planning team found evidence against, and, to protect public health, carefully
guarded against the false rejection decision error of incorrectly rejecting the baseline
condition.
Final
EPA QA/G-4 6 - 14 August 2000
-------
How was the gray region specified? The gray region, in this case, was specified in
terms of proportions. The planning team decided that the gray region should be from
0.98 to 0.995.
How were tolerable decision error limits set? The planning team determined that the
tolerable false rejection decision error rate should be 10% or less. While lowering the
tolerable bound on such error was desirable, the planning team, based on observed
PM25 daily concentration variability in other parts of the country, believed that
significantly smaller false rejection error rates were unobtainable for all but the most
extensive and costly network designs. The team also wished to protect against
implementing unnecessary and costly control strategies (i.e., incorrectly failing to reject
the baseline condition), but was willing to tolerate a somewhat larger probability of
making this false acceptance decision error. The planning team decided that the false
acceptance decision error rate should be not larger than 30%. These are shown in Figure
6.6.
a
._§ 0.90 -
O
o" oo 0.80 -
a" ^
§ £ 0.70 -
^ 1 0.60 _
*j «. o so
Js I "
e" < 0.40 -
2 ^
Q i
0 S 0.20 -
=" ^
ID 0.10 -
£ o oo
—
Gray Region
Relatively i .«!•<«>
Decision Error Rates are
Considered Tolerable
Tolerable False
Rejection Decision Error
Rate ^^
^^^^
>^^
i i i i i
4
0
0
s
0
0
0
0
0
0
0
0
0
0
0
.......
Tolerable
False
.cceptanci
Decision
Crror Rat<
1 '
l.UU
- 0.90
- 0.80
- 0.70
- 0.60
- 0.50
- 0.40
_ 0.30
- 0.20
0 10
0 00
0.950 0.955 0.960 0.965 0.970 0 975 0.980 0.985 0.990 0.995 1 000
t
Action Level
True Value of the Parameter
3
(True Proportion of Daily Concentrations Below 65 ug/m )
Figure 6-6. Decision Performance Goal Diagram for Example 2
Looking ahead to other DQO steps:
The information developed in Step 6 is then translated into the
requirements for a sampling plan in Step 7.
EPA QA/G-4
6- 15
Final
August 2000
-------
Final
EPA QA/G-4 6 - 16 August 2000
-------
CHAPTER 7
STEP 7. OPTIMIZE THE DESIGN FOR OBTAINING DATA
5.
the Problem
Identify the Decision
Identify the Inputs to the Decision
Define the Boundaries of the Study
Develop a Decision Rule
6. Specify Tolerable Limits on Decision
Errors
7. Optimize the Design for Obtaining
Data
7. Optimize the Design for
Obtaining Data
• Review the DQO outputs.
• Develop data collection design
alternatives.
• Formulate mathematical expressions
for each design.
• Select the sample size that satisfies the
DQOs.
• Decide on the most resource-effective
design, or agreed alternative.
• Document details in the QA Project
Plan
After reading this chapter you should have a broad understanding of
the steps that are needed to develop a sampling and analysis design to
generate data that meet the Data Quality Objectives and Decision
Performance Goals developed in Steps 1 through 6 of the DQO
Process.
7.1 Background
This step builds on the previous steps where you have:
identified members of the planning team, including decision makers;
• concisely described the problem;
developed a conceptual model of the environmental problem to be investigated;
• identified the decision that needs to be made;
determined the type of information required, the Action Level, and probable
measurement methods;
decided on the spatial/temporal boundaries of the decision and the scale of the
decision making;
decided on the theoretical "if...then" decision rule; and
• specified tolerable limits on decision errors.
EPA QA/G-4
7- 1
Final
August 2000
-------
The purpose of Step 7 is to develop a resource-effective sampling and analysis design for
generating data that are expected to satisfy the DQOs and DPGs developed in Steps 1 through 6
of the DQO Process.
7.2 Activities
In this final step you should:
• review existing environmental data;
evaluate operational decision rules;
• develop general data collection design alternatives;
calculate the number of samples to be taken; and
• select the most resource-effective data collection design.
Why should you review existing environmental data? Review existing data in more detail if it
appears that they can be used to support the data collection design (e.g., analyze the variability in
existing data if they appear to provide good information about the variance for the new data). If
no existing data are available, it may be cost-effective to conduct a limited field investigation to
acquire preliminary estimates of variability for determining the number of samples. If existing
data are going to be combined with new data to support the decision, then determine if there are
any gaps that can be filled or deficiencies that might be mitigated by including appropriate features
in the new data collection design. The existing data should also be reviewed for indications of
analytical problems, such as detection limits, that may rule out using certain statistical techniques.
Prior knowledge of the probability distribution (characteristics) exhibited by the data may also
have an effect on the choice of statistical tests.
How do you evaluate operational decision rules? The theoretical decision rule you developed in
Step 5 was based on the assumption that you knew the true value of the parameter of interest
(e.g., the true mean or median). Since you will be using measurements made on samples to make
your decision, an operational decision rule will be needed to replace the theoretical decision rule.
This operational decision rule will most likely be in the form of a statistical hypothesis test which
may involve some form of a statistical interval such as a confidence interval or tolerance interval.
The design team should evaluate the possible operational decision rules and choose one that best
matches the intent of the theoretical decision rule with the statistical assumptions. Each
operational decision rule will have a different formula for determining the number of samples
needed to meet your DPGs.
Some common statistical hypothesis tests and their sample size formulas are described in
detail in Guidance for Data Quality Assessment: Practical Methods for Data Analysis (EPA
QA/G-9), (U.S. EPA, 1997a). Most tests applied to environmental data can be broadly classified
as one-sample (single-site) tests or two-sample (two-site) tests. In one-sample cases, data from a
site are compared with an absolute criterion such as a regulatory threshold or an Applicable or
Relevant and Appropriate Requirement. In the two-sample cases, data from a site are compared
Final
EPA QA/G-4 7- 2 August 2000
-------
with data from another site or background (reference) area or from another time period at the
same site. In this case, the parameter of interest is usually the difference between the two means,
two medians, two proportions, or two percentiles, and the Action Level is often zero (i.e., no
difference).
How do you develop data collection design alternatives? A full explanation of the procedures
for developing a data collection design is beyond the scope of this guidance document. This
document provides a broad overview of the steps that need to be accomplished to reach a final
sampling plan. This section provides a general description of the activities necessary to generate
sampling design options and select the one that optimally satisfies the DPGs defined in Step 6. In
addition, it contains information about how outputs from the previous six steps of the DQO
Process are used in developing the most resource-effective sampling and analysis design.
The design team should develop alternative data collection and analysis designs based on
the DQO outputs and other relevant information, such as historical patterns of contaminant
deposition, estimates of variance, and technical characteristics of the contaminants and media.
The most important element of this step is to reduce the total variability through judicious choice
of a spatial and temporal sampling design and analytical measurement technique (see also Figure
6-1). If the total variability can be reduced to a value less than that specified in Step 6, the result
will be either a reduction in decision error rates (given a fixed number of samples) or reduction in
the number of samples (and, hence, resource expenditure) for a given set of decision error rates.
In general, the more complex the sampling design, the lower the total variability of the sample will
be.
Generally, the goal is to find cost-effective design alternatives that balance the number of
samples and the measurement performance, given the feasible choices for spatial and temporal
sample designs and measurement methods. In cases where there is relatively high spatial or
temporal variability, it may be more cost-effective to use less expensive and less precise analytical
methods so that a relatively large number of samples over space and time can be taken, thereby
controlling the sampling design error component of total study error. In other cases, where the
contaminant distribution over space and time is relatively homogeneous, or the Action Level is
very near the method detection limit, it may be more cost-effective to use more expensive more
precise and/or more sensitive analytical methods and collect fewer samples, thereby reducing the
analytical measurement error component of total study error. These alternatives should, at a
minimum, include the sample selection technique, the sample type, the number of samples, and the
number of analyses per sample. To generate alternative designs, the planning team may vary the
number and spatial/temporal locations of samples, the type of samples collected, the field
sampling or analytical methods used, or the number of replicate analyses performed on samples.
It should be remembered that the objective of the design is to estimate the parameter (mean,
median, percentile) with as much precision as possible such that the DPGs can be achieved.
How do you calculate the number of samples that satisfy the DPGs for each design
alternative and determine the cost for each design? You should use the formulas identified in
Final
EPA QA/G-4 7- 3 August 2000
-------
the previous activity to calculate the number of samples needed to meet the DPGs for each data
collection design alternative. You should then determine the associated cost for each design
alternative.
To assist the design team in their development of alternative designs and evaluation of
costs for a few select sampling designs and operational decision rules, EPA has developed the
software, Data Quality Objectives Decision Error Feasibility Trials (DEFT) Software (EPA
QA/G-4D), (U.S. EPA, 1994b). DEFT is a personal computer software package developed to
assist your planning team in evaluating whether the DQOs are feasible (i.e., can be achieved
within resource constraints) before the development of the final data collection design is started.
DEFT uses the outputs generated in Steps 1 through 6 of the DQO Process to evaluate several
basic data collection designs and determines the associated cost. DEFT presents the results in the
form of a Decision Performance Goal Diagram that overlays the desired Decision Performance
Curve of the sampling design.
If the DQOs are not feasible or not achievable within resource constraints, the DEFT
software allows you to relax some of the DQOs and DPGs until a feasible alternative is achieved.
The software allows the user to change the action level, the baseline condition, the width of the
gray region, the decision error rates, the estimate of the standard deviation, and the sample
collection and analysis costs. For each change, the software computes a new sample size and
total cost and shows the corresponding Decision Performance Curve in the Decision Performance
Goal Diagram.
How do you select the most resource-effective data collection design that satisfies all of the
DQOs? You should evaluate the design options based on cost and ability to meet the DQO
constraints and DPGs. The design that provides the best balance between cost (or expected
cost) and ability to meet the DQOs, given the non-technical, economic, and health factors
imposed on the project, is the most resource-effective (or the optimum design).
The statistical concept of a power function is extremely useful in investigating the
performance of alternative designs. The power function is the probability of rejecting the null
hypothesis (H0) when the null hypothesis is false (i.e., the alternative condition is true). If there
was no error associated with a decision, the ideal power function would be 0 if H0 were true, and
1 if H0 were false. Since decisions are based on imperfect data, however, it is impossible to
achieve this ideal power function. Instead, the power function will most likely yield values that
are small when H0 is true and large when H0 is false. A performance curve is based on the graph
of the power function.2 The performance curve can be overlaid into the Decision Performance
Goal Diagram to assess how well a test performs or to compare competing test. A design that
produces a very steep performance curve is preferred over one that is relatively flat.
In this guidance, the performance curve is based on either the power curve or the complement of the
power curve. This ensures that the performance curve always rises from left to right.
Final
EPA QA/G-4 7- 4 August 2000
-------
One simple method to improve the power of the statistical design is the use of
stratification to reduce the total variability in the data. Stratification is done by dividing the target
population into strata that are relatively homogeneous. The planning team may have made an
initial attempt at this in Step 4, Define the Boundaries of the Study. The strata may be physically
based (areas proximal to an incinerator, septic tanks, receptor wells, underground storage tanks)
or based on other factors (potential exposure, activity patterns, residences, ecological habitats,
agricultural sectors, historical or future use).
The advantages of stratification are:
• reducing the complexity of the problem by dividing it into manageable segments;
reducing the variability in subsets; and
• improving the efficiency of sampling.
Disadvantages of stratification include:
• difficulty in determining the basis for selecting strata (prior estimates of variability,
estimates of strata area may be needed);
• overstratifying may require more samples so increasing costs; and
stratifying areas that are not approximately homogeneous may result in developing
a design for collecting data that is inefficient or does not accurately reflect the
characteristics of the population.
If none of the data collection designs satisfies all of the DQOs and DPGs within the
resource constraints of the project, the planning team will need to review the outputs from the
entire DQO Process and alter one or more of the steps. Examples of adjustments that could be
made are:
• increasing the tolerable limits on decision errors;
increasing the width of the gray region;
• increasing the funding for sampling and analysis;
changing the boundaries (it may be possible to reduce sampling and analysis costs
by changing or eliminating subgroups that will require separate decisions); and
relaxing other project constraints.
For other sampling designs and/or operational decision rules, it will be necessary for the design
team to evaluate the design alternatives by other methods (perhaps computer simulation) and
possibly involve a statistical expert on sampling design and analysis.
Application of the DQO Process to remediation problems and integration with
geostratistical approaches to the analysis of soil contamination scenarios may be found in Myers
(1997). Once the final data collection design has been selected, it is important to ensure the
design and operational decision rule are properly documented. This improves efficiency and
Final
EPA QA/G-4 7- 5 August 2000
-------
effectiveness of later stages of the data collection and analysis process, such as the development
of field sampling procedures, QC procedures, and statistical procedures for data analysis. The
key to successful design documentation is in drawing the link between the statistical assumptions
on which the design and operational decision rule are based and the practical activities that ensure
these assumptions generally hold true.
For EPA programs, the operational requirements for implementing the data collection
design are documented in the Field Sampling Plan, Sampling and Analysis Plan, QA Project Plan
or other required document. Design elements that should be documented include:
• number of samples,
sample type (e.g., composite vs. grab samples),
• general collection techniques (e.g., split spoon vs. core drill, or activated charcoal
media vs. evacuated canister),
• physical sample (i.e., the amount of material to be collected for each sample),
sample support (i.e., the area, volume, or quantity that each individual sample
represents),
sample locations (surface coordinates and depth) and how locations were selected,
• timing issues for sample collection, handling, and analysis,
analytical methods (or performance-based measurement standards), and
• statistical sampling scheme.
Note that proper documentation of the model, operational decision rule, and associated
assumptions used for collecting and statistically analyzing data is essential to maintain the overall
validity of the study in the face of unavoidable deviations from the original design. Additionally,
the documentation will serve as a valuable resource for Data Quality Assessment (DQA) activities
after the data have actually been collected and the subsequent decision making process has been
completed.
7.3 Outputs
The outputs from this step are the full documentation of the final sampling design and
discussion of the key assumptions supporting the sampling design.
7.4 Examples
The examples presented here represent the initial final output of the DQO Process.
Example 1. Making Decisions About Incinerator Fly Ash for RCRA Waste Disposal
What was the selected sampling design ? The planning team 's statistician performed an
initial cost/benefit analysis that indicated a composite sample design was the best
sampling option to use to determine whether a container of ash should be sent to a
RCRA landfill or to a municipal landfill. Eight composite samples, each consisting of
Final
EPA QA/G-4 7- 6 August 2000
-------
eight grab samples, were taken from each container; and two subsamples from each
composite were sent to the laboratory for analysis. To form the composite samples, the
containers were divided into eight equal size areas and grab samples were taken
randomly within each area and composited. Each grab sample was a core that was
extracted, then mixed together to form the composite sample. From this composite
sample, two subsamples were sent to the laboratory for analysis.
What were the key assumptions supporting the selected design? The cost of this design
was based on the cost of collecting ($10) and analyzing ($150) a sample. Eight grab
samples were collected for each composite sample, for a sampling cost of $80; two
subsamples were analyzed from each composite sample for a cost of $300. Therefore,
each composite sample cost $380. The total cost of collecting and analyzing the eight
composite samples in one container was eight times the cost of one composite, for a total
of $3,040. The assumption that composite measurements were normally distributed was
made. This assumption was evaluated after the measurements were obtained. If the
assumption was not viable, then the planning team would recommend that additional
grab samples per composite be taken, or that a revised compositing process be used to
achieve normally distributed data. Based on the pilot study, the incineration company
determined that each load of waste fly ash was fairly homogeneous and estimated the
standard deviation in the concentration of cadmium among grab samples within loads of
ash to be 0.6 mg/L. It was assumed that the variability among sub-samples within a
composite sample was negligible. Data from the subsamples was used to test this
assumption and to collect additional subsamples, if necessary.
Example 2. Making Decisions about Urban Air Quality Compliance
What was the selected sampling design? Information from Step 6 indicated that
sampling everyday, regardless of the false rejection decision error rate tolerated, was
probably an inefficient use of resources and was unnecessary. This conclusion was
reached because sampling daily resulted in false acceptance decision error rates that
were far below those required in Step 6. In contrast, l-in-6-day or l-in-3-day sampling
could not satisfy the false acceptance decision error rate of 30% when the rather
restrictive constraint of a 1% false rejection decision error rate was used. The current
sampling scheme (1 in 3 days) performed at a satisfactory level as long as the false
rejection decision error rate allowed was in the range of 5% to 10%. If the planning
team decided that up to 10% false rejection decision error truly could be tolerated, then
information in Table 7-1 indicated it was possible to reduce sampling frequency from the
current rate to l-in-6-day sampling, thereby reducing costs while maintaining an
acceptable false acceptance decision error rate around 23%.
What were the key assumptions supporting the selected design? The monitoring
network was already in place, so the goal at this stage was to determine the performance
of the existing design, and to change the design, if needed, to achieve better
performance. Information in Table 7-1 showed the design performance (false
acceptance decision error rate) as a function of different false rejection error rate
allowances and alternative sampling frequencies (sample sizes) over a 3-year period of
data collection. In general, data in Table 7-1 indicated that the false acceptance
Final
EPA QA/G-4 7- 7 August 2000
-------
decision error rate decreased when a higher false rejection decision error rate was
tolerated. Similarly, false acceptance decision error rates decreased when sampling
intensity was increased from one-in-six-days sampling to every-day sampling.
Table 7-1. False Acceptance Decision Error Rates and Alternative Sampling Frequencies
Tolerable
False Rejection
Decision Error
Rates
1%
5%
10%
Sampling Frequency At Each Of Three Monitors
1 in 6 Days
>50%
>50%
23%
1 in 3 Days
(Current)
>50%
28%
11%
Every Day
1%
<1%
<1%
EPA QA/G-4
7-:
Final
August 2000
-------
CHAPTER 8
BEYOND THE DATA QUALITY OBJECTIVES PROCESS
After reading this chapter you should understand the kinds of
information that will be necessary to develop a QA Project Plan and the
role of Data Quality Assessment.
A project's life cycle consists of three principal phases: planning, implementation, and
assessment (described in Chapter 0 as the project tier of EPA's Quality System). Quality
assurance activities that are associated with each of these phases are illustrated in Figure 0-1.
Systematic planning (e.g., the DQO Process) and developing the sampling design comprises the
planning phase; the actual data collection process is the implementation phase; and an evaluation
(Data Quality Assessment) that the collected data met the performance criteria specified in the
DQOs is the final phase of a project.
8.1 Planning
During the planning stage, investigators specify the intended use of the data to be
collected and plan the management and technical activities (such as sampling) that are needed to
generate the data. Systematic planning and the DQO Process are the foundation for the planning
stage and lead to a sampling design, the generation of appropriate data quality indicators, and
standard operating procedures, which are all finally documented in the Agency's mandatory QA
Project Plan or similar document.
Environmental data for EPA programs may not be collected without having an approved
QA Project Plan in place (EPA Order 5360.1 A2). The mandatory QA Project Plan (EPA, 1998)
documents four main groups - project management, data generation and acquisition,
assessment/oversight, and data validation and usability (shown in Table 8-1).
Group A - Project Management
These elements address project management, project history and objectives, and roles and
responsibilities of the participants. These elements help ensure that project goals are
clearly stated, that all participants understand the project goals and approach, and that the
planning process is documented.
Group B - Data Generation and Acquisition
These elements cover all aspects of the project design and implementation (including the
key parameters to be estimated, the number and type of samples expected, and a
description of where, when, and how samples will be collected). They ensure that
appropriate methods for sampling, analysis, data handling, and QC activities are employed
and documented.
Final
EPA QA/G-4 8 - 1 August 2000
-------
Table 8-1. Elements of a Quality Assurance Project Plan
QA Project Plan Elements
A. Project Management
Al Title and Approval Sheet
A2 Table of Contents
A3 Distribution List
A4 Project/Task Organization
A5 Problem Definition/Background
A6 Project/Task Description
A7 Quality Objectives and Criteria
A8 Special Training /Certification
A9 Documents and Records
B. Data Generation and Acquisition
B1 Sampling Process Design
(Experimental Design)
B2 Sampling Methods
B3 Sample Handling and Custody
B4 Analytical Methods
B5 Quality Control
B6 Instrument/Equipment Testing, Inspection,
and Maintenance
B7 Instrument/Equipment Calibration and
Frequency
B8 Inspection/Acceptance of
Supplies and Consumables
B9 Nondirect Measurements
B1 OData Management
C. Assessment and Oversight
Cl Assessments and Response Actions C2 Reports to Management
D. Data Validation and Usability
Dl Data Review, Verification, and Validation D2 Verification and Validation Methods
D3 Reconciliation with User Requirements
Group C - Assessment and Oversight
These elements address activities for assessing the effectiveness of project implementation
and associated QA and QC requirements; they help to ensure that the QA Project Plan is
implemented as prescribed.
Group D - Data Validation and Usability
These elements address QA activities that occur after data collection or generation is
complete; they help to ensure that data meet the specified criteria.
EPA QA/G-4
-2
Final
August 2000
-------
Additional information on the preparation of QA Project Plans is provided in EPA's guidance
document, Guidance on Quality Assurance Project Plans (EPA QA/G-5) (U.S. EPA, 1998).
8.2 Implementation
During the implementation phase of the project, data are collected and samples are
analyzed according to the specifications of the QA Project Plan or the Sampling and Analysis Plan
depending on specific program requirements. These provide detailed specific objectives, QA and
QC specifications, and procedures for conducting a successful field investigation that is intended
to produce data of the quality needed to satisfy the performance criteria. QA and QC activities
(e.g., technical systems audits and performance evaluations) are conducted to ensure that data
collection activities are conducted correctly and in accordance with the QA Project Plan.
8.3 Assessment
During the final phase (assessment) of a project, data are verified and validated in
accordance with the QA Project Plan, and DQA is performed to determine if the performance
criteria have been satisfied.
DQA is built on a fundamental premise: data quality, as a concept, is meaningful only
when it relates to the intended use of the data. Data quality does not exist without some frame of
reference; you really should know the context in which the data will be used in order to establish a
yardstick for judging whether or not the data set is adequate. DQA is the scientific and statistical
process that determines whether environmental data are of the right type, quality, and quantity to
support a specific decision. DQA consists of five steps that parallel the activities of a statistician
analyzing a data set; and include the use of statistical and graphical tools that nonstatisticians can
apply to data sets (see Figure 8-1).
DQA involves the application of statistical tools to determine whether the data are of
appropriate quality to support the decision with acceptable confidence. To conclude the
assessment phase, it is necessary to document all the relevant information collected over all phases
of the project's life cycle. The conclusion from a DQA must be presented in a fashion that
facilitates the comprehension of the important points. Care should be taken to explain statistical
nomenclature and avoid the use of statistical jargon whenever possible. For more information on
Data Quality Assessment, see EPA's guidance document, Guidance for Data Quality
Assessment: Practical Methods for Data Analysis (EPA QA/G-9), (U.S. EPA, 1997a) and the
associated software Data Quality Assessment Statistical Toolbox (DataQUEST) (EPA QA/G-
9D), (U.S. EPA, 1997b).
Final
EPA QA/G-4 8 - 3 August 2000
-------
1. Review the DQOs and Sampling Designs
Review DQO outputs; if DQOs have not been developed, define the statistical
hypothesis and specify tolerable limits on decision errors; and review the
sampling design and the data collection documentation for consistency.
1
r
2. Conduct Preliminary Data Review
Generate statistical quantities and graphical representations that describe the
data and use this information to learn about the structure of the data and to
identify any patterns or relationships.
1
r
3. Select the Statistical Test
Select the most appropriate procedure for summarizing and analyzing the
data based on the preliminary data review and identify the underlying
assumptions of the test.
1
I
4. Verify the Assumptions of the Statistical Test
Examine the underlying assumption of the statistical test in light of the
environmental data actually collected.
1
r
5. Draw Conclusions from the Data
Perform the calculations of the statistical hypothesis tests and document the
inferences drawn as a result of these calculations; and evaluate the
performance of the sampling design if the design is to be used again.
Figure 8-1. Data Quality Assessment Process
EPA QA/G-4
-4
Final
August 2000
-------
APPENDIX A
DERIVATION OF SAMPLE SIZE FORMULA FOR TESTING MEAN
OF NORMAL DISTRIBUTION VERSUS AN ACTION LEVEL
This appendix presents a mathematical derivation of the sample size formula used in the DQO
Example 1.
Let Xb X2,...,Xn denote a random sample from a normal distribution with unknown mean (i and
known standard deviation o. The decision maker wishes to test the null hypothesis H0: (i = AL versus the
alternative HA: ^ > AL, where AL, the action level, is some prescribed constant; the false positive (Type I)
error rate is a (i.e., probability of rejecting H0 when (i = AL is a); and for some fixed constant U > AL
(where U is the other bound of the gray region), the false negative (Type II) error rate is (3 (i.e., probability
of rejecting H0 when (i = U is 1-p). Let X denote the sample mean of the Xs. It will have a normal
distribution with mean (i and variance o2/n. Hence the random variable Z, defined by
Z = v Hyv , (A - 1)
a
will have a standard normal distribution (mean 0, variance 1). Let zp denote the p* percentile of the
standard normal distribution (available in most statistics books). Recall that the symmetry of the standard
normal distribution implies that zp = -z^.
Case 1: Standard Deviation Known
The test of H0 versus HA is performed by calculating the test statistic.
T = (X-AL) Z]_o, the null hypothesis is rejected.
Note that
T = - - = Z+e(M) (A_3)
Final
EPA QA/G-4 A - 1 August 2000
-------
where
e(n) = . (A . 4)
Thus T has a normal distribution with mean e((i) and variance 1, and, in particular, e(AL) = 0. Hence the
Type I error rate is
Projecting HQH0} = Pr[T>Zl_^=AL\ = Pr[Z^(AL)>Zl_^ = i (A - 5)
Achieving the desired power 1-p when ^ = U requires that
Pr[reject HQ\y.= U] = 1 - P.
Therefore,
Pr[T40), the approximation is good. The particular
noncentral t distribution involved in the calculation depends on the sample size n. Thus, determining the
Final
EPA QA/G-4 A - 2 August 2000
-------
exact minimum n that will satisfy the Type I and Type II error rate conditions requires an iterative
approach in which the noncentral t probabilities are calculated for various n values until the desired
properties are achieved. With the aid of a computer routine for calculating such probabilities, this is not
difficult; however, a simple and direct approach for approximating n is available. This approach, whose
derivation is described in the paragraphs below, leads to the following approximate but very accurate
formula for n:
(A-8)
In practice, since o is unknown, a prior estimate of it must be used in Equation A - 8.
The approach is based on the assumption that, for a given constant k, the statistic X - kS
is approximately normal with mean (i-ko and variance (o2/n)(l+k2/2) (Guenther, 1977 and 1981).
The classical t-test rejects H0 when, T = \(X - AL}IS\fn)\>D where the critical value D is
chosen to achieve the desired Type I error rate a. The inequality can be rearranged as
X - kS >AL, where k = D\fn. Subtracting the mean (assuming H0) and dividing by the standard
deviation of X - kS on both sides of the inequality leads to
X-kS-(AL-ko) ^ AL-(AL-ko) kJn
- - = (A - 9)
By the distributional assumption on X - kS , the left side of Equation A - 9 is approximately
standard normal when \a = AL, and the condition that the Type I error rate is a becomes
Pr[Z>kJn/}/\ + k2/2] = a, (A -10)
i.e., ZI_K = k^l + k2/2. (A -11)
One can show that Equation A - 1 1 is equivalent to
P/2] = \-zl_J2n. (A- 12)
The condition that the Type II error rate is (3 (or that power is 1-p) when (i = U means that the event of
incorrectly accepting H0 given X - kS < AL should have probability p. Subtracting the mean
(U - ko) and dividing by the standard deviation of X - kS on both sides of this inequality yields
Final
EPA QA/G-4 A - 3 August 2000
-------
X-kS-(U-ka) AL-(U-ka)
(A-13)
Again, the left side is approximately standard normal and the Type II error rate condition becomes
Pr \Z < [AL - (U - £o)]/[(oV«)A/l
which implies
(AL-U)+ko
(A-14)
Subtracting Equation A - 14 from Equation A - 11 yields
. _ _ (U-AL)_
P/2
(A-15)
or
(U-AL)
Substituting Equation A - 12 into the denominator on the right side of Equation A - 16 yields
(A-16)
(U-AL)
= Jn< \-z_J2n.
(A-17)
Squaring both sides of Equation A - 17 and solving for n yields Equation A - 8.
References
Guenther, William C. 1977. Sampling Inspection in Statistical Quality Control. Griffin's Statistical
Monographs and Courses, No. 37, London: Charles Griffin.
Guenther, William C. 1981. Sample size formulas for normal theory T test. The American
Statistician 35, 4.
EPA QA/G-4
A-4
Final
August 2000
-------
APPENDIX B
BIBLIOGRAPHY
ANSI/ASQC E4-1994, Specifications and Guidelines for Environmental Data Collection and
Environmental Technology Program. 1994. American Society for Quality, Milwaukee,
WI.
40 CFR 50, Code of Federal Regulations, "National Ambient Air Quality Standards for
Paniculate Matter; Availability of Supplemental Information and Request for Comments,
Final Rule." Washington, DC.
40 CFR 261, Code of Federal Regulations, Appendix II, "Identification and Listing of Hazardous
Waste; Method 1311 Toxicity Characteristic Leaching Procedure (TCLP)." Washington,
DC.
40 CFR 745, Code of Federal Regulations, "Lead; Identification of Dangerous Levels of Lead;
Proposed Rule." Washington, DC.
ASTM El 613-99, "Standard Test Method for Determination of Lead by Inductively Coupled
Plasma Atomic Emission Spectrometry (ICP-AES), Flame Atomic Absorption
Spectrometry (FAAS), or Graphite Furnace Atomic Absorption Spectrometry (GFAAS)
Techniques," American Society for Testing and Materials, 1999.
ASTM E1644-98, "Standard Practice for Hot Plate Digestion of Dust Wipe Samples for the
Determination of Lead," American Society for Testing and Materials, 1998.
ASTM El728-99, "Standard Practice for Field Collection of Settled Dust Samples Using Wipe
Sampling Methods for Lead Determination by Atomic Spectrometry Techniques,"
American Society for Testing and Materials, 1999.
Gilbert, R.O. Statistical Methods for Environmental Pollution Monitoring. 1987. John Wiley,
New York.
Grumley, Thomas. 1994. Institutionalizing the Data Quality Objectives Process for EM's
Environmental Data Collection Activities. Department of Energy, Washington, DC.
Harbinger Communications. 1996. The Harbinger File: A Directory of Citizen Groups,
Government Agencies, and Environmental Education Programs Concerned with
California Environmental Issues. Santa Cruz, CA.
Final
EPA QA/G-4 B -1 August 2000
-------
HUD (Department of Housing and Urban Development). 1995. Guidelines for the Evaluation
and Control of Lead-based Paint Hazards in Housing. Office of Lead-based Paint
Abatement and Poisoning Prevention, Washington, DC.
Myers, J.C. Geostatistical Error Management: Quantifying Uncertainty for Environmental
Sampling and Mapping. 1997. Van Nostrand Reinhold. New York.
Thompson, S.K. Sampling. 1992. John Wiley. New York.
U.S. EPA (Environmental Protection Agency). 1994a. Guidance for the Data Quality
Objectives Process (EPA QA/G-4). EPA/600/R-96/055. Washington, DC.
U.S. EPA (Environmental Protection Agency). 1994b. Data Quality Objectives Decision Error
Feasibility Trials (DEFT) Software (EPA QA/G-4D). EPA/600/R-96/056. Washington,
DC.
U.S. EPA (Environmental Protection Agency). 1996. Soil Screening Guidance: User's Guide.
EPA/540/R-96/0180. Washington, DC.
U.S. EPA (Environmental Protection Agency). 1997a. Guidance for Data Quality Assessment:
Practical Methods for Data Analysis (EPA QA/G-9). EPA/600/R-96/084. Washington,
DC.
U.S. EPA (Environmental Protection Agency). 1997b. Data Quality Assessment Statistical
Toolbox (DataQUEST) (EPA QA/G-9D). EPA/600/R-96/085. Washington, DC.
U.S. EPA (Environmental Protection Agency). 1998. Guidance on Quality Assurance Project
Plans (EPA QA/G-5). EPA/600/R-98/018. Washington, DC.
U.S. EPA (Environmental Protection Agency). 1999. Data Quality Objectives for Hazardous
Waste Site Investigations (EPA QA/G-4HW). Washington, DC.
U.S. EPA (Environmental Protection Agency). 2000a. EPA Order 5360.1 A2: Policy and
Program Requirements for the Mandatory Agency-wide Quality System.. Washington,
DC.
U.S. EPA (Environmental Protection Agency). 2000b. EPA Order 5360: The EPA Quality
Manual for Environmental Programs. Washington, DC.
Final
EPA QA/G-4 B - 2 August 2000
-------
APPENDIX C
Data Quality Objectives; Household Dust Lead Hazard Assessment
This example concerns the use of the median in planning for environmental decision
making. The example is presented in a continuous format to show the seven-step DQO Process in
its entirety.
0. Background
The adverse health effects resulting from exposure to lead hazards (paint, dust, and soil)
have received increasing attention because chronic exposure to low levels of lead can cause
impairment of the central nervous system, mental retardation, and behavioral disorders. Young
children (below the age of six) are at a particularly high risk for these adverse effects. Concern
about the exposure to lead hazards in residential housing has led federal agencies, including the
EPA and Department of Housing and Urban Development, to develop programs to evaluate, and
ultimately control, lead hazards in housing.
A critical pathway for exposure to lead by a child is through the ingestion of household
dust because dust collects on hands, toys, and food and is easily transferred by hand-to-mouth
activities. As a result of the concern about the dust-to-mouth pathway, an important component
of risk assessment is dust sampling. Dust sampling offers a way of characterizing dust lead levels
at a property and determining if intervention is warranted. One of the preferred methods for
sampling residential dust is using baby wipes to wipe a specified surface area. A single area may
be sampled using an individual wipe; or multiple areas of a room may be sampled with individual
wipes, and the individual wipes combined, or composited, then submitted to the laboratory as a
single sample (40 CFR 745). The distribution of dust lead levels is such that normality cannot be
assumed and a 50th percentile (the median) is the appropriate risk assessment level. This example
demonstrates use of the median (i.e., 50th percentile) as the primary population parameter of
concern
1. State the Problem
How were the planning team members selected? The planning team included the
property owners, a certified risk assessor (to collect and handle dust samples and serve as a liaison
with the laboratory), a statistician, and a quality assurance specialist. The decision makers were
the property owners.
How was the problem described and a conceptual model of the potential hazard
developed? The problem was described as evaluating potential hazards associated with lead in
dust in a single-family residence because other residences in the neighborhood had shown levels of
lead in dust that might pose potential hazards.
Final
EPA QA/G-4 C- 1 August 2000
-------
The conceptual model described a single-family residence in a neighborhood where
hazardous levels of lead had been detected in other residences. Interior sources of lead in dust
were identified as lead-based paint on doors, walls, and trim, which deteriorated to form, or attach
to, dust particles. Exterior sources included lead in exterior painted surfaces that had deteriorated
and leached into the dripline soil, or lead deposited from gasoline combustion fumes that
accumulated in soil. In these cases, soil could be tracked into the house, and collected as dust on
floors, window sills, toys, etc. Because this dust could be easily ingested through hand-to-mouth
activities, dust was considered to be a significant exposure route. Levels of lead in floor dust
were to be used as an indicator of the potential hazard.
What were the available resources and relevant deadlines? The property owners were
willing to commit up to $1,000 for the study. To minimize inconvenience to the family, all
sampling would be conducted during one calendar day.
2. Identify the Decision
What was the Decision Statement? The decision statement was determining if there
were significant levels of lead in floor dust at the residence.
What were the alternative actions? If there were significant levels of lead in floor dust
at the residence, the team planned follow-up testing to determine whether immediately dangerous
contamination exists and the location of the contamination in the property. If not, the team
decided there was not a potential lead hazard, and testing was discontinued.
3. Identify the Inputs to the Decision
Identify the kind of information. The assessment of a dust lead hazard was evaluated
by measuring dust lead loadings by individual dust wipe sampling.
Identify the source of information. The EPA proposed standard stated that if dust lead
levels were above 50 |ig/ft2 on bare floors, a lead health hazard was possible and follow-up testing
and/or intervention should be undertaken (40 CFR 745).
What sampling and analytical methods were appropriate? Wipe samples were
collected according to ASTM standard practice E1728. These samples were digested in
accordance with ASTM standard practice El644 and the sample extracts were chemically
analyzed by ASTM standard test method E1613. The results of these analyses provided
information on lead loading (i.e., jig of lead per square foot of wipe area) for each dust sample.
The detection limit was well below the Action Level.
Final
EPA QA/G-4 C- 2 August 2000
-------
4. Define the Boundaries of the Study
What population was sampled? Dust contained in 1 ft2 area of floors of the residence
was sampled and sent to a laboratory for analysis.
What were the spatial boundaries? The spatial boundaries of the study area were
defined as all floor areas within the dwelling that were reasonably accessible to young children
who lived at, or visited, the property.
What was an appropriate time frame for sampling? The test results were considered
to appropriately characterize the current and future hazards. It was possible that lead contained in
soil could be tracked into the residence and collect on surfaces, but no significant airborne sources
of lead deposition were known in the region. The dust was not expected to be transported away
from the property; therefore, provided the exterior paint was maintained in intact condition, lead
concentrations measured in the dust were not expected to change significantly over time.
What were the practical constraints for collecting data? Permission from the residents
was required before risk assessors could enter the residence to collect dust wipe samples.
Sampling was completed within 1 calendar day to minimize the inconvenience to the residents.
What was the scale of the decision making? The decision unit was the interior floor
surface (approximately 1,700 ft2) of the residence at the time of sampling and in the near future.
5. Develop a Decision Rule
What was the decision rule and Action Level? From 40 CFR 745, the median was
selected as the appropriate parameter to characterize the population under study. The median
dust lead loading was defined to be that level, measured in |ig/ft2, above and below which 50% of
all possible dust lead loadings at the property were expected to fall. If the true median dust
loading in the residence was greater than 50 |ig/ft2, then the planning team required followup
testing. Otherwise, they decided that a dust lead hazard was not present and discontinued testing.
6. Specify Tolerable Limits on Decision Errors
How was the baseline condition set? The baseline condition adopted by the property
owners was that the true median dust lead loading was above the EPA hazard level of 50 |ig/ft2
due to the seriousness of a potential hazard. The planning team decided that the most serious
decision error would be to decide that the true median dust lead loading was below the EPA
hazard level of 50 |ig/ft2, when in truth the median dust lead loading was above the hazard level.
This incorrect decision would result in significant exposure to dust lead and potential adverse
health effects.
Final
EPA QA/G-4 C- 3 August 2000
-------
How was the gray region specified? The edge of the gray region was designated by
considering that a false acceptance decision error would result in the unnecessary expenditure of
scarce resources for follow-up testing and/or intervention associated with a presumed hazard that
did not exist. The planning team decided that this decision error should be adequately controlled
for true dust lead loadings of 40 |ig/ft2 and below.
How were tolerable decision error limits set? Since human exposure to lead dust
hazards causes serious health effects, the planning team decided to limit the false rejection error
rate to 5%. This meant that if this dwelling's true median dust lead loading was greater than 50
|ig/ft2, the baseline condition would be correctly rejected 19 out of 20 times. The false
acceptance decision, which would result in unnecessary use of testing and intervention resources,
was allowed to occur more frequently (i.e., 20% of the time when the true dust-lead loading is 40
|ig/ft2 or less). These are shown in Figure C-l.
H £
0 o
e l
Q -c
»i
.tS £ 0.20
1 w
.o o.io H
Tolerable False
Acceptance Decision
Error I
Tolerable False
Rejection Decision
Error Rates
Gray Region
Relatively Large
Decision Error Rates are
Considered Tolerable
Action Level
True Value of the Parameter (Median Dust-lead loading, ug/ft )
Figure C-l. Decision Performance Goal Diagram for Dust Lead Loading
7. Optimize the Design for Obtaining Data
What was the selected sampling design? The planning team determined that the cost of
sending a certified risk assessor to the property for collecting and handling dust wipe samples was
about $400. Also, an NLLAP-recognized laboratory was selected to analyze the collected wipe
samples at a cost of $10 per sample. Thus, a maximum of 60 samples could be obtained within
the study's cost constraint of $1,000. From Step 6 the initial gray region lower bound for the
study was set at 40 |ig/ft2, but, the team found that this requirement could not be met given the
specified decision errors (i.e., false rejection rate of 5% and false acceptance rate of 20%),
EPA QA/G-4
C-4
Final
August 2000
-------
assumed standard deviation (of the natural logarithms), range, and cost constraints of the study
(i.e., a maximum of 60 samples). The planning team decided they were unwilling to relax the
decision error rate requirements and elected to expand the width of the gray region from the
original 40 to 50 |ig/ft2 to the less restrictive range of 35 to 50 |ig/ft2. Further, the planning team
decided that a standard deviation (of the natural logarithms) value of o=1.0 was probably more
realistic than the more conservative estimate of o=1.5.
The planning team used the upper variability bound to develop Table C-l which presented
statistical sample size requirements across various assumed dust lead loading standard deviations
(of the natural logarithms) and various lower bounds of the gray region. This table indicated that
sample size requirements increased rather dramatically as variability increased and/or as the gray
region was made more narrow.
Therefore, based on Table C-l, the planning team decided that a total of 50 samples
should be collected by a certified risk assessor (all within 1 calendar day) using simple random
sampling throughout the residence. Samples were sent to the selected NLLAP-recognized
laboratory for analysis. The total study cost was approximately $900 to the property owners.
What were the key assumptions supporting the selected design? The dust lead
loading data was assumed to be log-normally distributed. The geometric mean was computed
using the data because the true median and true geometric mean are the same when log-normality
is assumed. The true variability in dust lead loadings was not known, but past data was used to
estimate a reasonable upper bound on variability.
Table C-l. Number of Samples Required for Determining
If the True Median Dust Lead Loading is Above the Standard
Gray Region
(jig/ft2)
20-50
25-50
30-50
35-50
40-50
45-50
Standard Deviation of Natural Logarithms
o=0.5
6
8
14
26
64
280
0=1.0
9
15
26
50
126
559
o=1.5
13
21
37
75
188
837
EPA QA/G-4
C-5
Final
August 2000
-------
Final
EPA QA/G-4 C- 6 August 2000
-------
APPENDIX D
GLOSSARY OF TERMS
acceptance criteria - specific limits placed on characteristics of an item, process, or service
defined in requirements documents.
action level - the numerical value that causes a decision maker to choose one of the alternative
actions (e.g., compliance or noncompliance). It may be a regulatory threshold standard, such as a
maximum contaminant level for drinking water; a risk-based concentration level; a technology
limitation; or a reference-based standard. Note that the action level defined here is specified
during the planning phase of a data collection activity; it is not calculated from the sampling data.
alternative condition - a tentative assumption to be proven either rue or false. When hypothesis
testing is applied to site assessment decisions, the data are used to choose between a presumed
baseline condition of the environment and an alternative condition. The alternative condition is
accepted only when there is overwhelming proof that the baseline condition is false. This is often
called the alternative hypothesis in statistical tests.
alternative hypothesis - see alternative condition.
baseline condition - a tentative assumption to be proven either true or false. When hypothesis
testing is applied to site assessment decision, the data are used to choose between a presumed
baseline condition of the environment and an alternative condition. The baseline condition is
retained until overwhelming evidence indicates that the baseline condition is false. This is often
called the null hypothesis in statistical tests.
bias - the systematic or persistent distortion of a measurement process that causes errors in one
direction (i.e., the expected sample measurement is different from the sample's true value.
boundaries - the spatial and temporal conditions and practical constraints under which
environmental data are collected. Boundaries specify the area of volume (spatial boundary) and
the time period (temporal boundary) to which a decision will apply.
confidence interval - the numerical interval constructed around a point estimate of a population
parameter, combined with a probability statement (the confidence coefficient) linking to the
population's true parameter value. If the same confidence interval construction technique and
assumptions are used to calculate future intervals, they will include the unknown population
parameter with the same specified probability.
data collection design - see sampling design
Final
EPA QA/G-4 D- 1 August 2000
-------
data quality assessment (DQA) - a statistical and scientific evaluation of the data set to
determine the validity and performance of the data collection design and statistical test, and to
determine the adequacy of the data set for its intended use.
data quality objectives (DQOs) - qualitative and quantitative statements derived from the DQO
Process that clarify study objectives, define the appropriate type of data, and specify tolerable
levels of potential decision errors that will be used as the basis for establishing the quality and
quantity of data needed to support decisions.
data quality objectives process - a systematic planning tool to facilitate the planning of
environmental data collection activities. Data quality objectives are the qualitative and
quantitative outputs from the DQO Process.
decision error - the error that occurs when the dat mislead the site manager into choosing the
wrong response action, in the sense that a different response action wold have been chosen if the
site manager had access to unlimited "perfect data" or absolute truth. In statistical test, decision
errors are labeled as false rejection or false acceptance depending on the concerns of the decision
maker and the baseline condition chosen.
decision performance curve - a graphical representation of the quality of a decision process. In
statistical terms it is know as a power curve (or a reverse power curve depending on the
hypotheses being tested).
decision performance goal diagram (DPGD) - a graphical representation of the tolerable risks
of decision errors. It is used in conjunction with the decision performance curve.
defensible - the ability to withstand any reasonable challenge related to the veracity or integrity of
project and laboratory documents and derived data.
detection limit (DL) - a measure of the capability of an analytical method of distinguish samples
that do not contain a specific analyte from sample that contain low concentrations of the analyte;
the lower concentration or among of the target analyte that can be determine to be different from
zero by a single measurement at a stated level of probability. DLs are analyte- and matrix-specific
and may be laboratory dependent.
distribution - (1) the appointment of an environmental contaminant at a point over time, over an
area, or within a volume; (2) a probability function (density function, mass function, or
distribution function) used to describe a set of observations (statistical sample) or a population
from which the observations are generated.
environmental conditions - the description of a physical medium (e.g., air, water, soil, sediment)
or a biological system expressed in terms of its physical, chemical, radiological, or biological
characteristics.
Final
EPA QA/G-4 D- 2 August 2000
-------
environmental data -any measurements or information that describe environmental processes,
location, or conditions; ecological or health effects and consequences; or the performance of
environmental technology. For EPA, environmental data include information collected directly
from measurements, produced from models, and compiled from other sources such as data bases
or the literature.
environmental processes - manufactured or natural processes that produce discharges to, or that
impact, the ambient environment.
environmental programs - work or activities involving the environment, including but not
limited to: characterization of environmental processes and conditions; environmental monitoring;
environmental research and development; and the design, construction, and operation of
environmental technologies; and laboratory operations on environmental samples.
environmental technology - an all-inclusive term used to describe pollution control devices and
systems, waste treatment processes and storage facilities, and site remediation technologies and
their components that may be utilized to remove pollutants or contaminants from, or to prevent
them from entering, the environment. Examples include wet scrubbers (air), soil washing (soils),
granulated activated carbon unit (water), and filtration (air, water). Usually, this term applies to
hardware-based systems; however, it can also apply to methods or techniques used for pollution
prevention, pollutant reduction, or containment of contamination to prevent further movement of
the contaminants, such as capping, solidification or vitrification, and biological treatment.
estimate - a characteristic from the sample from which inferences on parameters can be made.
false acceptance decision error - the error that occurs when a decision maker accepts the
baseline condition when it is actually false. Statisticians usually refer to the limit on the possibility
of a false acceptance decision error as beta (P) and it is related to the power of the statistical test
used in decision making. An alternative name is false negative decision error.
false negative decision error - see false acceptance decision error.
false positive decision error - see false rejection decision error.
false rejection decision error - the error that occurs when a decision maker rejects the baseline
condition (null hypothesis) when it actually is true. Statisticians usually refer to the limit on the
possibility of a false rejection decision error as alpha (a), the level of significance, or the size of
the critical region, and it is expressed numerically as a probability. An alternative name is false
positive decision error.
field variability - see sampling design error.
Final
EPA QA/G-4 D- 3 August 2000
-------
gray region - the range of possible parameter values near the action level where the cost of
determining that the alterative condition is true outweighs the expected consequences of a
decision error. It is an area where it will not be feasible to control the false acceptance decision
error limits to low levels because the high costs of sampling and analysis outweigh the potential
consequences of choosing the wrong course of action. It is sometimes referred to as the region
where it is "too close to call."
limits on decision errors - the acceptable decision error rates established by a decision maker.
Economic, health, ecological, political, and social consequences should be considered when
setting limits on decision errors.
mean - a measure of central tendency. A population mean is the expected value ("average"
value) from a population. A sample mean is the sum of all the values of a set of measurements
divided by the number of values in the set.
measurement error - the difference between the true or actual state and that which is reported
from measurements. Also known as measurement variability.
median - a measure of central tendency, it is also the 50th percentile. The sample median is the
middle value for an ordered set of n values; represented by the central value when n is odd or by
the average of the two most central values when n is even.
medium - a substance (e.g., air, water, soil) that serves as a carrier of the analytes of interest.
natural variability - the variability that is inherent or natural to the media, objects, or people
being studied.
null hypothesis - see baseline condition.
parameter - a description measure of a characteristic of a population. For example, the mean of
a population (ji).
percentile - a value on a scale of 100 that indicates the percentage of a distribution that is equal
to or below it. For example, if 100 ppm is the 25th percentile of a sample, then 25 percent of the
dat are less than or equal to 10 ppm and 75 percent of the dat are greater than 10 ppm.
planning team - the group of people who perform the DQO Process. Members include the
decision maker (senior manager), site manager, representatives of other data users, senior
program and technical staff, someone with statistical expertise, and a quality assurance and quality
control advisor (such as a QA Manager).
population - the total collection of objects or people to be studied and from which a sample is to
be drawn.
Final
EPA QA/G-4 D- 4 August 2000
-------
precision - a measure of mutual agreement among individual measurements of the same property,
usually under prescribed similar conditions expressed generally in terms of the standard deviation.
quality assurance (QA) - an integrated system of management activities involving planning,
implementation, documentation, assessment, reporting, and quality improvement to ensure that a
process, item, or service is of the type and quality needed and expected by the customer.
QA Project Plan - a document describing in comprehensive detail the necessary quality
assurance, quality control, and other technical activities that should be implemented to ensure that
the results of the work performed will satisfy the stated performance criteria.
quality control (QC) - the overall system of technical activities that measure the attributes and
performance of a process, item, or service against defined standards to verify that they meet the
stated requirements established by the customer; operational techniques and activities that are
used to fulfill requirements for quality.
quality system - a structured and documented system describing the policies, objectives,
principles, organizational authority, responsibilities, accountability, and implementation plan of an
organization or ensuring quality in its work processes, products (items), and services. The quality
system provides the framework for planning, implementing, documenting, and assessing work
performed by the organization and for carrying out required quality assurance and quality control.
range - the numerical difference between the minimum and maximum of a set of values.
sample - (a) a single item or specimen from a larger whole or group, such as any single sample of
any medium (e.g., air, water, soil); or (b) a group of samples from a statistical population whose
properties are studies to gain information about the whole. The definition is decided by context of
usage.
sampling - the process of obtaining a subset of measurements from a population.
sampling design - a design that specifies the final configuration of the environmental monitoring
effort to satisfy the DQOs. It includes what types of samples or monitoring information should be
collected; where, when, and under what conditions they should be collected; what variables are to
be measured; and what quality assurance and quality control components will ensure acceptable
sampling error and measurement error to meet the decision error rates specified in the DQOs.
The sampling design is the principal part of the QA Project Plan.
sampling design error - the error due to observing only a limited number of the total possible
values that make up the population being studied. Sampling errors are distinct from those due to
imperfect selection; bias in response; and mistakes in observation, measurement, or recording.
Also known as field variability.
Final
EPA QA/G-4 D- 5 August 2000
-------
stakeholder - a person or organization having an interest in the development of the project.
standard deviation - a measure of the dispersion or imprecision of a sample or population
distribution expressed as the positive square root of the variance and has the same unit of
measurement as the mean.
statistic - a function of the sample measurements (e.g., the sample mean or sample variance).
standard operating procedure - a written document that details the method for an operation,
analysis, or action with thoroughly prescribed techniques and steps and that is officially approved
as the method for performing certain routine or repetitive tasks.
total study error - the sum of all the errors incurred during the process of sample design through
data reporting. This is usually conceived asa sum of individual variances at different stages of
sample collection and analysis. Also known as total variability.
total variability - see total study error.
type I error - the statistical term for false rejection decision error.
type II error - the statistical term for false acceptance decision error.
variability - refers to observed difference attributable to heterogeneity or diversity in a
population. Sources of variability are the results of natural random processes and stem from
environmental differences among the elements of the population. Variability is not usually
reducible by further measurement but can be better estimated by increased sampling.
variance - a measure of a the dispersion of a set of values. Small variance indicating a compact
set of values; larger variance indicates a set of values that is far more spread out and variable.
Final
EPA QA/G-4 D- 6 August 2000
------- |