x>EPA
                  United States
                  Environmental Protection
                  Agency
              Industrial Environmental Research
              Laboratory
              Cincinnati OH 45268
EPA-600-'9-79-046
December 1979
                  Research and Development
Quality Assurance
Guidelines for lERL-Ci
Project Officers

-------
                                            EPA-600/9-79-046
                                            December 1979
        QUALITY ASSURANCE GUIDELINES
        FOR IERL-CI PROJECT OFFICERS
                     by

        C.L.  Stratton and J.D.  Bonds
 Environmental Science and Engineering,  Inc.
            Post Office Box 13454
         Gainesville, Florida  32604
           Contract No.  68-03-2656
               Project Officer

                Paul E. Mills
          Quality Assurance Branch
Industrial Environmental Research Laboratory
           Cincinnati, Ohio  45268
INDUSTRIAL ENVIRONMENTAL RESEARCH LABORATORY
     OFFICE OF RESEARCH AND DEVELOPMENT
    U.S.  ENVIRONMENTAL PROTECTION AGENCY
           CINCINNATI, OHIO  45268

-------
                                 DISCLAIMER
     This report has been reviewed by the Industrial Environmental Research
Laboratory, U.S. Environmental Protection Agency, and approved for publica-
tion.  Approval does not signify that the contents necessarily reflect the
views and policies of the U.S. Environmental Protection Agency, nor does
mention of trade names or commercial products constitute endorsement or
recommendation for use.
                                      ii

-------
                                  FOREWORD
     When energy and material resources are extracted, processed, converted,
and used, the related pollutional  impacts on our environment and even on  our
health often require that new and  increasingly more efficient pollution
control methods be used.  The Industrial Environmental Research Laboratory-
Cincinnati (lERL-Ci) assists in developing and demonstrating new and  improved
methodologies that will meet these needs both efficiently and economically.

     These quality assurance guidelines consist of three sections:

     (1)  Quality assurance guidelines for procurement of projects  requiring
          sampling and analysis
     (2)  Quality assurance guidelines for monitoring of projects requiring
          sampling and analysis
     (3)  Quality assurance guidelines for auditing of projects requiring
          sampling and analysis

     The first section provides guidelines and checklists to assist the
Project Officer in project conception, preparation of procurement requests,
evaluation of proposals, and recommendations for selection to the Contracting
Officer.  The second section presents quality assurance aspects and
checklists to assist in monitoring projects from project initiation through
the final report; and the third section presents information in relation  to
planning and conducting project audits.

     Further information may be obtained through the Industrial Environmental
Research Laboratory-1 s Quality Assurance Officer, Cincinnati, Ohio.

                                              David G. Stephan
                                                  Director
                                Industrial Environmental Research Laboratory
                                              Cincinnati, Ohio
                                     iii

-------
                                  ABSTRACT
     This report presents quality assurance guidelines to assist Project
Officers in the procurement, monitoring, and auditing phases of extramural
projects requiring sampling and analysis.

     The first section presents guidelines to  insure that quality  assurance
(QA) is adequately addressed during project conception and solicitation and
that prospective grantees are  informed of QA requirements.  A technical eval-
uation system is presented that should disqualify those offerers who do not
provide adequate sampling and  analysis QA for  the purposes of the  program.  A
checklist is provided for the  evaluation of the quality assurance  aspects of
proposals and grant applications.

     In the second section, the Project Officer's responsibilities  are
described for the initiation, monitoring, and  satisfactory conclusion of
contracts, research and demonstration grants,  and cooperative agreements of
the type normally funded by lERL-Ci.  The three basic foundations  of project
quality assurance are described:  (1) the contractor's/grantee's quality
assurance program, (2) the Project Work Plan,  and (3) quality assurance
monitoring of the performance  of the contractor/grantee.  Checklists are
included to assist the Project Officer in assessing the completeness of a
contractor's or grantee's overall QA program and his Project Work  Plan.

     The third section provides information concerning the scheduling and
performance of laboratory audits.  Checklists  are provided to assist in
performing the audit.

     These quality assurance guidelines were submitted in fulfillment of
Contract No. 68-03-2656 by Environmental Science and Engineering,  Inc. under
the sponsorship of the U.S. Environmental Protection Agency.  This  report
covers the period August 1978  to December 1979, and work was completed as of
December 1979.
                                     iv

-------
                                  CONTENTS
Disclaimer	    ii
Foreword	   iii
Abstract	    iv
Figures	   vii
Tables	viii
1.  QUALITY ASSURANCE GUIDELINES FOR PROCUREMENT OF PROJECTS REQUIRING
    SAMPLING AND ANALYSIS

     Introduction	     1
         Importance of Quality Assurance 	     1
         Purpose of These Guidelines 	     2
         Definitions of Quality Assurance and
         Quality Control 	     3
         QA in the Procurement Process	     5
     QA Criteria in Project Conception and Solicitation	     8
         Project Conception (Contracts)	     8
         Contract Solicitation 	    12
         Project Conception and Application for
         Research and Demonstration Grants 	    19
         Use of Performance Test Samples in
         Contractor/Grantee Selection	    20
         Cost Considerations	    21
         Preproposal Conference	    23
     QA Criteria in Technical Evaluation of Proposals and Grant
     Applications	    23
         Technical Evaluation of Contract Proposals	    23
         QA Evaluation Criteria and Scoring Procedures
         for Contracts	    24
         Written/Oral Discussions with Offerers	    25
         QA Evaluation of Grant Applications 	    28
         Evaluation of Performance Test Samples	    29
     Evaluation of Previous Performance History	    29
     Cost Evaluation	    30
     Best and Final Offer Evaluation 	    31
     Pre-Award Surveys 	    31

-------
                            CONTENTS (Continued)
2.  QUALITY ASSURANCE GUIDELINES FOR MONITORING OF
    PROJECTS REQUIRING SAMPLING AND ANALYSIS

     Introduction	    35
     Project Officer's QA Responsibilities 	    37
         Contracts	    37
         Grants and Cooperative Agreements 	    40
         Summary of the Project Officer's Role in
         Achieving QA	    42
     The Contractor's/Grantee's QA Program 	    43
         Elements of a QA Program	    43
         QA Program Checklist	    47
     The Project Work Plan	    48
         Elements of a Project QA Plan	    48
         Project Work Plan Checklist	    50
     Project QA Monitoring 	    50
         Methods of Monitoring Contractor/Grantee QA 	    50
3.  QUALITY ASSURANCE GUIDELINES FOR AUDITING OF
    PROJECTS REQUIRING SAMPLING AND ANALYSIS

     Introduction	    53
         Quality Assurance Audits	    53
         Quality Assurance in Audits 	    53
     Quality Assurance Audit Guidelines	    54
         When to Conduct a Quality Assurance Audit 	    54
         Audit Worksheet and Checklist	    54
         Use of Performance Test Samples	    58
     Conducting the Quality Assurance Audit	    58
         Review of Worksheet	    58
         Site Visit	    60
         Quality Assurance Audit Report	    62
4.  CONCLUSIONS

     Procurement	    67
     Monitoring	    67
     Auditing	    67
BIBLIOGRAPHY 	    68
                                     VI

-------
                            CONTENTS (Continued)
APPENDICES
     A.  Quality Assurance Evaluation Criteria Checklist for
         Proposals and Grant Applications Offering Sampling
         and Analysis Services 	    71
     B.  Quality Control Performance/Reference Test Samples	    75
     C.  EPA-Accepted Analytical Methods 	    79
     D.  Quality Assurance Program Checklist 	    93
     E.  Project Quality Assurance Plan Checklist	   105
     F.  Quality Assurance Pre-Audit Worksheet 	   113
     G.  Instrumentation, Equipment, and Personnel Skill
         Rating for Specific Methods 	   126
     H.  Sample Preservation Methods and Recommneded
         Holding Times 	   146
     I.  Quality Assurance Audit Checklist  	   150
GLOSSARY	   179
                                     Vll

-------
                                   FIGURES

Number                                                                  Page
       Quality assurance (QA) and quality control (QC)
       at lERL-Ci  	
       Processing sequence for contract source evaluation
       and selection 	
       Processing sequence for research and demonstration
4
S
fS
7
8
9
Example of use of QA evaluation criteria checklist
for scoring proposals 	
Example of technical evaluation scoring system . . . .
Foundations of project quality assurance 	
Contracts administration ..........
Grants and cooperative agreements administration.
Quality assurance audit report 	
.... 26
.... 27
.... 36
. ... 38
.... 41
.... 63
                                    viii

-------
                                   TABLES
Number                                                                  Page
  1    Assignment of Evaluation Points for QA 	    11

  2    Evaluation of the Importance of Sampling and Analysis QA .  .  .    13

  3    Example of the Assessment of the Importance of
       Sampling and Analysis QA	    15

  4    Example Technical Evaluation Criteria Incorporating
5
6
7
8
9
Suggested Criteria for Using Performance Test Samples
for Offerer Evaluation 	
Criteria for Conducting a Pre~Award Survey 	
Suggested Pre-Award Survey Agenda 	
Available Quality Assurance Audit Procedures 	
Recommended Agenda for QA Audit Site Visit 	
. . . . 22
. . . . 32
. . . . 34
. . . . 56
. . . . 61
                                     IX

-------
                                   SECTION  1

               QUALITY ASSURANCE  GUIDELINES  FOR  PROCUREMENT  OF
                  PROJECTS  REQUIRING  SAMPLING  AND  ANALYSIS
INTRODUCTION

Importance of Quality Assurance

     Quality assurance  (QA) encompasses  all  actions  taken  by  an organization
to achieve accurate and  reliable  results  for  programs  undertaken.   An estab-
lished QA program  is essential  for  any organization  to produce  valid  sampling
and analytical data to  support  research,  demonstration,  and monitoring
efforts.  QA requirements  are becoming increasingly  more  important  to the
Environmental Protection Agency (EPA) since:

     1.  The number of  commercial  laboratories  and research institutions
         participating  in  sampling  and analysis  programs  is increasing,
     2.  The sampling and  analytical equipment  and procedures used  are
         becoming more  varied and  complex,
     3.  There is an increasing movement  toward  consolidated  data bases,
     4.  More and more  data must  withstand  legal  scrutiny, and
     5.  Policy decisions  of national economic  importance  must  be made on  the
         basis of reliable data.

     It  is the responsibility of  lERL-Ci  to  assist in  developing and  demon-
strating new and improved  pollution control  technology that will meet our
nation's needs both efficiently and economically.  To  accomplish this,  our
research must conform to the highest practicable  quality  assurance  standards.
Sampling and analytical  programs  are an  important aspect  of most research  and
demonstration projects  conducted  by lERL-Ci.  Decisions  of substantial
technical and economic  importance  are often  based on the  data generated by
lERL-Ci  internal sampling  and analytical  programs, by  programs  conducted by
contractors, and by grantees.*

     The Contracting Officer and  the Project  Officer bear  the responsibility
of procuring and directing sampling and  analysis  services  for lERL-Ci
research and demonstration projects.  To  fulfill  this  responsibility,  they
must be knowledgeable about the basic principles  underlying currently

* Throughout this document, the definition of the word "grant"  includes
  cooperative agreements.

-------
acceptable QA practices.  These guidelines have been prepared  to  assist  the
Project Officer and the Contracting Officer  in procuring high  quality  sam-
pling and analytical services.  They establish standards for acceptable  QA
practices on lERL-Ci research and demonstration projects.  Project Officers
and Contracting Officers should use these guidelines during preparation  of
procurement requests and in the technical evaluation of offers and grant
applications.

     This is the first  in a series of three  sections describing:

     1.  Quality assurance guidelines for procurement of projects requiring
         sampling and analysis,
     2.  Quality assurance guidelines for monitoring projects  requiring
         sampling and analysis, and
     3.  Quality assurance guidelines for auditing projects requiring
         sampling and analysis.

     Together, these three sections should assist TERL-Ci Project Officers in
applying stringent, but equitable, quality assurance requirements throughout
the life of all research and demonstration projects that require  sampling and
analysis.  The requirements should apply to  contractor, grantee,  and in-house
programs equally.

     This report is intended as a working document and will be revised as
necessary to more accurately describe processes, legal obligations, and
current policy.

     The Director has assigned to the lERL-Ci Quality Assurance Officer  the
responsibility of overseeing the quality assurance aspects of  all lERL-Ci
contracts, research and demonstration grants, and in-house research and
demonstration projects.  The Quality Assurance Officer has developed the
guideline document to assist in fulfilling that responsibility.   He is
available to work with Project Officers, Contracting Officers, and in-house
program managers in any way possible to assure the quality of  lERL-Ci
projects.

Purpose of These Guidelines

     This section has been prepared to  insure adequate consideration of  QA
requirements in the procurement stage of lERL-Ci research and  demonstration
projects.  It provides  specific guidelines to:  (1) accurately reflect QA
requirements in the proposal solicitation stages of procurement;  (2) solicit
the necessary QA information from prospective contractors, grantees, and
bidders to allow accurate evaluation of capabilities; (3) evaluate proposals
received with respect to QA; (4) evaluate past performance history of  the
prospective contractor  or grantee, if available; and (5) determine the
overall acceptability of an offerer's QA program with respect  to  the research
or demonstration project under consideration.

-------
     These guidelines are intended to be used by lERL-Ci Project Officers
responsible for procuring sampling and analysis services in support of
lERL-Ci research and demonstration projects.  They should also be used  in the
evaluation of research and demonstration grant applications when sampling and
analysis aspects are involved.  Contracting Officers will find these guide-
lines of value in preparing Request for Proposal (RFP) packages when
soliciting service contracts that include sampling and analysis.

Definitions of Quality Assurance and Quality Control

     The terms quality assurance (QA) and quality control (QC) are often used
synonymously, although they represent two distinct concepts.  The following
definitions have been established by the Director, lERL-Ci.

     Quality Control (QC):    Actions taken by the Laboratory (lERL-Ci)
                              organization (on in-house projects) and by
                              contractors/grantees (on extramural projects)
                              in day-to-day activities to achieve desired
                              accuracy, reliability, and comparability  in the
                              results obtained from sampling and analysis
                              activities.  Review by contractors/grantees of
                              their overall quality control activities  is
                              "quality assurance" to them, but "quality
                              control" from the Laboratory's viewpoint.

     Quality Assurance (QA):  Actions taken by the Laboratory (lERL-Ci) line
                              organization under the specific auspices  of the
                              Office of the Director, to assure that quality
                              control policies and procedures are being
                              properly implemented and appropriate levels of
                              accuracy, reliability, and comparability  are
                              being achieved in the sampling and analysis
                              activities (including data reduction and
                              handling) of the Laboratory to fulfill the
                              Laboratory's assigned mission.

     In broad terms, QA is the overall program that specifies the quality
control practices applied to the many individual aspects of a program.
Figure 1 graphically presents QA and QC responsibilities at lERL-Ci for both
in-house and contractor/grantee sampling and analytical programs.

     Simplified definitions of quality assurance and quality control have
been adopted by the Office of Monitoring and Technical Support, Office  of
Research and Development to aid in eliminating confusion.  "Quality assurance
is the total program for assuring the reliability of monitoring data.
Quality control is the routine application of procedures for controlling the
measurement process."

-------
                         OFFICE OF THE DIRECTOR
    DIVISIONAL
QUALITY ASSURANCE
                       QUALITY ASSURANCE OFFICER
       DIVISIONAL
IN-HOUSE QUALITY CONTROL
                                                       IN-HOUSE
                                                 SAMPLING AND ANALYSIS
                        LABORATORY QUALITY ASSURANCE

                          EXTERNAL QUALITY CONTROL
CONTRACTOR/GRANTEE
QUALITY ASSURANCE
CONTRACTOR/GRANTEE
 QUALITY CONTROL
CONTRACTOR/GRANTEE
SAMPLING & ANALYSIS
SOURCE:  In-house memorandum from lERL-Ci Director, March 1978.
Figure 1.  Quality assurance (QA) and quality control (QC) at lERL-Ci.

-------
QA in the Procurement Process

     Quality assurance begins with, and  is most directly  controlled  in,  the
procurement process.  If the Contracting Officer  and  the  Project Officer  fail
to procure high quality sampling and analytical services,  then  successful
project accomplishment is seriously jeopardized from  the  beginning.   Hence,
prime emphasis must be placed on giving  adequate  attention  to QA during  pro-
curement of sampling and analysis services in support of  lERL-Ci programs.

Overview of the Contracts Procurement Process—
     Many lERL-Ci research and demonstration projects are  awarded  on  a
competitive contract basis to qualified  commercial contractors.  The
procurement process for contracts is described  in Procurement Information
Notice (PIN) 77-15, which is entitled "Source Evaluation  and Selection
Procedures." It is EPA policy that source selection and evaluation shall  be
conducted in accordance with standards and procedures that  insure  fair and
impartial treatment of all offerers, and further  insure the selection of
sources whose performance is expected to best meet EPA objectives.

     The contract procurement processing sequence is  illustrated in Figure  2.
This figure also indicates those points  in the  procurement  process where  QA
considerations should apply and the page of this  section  where  each step  in
the processing sequence is discussed.

     The contract procurement process begins with conception of the work  to
be performed.  This conception is translated into a solicitation by comple-
tion of a procurement request rationale  document  with accompanying EPA Form
No. 1900-8 (Procurement Request/Requisition).  QA requirements  must be clear-
ly specified within the scope of work sections  of the procurement  request
rationale document and must be given a proper weighting in  the  evaluation
criteria.  After advertisement for interested contractors,  a Request  for  Pro-
posal (RFP) is prepared for the Project Officer by the Contracts Management
Division.  The QA requirements delineated in the  procurement request  ration-
ale document are described in the RFP and will  eventually  become a part  of
the contract.  A preproposal conference may be  requested  by the Project
Officer and called by the Contracting Officer to  further  discuss procurement
requirements with prospective offerers.  If this  is the case, prospective
offerers should be advised of the level  of importance attached  to QA.  After
receipt of offers, a series of technical reviews  is conducted by the
Technical Evaluation Panel.  QA aspects must be given proper consideration
during each of the technical reviews.  The initial objective at this  stage  is
to disqualify those offerers who do not  propose the minimally required QA,
provided the evaluation requirements have been  clearly and  adequately stated
in the RFP.  Beyond that, it is of value to rank  the  responsive offerers  with
respect to QA and all other technical aspects as  stated in  the  proposal
evaluation criteria.  A checklist has been provided in these guidelines  to
assist in the ranking of offerers in this regard.  The Technical Evaluation
Panel may elect to provide performance test samples to the  responsive
offerers if this provision was stated in the RFP.

-------
          *PROJECT CONCEPTION  (Page 8)

       *PROCUREMENT REQUEST  (Pages 8, 12,  21)

   *DEVELOP EVALUATION CRITERIA  (Pages 8,  12,  20)

        PREPARE AND ISSUE SOLICITATION

    *PREPROPOSAL CONFERENCE  (OPTIONAL) (Page 23)

               RECEIVE OFFERS

*PRELIMINARY TECHNICAL REVIEW	^-TECHNICALLY  UNACCEPTABLE	»-REJECT  OFFER

     *TECHNICAL AND COST EVALUATION
           (Pages 23, 24, 30)

         DETERMINE COMPETITIVE RANGE	»- NOT IN COMPETITIVE RANGE	
     *CONDUCT WRITTEN/ORAL DISCUSSIONS (Page 25)

         REQUEST "BEST AND FINAL" OFFERS

        *FINAL EVALUATION
          (Pages 29, 31)
                       I
        *PRE-AWARD SURVEY (OPTIONAL)	
                (Page 31)

         SELECT SOURCE FOR NEGOTIATIONS

               CONDUCT NEGOTIATIONS

                 AWARD CONTRACT
*QA considerations are  important at these points  in the process.
Figure 2.  Processing sequence  for contract source evaluation and  selection.

-------
     Qualified offerers are  identified  on  the  basis  of  technical  review.   A
business and cost evaluation of  the offer  is conducted  at  the  same  time.
Since the cost to conduct an effort requiring  sampling  and  analysis  tasks  is
clearly affected by the QA methods employed, QA must  be given  proper consid-
eration in the business evaluation.  The costs of  the necessary QA  program
for the sampling and analysis aspects are  a  legitimate  charge  to  the project,
and the costs are covered in the contract  by EPA.  After both  the  technical
and business evaluation, the competitive range is  established, and  offerers
are notified by the Contracting Officer whether or not  they are  in  the
competitive range.

     According to agency regulations, written  or oral discussions must  be
conducted with all responsive offerers  who have submitted  proposals  within
the competitive range.  During these discussions,  any uncertainties  should be
resolved concerning the offerer's compliance with  the specified QA  aspects of
the effort.  Following the receipt of "best  and final"  offers, it  is fre-
quently of value to conduct  a pre-award survey of  the offerers in  the final
competitive range.  The pre-award survey provides  an  ideal  opportunity  for
the Project Officer to gain  an in-depth knowledge  of  the QA programs of the
various offerers.  For this  reason, pre-award  surveys should be encouraged
for projects in which sampling and analysis  constitute  a large portion  of  the
effort or are, in themselves, a major cost factor.

     After "best and final"  offers are  received, a final evaluation, which
may include a pre-award survey,  is conducted,  and  a  determination made  of  the
source to be selected for negotiations.  Negotiations are  conducted, and  the
business evaluation is completed by performance of a  financial audit or a
cost advisory.  When this step is successfully concluded,  a contract is
awarded to the successful bidder.

Overview of the Research and Demonstration Grant or Cooperative Agreement
Procurement Process—
     lERL-Ci research and demonstration projects are  also  conducted  by  the
granting of funds to qualified research institutions  for the entire  program
or for various portions of a program.   The Project Officer's role  in the
procurement process for research and demonstration grants  is described  in  the
"EPA Project Officer's Guide (Research  & Demonstration  Grants)"  (EPA,
undated).  The procurement process for  research and  demonstration  grants  is
entirely different than for  contracts.  Unlike contractors, grantees are  not
normally selected on a competitive basis.  The assigned Project Officer or
the prospective Project Officer  for a research and demonstration  grant  has a
much more interactive role with  the prospective grantee.   He,  in  fact,  may be
the principal EPA contact with the prospective grantee  during  the  application
and processing stages, whereas this is  specifically  prohibited for  contracts.
He has major authority in the selection of a grantee  and the awarding of  the
grant.  Assistance is provided to the Project Officer by the Grants
Administration Division (GAD) in Headquarters.  A  technician in  the  Grants
Operations Branch of GAD is  assigned to assist the Project  Officer  with
administrative matters during the procurement  process and  throughout the  life
of the project.

-------
     The processing  sequence  for  research  and  demonstration grants  and
cooperative agreements  is shown  in Figure  3.   QA  considerations  enter  the
procurement process  during preparation  and  technical  review of  the
application.

     After preliminary  discussions with  the  appropriate  program  office,  a
research or demonstration grant  application  is  normally  submitted  in an  area
of particular interest  to or  expertise  of  the  applicant.   The Project  Officer
may advise the applicant  in the  preparation  of  the  formal  grant  application.
One objective at this point in the process  is  to  assure  that  the applicant
gives adequate regard to  the  QA  aspects  of  his  work.   The  applicant  should  be
provided guidance on the QA requirements of  the Laboratory (lERL-Ci) and,  in
his application, should commit to meeting  these requirements  as  stated in
Section 2, "Quality  Assurance Guidelines for Monitoring  of Projects  Requiring
Sampling and Analysis."

     QA aspects should  be given  an appropriate  and  equitable  level  of  atten-
tion during the technical review  of  grant  applications.  Even though the
evaluation criteria  are more  subjective  than  for  contracts,  similar  QA cri-
teria can be used for guidance as discussed  in  this  section under  "QA  Eval-
uation of Grant Applications" on  page 28.   Pre-award  surveys  of  prospective
grantees may be conducted and, in fact,  are  strongly  encouraged  when possi-
ble.  Performance testing may also be used  to  evaluate an  applicant.

     During the decision  process  on  whether  or  not  to fund a  grant,  negotia-
tions with the applicant may  occur.  If  this  is the  case,  the Project  Officer
works with the applicant  to modify the  proposal to  meet  the objectives of  the
program office.  For example, if  the proposed  QA  procedures were found to  be
inadequate or inappropriate,  they may be adjusted by  the applicant  at  this
time.

QA CRITERIA IN PROJECT  CONCEPTION AND SOLICITATION

Project Conception (Contracts)

     QA must be given proper  consideration  in  the conceptual  stages  of a
project.  Project sampling and analysis  programs  may  range from  non-existent
to greatly important for  the  outcome of  the  work, regardless  of  the  magnitude
of the sampling and  analytical effort.   Correspondingly, the  resulting data
may be ancillary to  the final outcome of the  effort  or of  vital  importance.
Since cost effectiveness  is another  criterion  to  be  considered,  and  since  QA
programs affect cost, judgment must  be  used  in stipulating QA requirements.
It is not categorically true  that the more  QA  applied to a given sampling  and
analysis program, the more effective that  program will be.

     The level of importance  ascribed to QA in any  given project must  be
clearly transmitted  through the  Contracting Officer  to those  organizations
that will propose to accomplish  the  work.   QA  requirements are  to  be

-------
                PRE-APPLICATION
     *PRE-APPLICATION  DISCUSSIONS  (Page  19)


                     J
      *FORMAL GRANT  APPLICATION  (Page  19) •+	



      INITIAL ADMINISTRATIVE  PROCESSING



              RELEVANCE  REVIEW	»-NON-RELEVANT-
                 RELEVANT
     *TECHNICAL REVIEWS  (Pages 28, 29)



     *PRE-AWARD SURVEY (OPTIONAL)  (Page 31)
         FUNDING DECISION	^REJECT-
                     I   \
            -NOTIFY APPLICANT
                   FUND
NEGOTIATE
MODIFY APPLICATION'
                   AWARD
*QA considerations are important at these points  in the  process.
Figure 3.  Processing sequence for research and demonstration grants and
           cooperative agreements.

-------
specified  in  the  procurement  request  rationale  document  with  accompanying EPA
Form 1900-8 (Procurement Request/Requisition) and,  ultimately,  in the RFP
package prepared  by  the Contracting Officer.  The QA requirements must be
adequately defined either  in  the  scope  of  work  section  of the procurement
request document  or  by attachments  to  this document.  The proper level of
importance must be assigned  in  the  scoring system of the proposal evaluation
criteria.

     It is recommended that  specific  evaluation criteria and  points  be
assigned to the offerer's  QA  program when  sampling  and  analysis  is required
to accomplish  the project.  Guidance  is  provided in this section for assign-
ing evaluation points.  Many  other  criteria routinely applied in the evalua-
tion of proposals are closely allied with  QA considerations  and  should be
viewed in  this regard, however.   For  example, proposal  evaluation criteria
and points are generally assigned to  demonstrated qualifications and experi-
ence of key project  personnel.  The availability of highly competent person-
nel on the offerer's staff is clearly  a  benefit to  overall QA.   Furthermore,
an offerer must have or must  obtain suitable sampling equipment  and  analy-
tical instrumentation in order  to accomplish high quality work.   Since
facilities and personnel qualifications  are generally awarded evaluation
points, QA aspects will normally  enter  into the evaluation criteria  under
several headings.

     As general guidance,  when  QA of  the sampling and analytical program is
of critical importance to  the success  of a project,  up  to 30  percent of the
evaluation points should be assigned  specifically to this area.   If  an
offerer fails  to  show competence  in QA,  he should be technically disqualified
overall and should not be  privileged  to  receive further  technical review.  An
offerer may be disqualified by  this method only if  it is clearly stated in
the RFP.   Conversely, if the  sampling  and  analytical effort  is  only  of
ancillary  importance to the project, as  few as  5 percent  of the  total
evaluation points should be assigned to QA.  This should  be a minimum  point
assignment.  If QA of the  sampling and analysis  effort  is  not worth  at  least
5 percent of the total technical  evaluation points,  then  the  necessity  of the
sampling and analysis effort  should be questioned.   The  individual preparing
the procurement request must  make a judgment of  the  relative  importance of QA
efforts in order to assign the  proper weighting within the range of  5  to
30 percent of the total technical evaluation points.  The  point  range  should
be adjusted so as to disqualify offers where the proposed  QA  program is
totally inadequate for the purposes and objectives  of the  project.

     The relative importance  of QA efforts  should be  assessed by determin-
ation of the ultimate use  of  the  sampling  and analytical  data,  the methods
required,  and the magnitude of  the effort.  Table 1  is provided  to assist the
Project Officer, who prepares the procurement request,  in  assigning  technical
evaluation points for the  QA  program offered by  each  RFP  respondent.   The
point ranges in Table 1 are broad and overlapping.   The  Project  Officer must,
therefore, use his judgment in  selecting the appropriate  point assignment,
taking into consideration  all other evaluation  criteria  and technical  aspects
of the project.
                                     10

-------
             TABLE 1.  ASSIGNMENT OF EVALUATION POINTS FOR QA
Relative importance of
sampling and analysis data
 Suggested assignment
of points (% of total)
Sampling and analysis data of only ancillary
importance to the overall objectives of the
project or sampling and analysis are only a
very small portion of the total effort

Semi-quantitative data is adequate

Data will be used for screening purposes
and will probably be validated by other data

Data will be used to make economically-
important decisions on equipment design

Sampling and analysis is a substantial
portion of the total effort

Data will be used for regulatory support

All data will clearly be subject to legal
scrutiny and defense
       (minimum)



          5-10

         10-15


         15-25


         20-25


         25-30

           30
                                     11

-------
     Table 2 may be of  further  assistance  to  the  Project  Officer.   This  table
is a checklist that can be used  to assess  the  relative  level  of  importance  of
the sampling and analysis QA program by  rating  the  requirements  of  the  pro-
gram based on the questions listed.  Ten questions  are  provided, with  a  pos-
sible total score of 30 points.  The Project Officer  should rate his  response
to each question and add the total points.  The total point value placed on
these responses corresponds to  the suggested  approximate  assignment  of  points
in the technical evaluation criteria, on the  basis  of the percent of  the
maximum evaluation point score  for all technical  criteria.  The  suggested
point assignment arrived at by  use of Table 2  should  also correspond  with the
appropriate ranges suggested in  Table 1.   Hence,  either approach can  be  used
to determine the approximate value of technical evaluation points for  Quality
Assurance.

     Example:  Requirement to demonstrate  the  feasibility of  a new  type  of
waste treatment process in the  removal of  numerous  contaminants  from  waste-
waters produced by shale oil recovery processes.  Bench scale tests have
shown the treatment process works exceptionally well, and it  has been  decided
by EPA to fund a pilot  plant study.  If  the results  from  the  pilot  treatment
plant prove successful, further  scale-up will  be  attempted, and  the  shale oil
industry may be required to install this type  of  treatment facility  at  each
location at a cost of millions  of dollars.  The approved  funding for  this
contract is $450,000, and the Project Officer  estimates that  $50,000  will be
expended for analytical laboratory services.   Sampling  and analysis will be
conducted of wastewaters typical of the  shale  oil industry during pilot  plant
tests.

     The Project Officer, in this example, may  evaluate the importance  of
sampling and analytical QA as shown in Table  3.   A  total  of 23 points  has
been assigned.  Hence,  the Project Officer should assign  points  within  the
range of 20-25 percent  of the total evaluation  points to  sampling and
analytical QA.

     Alternatively, the Project  Officer  may have  drawn  the same  conclusion  by
assessing the evaluation points  in Table 1.   Since  the  data clearly will be
used to make economically-important decisions  and to  design major equipment,
the point total for QA  should lie in the high  end of  the  15-25 percent
range.

     Once the proper relative weighting  for QA  is established, the  evaluation
criteria for this solicitation  may appear  as  shown  in Table 4.   It  should
also be stated in the RFP that  a bidder  must  show adequate qualifications in
each Quality Assurance  area to  be qualified for consideration with  regard to
the other technical areas.

Contract Solicitation

     The Contracting Officer will prepare  a Request  for Proposal (RFP)  pack-
age based on the information submitted in  the  procurement request rationale
document with accompanying EPA  Form 1900-8 (Procurement Request/Requisition).
                                      12

-------
    TABLE 2.  EVALUATION OF THE IMPORTANCE OF SAMPLING AND ANALYSIS  QA

                                             Low                      High
                                              01         23
Evaluation considerations                  «10%)  (10-50%)  (50-90%)  (>90%)
What percent of the total project effort
is related to sampling and analysis?

What is the probability the resulting
data or conclusions drawn from the data
will be subject to legal scrutiny?

What is the probability the data will
be used to make economically-important or
policy decisions?

What is the likelihood the data will be
used to set compliance standards or
performance standards?

What is the probability the data will be
used for equipment or facilities design?

What possibility exists that the data will
be used for regulatory support functions?

Is the data to be used only for screening
purposes or for validation purposes?  If
for screening only, assign relatively low
importance (0 to 1).  If for validation,
assign relatively high importance (2 or 3).

Is semi-quantitative data adequate for the
needs of the project, or is a high degree
of accuracy required?  Assign importance
relative to the degree of quantitative
accuracy required:  0 if semi-quantitative
data is adequate; 3 if the highest possible
accuracy is necessary.
(continued)
                                      13

-------
                            TABLE 2 (continued)
                                             Low                     High
                                              01        23
Evaluation considerations                  «10%)  (10-50%)  (50-90%)  (>90%)
Are a variety of sampling and analytical
methods required, including complex
methods?  Assign relative importance,
compared to other lERL-Ci projects,
based on the degree of variety and
complexity of the methods required.

What is the possibility that the sampling
and analytical data will be included in a
consolidated data base?
                                           TOTAL POINTS

                                           Approximate percent of proposal
                                           technical evaluation points to
                                           be assigned to QA procedures.
                                     14

-------
           TABLE 3.  EXAMPLE OF THE ASSESSMENT  OF  THE  IMPORTANCE
	OF SAMPLING AND ANALYSIS QA	

                                             Low                      High
                                              01         23
Evaluation considerations                   «10%)  (10-50%)  (50-90%)  (>90%)
What percent of the total project effort
is related to sampling and analysis?

What is the probability the resulting
data or conclusions drawn from  the data
will be subject to legal scrutiny?

What is the probability the data will
be used to make economically-important or
policy decisions?

What is the likelihood the data will be
used to set compliance standards or
performance standards?

What is the probability the data will be
used for equipment or facilities design?

What possibility exists that the data will
be used for regulatory support  functions?

Is the data to be used only for screening
purposes or for validation purposes?  If
for screening only, assign relatively low
importance (0 to 1).  If for validation,
assign relatively high importance (2 or 3).

Is semi-quantitative data adequate for the
needs of the project, or is a high degree
of accuracy required?  Assign importance
relative to the degree of quantitative
accuracy required:  0 if semi-quantitative
data is adequate; 3 if the highest possible
accuracy is necessary.
(continued)
                                      15

-------
                              TABLE 3  (continued)
                                             Low                     High
                                              01        23
Evaluation considerations                   «10%)  (10-50%)  (50-90%)  (>90%)
Are a variety of sampling and analytical
methods required, including complex
methods?  Assign relative importance,
compared to other lERL-Ci projects,
based on the degree of variety and
complexity of the methods required.

What is the possibility that the sampling
and analytical data will be included  in a
consolidated data base?
                                           TOTAL POINTS        |23|

                                           Approximate percent of  proposal
                                           technical evaluation  points  to
                                           be assigned to QA procedures.
                                     16

-------
               TABLE 4.  EXAMPLE TECHNICAL EVALUATION CRITERIA
                INCORPORATING QUALITY ASSURANCE REQUIREMENTS
Evaluation criteria                                 Numerical weight
I.     Adequacy of the Technical Proposal              60
       a.  Logic of approach to the study                      20
       b.  Proposed pilot plant design                         20
       c.  Presentation of findings                            20

II.    Project Management                              50
       a.  Previous experience the Project                     15
           Manager has had in this type of
           effort
       b.  Company resources available to the                  25
           Project Manager
       c.  Project management organization and                 10
           plan

III.   Quality Assurance                               50
       a.  Quality assurance management policy/                10
           written procedures
       b.  Quality assurance procedures for                    15
           sampling
       c.  Quality assurance procedures for                    15
           analysis
       d.  Quality assurance procedures for                    10
           data management

IV.    Personnel Qualifications                        40
       a.  Technical experience of the principal              20
           project staff related to the project
       b.  Educational background of principal                20
           project staff                              	

                                               TOTAL  200
                                     17

-------
The RFP must be written  in  such  a manner  as  to  elicit  adequate  information
from the offerers so  the Technical Evaluation Panel  can  evaluate  each
offerer's technical proposal  as  well  as  the  level  of QA.   In addition,  the
evaluation criteria must be clearly  stated  so that all offerers will  attach
the appropriate level of importance  to QA aspects  of the  effort.

     There are two conceptual  approaches  to  defining QA  requirements  in the
RFP.  One is to state clearly  and unequivocally in the scope of work  section
the QA and QC steps required  of  the  contractor.  It  may  be stated,  for
example, that all samples are  to be  collected in duplicate using  a  specified
technique and that certain  analytical precision must be  achieved  and
demonstrated.  Another approach  is to state  clearly  that  QA and QC  are  of
importance and that the offerer  is to explain his  proposed procedures.
Either approach is acceptable, but both  have shortcomings.

     If the required  QA and QC steps  are  specifically  delineated,  it  is
likely that proposals will  be  submitted  that simply  "parrot" those  require-
ments.  The offeror cannot  very  well  do  otherwise.   As a result,  technical
evaluation of proposals  is  simplified and cost  comparisons are more easily
made.  The Project Officer  and the Contracting  Officer will know  exactly the
level of QC that will be applied, since  it has  been  specified.  This  approach
is recommended for selection  of  contractors  when the scope of work  is clearly
defined, and selection can  be made on costs  proposed by  basically  qualified
bidders.  It is generally impractical, however,  to apply this approach  to
less well-defined efforts.  In such  cases,  statements  should be made  as to
the importance of QA  to  the overall  success  of  the program, and the offeror
should be free to propose QA  procedures  appropriate  to the nature  of  the
sampling and analytical  effort.  This will  allow a better evaluation  of the
offerer's knowledge of QA procedures.  The  offeror should be asked  to provide
a description of his  QA administration and  procedures, to show an  awareness
of proper sampling and analytical techniques, or to  describe his  methods of
documenting data quality and  of  handling  data.   The  Project Officer may
choose to restate the applicable questions  listed  in Table 2 as statements of
fact so that the offeror will  understand  lERL-Ci's interest in QA  on  the
project.

     In the case of the  pilot  plant  example  previously discussed,  a statement
may be made in the RFP to the  effect  that:   "The performance results  of the
pilot plant to be constructed  and evaluated  under  this contract may be  used,
in part, to determine EPA policy on  the  applicability  of this method  of waste
treatment for the shale oil industry.  For  this reason,  considerable  impor-
tance is attached to  quality  assurance of the sampling and analytical aspects
of the scope of work.  Offerers  are  requested to describe, in detail, their
quality assurance procedures  as  they  would  apply to  this  program."

     Specific provisions may  also be  used in a  solicitation to define the
level of QA required  and to reduce the number of non-responsive proposals.
Some typical provisions  are as follows:
                                      18

-------
     1.  "If subcontractors are  to be  used  in  this  effort,  the  prime
         contractor has  full responsibility  for  quality  assurance  of  the
         subcontractor's efforts as well  as  his  own efforts."
     2.  "Offerers whose proposals are determined to be  technically accep-
         table under the initial evaluation  criteria stated  herein shall
         (may) be required to demonstrate acceptable performance by analyzing
         not more than	 unknown samples  for 	 parameters per sample.
         The cost for analysis of these samples  shall  be  at  the offerer's
         expense.  Sample analysis results  shall be compared to known EPA
         results for scoring purposes."
     3.  "Split sample analysis will be required from time  to  time during  the
         course of the project to validate  the quality of the  data.   Split
         samples shall not exceed a maximum  of one  percent  of  the  total
         samples analyzed."
     4.  "lERL-Ci has established basic quality  assurance guidelines  relating
         to the sampling and analysis  required under this contract.   These
         guidelines are  attached.  The successful offerer's  quality assurance
         procedures should generally conform to  these  guidelines or
         significant deviations must be justified and  approved  by  the Project
         Officer."

     The Quality Assurance Officer is  familiar with the  applicability of
typical provisions that may be used in any  particular  case.  He should be
contacted for guidance in this regard.

Project Conception and Application for Research  and Demonstration  Grants

     As discussed previously, the Project Officer may  be  the key contact
person between EPA and a prospective grantee.  The  Agency,  through the Pro-
ject Officer, should make sure that each  applicant  for grant funds receives
full and impartial consideration of his application and  early  notification of
its disposition.  Although grant applications may be submitted  at  any time on
any research and development project,  EPA encourages pre-application  contact,
which can benefit both the applicant and  the Agency by avoiding time-
consuming preparation and review of applications that  cannot be funded and by
sharpening the focus of  proposed projects in terms  of  EPA needs.   It  is  the
responsibility of the Project Officer  to  assure  that grant  allocations are
provided only to those organizations where  work  of  acceptable  quality will be
performed.

     QA should be given  as much consideration  in the allocation of grant
funding as in contracts  or in-house research projects.   It  is  somewhat more
difficult, however, to accord sufficient QA  consideration for  grants, since
competitive selection based on defined evaluation criteria  is  not  generally
practiced.  The Project Officer, therefore,  must assure  that QA is adequately
addressed in the application stages of the  grant process  (see  Figure  3).

     In those cases in which the Project Officer is involved in pre-
application activities, he may be called  upon  to work  with  the  applicant  in
the development of an adequate application  that  meets  the requirements of  EPA
                                     19

-------
and falls within general EPA objectives.  This would  include  assisting  the
applicant in addressing the necessary QA aspects  of sampling  and  analysis
efforts as required by lERL-Ci.  It  is appropriate, if  possible,  for  the
Project Officer to make site visits  or otherwise  assist  the applicant  in
developing and documenting the application  in accordance with  available
criteria for application evaluation.

     When the Project Officer  is involved in pre-application  activities, he
should provide the prospective grantee with the QA guidelines  that  will be
used to evaluate the formal grant application.  These guidelines  are
described on page 28.  The applicant should also  be made aware  of lERL-Ci
requirements as described in Section 2 of this document, "Quality Assurance
Guidelines for Monitoring of Projects Requiring Sampling and Analysis."  If
an ongoing project is under consideration for further grant funding,  the
Project Officer can conveniently review QA requirements with  the  applicant at
that time.

     The prospective grantee's pre-application should be reviewed with  regard
to his proposed QA program by  using  essentially the same criteria for con-
tracts.  In review of the pre-application and response  to  the  applicant, any
QA shortcomings should be noted and  guidance provided on the  general  level of
QA expected.  It may be of value to  inform  the applicant that  a pre-award
survey will be conducted as part of  the formal grant  review process and that
the applicant's QA procedures  will be evaluated during  this survey.

Use of Performance Test Samples in Contractor/Grantee Selection

     Performance test samples  can be used by the  Project Officer  to assist  in
selecting qualified contractors or grantees for projects requiring  analytical
services.  These are samples of known concentration (known only to  EPA) that
are provided to the prospective contractor  or grantee for  analysis.   Analysis
is conducted and the analytical results are reported  within a  specified time.
The results are then compared  with the EPA  known  values  to determine  the
accuracy of the laboratory's results.

     If it has been decided to use performance test samples to  evaluate
offerers, the lERL-Ci Quality  Assurance Officer should  be  notified.   He will
arrange for delivery of the samples  and evaluation of the  results.  The
procedure for obtaining test samples is described on  page  29.

     Performance test samples  have been prepared  and  made  available to
Project Officers by the EPA Environmental Monitoring  Systems  Laboratories
(Cincinnati, Las Vegas, and Research Triangle Park) through the lERL-Ci
Quality Assurance Officer.  Reference standards can also be obtained  from  the
National Bureau of Standards (NBS) through  the lERL-Ci  Quality  Assurance
Officer.  Appendix B contains  information concerning  the availability of
performance test samples.

     Indiscriminate or excessive use of performance test samples  for  the
evaluation of offerers is discouraged.  The samples are  costly  to both  EPA
                                     20

-------
and to the offerer  if he must conduct many  such  tests.  As  general  guidance,
performance test samples should be considered  for use  only  when  two or  more
of the criteria listed  in Table 5 apply.  Some of these criteria  cannot  be
assessed, of course, until all responsive offers have  undergone  technical
evaluation and have been scored according to the stated evaluation  criteria.
Hence, it is suggested  that  if the Project  Officer  is  uncertain  about  the
need  for performance test samples when  preparing the procurement  request, he
state the requirement for analyzing performance  test samples  in  the RFP  and
choose later whether to exercise the option.  This  can be accomplished  by
inserting the following provision in the procurement request:

     "Offerers whose proposals are determined  to be technically  acceptable
     under the initial  evaluation criteria  stated herein shall  (may) be
     required to demonstrate acceptable performance by analyzing  not more
     than   _   unknown samples for 	 parameters per sample.  The  cost
     for analysis of these samples shall be at the  offerer's  expense.   Sample
     analysis results shall  be compared to  known EPA results  for  scoring
     purposes."

     The final determination of whether or  not performance  test  samples  would
be of sufficient value  in the assessment of offeror(s) performance  to  warrant
their cost must be made by the Project Officer.  If the Project Officer
chooses to use performance test samples, it must be made clear  at what  point
in the selection process these samples will be used.  Will  they  be  sent  to
all offerers,  or just those  in the running  at  the final evaluation  stage?
Furthermore, the proposal evaluation criteria must  allow for  the  scoring of
test results.   The Project Officer is encouraged to seek the  advice of  the
lERL-Ci Quality Assurance Officer on the proper  use of performance  test
samples.

Cost Considerations

     To complete the procurement request rationale  document with  accompanying
EPA Form 1900-8 (Procurement Request/Requisition) for  a project,  the Project
Officer must prepare an estimate of the anticipated cost of the  required
scope of work.  For cost estimating, proper consideration should  be given to
the cost incurred for QA procedures on  sampling  and analysis  programs.   It  is
recommended by EPA in "Handbook for Analytical Quality Control  in Water  and
Wastewater Laboratories" (EPA, 1979) that daily  control of  analytical  per-
formance in the laboratory requires approximately 15 to 20  percent  of  the
analyst's time.  The good laboratory QA program will also call  for  the  main-
tenance of analytical control charts, training of personnel,  the  frequent
analysis of performance test samples, and extensive recordkeeping.   Hence,  an
adequate QA program will normally require a minimum of 20 percent of the
total analytical budget.  When sampling is  required, a minimum  of 10 percent
of the samples should normally be collected in duplicate.   Samples  can  be
split after collection, or several aliquots can  be  taken from the same
sample.  This increases both the sampling and  analytical costs  by 2 to
3 percent, depending, of course, on the scope  and complexity  of  the sampling
effort.
                                     21

-------
                  TABLE  5.   SUGGESTED  CRITERIA  FOR  USING
             PERFORMANCE TEST  SAMPLES  FOR OFFEROR EVALUATION*
	  The sampling  and  analysis  effort will  exceed  $25,000  in  value.

	  Analytical QA is  very  important  to  the  overall  success of  the
       project.  See Table  2.

_-___  There  is some doubt  concerning the  performance  capability  of the
       offerers.

	    The offeror(s) cannot  provide documented  results  of previous
       performance tests.
*Assuming appropriate  test  samples  are  available  (see Appendix B).
                                     22

-------
Preproposal Conference

     Preproposal conferences are an optional but  important  part  of  the
solicitation process for contracts.  They can serve an  information  function
similar to pre-application discussions with prospective grantees.   However,
they must be conducted in a fair and impartial manner that  will  not  give  any
prospective offerer an unfair advantage over another.  The  determination  to
conduct a preproposal conference may be made by the Project Officer  through
the Contracting Officer for any of the following  reasons:

     1.  To clarify or explain complex specifications,  statements of work, or
         proposed contractual provisions (e.g., QA program  requirements),
     2.  To discuss or emphasize the importance of any  qualification
         requirements, such as QA requirements,
     3.  To provide additional background material to prospective offerers,
         such as documents that are too voluminous to include with  the
         solicitation package, a site tour, or visits to  the place  of
         performance,
     4.  To respond to numerous questions of potential  offerers  regarding the
         solicitations, or
     5.  To comply with the request of an important segment of  the  industry.

     If the QA aspects of the proposed research effort  are  particularly
extensive, difficult to convey, or important to the project, the Project
Officer may elect to have the Contracting Officer call  a  preproposal con-
ference to provide further explanation of the QA  requirements to potential
offerors.  The preproposal conference must be conducted in  accordance with
PIN 77-15.  A record must be kept on the proceedings of the preproposal
conference; amendments to the RFP may be necessary as a result  of the con-
ference.  Proceedings of the preproposal conference may be  made  available to
all potential offerors.

QA CRITERIA IN TECHNICAL EVALUATION OF PROPOSALS  AND GRANT  APPLICATIONS

Technical Evaluation of Contract Proposals

     Technical evaluation of proposals must be made by  using only the
criteria set forth in the RFP in accordance with  "Source  Evaluation  and
Selection Procedures" (PIN 77-15).  In order to assist  in applying  QA in  the
technical evaluation of proposals, the proposal evaluation  checklist in
Appendix A provides a mechanism for the Technical Evaluation Panel  to deter-
mine if an offerer meets basic QA program requirements.   This checklist
describes the basic requirements for any adequate QA program for sampling and
analysis.  Questions on the checklist are grouped into  the  following four
areas:

     1.  QA management policy and written procedures,
     2.  QA procedures for sampling,
     3.  QA procedures for analysis, and
     4.  QA procedures for data management.
                                     23

-------
     Several questions are provided  in each  of  these  areas with  which  to  test
the offerer's proposal.  The questions are weighted according  to their  impor-
tance to a basic QA program.  Proper use of  this checklist when  evaluating
proposals should allow the Technical Evaluation Panel  to  identify  organi-
zations that do not offer an acceptable  level of QA and those  that  do.
Furthermore, the checklist should assist in  the technical ranking  of offerers
who meet and surpass the basic QA program requirements.

QA Evaluation Criteria and Scoring Procedures for Contracts

     The technical reviewers should  evaluate each proposal with  respect to
the offerer's demonstrated understanding of  the basic  requirements  of  a QA
program.  Assuming the guidelines for project conception  and solicitation on
page 10 have been followed in preparation of the procurement request,  and it
has been indicated in the RFP that QA is important to  the technical
reviewers, then they can expect adequate discussion of the proposed QA
program from responsive offerors.  Each  proposal should be evaluated with the
checklist (see Appendix A).  This checklist  can be copied and  used  by  the
Technical Evaluation Panel.  The following scoring plan (PIN 77-15) should  be
used:
                              SCORING PLAN

     Scoring      Percent of
      Value    Available Points           Evaluation

        0              0         Not addressed  in the  offer.
        1             20         Addressed,  but totally deficient.
       2a             40         Deficient,  but appears to be  capable  of
                                   improvements to adequate or better
                                   without adopting a  new approach.
       2b             —         Appears to  be  deficient; however,  final
                                   scores will  be determined subsequent
                                   to written questions and/or oral
                                   discussions.
        3             60         Adequate; overall it  meets the
                                   specifications.
        4             80         Good; has some superior  features.
        5            100         Generally superior in most features.

     The relationship of the scoring plan to written  or oral discussions  and
to subsequent negotiations is as follows:

     1.  Value of "0," "1," or "2a"—The element or sub-element  clearly is
         deficient and is not to be  questioned  or discussed during  written  or
         oral discussion.  Such values are solely for  the purpose  of scoring,
         ranking, and determination  of the technical  competitive range.   If,
         however, the offer attains  an overall  score,  because  of other  fac-
         tors, that places it in a sufficiently high  position  to be selected
         for negotiations, the offerer shall be allowed to correct  these
         deficiencies during negotiations.
                                     24

-------
     2.  Value of "2b"—The  element  or  sub-element  contains  uncertainties
         which must be  resolved  before  the  offer  is  fully  understood.   Such
         uncertainties  are to be  resolved during  written or  oral  discussions,
         and the offer  is to be  given a final  score  that is  based on the
         offerer's clarifications.

     3.  Values of "3", "4," or  "5"—The element  or  sub-element  is fully
         understood and there is  no  need for clarification by the offerer.
         However, discussions involving any such  elements  or sub-elements are
         not precluded.

     Each offerer's proposal is  evaluated in the  four  QA areas which have
been identified in the  technical  evaluation criteria included in  the RFP (see
Table 4).  In the area  of QA procedures for analysis,  one  of the  criteria is
an evaluation of the use of  standard analytical methods (a listing of  these
methods  is presented in Appendix  C).

     If  there is insufficient information provided  in  the  proposal to  assess
any particular question(s) in the checklist, the  reviewer  should  assign a
score of 2b on that question.  Those questions assigned a  value of 2b  should
be specifically addressed during  subsequent written  or oral  discussions with
technically competitive offerers  in  accordance with  PIN 77-15 procedures.

     Example;  To return to  the  previous example  of  a  wastewater  treatment
pilot plant for the shale oil industry, one offerer  may have scored as shown
in Figure 4 on QA procedures for  data management, one  sub-element of the
proposal evaluation criteria.  This  offerer scored  a total of 50  out of a
possible 75 points on this sub-element.  Hence, he was awarded 66 percent of
the total points.  Since this is  closest to 60 percent of  the points avail-
able for this sub-element, this  offerer is  assigned  a  score  of 3  for sub-
element  Hid (Table 4)  of the proposal  evaluation criteria.   This sub-element
is weighted according to the original proposal evaluation  criteria.  The
offerer, therefore, may have achieved a total  technical score as  shown in
Figure 5.  In this way, the  scoring  of  the  QA  evaluation criteria checklist
is integrated with the  proposal  evaluation  criteria.

     This offerer was awarded a  2b on one question  on  the  checklist
(Figure  4).  It was not clear in  the proposal  whether  data and other records
associated with the project would be retained  for a  minimum  of three years.
This point should be resolved during written or oral discussions  with  this
offerer  if he is in the technical competitive  range.

Written/Oral Discussions with Offerers

     Public law requires written  or  oral discussions  in negotiated procure-
ments with all responsible offerers who submit proposals within the  competi-
tive range (PIN 77-15).  The Project Officer and  the Contracting  Officer will
participate in these discussions.  The  purposes of  these discussions are to:
                                     25

-------
                    QUALITY ASSURANCE EVALUATION CRITERIA CHECKLIST  FOR
                     PROPOSALS OFFERING SAMPLING AND ANALYSIS  SERVICES'
                  Criteria
                                      Scoring   Numerical    Individual
                                       Value  x  Weight    •    Score
       D.   Quality  assurance  procedures for data management.
Id.   Does the offerer  possess
     appropriate  data  handling,
     processing,  and  retrieval
     capabilities?

2d.   Will QC data (e.g.,  standard
     curves, duplicate results, spike
     sample results)  be maintained and
     be accessible  to  the Project
     Officer?

3d.   Does the organization  routinely
     maintain analytical  performance
     records such as  quality control
     charts?

4d.   Are all. laboratory results and
     QC data reviewed  by  laboratory
     supervisory  personnel?

Sd.   Are all data and  records  retained
     for a minimum of  3 years?

6d.   Are field notebooks  used  to
     record sampling  and  engineering
     data (e.g.,  sample number, date/
     time of collection,  flow,
     operating conditions,  etc.)?
                                               *f
                                               */
                                     Total Score for Sub-Element D

                                     Maximum Possible Score
                                                                 75
       Percent of maximum possible score awarded for sub-element  D  (circle
       closest value).                           —^
                         OZ    20Z   40Z   —   (60ZJ  80Z   100Z

       Score for this  sub-element of the proposal evaluation criteria  (circle
       corresponding value).                     -•.
                         0      1    2a    2b    U)    4     5
Figure 4.   Example  use of  QA evaluation criteria checklist for
              scoring  proposals.
                                         26

-------
           Evaluation Criteria
                                  Numerical  Scoring  Individual
                                   Weight     Plan     Score
           I.
           II.
           III.
           IV.
Adequacy of the Technical Proposal    60
                     principal  project staff  related
                     to the project
                     Educational  background of
                     principal  project staff
                                                                          40
a.
b.
c .
Logic of approach to the study
Proposed pilot plant design
Presentation of findings
20
20
20
3
2b
5
Project Management 50
a.


b.

c.

Previous experience the
Project Manager has had in
this type of effort
Company resources available to
the Project Manager
Project management organization
and plan
15


25

10

2


5

3

Quality Assurance 50
a.

b.

c .

d.

Quality assurance management
policy/written procedures
Quality assurance procedures
for sampling
Quality assurance procedures
for analysis
Quality assurance procedures
for data management
10

15

15

10

4

3

4

3

Personnel Qualifications 40
a.
Technical experience of the
20
4
12
8
20
37
6


25

6

35
8

9

12

6

28
16
                                        20     3        12
                                                             TOTAL  SCORE  140
Figure  5.   Example of technical  evaluation scoring system.
                                         27

-------
     1.  Provide offerers an opportunity  to  further  explain  their  offers,
     2.  Afford the Contracting Officer and  the Project Officer  an oppor-
         tunity to understand  fully what  is  being  offered,
     3.  Arrive at preliminary agreements  regarding  price, cost, performance,
         contract terms and conditions, and
     4.  Resolve minor technicalities  in  offers.

     This is the point in the  procurement  process  where any  of  the offerers
may be asked to clarify their  proposed QA  procedures.  If, during  the  course
of the scoring of an offerer's proposal,  a score of  2b is  given  for  any  of
the QA Evaluation Criteria Checklist questions, clarification or more
information should be requested during the written or oral discussions.  The
response to these questions will be used  in  the final review (Figure 4).
Upon conclusion of the written or oral discussions,  the Project Officer  and
the Contracting Officer will determine of  whom a "best and final"  offer  is  to
be requested.

QA Evaluation of Grant Applications

     If a grant application is judged  relevant to  the lERL-Ci program  mission
and funding can be provided, then arrangements are made for  technical
reviews.  A minimum of one intramural  and  two extramural reviews of  tech-
nical and scientific merit are required for  new grant applications.

     The technical evaluation  criteria for research  and demonstration  grant
applications are much more subjective  than those for competitive procure-
ments.  There is no standard set of criteria that  can apply  in  each  case
because of the diversity of the programs,  the sampling and analytical
requirements, and the organizations performing the work.   This  should  not,
however, be used as an excuse  for failing  to recognize the importance  of a
sound QA program.

     When the project involves sampling and  analysis activities, each  tech-
nical reviewer should be requested to  address specifically the  applicant's
proposed QA program or plan during his evaluation  of the technical and scien-
tific merit of the proposal.   To aid project officers and  reviewers  in this
regard, the checklist in Appendix A should be used.  It is not  necessary,
however, to score the checklist results since grants are not awarded on  a
competitive basis.  For grants, the scoring  plan should be disregarded and
each checklist question answered in the affirmative  or negative.   The  Project
Officer should assure that a copy of this  checklist, which may  be  reproduced
from Appendix A, is forwarded  along with  the grant application  in  the  review
package provided to each reviewer.  The transmittal  letter that  accompanies
the review package should request that the reviewer  consider use of  this
checklist to evaluate the technical and scientific merit of  the  proposed
study.  Each reviewer's comments should be used to support the  funding
decision (see Figure 3) and/or to improve  the technical or scientific  merit
of the proposed study.
                                      28

-------
Evaluation of Performance Test Samples

     If it has been decided to use performance  test  samples  to  help  in  selec-
tion of a contractor,  in accordance with  the guidelines  provided  on  page  20,
then they must be provided to all offerers whose  proposals  are  technically
acceptable and are in  the competitive range.  Select  those  performance  test
samples from the list  of available samples, Appendix  B,  that  evaluate  analy-
tical capabilities that will be required  on the project.  Inform  the lERL-Ci
Quality Assurance Officer of your needs  for performance  test  samples,  and he
will arrange to have them sent to the competitive  offerers.   Instructions
will be provided for return of the analytical results  to  the  Quality
Assurance Officer.

     Allow each offerer sufficient time  to receive  the test  samples  from  EPA,
to conduct the analyses, and to report the results  along  with his  normal  lab-
oratory routine.  A minimum of one month  will normally be required.  It may
be advisable to ask the offeror(s) what  period  of  time he requires to  com-
plete the analyses and report the results.  The lERL-Ci Quality Assurance
Officer can also provide recommendations  in this  regard.  If  insufficient
analysis time is allotted, special handling of  these  samples  will  be
required, and this would result in an increased cost  to  the  offeror(s).

     After receiving all results  from responsive  offerers within  the
specified period of time, the Quality Assurance Officer will  assist  in
interpreting the results.

     Performance test  samples become part of the  technical  evaluation  and may
be discussed during post-award debriefings.  The  Project  Officer  is  to  inform
the Contracting Officer of the results of the performance test  sample  analy-
sis, and the Contracting Officer will convey the  results  to  the offeror(s).
Each parameter should  be defined as acceptable  or  unacceptable, and  the
evaluation criteria should be provided.   Any unacceptable results  should  be
described as either high or low, as appropriate,  since true  values cannot be
provided.  Since offerers must bear the  cost of performance  test  sample
analysis, they should  be informed of their performance as soon  as  possible  so
that any necessary corrective action may  be taken.

     Performance test  samples can be used to evaluate  research  and demon-
stration grant applicants in the same manner as prospective  contractors.

EVALUATION OF PREVIOUS PERFORMANCE HISTORY

     One indication of the probability of successful  project  accomplishment
on the part of a prospective contractor  or grantee  is  his performance  record
on previous projects.  The offerer is required  to  list current  and recently
completed EPA contracts in his proposal.  The interested  Project Officer  can
request a list of the  Project Officers on these contracts.  As  part  of  the
technical evaluation of the offerers, Project Officers on other government
projects can be contacted.  These Project Officers  could  provide  useful
information concerning their level of satisfaction  with  the  prospective
                                     29

-------
contractor's sampling and analytical QA,  if  the  project  calls  for  these
services.  Likewise, a prospective grantee may be evaluated by  contacting
Project Officers on past or present grants he holds.

     If the offerer has performed on a  previous  EPA contract,  the  Contracting
Officer should have access to records regarding  his performance.   If  the
offerer has performed past work for other agencies of  the government,  the
Contracting Officer may be able to obtain this information from the
Contractor Relations Section, Contracts Management Division, Washington,
D.C., 20460.

     Each Project Officer is required to  complete a Project Officer's  Eval-
uation of Contractor Performance (EPA Form 1900-27) at the conclusion  of a
contract.  Review of this form may provide valuable information to help
assess the likelihood that a given offerer may perform well on  your project.
Several points must be remembered, however,  when reviewing these forms
reporting on past project performance:

     1.  The work performed should be similar to that  proposed  in  the  current
         scope of work,
     2.  Only performance in the recent past should be considered, since
         changes in the organization may  invalidate previous performance
         information, and
     3.  Past performance, whether good or poor, may not be used as pass/fail
         criteria.

COST EVALUATION

     The cost evaluation of a proposal  is conducted by the Contracting
Officer concurrently with the evaluation  of  the  technical aspects  by  the
Project Officer.

     The Contracting Officer reviews each element of the offerer's business
and management proposal for the following:

    1.  Reasonableness of price or estimated cost with respect  to  the
        requirement,
    2.  Investigation of unreasonably high or low cost elements,
    3.  Evaluation of the proposed management structure  to be  utilized for
        performance,
    4.  Indirect cost management, and
    5.  Analysis of manhours and materials.

     The Project Officer is notified of those cost proposals considered to be
acceptable and receives copies of the cost proposals to  be compared against
those offerers' technical proposals which have been judged to  be within the
acceptable technical range.

     The Project Officer compares the cost and business  proposals  and  decides
which proposals merit additional consideration.  QA cost considerations are
presented on page 21.


                                     30

-------
     After a thorough review of the 5 specific  items mentioned  previously,
the Contracting Officer, with  input from  the Project Officer, makes  the
decision as to which cost proposals are considered  to be  in  the  acceptable
range.

BEST AND FINAL OFFER EVALUATION

     The evaluation of cost and technical proposals results  in  the selection
of candidates who are asked to present a  best and final offer.   Negotiations
are begun with the successful  offeror(s)  based  on their "best and  final"
offers.  The funding decision  serves the  same purpose for non-competitive
grants.  The extent of this final evaluation is determined by the  Contracting
Officer.  All scores of 2b must be reassessed and rescored.  The Project
Officer may choose to conduct  a pre-award survey of one or more  of the
offerers following the "best and final" offer evaluation.  The  Project Offi-
cer may be accompanied by the  Contracting Officer during  pre-award surveys.

     Following completion of final evaluation,  negotiations, and award,  all
records pertaining to the selection process (e.g.,  checklists)  are provided
to and retained by the Contracting Officer.  The contractor  or  the grantee
should be informed upon award  that quality assurance procedures  will  be  moni-
tored throughout the life of the project  by the Project Officer.   (Section  2
of this document, "Quality Assurance Guidelines for Monitoring  Projects
Requiring Sampling and Analysis", describes project monitoring  procedures.)

PRE-AWARD SURVEYS

     It was stated on page 25  that written or oral  discussions  with  each of
the offerers who submitted proposals in the competitive range were mandatory.
At the option of the Contracting Officer  and the Project Officer,  these
discussions may include a visit to the prospective  contractor's  or grantee's
facilities, but only after the "best and  final" offer has been  received.
Such visits offer an excellent way to assess thoroughly an offerer's  QA
program and awareness.  Some of the criteria that may justify a pre-award
survey are listed in Table 6.

     Pre-award surveys must be conducted  in such a  manner as to  avoid giving
an unfair advantage to any offerer on competitive procurements.  All  com-
munications must be conducted  in accordance with the guidelines  stated  in
PIN 77-15.  The Contracting Officer must  accompany  the Project  Officer on
pre-award surveys.  At the request of the Project Officer, the  lERL-Ci
Quality Assurance Officer or an lERL-Ci contractor  may participate in the
pre-award survey.

     The pre-award survey is a good time  to resolve any questions  that remain
unresolved from the quality assurance evaluation criteria checklist.  Any
scores of 2b on this checklist should, in particular, be  addressed in the
pre-award survey if they have  not been resolved in  previous  discussions.
Each of the QA criteria should be rescored during or after the  visit.
Table 7 is a suggested agenda  for a pre-award survey.
                                     31

-------
           TABLE 6.  CRITERIA FOR CONDUCTING A PRE-AWARD  SURVEY*	

       After technical evaluation of the proposal(s),  there  remain  suffi-
       cient unresolved questions to merit a face-to-face meeting with  the
       prospective contractor or grantee after submission of  the "best  and
       final" offer.

       A greater in-depth knowledge is required of  the  prospective
       contractor's or grantee's QA procedures.

       An inspection of the prospective contractor's or grantee's facil-
       ities would assist in the evaluation process after submission  of
       the "best and final" offer.

       Face-to-face discussions with key project  personnel would assist in
       the evaluation process.

       Past performance history of the prospective  contractor  or grantee
       has not been particularly good; however, there  is  reason to  believe
       improvements have been made.

       A decision among the top contenders cannot be conclusively justi-
       fied without further information.
*If any of these questions are answered  in the affirmative,  then  a  pre-award
 survey should be strongly considered.
                                     32

-------
     A pre-award survey may also precede the  funding decision  for  a  research
and demonstration grant application.  The Project Officer  is encouraged  to
invite the participation of the lERL-Ci Quality Assurance Officer  in  this
survey.  During the survey, any unresolved questions that  remain  from the
recommended quality assurance criteria discussed on page 28 pertaining to  the
evaluation of grant applications should be addressed and settled.
                                     33

-------
                TABLE 7.  SUGGESTED PRE-AWARD SURVEY AGENDA
A.  Discussions with key project personnel

    1.  Project management organization

    2.  Administrative management/project management  interface

    3.  Approach to accomplishing the scope of work

    4.  QA aspects of the program

    5.  Establish communications


B.  Review prospective contractor's/grantee's  facilities

    1.  Discuss equipment/instrumentation that will be  used

    2.  Evaluate overall appearance, resources,  and scope  of  capabilities


C.  Review prospective contractor's/grantee's QA program

    1.  Ask for copy of written QA  plan

    2.  Discuss how this plan will  apply to the  project under discussion

    3.  Review QA documentation on  typical work

    4.  Discuss QA procedures for the sampling effort*

    5.  Discuss QA procedures for analysis*
*To assist  in this regard, the Project Officer  is  encouraged  to  use  the  QA
 Evaluation Criteria Checklist (Appendix A) and  the QA  Audit  Checklist
 (Appendix  I).
                                     34

-------
                                  SECTION  2

               QUALITY ASSURANCE GUIDELINES FOR MONITORING OF
                  PROJECTS REQUIRING SAMPLING AND ANALYSIS
INTRODUCTION
     These guidelines have been prepared  to  assist  the  lERL-Ci  Project
Officer in directing contractor's efforts and  in  overseeing  research  and
demonstration grants which incorporate sampling and  analysis  activities.   The
Project Officer's QA responsibilities are described  beginning on  page 37  for
both contracts and grants.  These responsibilities  begin with project
initiation and extend through the monitoring of all  sampling  and  analysis
activities.  The Project Officer's QA responsibilities  do  not end with  the
field and the laboratory work, however; they continue through close review of
the results of the sampling and analysis  program, the interpretation  of the
quality control data, and the use to which the sampling and  analytical  data
are applied, including the conclusions and recommendations of the project.
                 l
     There are three basic foundations upon  which project  quality assurance
is built.  They are:

     1.  The contractor's or grantee's QA program,
     2.  The Project QA Plan, and
     3.  QA monitoring of project activities and  results.

These foundations of project quality assurance are  depicted  in  Figure 6,
along with the elements of each.

     All contractors and grantees should have  a QA  program.   This QA  program
describes the in-house procedures used by the  contractor/grantee  to guaran-
tee, to the extent possible, the quality  of  his work.   The elements of  a
suitable and acceptable QA program are described  under  Contractor's/Grantee's
QA Program on page 43.  A checklist is provided in Appendix D to  assist the
Project Officer in reviewing a contractor's  or a  grantee's QA program.  This
checklist can also be used to help the contractor or the grantee  to develop
an acceptable QA program if his present program does not meet lERL-Ci
requirements.

     Each project should begin by preparation  of  a Project Work Plan  which
describes how the effort will be accomplished.  A key part of any Project
Work Plan is a discussion of the sampling and  analysis  program  and the
                                      35

-------
                           CONTRACTOR'S/GRANTEE'S
                              QUALITY ASSURANCE
                                   PROGRAM
                                  (Page 43)
                                                 Organization and Personnel
                                                 *Facilities and Equipment
                                                 *Analytical Methodology
                                                 *Samp ling and Sample
                                                    Handling Procedures
                                                 *Quality Control
                                                 *Data Handling
                                   PROJECT
                                   QUALITY
                                  ASSURANCE
PROJECT WORK
    PLAN
 (Page 48)
                EXTERNAL QUALITY
              ASSURANCE MONITORING
                     (Page 50)
* Project Objective
* Project Staff
* Facilities/Equipment
* Sampling Plan/Methods
* Analytical Methods
* Quality Control
* Data Management
* Project Schedule
* Review of Project Reports
* Conferences/Project Reviews
* Site Visits/QA Audits
* Performance Tests
* Sub-Contracts
            Figure 6.  Foundations of project quality assurance.
                                  36

-------
procedures that will be  followed  to validate  the  quality  of  the  resulting
data.  A checklist  is provided  in Appendix E  to .assist  the Project  Officer  in
reviewing project work plans and  to assure that the QA  aspects of  the  plan
are given adequate  attention.

     Project QA monitoring procedures are discussed on  page  50.  The Project
Officer has full responsibility for monitoring of  project QA during the
course of project accomplishment.  Several techniques are described for
fulfilling this key project responsibility, including:  (1)  the  effective use
of project conferences and project reviews, (2) a  number  of  independent
performance tests (e.g., QA audits, performance test samples,  split samples),
and (3) review of project reports.

PROJECT OFFICER'S QA RESPONSIBILITIES

     This section briefly describes the QA-related responsibilities of the
Project Officer and discusses how he can  influence the  quality of  the  work
during the course of the project.

Contracts

     Figure 7 is a  diagram of the major aspects of contract  administration.
After award of the  contract to  the contractor, the Project Officer  assumes
technical responsibility for project initiation,  project  monitoring, conclu-
sion, and closeout.  The Project Officer  fulfills  his QA  role  by interacting
with the contractor in each of  these five phases  of the work.

Project Initiation—
     After completion of all negotiations and award of  a  contract,  the
Project Officer should discuss w'ith the contractor their  respective roles in
project quality assurance and the necessity for cooperative  efforts to
achieve a high-quality result.

     The Project Officer should discuss with  the  contractor  how  he  intends  to
maintain quality assurance on this project commensurate with the objectives
of the effort.  He  should discuss the three basic  foundations  of project  QA
with the contractor.  They are:   (1) the  contractor's QA  program,  (2)  the
Project Work Plan,  and (3) QA monitoring  by the Project Officer  or  the
lERL-Ci Quality Assurance Officer.

     At this time,  if not provided during procurement,  the Project  Officer
should request a review copy of the contractor's QA program  manual.  Guide-
lines for reviewing the contractor's QA program are discussed  under "The
Contractor's/Grantee's QA Program" (page  43); Appendix  D  is  a  checklist  to
assist the Project Officer in this review.  The Project Officer  and the  con-
tractor should also discuss quality assurance procedures  to  be addressed  in
the Project Work Plan, which will be prepared by  the contractor  before the
work begins.  The elements of a Project Work Plan  are described  on  page 48. A
checklist is also provided in Appendix E  to assist the  Project Officer in
review of the QA aspects of the plan.
                                     37

-------
Contract Procurement
Contract Award
Project Initiation
Project Monitoring
Project Conclusion-
Project Closeout-
Initial Contact









Project Plan









Reports




Conferences and Project Reviews




Site Visits/QA Audit




Performance Tests




Sub-Contracts









Final Project Report









Project Evaluation
          Figure 7.  Contracts administration.
                               38

-------
Project Monitoring—
     The Project Officer serves  in  a key  position  in QA monitoring  of the
contractor's project work.  There are a number  of  ways  in  which  he  can
fulfill this responsibility (Figure 7).   He  can review  all project  reports,
conduct project conferences and  reviews,  conduct site visits  and QA audits,
use performance tests when appropriate, and  assure  that the  same level of  QA
is applied to all subcontracted  work.  The lERL-Ci  Quality Assurance Officer
can assist the Project Officer  in accomplishing these responsibilities.

     The Project Officer is responsible for  the technical  content and QA
review of all project reports.   This includes periodic  progress  reports,
interim reports, task reports, QA reports, and  final project  reports.   Since
the Project Officer must approve the Project Work  Plan, he should be aware of
all sampling, analysis, and QA procedures that  are  to be employed in develop-
ment of the data.  There remain  two areas, then, where  the Project  Officer
should focus his attention in the review  of  reports:  (1)  Has  all the QA
information been documented in the  report?,  and (2) Are the  conclusions  which
are drawn from the work supported by the  sampling  and analysis data?  Each
task report, interim report, and certainly the  final report  should  be
reviewed by one or more persons  who are technically qualified  to assess  the
quality of the work and the conclusions which are  drawn.

     In his role as technical representative for EPA, the  Project Officer
must insure that the contractor  is  complying with  the procedures delineated
in his QA program and in the Project Work Plan.  One of the  most effective
ways of doing this is to conduct occasional  visits  to the  study  site and to
the contractor's laboratories.   The contractor  should be contacted  as early
in advance of the visit as practicable and advised  of any  specific  matters to
be discussed.  The Project Officer  must be knowledgeable of  the  status of  the
work and of the procedures being used.  The  visit  and observations  from the
visit should be documented in the project file.

     At least one site visit, preferably  the first  one, should incorporate a
QA audit.   Procedures and checklists for  conducting a QA audit may  be found
in Section 3 of this document, "Quality Assurance  Guidelines  for Auditing  of
Projects Requiring Sampling and  Analysis."   The Quality Assurance Officer  may
participate in site visits whenever possible and,  in particular, when a QA
audit is to be conducted.

     The Project Officer may call upon the contractor to conduct one or  more
performance tests.   Performance  test samples for many parameters are made
available by EPA.  Procedures for obtaining  these  materials  are  found on
page 50.  The Project Officer may request the contractor to  analyze reference
materials in order to establish  analytical accuracy and precision of the
analytical methods, or he may have  split  samples analyzed  by  another labora-
tory for purposes of comparison.  Independent sampling  by  lERL-Ci or a third
party may also be called for if  it  is important to  document  certain param-
eters upon which critical decisions are to be made.
                                     39

-------
     When sampling and/or analytical  activities  are  sub-contracted,  it  must
be made clear that the prime contractor bears  full  responsibility  for  the
application of QA procedures and  for  the  quality  of  the  final  product.   The
Project Officer should exert the  same control  as  when  all work  is  done  only
by the prime contractor and no sub-contractor  is  used.   All  parties  should
understand the QA requirements of  the program  and agree  to cooperate  in the
QA effort.  The Project Officer should have  direct  access to the sub-
contractor or, at least, he should  feel confident that someone  on  the  prime
contractor's project staff  is knowledgeable  of the  technical aspects  of the
sampling and analysis activities  and  can  represent  this  area of project
activity.

Project Conclusion and Closeout—
     Most projects conclude with  a  final  project  report.  A  draft  final
report is submitted to the Project Officer  for review  and approval.   This  is
the culmination of the Project Officer's  QA  responsibilities.   He  should
assure all sampling and analytical data are  thoroughly documented  and  sub-
stantiated.  Independent review of  the final report  by other qualified  Agency
personnel is required.

     Each Project Officer is required to  complete a  Project  Officer's  Evalu-
ation of Contractor Performance (EPA  Form 1900-27) at  the conclusion  of a
contract.  The degree to which the contractor  has complied with the  con-
tracted work scope and the required QA should  be  reflected in  this evaluation
so that the Agency may develop an  accurate historical  performance  record of
contractors.

Grants and Cooperative Agreements

     Unlike contracts management,  the Project  Officer's  role is not  one of
directing the activities of a grantee.  He still  has the responsibility, how-
ever, of assuring a quality product is delivered  to  the  Agency  upon  project
completion.  As shown in Figure 8, he has many of the  same responsibilities
for project initiation, project monitoring,  and  project  conclusion as  he does
for a contract (see Figure 2).  His basic methods of fulfilling these  respon-
sibilities would be the same as described above  for  contracts,  but his
approach would be somewhat different, since  he is not  expected  to  "direct"
the grantee but to "assist" him in achieving the  quality requirements  of the
Agency.

Project Initiation—
     Upon grantee acceptance of the grant or cooperative agreement,  the Pro-
ject Officer should inform the grantee that  the Agency is concerned  that the
quality of all sampling and analysis  can  be  documented.  He  should describe
his role in monitoring project QA  procedures.

Monitoring of Project QA—
     The key responsibility of the Project Officer  is  the technical
monitoring of the grant, basically as a means  of  assuring that  the grantee
                                     40

-------
     Grant Application


            I
     Grant Award
•Study  Plan
     Project Initiation-
     Project Monitoring
     Project Conclusion-
     Project Closeout
 Initial  Contact








 Continuing  Communication



 Encourage Grantee  Performance



 Site  Visits/QA Audit



 Sub-Agreements



 Progress Reports







 Final Project  Report
Figure 8.  Grants and cooperative agreements administration.
                                41

-------
carries out the approved scope of work  on  schedule  and  in  conformance  with
applicable rules, regulations, and  any  special  conditions  that  may have been
imposed by the grant agreement.  The  grantee  should  use  currently  accepted  QA
procedures commensurate with  the objectives of  the  effort.

     As shown in Figure 8, the study  plan  is  normally approved  before  project
initiation as a result of  funding of  the grantee's  project.   Hence,  it  is
critically important that  QA  be a major consideration in the  technical  evalu-
ation of grant and cooperative agreement applications.  Guidance  is  provided
in the first section of this  document,  "Quality Assurance  Guidelines  for
Procurement of Projects Requiring Sampling and Analysis."

     In carrying out the project monitoring function  for a grant or  a  cooper-
ative agreement, the Project  Officer  should:  (1) conduct  site  visits  and a
QA audit as appropriate, (2)  monitor  the QA practices on any  sub-agreements,
and (3) review the QA aspects of all  progress reports and  the  final  report
submitted by the grantee.

     The Project Officer may  have access to grantee  records and to records  of
contractors under the grant,  as well  as to the  performance  site, and  should
use such access as one tool in monitoring  grant QA  activities.  The  frequency
of visits is a matter of Project Officer judgment,  although it  may be
affected by the availability  of travel  funds.  As a  general rule,  a  project
QA audit should be conducted  once during each project.  The QA  audit  is most
effectively accomplished early in the project when  the  first  sampling  and
analysis activities are begun.  Recommendations for  conducting  a QA  audit are
discussed beginning on page 50.  QA audit  procedures  and checklists  may be
found in Section 3, "Quality  Assurance  Guidelines for Auditing  of  Projects
Requiring Sampling and Analysis."

     The Project Officer reviews any  sub-agreements  or  sub-contracts  under  a
grant or cooperative agreement for  technical  content.   He  should make  sure
that all sub-agreements are written to  include  necessary and  reasonable QA
requirements.  The Project Officer  should  not approve the  award of a sub-
agreement until he is satisfied that  QA requirements  equivalent to those
expected of the prime grantee have  been accepted.

Summary of the Project Officer's Role in Achieving  QA

     The role of the Project  Officer  in achieving QA  can be best  fulfilled  by
keeping in mind a few working principles.

     1.  Have a good basic understanding of conventional QA practices  and the
         principles behind project  QA.  Work  with the Quality Assurance
         Officer whose job it is to be  thoroughly knowledgeable of ways to
         achieve and document good  quality work.

     2.  Be knowledgeable  of  the sampling  and analytical procedures  used  or
         ask for assistance from someone who  is.  Understand  the precision
                                     42

-------
         and accuracy which  is  achievable with  the methods  and  what  can most
         affect the quality  of  the work.

     3.  Know the level of QA required  to achieve the  objectives  of  the
         project.  Is qualitative or  screening  information  adequate  or  must
         the highest possible accuracy  be achieved?

     4.  Be certain the contractor/grantee has  adequate  funds budgeted  to
         accomplish the work at the level of  quality desired without taking
         unjustified shortcuts  to the detriment  of QA.

     5.  Use the lERL-Ci Quality Assurance Officer for  independent review of
         the QA procedures,  quality control data, and  project reports.   Ask
         him to point out any QA weaknesses in  the program.

     6.  Recognize that the  purpose of  a QA program  is  to discover and
         correct errors.  Use well-founded QA practices  to  identify  any
         sampling or analytical problems before  effort  is lost  in producing
         unreliable data.

THE CONTRACTOR'S/GRANTEE'S QA PROGRAM

     Each contractor or grantee should  have a written  QA program  which
describes the in-house procedures that  he employs to guarantee, to the  extent
possible, the quality of all sampling and analysis activities.  The
contractor's/grantee's QA program is  not project specific.  It  describes the
quality assurance and the quality control procedures used on any  project
which requires sampling and  analysis.

     There is no standard, universally-accepted  QA program.  Each contractor
and each grantee prepares his own to  suit the needs of his  organization as
he has best determined.  There  are, however,  a  number  of publications which
discuss the recommended contents of a comprehensive QA program.   These
publications are listed in the  Bibliography.  Although  each author describes
the essential elements of a QA  program  in his own manner and with his own
perspective, there is general agreement on what  a QA program should  contain.
These essential elements are described  in this  chapter.  In addition, a
checklist has been prepared  to  aid the  Project Officer  in reviewing  a QA
program and to facilitate his feedback  to the contractor or grantee  on  the
suitability of his program with respect to lERL-Ci requirements.

Elements of a QA Program

     A comprehensive QA program will  address  the six major  elements  upon
which the final quality of the  laboratory's work depends.  Generally, these
include:

     1.  Organization and personnel,
     2.  Facilities and equipment,
     3.  Analytical methodology,
                                     43

-------
     4.  Sampling and  sample handling  procedures,
     5.  Quality control, and
     6.  Data handling.
     A comprehensive, well-written QA  program will  address  each  of  these
essential aspects of QA.  In  the  following  descriptions  these  six major areas
have been divided into sub-elements, where  applicable.

Organization and Personnel—

QA policy and objectives—Each organization  should  have  a written quality
assurance policy that should  be made known  to all organization personnel.
Objectives should be established  to produce  data  that meet  project  require-
ments in terms of completeness, precision,  accuracy, representativeness,
documentation, and comparability.  The QA program should require the  prepara-
tion of a project-specific QA plan for each  major project.

QA organization—The organization and  management  of the  QA  function  should be
described in the contractor's/grantee's QA  document.  Reporting  relationships
and responsibilities should be clearly defined.  A  QA Coordinator or  Super-
visor should be appointed and his responsibilities  established.  A  descrip-
tion of the QC paperwork flow should be available.   There should be  a clear
designation of those who are  authorized to  approve  data  and  results.   Respon-
sibilities for taking corrective  action should  be assigned  to  appropriate
management personnel.

Personnel training—It is highly  desirable  that there be a  programmed
training system for employees.  This system should  include  motivation toward
producing data of acceptable  quality and should involve  "practice work" by
the new employee.  The quality of this work  can be  immediately verified and
discussed by the supervisor,  with appropriate corrective action  taken.

Document cpntrol and revisions—A QA program should include  a  system  for
documenting:(1)sampling procedures, (2)  calibration procedures,  (3) analy-
tical procedures, (4) computational procedures, (5) quality  control  proce-
dures, (6) field data, and (7) operating procedures, or  any  changes  to these
procedures.

     Procedures for making revisions to technical documents  must be  clearly
defined, with the lines of authority indicated.  Procedural  revisions should
be written and distributed to all affected  individuals,  thus  insuring imple-
mentation of changes.

     Documentation control becomes increasingly important in  field  studies,
since procedures must often be adapted to the particular situation,  tested,
and approved by project management authority.  Any  revisions  to  the  sampling
program must be strictly documented and approved by the  Project Officer.
Undocumented changes in either the sampling  or  analysis  program  can  seriously
affect the substantiation of  the  final project  conclusions.
                                     44

-------
Facilities and Equipment—

Procurement and inventory procedures—Purchasing guidelines  for  all  equipment
and reagents having an effect on data quality  should be well-defined and
documented.  Similarly, performance specifications  should  be  documented  for
all items of equipment having an effect on data quality.   Once any  item which
is critical to the sampling and analysis program such  as a flowmeter,  in  situ
instrument, or reagent is received and accepted by  the organization,  docu-
mentation should be retained of the type, age, and  acceptance status of  the
item.   Reagents should be dated upon receipt  in order  to establish  their
order of use and to minimize the possibility  of exceeding  their  useful shelf
life.

Preventive maintenance—Preventive maintenance procedures  should  be  clearly
defined and written for each measurement system and required  support equip-
ment.   When maintenance activity is necessary, it should be  documented on
standard forms maintained in logbooks.  A history of the maintenance record
of each system serves as an indication of the  adequacy of  maintenance sched-
ules and parts inventory.

Analytical Methodology—

Calibration and operating procedures—Calibration is the process  of  estab-
lishing the relationship of a measurement system output to a  known  stimulus.
In essence, calibration is a reproducible reference point  to  which  all sample
measurements can be correlated.  A sound calibration program  should  include
provisions for documentation of frequency, conditions, standards, and records
reflecting the calibration history of a measurement system.

     The accuracy of the calibration standards is an important point to  con-
sider since all data will be in reference to  the standards used.  A  program
for verifying the accuracy of all working standards against  primary  grade
standards should be routinely followed.

Feedback and corrective action—The QA program should  specify the corrective
action that is to be taken when an analytical  or sampling  error  is  dis-
covered.  The program should require documentation  of  the  corrective action
and notification of the analyst or sample collector of the error  and correct
procedures.

Sampling and Sample Handling Procedures—

Configuration control—Some sampling and analysis programs require  a more or
less elaborate array of sampling equipment, sampling systems, or  in  situ
instrumentation.  It is important for quality  assurance that  once such an
array has been established, the configuration  of the array should be docu-
mented.  Furthermore, any changes in the configuration must  be made  only
after due consideration of the effects on the  data  which are  gathered.  Any
                                     45

-------
changes in configuration or design changes  in  the  sampling  or  analysis  system
must be thoroughly documented.

Reliability—The reliability of each  component  of  a  measurement  system  is
related directly to the probability of  obtaining valid  data  from that system.
It follows that procedures  for reliably  collecting,  processing,  and  reporting
data should be clearly defined and in written  form for  each  system component.
Reliability data should be  recorded on  standard  forms and kept  in a  logbook
for easy reference.  If this procedure  is  followed,  the  data can be  used  in
revising maintenance and/or replacement  schedules.

Quality Control—

Quality control procedures—The quality  control  procedures  used  during  samp-
ling and analysis should be described.   The quality  control  checks routinely
performed during sample analysis  include reagent blank  analysis  to establish
background absorbance, duplicate  analysis  to establish  analytical precision,
and spiked and blank sample analysis  to  determine  analytical accuracy.  The
frequency of these quality  assurance  checks should be specified.  Limits  of
acceptance or rejection should be defined  for  analysis  and  control charts
should be used where practicable.  Gas  chromatography confirmation procedures
should be discussed.

Control checks and internal audits—A good QA  program will make  provision  for
periodic control checks and internal  audits by  the performing  organization.
Several approaches are commonly used  for control checks.  These  include:

     1.  Reference material analysis.  Analytical  reference materials are
         available from several commercial and  government sources, or they
         may be prepared in-house.  The  National Bureau  of  Standards  (NBS)
         has made available a series  of  standard reference materials  which
         may be purchased.  The chemical analysis  of these materials  has  been
         well established.  Such  materials can  be  analyzed  alongside  routine
         samples and the results  used to check  the accuracy  of  analytical
         procedures.

     2.  Split sample analysis.   The  analysis  by two or  more analysts of  a
         single sample that has been  well mixed  and  divided  is  useful for
         establishing precision of the  analytical  techniques and the  perfor-
         mance of the analysts.

     3.  Spiked sample analysis.  The analysis  of  a  routine  sample which  has
         been spiked with a known amount of the measured material should  be
         commonly employed  to establish  the recovery of  an  analytical
         method.

     4.  Side-by-side analysis.   Under  particularly  intractable  conditions
         where it is important to acquire useful data but difficult  to  con-
         trol all important variables,  it may  be useful  to  conduct
                                     46

-------
         side-by-side  analysis  by  two  analysts.   These  analysts  may use the
         same or  perhaps  different  analytical  methods  to  acquire comparable
         data.

     5.  Reference devices.   Some  measurement  systems,  in particular flow
         measurement systems, can  be checked by  the  use of available refer-
         ence devices.  The  use  of  these  devices are generally disruptive to
         the operation  and can  seldom  be  employed without the knowledge of
         the operator  or  the  analyst.

     Internal audits should  be  periodically conducted  to  evaluate the
functioning of  the QA  program.   This involves  an independent  check of the
performance of  the field  crew or the laboratory  analysts  to determine if
prescribed procedures  are closely  followed.

Data Handling—

Data handling,  reporting, and recordkeeping—Data handling,  reporting,  and
recordkeeping procedures  should  be  described.  Data  handling  and reporting
includes all procedures used  to  record  data on standard forms,  laboratory or
field notebooks.  This  includes  bench  data and field data.  The  reporting
format  for different types of bench data  should  be discussed  and the forms
provided.  The  contents of field notebooks should be specified.

     Recordkeeping of  this type  serves  at  least  two  useful functions:   (1) it
makes possible  the reanalysis of a  set  of data at a  future time,  and (2) it
may be  used in  support  of the experimental conclusions  if various aspects of
the study are called into question.

Data validation—Data  validation procedures, defined ideally  as  a set  of
computerized and manual checks  applied  at various appropriate levels of the
measurement process, should  be  in written  form and clearly defined for  all
measurement systems.   Criteria  for  data validation must be documented and
include  limits  on:  (1) operational parameters such  as  flow rates; (2)  cali-
bration  data; (3) special checks unique to each  measurement,  e.g., successive
values/averages; (4) statistical tests, e.g.,  outliers; and (5)  manual  checks
such as  hand calculations.  The  limits must be adequate to insure a high
probability of  detecting  invalid data  for either all or the majority of the
measurement systems.  The required  data validation activities (flow rate
checks,  analytical precision, etc.) must  be recorded on standard forms  in a
logbook.

QA Program Checklist

     An  important responsibility of the Project  Officer,  as described  on
page 37  for contracts  and page 40  for  grants and cooperative  agreements, is
to review the QA program  of  the  contractor or  grantee before  project work
begins.  A checklist has  been provided  in Appendix D to assist  the Project
Officer  in reviewing a  contractor's or  grantee's QA  program.   In the event a
contractor/grantee does not have a  written QA  program,  this checklist  may
                                     47

-------
still be used to evaluate  the QA  procedures  employed  or  to  assist  the
contractor/grantee to develop a suitable QA  program.

     This checklist applies  to a  "model" QA  program;  hence, many  programs
will have deficiencies.  It  is up to  the judgment  of  the  Project  Officer,
after the review of the QA program and  possibly  clarifying  discussions  with
the contractor/grantee, whether to accept  the  program or  require  upgrading of
the program before project work begins.  The Quality  Assurance  Officer  can
help the Project Officer in  applying  standards that conform with  those  used
on other lERL-Ci projects  of a similar  nature.

THE PROJECT WORK PLAN

     A work plan should be prepared before commencement  of  each project.   The
work plan normally consists  of:
     1.  A statement of  project objectives,
     2.  Description of  the  project  staff,
     3.  Facilities and  equipment,
     4.  A sampling plan,
     5.  An analytical plan,
     6.  The Project QA  Plan, and
     7.  The project schedule.
     It  is Agency policy  that  all  extramural  projects  involving environmental
measurements must have  a  Project QA  Plan  (Costle,  1979).   Although  the  Proj-
ect QA Plan, which  is generally a  section  of  the Project  Work  Plan,  specifi-
cally addresses quality assurance  of  the  sampling  and  analysis effort,  in the
broader  sense the total work  plan  encompasses  aspects  of  project  quality
assurance.  For contracts,  the QA  Plan  is  an  integral  part  of  the Project
Work Plan, which  is prepared  after award  of  the contract  (see  Figure 2) and
submitted to the Project  Officer for  review  and approval  before sampling and
analysis begins.  For research and demonstration grants or  cooperative
agreements, the QA  Plan should be  a  key  section of the Study Plan which
accompanies the grantee's formal application.  If  this is not  the case, the
Project  Officer should  request that  the  grantee prepare a QA Plan for the
project  and submit  it for review in  the  initial stage  of  the effort.

Elements of a Project QA  Plan

Project  Objective—
     The objective  of the project  and of  the  sampling  and analysis  effort
should be stated  in clear and  concise terms.   From the statement  of the
objective, one should derive  the level  of  QA  to be applied  to  sampling  and
analysis activities and this,  also,  should be  clearly  stated.

     The Project Officer  should review  the recommendations  of  the contractor
or grantee and assure that  he  is in  agreement  with the stated  use of the data
and the  appropriateness of  the level  of  QA to  be applied  to the data collec-
tion activities.
                                     48

-------
Project Staff—
     Project reporting relationships  should be  shown.  Normally  on  projects
requiring sampling and analysis, a Quality Assurance Supervisor  should  be
assigned.  His responsibility  should  be  to monitor  the quality (internal
audit) of the sampling and analysis program and  to  assure  that stated proce-
dures are, in fact, being employed.   He  should  initiate  corrective  action
when problems are detected.

Facilities and Equipment—
     The facilities, instrumentation, and equipment that are  unique or  cri-
tical to the success of the effort should be described.  This would include:
(1) on-site and in situ instrumentation, (2) sampling equipment,  (3) mobile
facilities, (4) temporary facilities, (5) subcontractor- or vendor-supplied
equipment, (6) government equipment or  facilities,  and (7) special  safety
equipment.  Operation and maintenance procedures  should  be described.

Sampling Plan and Methods—
     The sampling plan should  be discussed in the detail necessary. Sample
points should be precisely located on a  site or  system diagram and  the  meth-
ods of collecting the sample(s) should be described.  The  sampling  schedule
should be established, as should the  procedures  to  guarantee  representative
samples.  The containers and preservatives that  are to be  used should be
described as well as methods to avoid sample contamination.   The  sampling
plan should describe how maximum sample  holding  time limitations  can be met
and how samples will be transported to  the laboratory.

Analytical Methods—
     Standard analytical methods should  be employed when applicable.  When
standard methods are not available, the  methods  employed should  be  documented
in step-by-step detail.  For example, method validation would be  required  for
adoption of a standard method  to a sample matrix  other than that  for which  it
was designed, or development of a non-standard method.  Analytical  detection
limits and an assessment of the anticipated variance, precision,  and accuracy
should be stated.

Project QA Plan—
     The Project QA Plan should describe the application of the  grantee's
general QA program to the project.  It should be  project-specific and should
describe what steps are to be  taken to assure that  the resultant  sampling  and
analytical data are reliable and suitable to project needs.   The  Project QA
Plan should specify such items as:

     1.  The steps taken to avoid sample contamination,
     2.  The method chosen to  assure  each sample  is representative  of the
         source,
     3.  The collection of background or baseline samples,
     4.  The frequency of duplicate sample collection,
     5.  The use of blank samples and field spiked  samples,
     6.  Split sample analysis,
                                     49

-------
     7.  The method of establishing data precision,  and
     8.  The method of establishing data accuracy.

Project Schedule—
     All project plans require a schedule of accomplishment and milestones.
The schedule should allow adequate time for the sampling and  analysis  pro-
gram, for QA review of the results, and for corrective actions.

Project Work Plan Checklist

     Appendix E is a checklist which has been developed to assist  the  Project
Officer in reviewing the Project Work Plan submitted by the contractor or
grantee.  This checklist is  intended to represent  the degree  of detail that
may be expected of a "model" Project Work Plan.  The results  of the  checklist
responses should help the Project Officer to specify improvements  that are
necessary or to have confidence that the project plan adequately addresses QA
for the purposes of the effort.  Before use of this  checklist, the Project
Officer should have a clear  understanding of the level of QA  required  to meet
project requirements.  The lERL-Ci Quality Assurance Officer  can assist the
Project Officer in this review, and he can help the  Project Officer  apply
standards that conform with  those used on other lERL-Ci projects of  a  similar
nature.

PROJECT QA MONITORING

     The third foundation of project quality assurance is external QA
monitoring.  To be thoroughly effective, a QA program and a Project  QA Plan
should be tested from time to time.  This is a responsibility of the Project
Officer.  He is encouraged,  however, to request the  assistance of  the  lERL-Ci
Quality Assurance Officer.

     External QA monitoring  may be accomplished by use of some of  the  same
techniques used for internal audits (page 43), namely:  (1) reference
materials analysis, (2) split sample analysis, (3) spiked sample analysis,
(4) side-by-side sample analysis, and (5) reference  devices.   In addition to
these, the Project Officer may make use of available performance test
samples, or he may conduct a QA audit of the project.

     The lERL-Ci Quality Assurance Officer is prepared to assist the Project
Officer in using any of these methods to monitor the quality  of the  sampling
and analysis aspects of the  project.  At the request of the Project  Officer,
he will arrange for reference or performance test  samples to  be sent to the
contractor/grantee, he will  arrange for a project  QA audit, or he  will
arrange for split or side-by-side sample analysis.

Methods of Monitoring Contractor/Grantee QA

     The QA performance of a contractor or grantee can be monitored  from time
to time throughout the project by:
                                      50

-------
     1.  Review of project  reports,
     2.  Conference and project  reviews,
     3.  Site visits,
     4.  Performance tests, and
     5.  Monitoring any sub-contracts.

The Project Officer should  apply these methods  throughout  the  course  of a
project.

Review of Project Reports—
     The Project Officer has  the responsibility of  reviewing all  technical
project reports submitted by  the contractor  or  grantee.  This  would  include
progress reports, interim reports, task reports, QA reports, and  final
project reports.  In reviewing project technical reports,  the  Project Officer
should expect to see a summary of  the quality control  data in  accordance with
the approved Project QA Plan.  Any conclusions  resulting from  the project
should be supported by the  sampling and analysis results when  one takes the
quality control data into consideration.  For example,  the author should not
attempt to interpret different values for a  given parameter measured  at two
points in a system when the difference between  the  values  is not  greater than
the parameter variance with a high level of  confidence.

     The analytical methods should be discussed in  the  final project  report.
If standard methods have been used, references  should  be provided to  their
source.  Non-standard methods should be described in step-by-step detail.

     The data should be tabulated  or displayed  in a logical and understand-
able manner with the appropriate number of significant  figures and appropri-
ate units.  Detection limits  should be indicated for parameters where values
are below the detection limit.  Mean values  should  be  reported, but  supported
by the range and standard deviation determined  by replicate analysis.

     The final report should  describe the sampling  program and methods.
Sampling points should be shown  on an appropriate schematic diagram.  Any
specialized sampling techniques  should be discussed in  detail  while  standard
methods should be properly  referenced.

Conferences and Project Reviews—
     Whenever a conference  or project review meeting is held,  the Project
Officer should make a special effort to review  project  QA.  He should review
field notebooks or laboratory notebooks for  completeness.  A summary  should
be provided of the quality  control data and  the results reviewed.  Are  accu-
racy and precision adequate to accomplish the objectives of the program or
must changes be made in either sampling or analytical  techniques? Are  sample
points properly located to  collect representative samples? Are short holding
time parameters being analyzed quickly enough after sample collection?   Is
the Project QA Plan being followed or have undocumented changes been  made?
Questions of this type should be asked by the Project Officer  in  order  to
test the efficacy of both the contractor's/grantee's QA program and  the
Project QA Plan.
                                     51

-------
Site Visits and QA Audits—
     Site visits to the facilities of  the contractor  or  grantee  and  to  the
sampling or demonstration site can be  conducted  as often  as  practicable  and
necessary to keep abreast of technical progress.  Such visits  can  serve  as a
major QA function.  At least one site  visit during the course  of the  techni-
cal work on a project should incorporate a QA audit.

     QA audit procedures and checklists are provided  in Section  3, "Quality
Assurance Guidelines for Auditing of Projects Requiring Sampling and
Analysis."  The Project Officer is referred to this document  for further
information on this very important tool for project QA monitoring.

Performance Tests—
     Performance tests may consist of:  (1) analysis  of performance  test
samples provided by the Agency, (2) analysis of  split samples  by another
laboratory, (3) independent sampling and analysis by  another  laboratory, or
(4) reference sample analysis.

     Performance test samples are prepared samples of known  concentration.
They have been prepared and are made available to Project Officers by the EPA
Environmental Monitoring Systems Laboratories (Cincinnati, Las Vegas, and
Research Triangle Park) through the lERL-Ci Quality Assurance  Officer.
Reference standards can also be obtained from the National Bureau  of
Standards (NBS) through the lERL-Ci Quality Assurance Officer.

     If it has been decided to use performance test samples,  inform  the
lERL-Ci Quality Assurance Officer of your needs  for test  samples,  and he
will arrange to have them sent to the  contractor or grantee.

     If the Project Officer wishes to  have one or more samples split  and
analyzed by another laboratory, he should contact the lERL-Ci  Quality Assur-
ance Officer who can make arrangements for the independent analysis.  In a
similar manner, he may arrange for independent sampling and  analysis  if
this is considered necessary by the Project Officer as a QA  check  on  the work
of the contractor/grantee.

Sub-Contracts and Sub-Agreements—
     In the event sub-contracted services are used for either  sampling  or
analysis, the Project Officer should review the  sub-contractor's QA  program
using the checklist in Appendix D.  It should meet the same  requirements that
would be expected of the prime contractor/grantee if  he were  doing the  work
in-house.  The prime contractor/grantee has full responsibility  for  the
quality of the work just as if it were conducted within his  own  organization.
When a substantial amount of the effort or a particularly critical part  of
the sampling or analysis is to be accomplished by a sub-contractor or by a
sub-agreement, a QA audit should be conducted.
                                     52

-------
                                   SECTION  3

                QUALITY ASSURANCE  GUIDELINES  FOR  AUDITING  OF
                  PROJECTS REQUIRING  SAMPLING AND ANALYSIS

INTRODUCTION

Quality Assurance Audits

     This section provides guidelines  for  the evaluation of  the  performance
of a contractor, grantee, or agency sampling  and  analysis  program and
describes the QA audit process.  Specific  guidelines  are provided to:

     1.  Determine when to conduct a  QA  audit,
     2.  Assist Project Officers in conducting QA audits of  grants  and
         contracts requiring sampling  and  analysis,
     3.  Assist those who conduct  the  QA audit  in evaluation  of  the results,
         and
     4.  Outline the procedures  for the  reporting of  information to the con-
         tractor, grantee, or  agency  doing the work  thereby  allowing them to
         correct any significant QA deficiencies.

In addition, the Quality Assurance Officer is available to conduct  or  to
assist in conducting quality assurance audits.  A checklist  is  provided to
facilitate the QA auditing process.

Quality Assurance in Audits

     In the procurement phase  of contracting, the EPA Project Officer  uses
quality assurance criteria to  aid  in  the selection of the  most  qualified
contractor.  Selection of a contractor or  specifying  QA criteria for
in-house projects and grants,  however, does not guarantee  that  the  overall
performance on the project will meet  the QA requirements.  Therefore,  it is
necessary at times to perform  laboratory quality  assurance audits.   These
audits help to assure the Project  Officer  that all the necessary quality
assurance is being applied by  the  project  team  in order to deliver  a quality
product.  Quality assurance audits allow the Project  Officer  to  determine
that:

     1.  The Organization and  Personnel  are qualified to perform assigned
         tasks;
     2.  Adequate Facilit ies are available;
     3.  Proper Analytical Methodology is  being used;
     4.  Proper Sample and Sample  Handling Procedures, including chain-
         of-custody of samples, are being  used;
                                      53

-------
     5.  Adequate analytical Quality Control,  including reference samples,
         control charts, and documented  corrective  action  measures,  is  being
         provided; and
     6.  Acceptable Data Handling and  documentation techniques  are  being
         used.

QUALITY ASSURANCE AUDIT GUIDELINES

When to Conduct a Quality Assurance Audit

     A QA audit should be conducted on all  projects requiring  sampling  and
analysis services, especially those which provide data  that  are:

     1.  Used to make economically  important decisions,
     2.  Used for regulatory monitoring,
     3.  Used for regulation promulgation,  or
     4.  Used for enforcement activities.

     A QA audit of an on-going project can  assure the Project Officer that
adequate QA measures are being taken to  yield  data  of acceptable  quality.
It will also convey to the  laboratory  organization  audited  that  the  Project
Officer places a high degree of  importance  on  the quality  of sampling and
analytical effort.  A QA audit should  not be used as a  punitive  device  to
express displeasure with the performance of an organization, but  should  be
used to jointly establish acceptable QA  procedures  for  the  given  project.
The Project Officer is dependent on reliable sampling and  analytical data  to
accomplish the objectives of the project.   The contractor,  grantee,  or  gov-
ernment organization conducting  the sampling and analysis  is also interested
in providing reliable data.  It  should be remembered that  both have  the  same
objectives.  Hence, cooperative  efforts  toward this end are  likely  to be
highly successful.

     Ideally, a QA audit should  be  conducted early  in a project,  preferably
before the first sampling and analysis effort  is completed.  This is the best
time to influence the outcome of the entire project.  If a  project  plan  calls
for periodic sampling and analysis, it is recommended that  the QA audit  be
scheduled during the first  or second phase  of  the sampling  and  analytical
effort.  For example, if a  project  is  of 12-month duration  and  requires
monthly sampling, the QA audit should  be conducted  during  the  first  or  second
month of the project.  Of course, QA audits may be  conducted more than  once
during a project or may be  repeated when necessary  to affirm that acceptable
QA procedures are in use.

Audit Worksheet and Checklist

     There are a large number of quality control procedures  which can be
observed during the course  of a  laboratory  audit.   Since an  audit rarely
lasts longer than one day and can often  be  as  short as  a few hours,  an
auditing procedure must be  used  which  can provide the Project Officer or his
                                     54

-------
designee with the greatest amount of  relevant  information  in  the  shortest
amount of time.

     The approach taken by these guidelines  is  to  combine  all  applicable
aspects of a laboratory quality assurance  program  into  the general  categories
listed below.

     1.  Organization and Personnel
     2.  Facilities
     3.  Sampling and Sample Handling Procedures
     4.  Analytical Methodology
     5.  Quality Control
     6.  Data Handling

     Two forms incorporating the preceding information  have been  prepared  to
assist the Project Officer in performing a laboratory audit.   The first  form
is a worksheet which is sent to the  laboratory  scheduled  to be audited.  This
form should be mailed to the project manager of  the  laboratory approximately
one month before the anticipated audit.

     The laboratory should complete  the worksheet  and return  it  to  the EPA
Project Officer within two weeks of  the anticipated  audit.  The  Project  Offi-
cer will then screen the worksheet for areas which may  require more detailed
explanations.  This screening process should be  limited to those  areas of
importance in the sampling and analysis efforts  of the  project.

     Some information which is requested in  the  worksheet  may  be  available in
the original proposal or grant application.  The use of this  information in
completing the worksheet is encouraged to  minimize the  QA  audit  cost.

     The second form which has been  developed  is a checklist  for  use by  the
Project Officer or his representative in performing  the audit.   The checklist
has been developed to:

     1.  Verify a representative number of responses received  in  the
         worksheet,
     2.  Elicit more project-specific information, and
     3.  Provide the Project Officer  additional  information to use  in
         evaluating the laboratory.

     The worksheet and checklist have been developed for  performing QA audits
on projects with either:  (1) a scope of work  encompassing several  scientific
areas, or (2) projects with a limited scope  of  work  which  do  not  have  specif-
ic audit procedures available.  Specific QA  elements for  all  IERL-Ci projects
are addressed in Section 2.  Available auditing  procedures are listed  in
Table 8.  Since auditing procedures  are project-specific  and  can  be used to
perform an audit in much greater detail, they  should be employed  where
available.  The laboratory worksheet  and the checklist  are appended.
                                     55

-------
         TABLE 8.  AVAILABLE QUALITY ASSURANCE AUDIT PROCEDURES
        Parameter
            Reference
Ambient Sulfur Dioxide
(Pararosaniline)
Lab and Field
Ambient Nitrogen Dioxide
(Sodium Arsenite)
Lab and Field

Total Suspended Particulate
(Hi-Volume Method)

Photochemical Oxidants
(Chemiluminescent)

Carbon Dioxide
(Non-Dispersive Infrared)

Metals, Organics, and Inorganics
  in Drinking Water
Radiation  in Drinking Water

Bacteria in Drinking Water
U.S. Environmental Protection
Agency.  1976.  Criteria and
Procedures for the Evaluation of
Ambient Air Monitoring Programs
—Laboratory and Field.

(same as above)
(same as above)


(same as above)


(same as above)
U.S. Environmental Protection
Agency.  1977.  Manual for the
Interim Certification of Labora-
tories Involved in Analyzing
Public Drinking Water Supplies.
Washington, D.C.
EPA-600/8-78-008.

U.S. Environmental Protection
Agency.  1979.  Handbook for
Analytical Quality Control.
Cincinnati, Ohio.

(same as above)

U.S. Environmental Protection
Agency.  National Environmental
Research Laboratory.  1975.
Handbook for Evaluating Water
Bacteriological Laboratories.
Cincinnati, Ohio.
EPA-670/9-75-006.
 (continued)
                                   56

-------
                         TABLE 8 (continued)
        Parameter                                 Reference
Bacteria in Drinking Water (Cont.)    U.S.  Environmental Protection
                                      Agency.   Environmental  Monitor-
                                      ing and  Support  Laboratory.
                                      1978.  Microbiological  Methods
                                      for Monitoring the Environment,
                                      Cincinnati,  Ohio.
                                      EPA-600/8-78-017.

Biological Sampling                   U.S.  Environmental Protection
                                      Agency.   Environmental  Monitor-
                                      ing and  Support  Laboratory.
                                      1978.  Quality Assurance  Guide-
                                      lines  for Biological  Testing.
                                      Las Vegas,  Nevada.
                                      EPA-600/4-78-043.
                                   57

-------
Use of Performance Test Samples

     Performance test samples, when available,  present  the  Project  Officer
with an alternate method of performing  a  laboratory  audit on  a  project  which
requires sampling and analysis.  EPA has  prepared  and has available test
samples for numerous parameters  in many different  matrices.

     Certain projects will not warrant  a  full-scale, on-site  audit  even
though the data produced by the  sampling  and  analysis effort  meets  the  cri-
teria established (see page 54).  Table 5,  page  22,  establishes  the criteria
to be used in determining the necessity of  requiring performance test
samples.  In these cases, test samples  can  be  used to audit the  analytical
performance of the laboratory.

     Other projects may be of sufficient  importance  to  require  both an
on-site audit and performance test samples.  When  problems  in analytical
procedures are identified during an on-site audit, performance  test samples
can be used to verify that discrepancies  have  been corrected  by  the
laboratory.

     Performance test samples should not  be used indiscriminately.   They  are
generally expensive for EPA and  the laboratory to  prepare,  analyze, and
evaluate.  The supply of test samples  is  not  unlimited;  however, in cases
where their value can be established,  they  should  be used.  Performance test
samples for audit purposes can be obtained  through the  Quality  Assurance
Officer (QAO), lERL-Ci.  The QAO can also provide  guidelines  in  the inter-
pretation of the results.  A listing of currently  available performance test
samples is found in Appendix B.

     After performance test samples are analyzed and the results evaluated,
the performing laboratory should be notified  of  the  true value  of the  samples
and the range of acceptance.  In cases  where  limited sets of  samples are  yet
unknown by laboratories, it may  not be  possible  to provide  the  absolute true
values.  When it is inadvisable  to inform the  laboratory of the  true value,
the EMSL through the lERL-Ci Quality Assurance Officer  should provide  infor-
mation, in writing, as to the acceptability/unacceptability and  whether the
results tended to be high or low.

CONDUCTING THE QUALITY ASSURANCE AUDIT

Review of Worksheet

     This section describes the  three  basic steps  for performing a QA audit.
These steps are:  (1) evaluating the returned  worksheet to  determine areas  to
be examined during the site visit, (2)  conducting  the site  visit, and
(3) evaluating the QA audit results, including preparation  of an audit
report.  This section describes  how to  evaluate  the  worksheet returned  by the
organization to be audited.  This worksheet is in  Appendix  F.
                                      58

-------
Organization and Personnel—
     This section of the worksheet  is designed  to:   (1)  familiarize  the
person performing the audit with the laboratory's  organizational  structure,
(2) identify the key personnel  involved  in  the  project,  and  (3) acquaint  the
QA auditor with the skills and  training  of  personnel  actually  involved  in the
sampling and analysis effort.   The QA auditor should  familiarize  himself  with
the Project Manager, the person responsible  for  field  sampling (if required
on this project), and the person responsible  for  the  overall analytical
program.

     Appendix G has been prepared to aid  the QA  auditor  in evaluating  the
skills and training of the personnel responsible  for  sample  collection  and
analyses.  Individuals with assigned responsibilities, as  indicated  in  the
worksheet, should be compared with  the qualifications  noted Appendix G  to
ascertain that the laboratory is placing  proper  emphasis on  quality  assur-
ance.  Any areas which may appear weak should be  noted for discussion  during
performance of the QA audit.

Facilities and Equipment—
     The instrumentation necessary  for the  successful  performance of a  labo-
ratory on a given project is highly dependent on  the  nature  of the project.
The laboratory has been selected based on skills  in  the  project field  and,
therefore, will provide a comprehensive  list of  the  necessary  instrumenta-
tion.  The QA auditor should review the  list of major  instrumentation  in
order to:  (1) determine the type of instrumentation  he  will encounter  during
his site visit, (2) note any instrumentation which may have  been  overlooked
in the preparation of the worksheet, and  (3) acquaint himself  with any  unfam-
iliar instrumentation in order  to perform a  proper evaluation  during the  QA
audit.  A listing of instrumentation and  equipment needed  for  various
sampling and analysis functions is  listed in Appendix G.

     Other equipment and facilities, such as analytical  balances, lighting,
etc., which were not covered by the worksheet will be  evaluated on the  check-
list during the QA audit.  As a general  rule, laboratory space for analytical
requirements should include approximately 120 square  feet of working space
and 6 linear feet of unencumbered bench  space for  each analyst.

Analytical Methodology—
     Methods of analysis vary according  to  the  sample matrix (e.g.,  air,
soil, sediment, wastewater, drinking water,  seawater,  etc.).   A list of
accepted methods for various media and scientific  fields is  presented  in
Appendix C.  These accepted methods should  be compared with  the analytical
methodologies listed in the completed worksheet.  Any  differences between
these methods and the methods listed by  the  contractor/grantee should  be
noted and discussed during the  on-site QA audit.

Sampling and Sample Handling Procedures—
     If sample collection constitutes a major portion  of the work effort,  the
QA auditor should consider it as part of  the QA audit  program  and should
                                     59

-------
study the sampling plan,  sample handling,  and  chain-of-custody procedures
proposed by the laboratory.  The QA auditor  should  make  every attempt  to meet
with laboratory personnel at the sampling  location  and  to  observe  the  quality
control procedures practiced during sampling.  Appendix  H  has been compiled
to aid the auditor in determining  correct  sample  preservation methods.

Quality Control—
     The worksheet poses many basic quality  control  questions.   These
questions should be screened by the QA  auditor to note:   (1)  any questions
answered no; and (2) if QA manuals, standard operating  procedures,  etc.,  are
available for inspection during the on-site  survey.   Each  of  the questions
answered yes is a potential area where  the depth  of  the  QA applied by  the
laboratory can be probed  to determine the  degree  of  commitment  to  quality
assurance on the project.

Data Handling—
     This area covered by the worksheet  allows the QA auditor to determine
the availability of such  items as  field  sampling  notebooks and  past  data.   It
also gives some idea as to the degree of quality  assurance practiced by  the
laboratory in:  (1) analyzing samples within recommended holding times,
(2) rechecking of calculations, (3) rejecting  data,  and  (4) cross-checking
data to reduce the possibility of  confusing  data  and  sample numbers.

Site Visit
     QA audits should be planned  to minimize  the  amount  of  time  on  site  and
to maximize the amount of  information which can be  obtained.   Implementation
of these guidelines should assure that  the QA auditor's  portion  of  the on-
site inspection can be completed  in a maximum of  six  hours.   In  meeting  this
goal, the following procedures should be  used:

     1.  Contact the Project Manager and  arrange  a  mutually  agreeable  time
         and date for the  site visit.   If  field sampling will  be included  in
         the audit, the necessary arrangements should  be made  at this  time.

     2.  Study the worksheet completed  by  the laboratory and  identify  the  key
         personnel.  Note  any questions unresolved  by  the pre-audit  worksheet
         on the QA audit form.  Ask that  the  key  project personnel  be
         available for discussions.

     3.  Prepare an agenda for the meeting which  includes:   (a)  meeting  with
         key personnel in  the laboratory  organization, (b)  a  brief  discussion
         of the purpose of the QA audit,  (c)  a tour of the  facilities, and
         (d) discussions with all personnel who can assist  in  completing the
         QA checklist form (Appendix I).  Table 9 is  a proposed  agenda.

     4.  If a field audit  is to be performed, an  agenda  should be arranged
         with the laboratory to minimize  any  interference that the  audit
         might have on the field  sampling  effort.
                                     60

-------
	TABLE 9.  RECOMMENDED AGENDA FOR QA AUDIT SITE VISIT*

A.  Meet with Key Project Personnel

    1.  Introductions
    2.  Describe purpose of visit

B.  Description of the Sampling and Analytical Effort

    1.  Sampling
    2.  Analysis
    3.  Data presentation and interpretation

C.  Quality Assurance Program

    1.  Description
    2.  Discussions with key project personnel
    3.  Discussion of worksheet review

D.  Tour of Facilities

    1.  Checklist
    2.  Discussions with staff personnel

E.  Debriefing

    1.  Preliminary evaluation
    2.  Serious problems requiring  immediate corrective action
    3.  Schedule for QA audit
*A minimum of one hour should be devoted to each section of B through E.
                                  61

-------
Quality Assurance Audit Report

     A written report of  the QA audit  results must  be  prepared.   This  report
should include:

     1.  An assessment of the performing organization's  status  in each of  the
         major areas addressed, including  personnel,  facilities,  procedures,
         and quality control.

     2.  A clear statement of areas requiring improvement  or  problems  to be
         corrected.

     3.  A timetable for  correction of problem  areas.

     To effectively achieve QA audit objectives,  the  report  should be  sent  to
the performing organization as soon as possible,  preferably  within 15  working
days of completion of the audit evaluation.  Corrective  action, when
required, should be scheduled to meet  the  timetable of the project.

     Figure 9 is a suggested format for the QA  audit  report.  Both the Pre-
Audit Worksheet and the Audit Checklist should  be appended to this report.

     The original QA Audit Report  is maintained by  the lERL-Ci  Quality
Assurance Officer, and copies should be distributed as follows:

     Copies

       1                  Project Officer
       1                  Extramural Laboratory  Project Manager

     If the QA audit was  satisfactory  and  no specific  followup  action  is
required, the QA audit is complete.  If deficiencies have  been  noted and cor-
rective actions specified, then the Project Officer must assure that correc-
tions have been made in accordance with the specified  timetable.   Written
communications should be  entered into  the  project file to  document the cor-
rection of noted deficiencies.  The principal investigator should be invited
to submit any response he desires  for  inclusion into  the project  file.
                                     62

-------
                     QUALITY ASSURANCE AUDIT REPORT

Project:	
Contract/Grant No.:	 Project Officer:	
Laboratory Audited:	
         City:	 State:
Laboratory Director:	
Audit Conducted By:
         Agency:	
         Date:
PURPOSE AND OBJECTIVES OF THE PROJECT:
BRIEF DESCRIPTION OF THE SAMPLING AND ANALYTICAL REQUIREMENTS:
               Figure 9.  Quality Assurance Audit Report.

                                 63

-------
                 RESULTS OF THE QUALITY ASSURANCE AUDIT
Organization and Personnel
Facilities
Analytical Methodology
Figure 9 (continued)
                                 64

-------
                 RESULTS OF THE QUALITY ASSURANCE AUDIT
Sampling and Sample Handling
Quality Control
Data Handling
Figure 9 (continued)
                                 65

-------
QUALITY ASSURANCE DEFICIENCIES:
RECOMMENDED CORRECTIVE ACTIONS AND TIMETABLE:
                                         Signed                 Date
                                                     Title

Distribution:

     Project Officer
     Quality Assurance Officer
     Contractor/Grantee


Figure 9 (continued)

                                 66

-------
                                  SECTION 4

                                 CONCLUSIONS

PROCUREMENT

     Section 1 of this document provides suggestions,  recommendations,  and
procedures whereby Project Officers can  introduce QA considerations  into  the
procurement process when sampling and  analysis  is involved  in  the  work.
Specific recommendations are given  for determining  the  appropriate weighting
to apply to QA in the proposal evaluation criteria  for  contracts.  Contrac-
tors and grantees must comply with  lERL-Ci quality  assurance requirements,  if
the Laboratory is to achieve its assigned mission of developing  new  and
improved pollution control technology.  These quality  assurance  program
requirements are intended to be stringent but justifiable.  Conscientious
application by lERL-Ci Project Officers of the  procedures described  in
Section 1 should result in equitable treatment  of all  offerers and grant
applicants.

MONITORING

     Section 2 of this document has been prepared by the lERL-Ci Quality
Assurance officer to describe procedures to maintain the quality of  sampling
and analysis activities conducted by contractors and grantees  under  the
direction or funding of lERL-Ci.  Of paramount  importance in achieving  and
maintaining quality work is the role and responsibilities of the Project
Officer.  These responsibilities have  been defined  and  discussed for
contracts, grants, and cooperative  agreements.

     The three foundations of project  QA:  (1)  the  contractor's/grantee's QA
program, (2) the Project Work Plan, and (3) project QA  monitoring  have  been
discussed from the standpoint of the review responsibilities of  the  Project
Officer.  Checklists are provided to assist the Project Officer  in reviewing
contractor/grantee QA programs and  project work plans.

     If the Laboratory is to achieve its assigned mission of developing new
and improved pollution control technology in a  cost-effective manner, then
each project must incorporate sound QA practices.   Contractors and grantees
must be informed of and comply with lERL-Ci quality assurance  requirements.

AUDITING

     Procedures and a checklist are provided in Section 3 for  conducting
quality assurance audits of on-going lERL-Ci projects which require  sampling
and analysis services.  It is recommended that  a QA audit be conducted during
the beginning stages of all projects which require  significant amounts of
sampling and analysis.
                                     67

-------
                                BIBLIOGRAPHY
American Public Health Association, American Water Works Association, and
     Water Pollution Control Federation.  1975.  Standard Methods for the
     Examination of Water and Wastewater.  14th Edition.  Washington, D.C.

American Society for Testing and Materials.   1978.  Annual Book of ASTM
     Standards, Part 31:  Water.  Philadelphia, Pennsylvania.

Bicking, C., Olin,  S., and King, P.  1978.  Procedure for the Evaluation of
     Environmental Monitoring Laboratories.   U.S. Environmental Protection
     Agency, Environmental Monitoring and Support Laboratory, Office of
     Research and Development, Cincinnati, Ohio.  EPA-600/4-78-017.

Costle, D.  June 14, 1979.  Memorandum—Quality Assurance Requirements for
     All EPA Extramural Projects Involving Environmental Measures.

Environmental Science and Engineering, Inc.   1977.  Water Quality Field
     Sampling Manual.  Gainesville, Florida.

Fairless, B.  1977.  Quality Assurance Practices and Procedures.
     U.S. Environmental Protection Agency, Region V, Surveillance and
     Analysis Division, Chicago, Illinois.  EPA-905/4-77-004.

Geldreich, E.E.  1975.  Handbook for Evaluating Water Bacteriological
     Laboratories.   Second Edition.  U.S. Environmental Protection Agency,
     Office of Research and Development, Municipal Environmental Research
     Laboratory, Cincinnati, Ohio.  EPA-670/9-75-006.

Kittrell, F.W.  1969.  A Practical Guide to Water Quality Studies of Streams.
     U.S. Department of Interior, Federal Water Pollution Control
     Administration.  CWR-5.

Krawczyk, D.F.  n.d.  Analytical Quality Control.  U.S. Environmental
     Protection Agency, Pacific Water Laboratory, Corvallis, Oregon.

Linch, A.L.  1973.   Quality Control for Sampling and Laboratory Analysis.
     In:  The Industrial Environment—Its Evaluation and Control,
     pp. 277-297.  U.S. Department of Health, Education, and  Welfare, Public
     Health Service, Center for Disease Control, National Institute for
     Occupational Safety and Health.
                                     68

-------
                          BIBLIOGRAPHY (Continued)
Sherraa, J.  1976.  Manual of Analytical Quality Control for Pesticides and
     Related Compounds in Human and Environmental Samples:   A Compendium of
     Systematic Procedures Designed to Assist in the Prevention and Control
     of Analytical Problems.  Prepared for U.S. Environmental Protection
     Agency, Office of Research and Development, Health Effects Research
     Laboratory, Research Triangle Park,  North Carolina.   EPA-600/1-76-017.

U.S. Environmental Protection Agency,   n.d.   EPA Project  Officer's Guide
     (Research & Demonstration Grants).  U.S. Environmental Protection
     Agency, Office of Planning and Management, Office of Administra-
     tion, Grants Administration Division, Washington, D.C.

U.S. Environmental Protection Agency,   n.d.   Guidance Package for Evaluation
     of State Laboratories (Source Sampling)—Draft.  Cincinnati, Ohio.

U.S. Environmental Protection Agency.   1976.  Minimal Criteria and Procedures
     for the Evaluation of Ambient Air Monitoring Programs—Laboratory and
     Field.  Draft III.

U.S. Environmental Protection Agency.   1977.  Materials from Course 470—
     Quality Assurance for Air Pollution Measurement Systems.  Research
     Triangle Park, North Carolina.

U.S. Environmental Protection Agency.   1977.  Procurement Information
     Notice PIN 77-15—Source Evaluation and Selection Procedures.

U.S. Environmental Protection Agency.   1978.  Project Management and the
     Procurement Process:  A Seminar Workshop for Project Officers and
     Other Technical Personnel.  Washington, D.C.  203 pp.

U.S. Environmental Protection Agency.   Health Effects Research Laboratory.
     Environmental Toxicology Division.  1974, 1977 rev.  ed.  Analysis of
     Pesticides Residues in Human and Environmental Samples:  A Compilation
     of Methods Selected for Use in Pesticide Monitoring  Programs.  Edited by
     J.F. Thompson.  Research Triangle Park, North Carolina.

U.S. Environmental Protection Agency.   Office of Research and Development.
     Environmental Monitoring and Support Laboratory.  1976.  Quality
     Assurance Handbook for Air Pollution Measurement Systems:  Volume I—
     Principles.  Research Triangle Park, North Carolina.  EPA-600/9-76-005.

U.S. Environmental Protection Agency.   Office of Research and Development.
     Environmental Monitoring and Support Laboratory.  1977.  Quality
     Assurance Handbook for Air Pollution Measurement Systems:  Volume II—
     Ambient Air Specific Methods.  Research Triangle Park, North Carolina.
     EPA-600/4-77-027a.
                                     69

-------
                          BIBLIOGRAPHY (Continued)
U.S. Environmental Protection Agency.  Office of Research and Development.
     Environmental Monitoring and Support Laboratory.  1978.   Environmental
     Radioactivity Laboratory Intercomparison Studies Program, 1978-1979.
     Las Vegas, Nevada.  EPA-600/4-78-032.

U.S. Environmental Protection Agency.  Office of Research and Development.
     Environmental Monitoring and Support Laboratory.  1979.   Handbook for
     Analytical Quality Control in Water and Wastewater Laboratories.
     EPA-600/4-79-019.

U.S. Environmental Protection Agency.  Office of Research and Development.
     Environmental Monitoring and Support Laboratory.  1979.   Methods for
     Chemical Analysis of Water and Wastes.  Cincinnati,  Ohio.
     EPA-600/4-79-020.

U.S. Environmental Protection Agency.  Office of Water Planning and
     Standards.  Monitoring and Data Support Division and Environmental
     Monitoring and Support Laboratory.  1975.  Minimal Requirements for a
     Water Quality Assurance Program, Cincinnati, Ohio.  EPA-440/9-75-010.

U.S. Geological Survey.  Office of Water Data Coordination.   1977.  National
     Handbook of Recommended Methods for Water-Data Acquisition.   Reston,
     Virginia.

Water Supply Quality Assurance Work Group.  1977.  Manual for the Interim
     Certification of Laboratories Involved in Analyzing Public Drinking
     Water Supplies—Criteria and Procedures.  Prepared for  the
     U.S. Environmental Protection Agency, Washington, D.C.

Weber, C.I., ed.  1973.  Biological Field and Laboratory Methods  for Meas-
     uring the Quality of Surface Waters and Effluents.  U.S. Environmental
     Protection Agency, National Environmental Research Center, Office of
     Research and Development, Cincinnati, Ohio.  EPA-670/4-73-001.
                                     70

-------
                               APPENDIX A

          QUALITY ASSURANCE EVALUATION CRITERIA CHECKLIST FOR
                PROPOSALS AND GRANT APPLICATIONS OFFERING
                     SAMPLING AND ANALYSIS SERVICES
                                       Scoring   Numerical   Individual
           Criteria                     Value  x  Weight   =   Score
A.  Quality assurance management policy/written procedures.

la.  Does the offerer have an           	  x     5     =    	
     on-going QA program?

2a.  Does the offeror have a written    	  x     4     =    	
     QA manual that he will make
     available for review?

3a.  Has the offeror designated         	  x     3     =    	
     a QA coordinator or a QA
     supervisor who reports to senior
     management levels?

4a.  Does the proposed project manage-  	  x     2     =    	
     ment structure provide for
     adequate QA?

5a.  Will a project specific QC plan    	  x     1     =    	
     be prepared before commencement
     of sampling and analysis?

                               Total Score for Sub-element A.   	

                               Maximum Possible Score             75

Percent of maximum possible score awarded for Sub-element A. (circle
closest value).
                  0%   20%   40%   —    60%   80%   100%

Score for this sub-element of the proposal evaluation criteria (circle
corresponding value).
                  0     1    2a    2b     3     4     5
                                   71

-------
                         APPENDIX A (continued)
Criteria
        Scoring   Numerical
         Value  x  Weight
                                                             Individual
                                                               Score
B.  Quality assurance procedures for sampling.

Ib.  Are sampling locations chosen      	
     to assure representative samples
     will be taken?

2b.  Will the proposed sampling         	
     program yield data of statisti-
     cal significance as appropriate
     to the objectives of the project
     (e.g., replicate samples,
     background samples, etc., should
     be discussed)?

3b.  Does the offerer show an under-    	
     standing of the proper techniques
     used to collect representative
     samples while avoiding sample
     contamination?

4b.  Does the offeror have access to    	
     the appropriate sampling
     equipment?

5b.  Are samples to be shipped          	
     promptly to the laboratory to
     meet maximum sample holding
     time limitations?

6b.  Are appropriate sample preserva-   	
     tion methods proposed?

7b.  Are sample chain-of-custody        	
     procedures described?
Total Score for Sub-element B.

Maximum Possible Score
                                                                  75
Percent of maximum possible score awarded for Sub-element B. (circle
closest value).
                  0%   20%   40%   —    60%   80%   100%

Score  for this sub-element of the proposal evaluation criteria (circle
corresponding value).
                  0      1    2a    2b     3     4     5
                                   72

-------
                         APPENDIX A (continued)
                                       Scoring   Numerical   Individual
           Criteria                     Value  x  Weight   =   Score
C.  Quality assurance procedures for analysis.

Ic.  Does the offerer intend to use     	
     standard analytical methods*
     where available?  If standard
     methods are not available, will
     the methods used be documented?

2c.  Does the offeror have a labora-
     tory QC program which specifies
     at least 5-10 percent sample
     replication and 5 percent spiked
     sample analysis?

3c.  Is high quality analytical         	
     instrumentation available for
     use on the project?
4c.   Are laboratory facilities          	  x     2     =    	
     adequate?

5c.   Are analytical detection limits    	  x     2     =    	
     adequate for the purposes of
     the project?

6c.   Does the offeror participate       	  x     2     =    	
     in EPA and/or other interlabora-
     tory QC programs?

                               Total Score for Sub-element C.   	

                               Maximum Possible Score             75

Percent of maximum possible score awarded for Sub-element C. (circle
closest value).
                  0%   20%   40%   —    60%   80%   100%

Score for this sub-element of the proposal evaluation criteria (circle
corresponding value).
                  0     1    2a    2b     34     5


*Appendix C provides a  listing of acceptable EPA Analytical Methods.
                                   73

-------
                         APPENDIX A (continued)
                                       Scoring   Numerical   Individual
           Criteria                     Value  x  Weight   =   Score
D.  Quality assurance procedures for data management.

Id.  Does the offeror possess           	  x     '.
     appropriate data handling,
     processing, and retrieval
     capabilities?

2d.  Will QC data (e.g., standard       	  x     '.
     curves, duplicate results, spike
     sample results) be maintained and
     be accessible to the Project
     Officer?
3d.  Does the organization routinely    	  x     2     =    	
     maintain analytical performance
     records such as quality control
     charts?

4d.  Are all laboratory results and     	  x     3     =    	
     QC data reviewed by laboratory
     supervisory personnel?

5d.  Are all data and records retained  	  x     1     =    	
     for a minimum of 3 years?

6d.  Are field notebooks used to        	  x     3     =    	
     record sampling and engineering
     data (e.g., sample number, date/
     time of collection, flow,
     operating conditions, etc.)?

                               Total Score for Sub-element D    	

                               Maximum Possible Score             75

Percent of maximum possible score awarded for Sub-element D (circle
closest value).
                  0%    20%   40%   —    60%   80%   100%

Score  for this sub-element of the proposal evaluation criteria  (circle
corresponding value).
                  0     1    2a    2b     3     4     5
                                   74

-------
                                 APPENDIX B

             QUALITY CONTROL PERFORMANCE/REFERENCE TEST SAMPLES
     The information contained in this Appendix consists of a listing of
performance/reference test samples and their sources.  In addition to those
samples currently available,  new materials expected to be available in FY 79
are also listed.
WATER QUALITY PARAMETERS (Available
lERL-Ci)

Mineral/Physical Analyses

   Calcium
   Magnesium
   Sodium
   Potassium
   Alkalinity
   Sulfate
   Chloride/Fluoride
   Residue, Total Filterable
   Hardness, Total
   pH
   Conductance

Demand Analyses

   Organic Carbon
   Chemical Oxygen Demand
   Biochemical Oxygen Demand

Nutrients

   NH3 - N
   N03 - N
   P04 - P
   Kjeldahl - N
   Phosphorous, Total

Nitrilotriacetic Acid

   NTA
                                    from EMSL-Ci through the QA Officer,
                                         Trace Metals
                                            Aluminum
                                            Arsenic
                                            Beryllium
                                            Cadmium
                                            Cobalt
                                            Chromium
                                            Copper
                                            Iron
                                            Lead
                                            Manganese
                                            Mercury
                                            Nickel
                                            Selenium
                                            Vanadium
                                            Zinc

                                         Mercury (Organic & Inorganic)

                                            Mercury,  Total

                                         Linear Alkylate Sulfonate

                                            EPA/SDA Standard Solution,
                                              5% Active
                                     75

-------
                           APPENDIX B (continued)
Residue Analysis

   Residue, Total Non-Filterable—
     fine particles
   Residue, Total Non-Filterable—
     fibrous material
   Residue, Total Non-Filterable—
     coarse particles

Chlorophyll

   Chlorophyll, a, b, c, and
     pheophytin for Spectrophoto-
     metric Analyses
   Chlorophyll, a and pheophytin
     for Fluorometric Analyses

Volatile Organics

   1,2, Dichloroethane
   Chloroform
   1,1,1 Trichloroethane
   1,1,2 Trichloroethylene
   Carbon Tetrachloride
   1,1,2,2 Tetrachloroethylene
   Broraodichloromethane
   Dibromochloromethane
   Bromoform

Polychlorobiphenyls

   Aroclor 1016
   Aroclor 1254

Petroleum Hydrocarbons for
Characterization

   Two Crude Oils
   #2 Fuel Oil
   Bunker C

Pesticides, Chlorinated Hydrocarbon

   Aldrin
   Chlordane
   Dieldrin
   Heptachlor
   DDT
   DDD
   DDE
Acid Extracts (Phenolics)

Municipal Digested Sludge
Cyanide
Oil/Gre~ase
Trihalomethanes
Turbidity
Free Chlorine

Base/Neutral Extracts

   Phthalates
   Nitro- and Chloro-benzenes
   PNAs
   Benzidine (Standard Only)
   2,4-Dinitrotoluene
   2,6-Dinitrotoluene
   Isophorone
   Nitrobenzene
   Acenaphthene
   1,2,4-Trichlorobenzene
   Hexachlorobenzene
   Hexachloroethane
   2-Chloronaphthalene
   1,2-Dichlorobenzene
   1,3-Dichlorobenzene
   1,4-Dichlorobenzene
   Fluoranthene
   Hexachlorobutadiene
   Hexachlorocyclopentadiene
   Naphthalene
   Dimethylphthalate
   Di ethyIphthaiate
   Di-n-butylphthalate
   Butylbenzylphthalate
   bis(2-ethylhexyl)phthalate
   1,2-Benzanthracene
   Benzo(a)pyrene
   Chrysene
   Acenaphthylene
   Anthracene
   1,12-Benzoperylene
   Fluorene
   Phenanthrene
   1,2,5,6 Dibenzanthracene
   Indeno(l,2,3 c,d)pyrene
   Pyrene
                                     76

-------
                           APPENDIX B (continued)
WATER SUPPLY QC SAMPLES (Available
  from QA Coordinators through QA
  Officer, lERL-Ci)

Metals

   Arsenic
   Barium
   Cadmium
   Chromium
   Lead
   Mercury
   Selenium
   Silver

Nitrate/Fluoride

Trihalomethanes

Pesticides

   Endrin
   Lindane
   Methoxychlor
   Toxaphene

Herbicides
   2,4 D
   2,4,5-TP Silvex

PRIORITY POLLUTANTS (Available from
  EMSL through QA Officer, lERL-Ci)

Pesticides

   Toxaphene
   Chlordane
   Endrin
   Heptachlor
   Aldrin
   Dieldrin
   4,4-DDT
   4,4-DDE
   4,4-ODD
   BHC
PRIORITY POLLUTANTS (continued)

Aroclors

   PCB-1016
   PCB-1254

Purgeable Compounds

   Benzene
   Carbon Tetrachloride
   Chlorobenzene
   1,2-Dichloroethane
   1,1,1-Trichloroethane
   1,1-Dichloroethane
   1,1,2-Trichloroethane
   1,1,2,2-Tetrachloroethane
   Chloroform
   1,1-Dichloroethylene
   1,2-trans-Dichloroethylene
   1,2-Dichloropropane
   1,3-Dichloropropylene
   Ethylbenzene
   Methylene chloride
   Bromoform
   Dichlorobromomethane
   Trichlorofluoromethane
   Tetrachloroethylene
   Toluene
   Trichloroethylene
   o,  m, and p Xylene
                                     77

-------
                           APPENDIX B (continued)
RADIATION (Available from EMSL— Las
  Vegas through QA Officer, lERL-Ci)

Water

   gross a
   gross
                ,
      Cs134,137)
   H3
   PU239
   Ra226,228
   Sr89,90

Water Laboratory Certification Samples

Blind Sample
Milk
Sr89,90>
                  Cs137> Ba140
Air
gross a ,  gross 0
                           ancj
Soil
   Pu238,239> Th228,230,232
Diet
8,89,90,
                         Ba140 and
Urine
Gas
AIR PARAMETERS (Available from EMSL—
  RTP through QA Officer, lERL-Ci)
    N02 (ambient)
    S02 (ambient)
    S04/N03 (glass-fiber filter)
    NH4S04 (glass-fiber and teflon filter)
    Pb (glass-fiber and membrane filter)
    As (glass-fiber filter)
AIR PARAMETERS (continued)

   NO (cylinders,  3 levels)
   N02 (cylinders)
   S02 (cylinders)
   CO (cylinders,  3 levels)
   CH^ (cylinders, 9 levels)
   Zero Air (for HC and NOX)
   S02 (source)
   NOX (source)
   Benzene (cylinders,  2 levels)
   Ethylene (cylinders, 3 levels)
   Methane/Ethane  (cylinders,
     4 levels
   Propane (cylinders,  2 levels)
   Propylene (cylinders, 2 levels)
   Toluene (cylinders,  2 levels)
   Methyl Acetate  (cylinders,
     2 levels
   Vinyl Chloride  (cylinders,
     2 levels
   Hydrogen Sulfide (cylinders,
     3 levels
   m-Xylene (cylinders, 2 levels)
   Chloroform (cylinders, 2 levels)
   Perchloroethylene (cylinders,
     2 levels)
   Butadiene (cylinders)
   Carbonyl Sulfide (cylinders,
     4 levels)
   Hexane (cylinders,  4 levels)
   Methyl Mercaptan (cylinders,
     4 levels)
   Methyl Ethyl Ketone (cylinders)
   Trichloroethylene (cylinders,
     2 levels)
   Vinylidene Chloride (cylinders,
     2 levels)
   1,2 Dichloroethane  (cylinders,
     2 levels)
   Acetaldehyde (cylinders, 2  levels)
   Propylene Dichloride (cylinders,
     2 levels)
   1,2 Dibromoethylene (cylinders
     2 levels)
   Acrylonitrile (cylinders,
     2 levels)
   Cyclohexane (cylinders)
   Methanol (cylinders)
                                     78

-------
                                             APPENDIX C
                                  EPA ACCEPTED ANALYTICAL METHODS

                            TABLE C-l.  WATER QUALITY MEASUREMENT METHODS
       Parameter
             Method
Reference*
Acidity, as CaC03, mg/1

Alkalinity, as CaC03, mg/1


Ammonia (as N), mg/1
Bacteria
Coliform (fecal), no./lOO ml
Coliform (total), no./lOO ml
Fecal streptococci, no./lOO ml
Benzidine,  mg/1
Biochemical oxygen demand 5-d
  (BOD5), mg/1
Bromide, mg/1
Chemical oxygen demand
  (COD), mg/1
Chloride, mg/1

Chlorinated organic compounds
  (except pesticides), mg/1
Chlorine—total residual, mg/1
Color, platinum cobalt units
  or dominant wavelength, hue,
  luminance, purity
(continued)
Electrometric end point (pH of 8.2)           1, 2, 3, 4
or phenolphthalein end point
Electrometric titration (only to pH 4.5)      1, 2, 3, 4
manual or automated, or equivalent
automated methods
Manual distillation  (at pH 9.5)               1, 2, 3, 4
followed by nesslerization, titration,
electrode, Automated phenolate

MPN; membrane filter                          2, 4
                                              2, 4
MPN; membrane filter, plate count             2, 4
Oxidation—colorimetric                       5
Winkler (Azide modification) or electrode     2, 4
method
Titrimetric, iodine-iodate                    1, 3, 4
Dichromate reflux                             1, 2, 3, 4

Silver nitrate; mercuric nitrate; or          1, 2, 3, 4
automated colorimetric-ferricyanide
Gas chromatography                            6

lodometric titration, amperometric or         1, 2, 3
starch-iodine end-point; DPD colorimetric
or Titrimetric methods (these last 2 are
interim methods pending laboratory testing)
Colorimetric; spectrophotometric; or          1, 2, 4
ADMI procedure

-------
                           TABLE C-l.   WATER QUALITY  MEASUREMENT METHODS (Continued)
             Parameter
             Method
Reference*
oc
c
      Cyanide,  total,  tng/1


      Dissolved oxygen,  mg/1

      Fluoride, mg/1

      Hardness—total,  as CaC03,  mg/1
      Hydrogen ion (pH),  pH units
      Kjeldahl nitrogen  (as N),  mg/1
      Metals
      All  dissolved metals

      Aluminum—total,  mg/1


      Antimony—total,  mg/1
      Arsenic—total, mg/1

      Barium—total,  mg/1
      Beryllium—total,  mg/1

      Boron—total, mg/1

      (continued)
Distillation followed by silver nitrate       1,  2,  3, 4
titration or pyridine pyrazolone (or
barbituric acid) colorimetric
Winkler (Azide modification) or               1,  2,  3, 4
electrode method
Distillation followed by ion electrode;       1,  2,  3, 4
SPADNS; or automated complexone
EDTA titration; automated colorimetric;       1,  2,  3, 4
or atomic absorption (sum of Ca and Mg
as their respective carbonates)
Electrometric measurement                     1,  2,  3, 4
Digestion and distillation followed           1,  2,4
by nesslerization, titration, or
electrode; automated digestion
automated phenolate

0.45 micron filtration' followed by           1,2,4
referenced method for total metal
Digestion® followed by atomic absorption^     1,  2,4
or by colorimetric (Eriochrome
Cyanine R)
Digestion" followed by atomic absorption'     1
Digestion followed by silver diethyl-         1,  2,  4
dithiocarbamate; or atomic absorption9
Digestion" followed by atomic absorption'     1,  2,  4
Digestion® followed by atomic                 1,  2,  4
absorption' or by colorimetric (Aluminon)
Corimetric (Curcumin)                         1,  2

-------
                            TABLE C-l.   WATER QUALITY MEASUREMENT METHODS (Continued)
             Parameter
                                                   Method
                                             Reference*
00
Cadmium—total, mg/1

Calcium—total, mg/1

Chromium VI, mg/1

Chromium—total, mg/1

Cobalt—total, mg/1

Copper—total, mg/1


Gold—total, mg/1
Iridium—total, mg/1
Iron—total, mg/1

Lead—total, mg/1

Magnesium—total, mg/1

Manganese—total, mg/1

Mercury—total, mg/1
Molybdenum—total, mg/1
Nickel—total, mg/1

Osmium—total, mg/1

(continued)
Digestion" followed by atomic absorption^     1, 2, 3, 4
or by colorimetric (Dithizone)
Digestion** followed by atomic                 1, 2, 3, 4
absorption; or EDTA titration
Extraction and atomic absorption;             1, 2,4
colorimetric (Diphenylcarbazide)
Digestion" followed by atomic absorption"     1, 2, 3, 4
or by colorimetric (Diphenylcarbazide)
Digestion** followed by atomic                 1, 2, 3, 4
absorption"
Digestion" followed by atomic                 1, 2, 3, 4
absorption^ or by colorimetric
(Neocuproine)
Digestion** followed by atomic absorption^
Digestion** followed by atomic absorption^
Digestion" followed by atomic absorption^     1, 2, 3, 4
or by colorimetric (Phenanthroline)
Digestion** followed by atomic absorption9     1, 2, 3, 4
or by colorimetric (Dithizone)
Digestion** followed by atomic absorption^     1, 2, 3, 4
or gravimetric
Digestion** followed by atomic absorption^     1, 2, 3, 4
or by colorimetric (Persulfate or periodate)
Flameless atomic absorption                   1, 2, 3, 4
Digestion** followed by atomic absorption9     1, 3
Digestion** followed by atomic absorption^     1, 2, 3, 4
or by colorimetric (Heptoxime)
Digestion" followed by atomic absorption'

-------
                             TABLE C-l.   WATER QUALITY MEASUREMENT METHODS (Continued)
             Parameter
                                                   Method
                                              Reference*
00
10
      Palladium—total,  mg/1
      Platinum—total,  mg/1
      Potassium—total,  mg/1
Rhodium—total, mg/1
Ruthenium—total, mg/1
Selenium—total, mg/1
Silica—dissolved, mg/1

Silver—total, mg/1

Sodium—total, mg/1

Thallium—total, mg/1
Tin—total, mg/1
Titanium—total, mg/1
Vanadium—total, mg/1

Zinc—total, mg/1

Nitrate (as N), mg/1

Nitrite (as N), mg/1

Oil and grease, mg/1
Digestion" followed by atomic absorption^
Digestion** followed by atomic absorption^
Digestion** followed by atomic absorption,     1,  2,  3,  4
colorimetric (Cobaltinitrite), or by  flame
photometric
Digestion** followed by atomic absorption^
Digestion" followed by atomic absorption*
Digestion** followed by atomic absorption^     1,  2
0.45 micron filtration? followed by           1,  2,  3,  4
colorimetric (Molybdosilicate)
         ft                 •              Q
Digestion0 followed by atomic absorption'     1,  2,  4
or by colorimetric (Dithizone)
Digestion" followed by atomic absorption      1,  2,  3,  4
or by flame photometric
Digestion** followed by atomic absorption^     1
Digestion** followed by atomic absorption^     1,  4
Digestion** followed by atomic absorption^     1
Digestion** followed by atomic absorption^     1,  2,  3,  4
or by colorimetric (Gallic acid)
Digestion** followed by atomic absorption9     1,  2,  3,  4
or by colorimetric (Dithizone)
Cadmium reduction; brucine sulfate;           1,  2,  3,  4
automated cadmium or hydrazine reduction^
Manual or automated colorimetric              1,  2,  4
(Diazotization)
Liquid-liquid extraction with                 1,  2
trichloro-trifluoroethane-gravimetric
      (continued)

-------
                      TABLE C-l.   WATER QUALITY MEASUREMENT METHODS (Continued)
       Parameter
Method
                                                                                         Reference
      Organic carbon—total (TOC), mg/1
      Organic nitrogen (as N), mg/1

      Orthophosphate (as P), mg/1

      Pentachlorophenol, mg/1
      Pesticides, mg/1
      Phenols, mg/1

      Phosphorus (elemental), mg/1
      Phosphorus—total (as P), mg/1

      Radiological
oo     Alpha—total, pCi per liter
      Alpha—counting error,
        pCi per liter
      Beta—total, pCi per liter
      Beta—counting error,
        pCi per liter
      Radium—total, pCi per liter
      Ra, pCi per liter
      Residue
      Total, mg/1
      Total dissolved (filterable), mg/1
      Total suspended (nonfilterable),
        mg/1
      Settleable, ml/1 or mg/1
      Total volatile, mg/1
      Specific conductance, micromhos
        per centimeter at 25°C
                                      Combustion—infrared method**
                                      Kjeldahl nitrogen minus ammonia
                                      nitrogen
                                      Manual or automated ascorbic acid
                                      reduction
                                      Gas  chroraatography^
                                      6
                                      Distillation followed by colorimetric
                                      (4AAP)
                                      Gas  chromatography
                                      Persulfate digestion followed by manual or
                                      automated ascorbic acid reduction

                                      Proportional or scintillation counter
                                      Proportional counter
                                      Scintillation counter

                                      Gravimetric,  103 to 105°C
                                      Glass fiber filtration, 180°C
                                      Glass fiber filtration, 103 to 105°C

                                      Volumetric or gravimetric
                                      Gravimetric,  550°C
                                      Wheats tone bridge conductimetry
                                1, 2, 3
                                1, 2, 4

                                1, 2, 3, 4

                                2, 3

                                1, 2, 3

                                12
                                1, 2, 3, 4
                                2, 3, 4
                                2, 3, 4

                                2, 3, 4
                                2, 3, 4

                                2, 3
                                1, 4

                                1, 2
                                1, 2
                                1, 2

                                2
                                1, 2
                                1, 2, 3, 4
(continued)

-------
                            TABLE  C-l.   WATER QUALITY MEASUREMENT METHODS (Continued)
             Parameter
             Method
Reference*
      Sulfate (as  804),  mg/1
      Sulfide  (as  S),  mg/1
      Sulfite (as  803),  mg/1
      Surfactants,  mg/1
      Temperature,  °C

      Turbidity, NTU
Gravimetric; turbidimetric; or
automated colorimetric (barium
chloranilate)
Titrimetric—iodine for levels
greater than 1 mg per liter;
Methylene blue photometric
Titrimetric, iodine-iodate
Colorimetric (Methylene blue)
Calibrated glass or electrometric
thermometer
Nephelometric
2, 3


1, 3, 4
1, 2, 3
1, 2, 3, 4
1, 2, 4

1, 2, 3, 4
00
      *References:

           1  Methods  for  Chemical  Analysis  of Water  and  Wastes.   1979.   U.S.  Environmental Protection
      Agency.  Office  of  Research  and Development.   Environmental Monitoring  and Support Laboratory.
      Cincinnati,  Ohio.

           2  Standard  Methods  for  the Examination  of Water  and Wastewater,  14th  edition.  1975.
      American Public  Health Association,  American Water Works Association  and Water Pollution Control
      Federation.   Washington,  D.C.
           3 Annual  Book of ASTM Standards,  Part  31:
      Materials.   Philadelphia,  Pennsylvania.
          Water.  1978.  American Society  for Testing  and
           ^ All  references  for  USGS methods,  unless  otherwise noted,  are to Brown,  E.,  Skougstad, M.W.,
      and Fishman,  M.J.,  "Methods  for Collection  and  Analysis  of Water Samples  for Dissolved Minerals and
      Gases."  U.S.  Geological Survey Techniques  of Water-Resources  Inv., Book  5,  Ch.  Al (1970).

-------
                            TABLE C-l.  WATER QUALITY MEASUREMENT METHODS  (Continued)
           -* Adequately tested methods for benzidine are not  available.  Until  approved  methods  are
      available, the following interim method can be used  for  the  estimation  of benzidine:   (1)  "Method for
      Benzidine and Its Salts in Wastewaters," available from Environmental Monitoring  and  Support
      Laboratory, U.S. Environmental Protection Agency, Cincinnati, Ohio   45268.

           ^Procedures for pentachlorophenol, chlorinated  organic  compounds,  and  pesticides  can  be
      obtained from the Environmental Monitoring and Support  Laboratory, U.S. Environmental  Protection
      Agency, Cincinnati, Ohio  45268.

           'Dissolved metals are defined as  those constituents which  will  pass  through  a 0.45  urn
      membrane filter.  A prefiltration is permissible  to  free the sample  from  larger  suspended  solids.
      Filter the sample as soon as practical after collection using the  first 50  to  100  ml  to  rinse the
      filter flask.  (Glass or plastic filtering apparatus  are recommended to avoid  possible
      contamination.)  Discard the portion used to rinse the  flask and collect  the  required  volume  of
      filtrate.  Acidify the filtrate with 1:1 redistilled  HN03  to a  pH of 2.   Normally,  3 ml  of
      (1:1) acid per liter should be sufficient to preserve the  samples.
oo
01          "For the determination of total metals the sample  is  not filtered  before  processing.
      Because vigorous digestion procedures  may result  in  a loss of certain metals  through  precipitation,
      a less vigorous treatment is recommended as given on  page  83 (4.1.4) of "Methods  for  Chemical
      Analysis of Water and Wastes"  (1974).  In those instances  where a more  vigorous  digestion  is
      desired, the procedure on page 82 (4.1.3) should be  followed.   For the  measurement  of  the  noble
      metal series (gold, iridium, osmium, palladium, platinum,  rhodium, and  ruthenium),  an  aqua regia
      digestion is to be substituted as follows: Transfer  a representative aliquot  of  the well-mixed
      sample to a Griffin beaker and add 3 ml of concentrated redistilled  HN03.   Place  the  beaker on
      a steam bath and evaporate to dryness.  Cool the beaker and  cautiously  add  a  5 ml  portion  of  aqua
      regia.  (Aqua regia is prepared immediately before use  by  carefully  adding  3 volumes  of
      concentrated HC1 to one volume of concentrated HN03).   Cover the beaker with  a watch  glass and
      return to the steam bath.  Continue heating the covered beaker  for 50 min.  Remove cover and
      evaporate to dryness.  Cool and take up the residue  in  a small  quantity of  1:1 HC1.   Wash  down the
      beaker walls and watch glass with distilled water and filter the sample to  remove  silicates and
      other insoluble material that could clog the atomizer.   Adjust  the volume to  some  predetermined
      value based on the expected metal concentration.  The sample is now  ready for  analysis.

      (continued)

-------
                           TABLE C-l.  WATER QUALITY MEASUREMENT METHODS (Continued)
         the various furnace devices (flameless AA) are essentially atomic absorption techniques,
they are considered to be approved test methods.   Methods of standard addition are to be  followed as
noted in page 78 of "Methods for Chemical Analysis of Water and Wastes," 1974.
          automated hydrazine reduction method is available from the Environmental Monitoring
and Support Laboratory, U.S. Environmental Protection Agency,  Cincinnati, Ohio  45268.

     ^Goerlitz, D. , Brown, E., "Methods for Analysis of Organic Substances in Water":
U.S. Geological Survey Techniques of Water-Resources Inv. ,  book 5, ch .  A3 (1972).

     l^R.p. Addison and R.G. Ackman, "Direct Determination of Elemental Phosphorus by
Gas-Liquid Chromatography, " "Journal of Chromatography, " vol.  47, No. 3, pp. 421-426, 1970.

-------
                                TABLE C-2.   DRINKING WATER MEASUREMENT METHODS
         Parameter
                                   Method
Reference
00
Organics
  (a) Chlorinated Hydrocarbons:
      Endrin
      Lindane
      Methoxychlor
      Toxaphene
  (b) Cholorophenoxys:
      2,4-Dichlorophenoxyacetic
        acid
      2,4,5-Trichloro-
        phenoxypropionic acid

Radiation
         Inorganic Chemicals
         Physical Measurements
         Microbiological Measurements
                                            Gas chromatography with
                                            electron capture detector
Methods for Organochlorine
Pesticides in Industrial
Effluents, MDQARL, EPA,
Cincinnati, Ohio, 1973
                                                                        Methods  for Chlorinated Phenoxy
                                                                        Acid Herbicides in Industrial
                                                                        Effluents,  MDQARL, EPA,
                                                                        Cincinnati, Ohio,  1973
Code of Federal
40(Parts 100 to

Code of Federal
40(Parts 100 to

Code of Federal
40(Parts 100 to

Code of Federal
40(Parts 100 to
Regulations,
399): 169-197

Regulations,
399): 169-197

Regulations,
399): 169-197

Regulations,
399): 169-197

-------
                                     TABLE  C-3.  AMBIENT AIR MEASUREMENT METHODS
         Pollutant
   Measurement method
      or principle
       Reference
oo
00
      Suspended  Particulates
      Sulfur  Dioxide
      Carbon  Monoxide
      Photochemical  Oxidants
      Nitrogen  Dioxide
High volume sampler
Tape sampler

Pararosaniline or equivalent
Nondispersive infrared
or equivalent

Gas phase chemiluminescence
or equivalent

Gas phase chemiluminescence
or equivalent
CFR 40, Part 50, Appendix B,
July 1, 1979

CFR 40, Part 50, Appendix A,
July 1, 1979

CFR 40, Part 50, Appendix C,
July 1, 1979

CFR 40, Part 50, Appendix D,
July 1, 1979

CFR 40, Part 50, Appendix F,
July 1, 1979

-------
                                      TABLE C-4.  SOURCE AIR METHODS
      Determination
                                   Description of method
                            Reference
00
      Sample and Velocity Traverses
      for Stationary Sources
Stack Gas Velocity

Dry Molecular Weight of Gas

Stack Gas Moisture

Particulate Emissions

Sulfur Dioxide



Nitrogen Oxide
      Sulfuric Acid Mist and
      Sulfur Dioxide
      Visible Emissions

      Carbon Monoxide

      Hydrogen Sulfide'


      (continued)
       Pitot

       Orsat

Volumetric & Gravimetric

    Gravimetric

Collection by impinger,
analysis by barium
perchlorate titration

Collection by evacuated
flask, colorimetric
analysis

Collection by impinger,
analysis by barium
perchlorate titration

Certified observer

Non-dispersive infrared

Collection by impinger,
iodimetric titration
EPA Method 1 Environmental
Protection Agency Performance
Test Methods, page 1-1,
EPA-340/1-78-011

EPA Method 2   1-13

EPA Method 3   1-45

EPA Method 4   1-61

EPA Method 5   1-79

EPA Method 6   1-119



EPA Method 7   1-135



EPA Method 8   1-157



EPA Method 9

EPA Method 10

EPA Method 11

-------
                            TABLE C-4.   SOURCE AIR METHODS (Continued)
Determination
Description of method
Reference
Fluoride
Sulfur Compounds
Sulfur Compounds
Particulate Matter
Collection by impinger,     EPA Method 13A (Colorimetric)
colorimetric, or specific   or 13B (Specific Ion Electrode)
ion electrode
Gas chromatographic
determination of sulfur
gases emitted by a
Claus Sulfur Recovery Unit

Gas chromatographic
determination of reduced
sulfur compounds emitted
by paper mills

In-stack filter
determination of parti-
culate matter
EPA Method 15
EPA Method 16
EPA Method 17

-------
             TABLE C-5.  PRIORITY POLLUTANT MEASUREMENT METHODS
     Recommended analytical methods for priority pollutants are described
in "Sampling and Analysis Procedures for Screening of Industrial Effluents
for Priority Pollutants".

     These guidelines for sampling and analysis of industrial wastes have
been prepared by the staff of the Environmental Monitoring and Support
Laboratory - Cincinnati, at the request of the Effluent Guidelines
Division, Office of Water and Hazardous Wastes, and with the cooperation
of the Environmental Research Laboratory, Athens, Georgia.  The procedures
represent the current state of the art, but improvements are anticipated
as more experience with a wide variety of industrial wastes is obtained.
Users of these methods are encouraged to identify problems encountered and
to assist in updating the test procedures by contacting the Environmental
Monitoring and Support Laboratory, EPA, Cincinnati, Ohio 45268.  These
methods were first made available in March 1977 and were revised in April
1977.
                                     91

-------
                                TABLE C-6.   RADIATION MEASUREMENT METHODS
Parameter and units
Method
Sample matrix
Reference
Alpha - total pCi per liter




Beta - total pCi per liter


Radium-226 - pCi per liter


Strontium 89, 90 - pCi
  per liter

Tritium - pCi per liter

Cesium 134 - pCi per liter



Uranium - pCi per liter
Others (various units
depending on media)
Proportional or    Water
scintillation
counter
Proportional
counter

Scintillation
counter
Fluorometric
    Water


    Water


    Water


    Water

    Water



    Water


    Various
Interim Radiochemical Methodology
for Drinking Water EPA-600/4-75-008
Standard Methods  for the Examination
of Water and Wastewater, 14th Ed.

(same as above)
(same as above)
                                  (same as above)
(same as above)

(same as above)
ASTM D-2459 Gamma Spectroscopy  in
Water, 1975

ASTM D-2907 Micro Quantities of
Uranium in Water by Fluorimetry, 1975

HASL Procedure Manual, HASL  300,
ERDA Health and Safety Laboratory,
New York, NY, 1973

-------
                                 APPENDIX D

                    QUALITY ASSURANCE PROGRAM CHECKLIST
Contractors and grantees must follow accepted quality assurance  procedures
to establish the quality of sampling and analytical data which are  used  in
lERL-Ci programs.  Each organization should have written QA procedures which
describe the routine steps taken to guarantee,  to the extent possible, the
quality of all sampling and analytical data reported by the laboratory.   The
purpose of this checklist is to provide guidance to the Project  Officer  in
reviewing the QA program of a contractor or grantee in order to  determine if
it is in general conformance with lERL-Ci requirements.  The questions in
this checklist address each major area of quality assurance that  should  be
encompassed by a comprehensive or "model" QA program.

This checklist applies to a contractor's general Quality Assurance  manual
which encompasses all aspects of work performed by the contractor.   It
should be used to check the contractor's overall QA program to establish
that all QA aspects are discussed.  This checklist should be completed
before project work begins.

This checklist differs from Appendix E in that  it covers the contractor's
overall QA program and philosophy.  Appendix E  is the checklist  to  be used
in determining whether any weakness exists in the Quality Assurance  as
applied to any specific work plan the contractor may submit.

This checklist should be completed by the Project Officer.
                                      93

-------
                           APPENDIX D (Continued)

                    QUALITY ASSURANCE PROGRAM CHECKLIST

                            GENERAL INFORMATION
Laboratory:
Street Address:
Mailing Address (If Different):
City  	   State  	  Zip

Laboratory Telephone No.:  Area Code 	  No.  	

Laboratory Director:  	

Project Manager:  	
Contract/Grant Number:

Contract/Grant Title:

Project Officer:  	
Review Conducted By:

Agency and Address:
Telephone No.:  Area Code 	  No.
                                   Reviewer Signature               Date


QA manual returned to contractor/grantee 	.
                                            Date
                                     94

-------
                         APPENDIX D (Continued)
QA POLICY AND OBJECTIVES
Item
Does the organization maintain written QA
procedures which describe the routine steps
taken to assure the quality of sampling and
analytical data?
Is there a clear statement of quality
objectives by top management levels?
Does the organization have a stated QA policy
consistent with the program requirements of
lERL-Ci?
Are QA plans required for all major projects or
programs requiring extensive sampling and/or
analysis?
Does the appearance, format, and content of the
QA program manual(s) reflect a conscientious
concern for the quality of data produced by
the organization?
Yes





No





Comment





QA ORGANIZATION
Item
Are QA responsibilities and reporting relation-
ships within the organization clearly defined?
Does the QA program provide for a QA Supervisor
and are his responsibilities and authority
defined?
Does the QA Supervisor have access to senior
levels of management?
Are the QA-related responsibilities of
laboratory and field sampling personnel
clearly defined?
Yes




No




Comment




                                   95

-------
                         APPENDIX D (Continued)
PERSONNEL TRAINING
Item
Has the designated QA Supervisor had formal
training in QA procedures (e.g., the EPA train-
ing course offered by EPA-Ci or EPA-Athens)?
Is there a training program for sampling
personnel?
Is there a training program for laboratory
personnel?
Does the training program go beyond "on the
job training" to include practice tasks,
seminars, instruction sessions, etc.?
Yes




No




Comment




DOCUMENT CONTROL AND REVISION
Item
Are there established procedures for documenting
sampling methods and for updating these methods
when required?
Are there established procedures for documenting
analytical methods and for updating these
methods when required?
Is there a system for documenting instrument
calibration procedures?
Does the QA program require documentation of
all computer programs that are used to calculate
or process data?
Are "standard" analytical methods used whenever
available and appropriate?
Yes





No





Comment





                                   96

-------
                         APPENDIX D (Continued)





                        FACILITIES AND EQUIPMENT




PROCUREMENT AND INVENTORY PROCEDURES
Item
Does the organization have procedures for
procurement quality control for reagents,
glassware, etc. ?
Are analytical reagents dated upon receipt?
Are reagent inventories maintained on a
first-in, first-out basis?
Are analytical reagents checked out before use?
Is new analytical instrumentation tested
before use?
Yes





No





Comment





PREVENTIVE MAINTENANCE
Item
Does the laboratory have a preventive
maintenance program or schedule for
laboratory instrumentation?
Does the QA program require the maintenance of
instrument log books and specify their general
content?
Are preventive maintenance activities
documented in instrument log books?
Yes



No



Comment



                                   97

-------
                         APPENDIX D (Continued)



                         ANALYTICAL METHODOLOGY



CALIBRATION AND OPERATION PROCEDURES
Item
Are general instrument calibration requirements
described in the QA program?
Does the QA program specify that calibration
standards should be traceable to primary (e.g.,
NBS) standards that are available?
Are instrument calibration data recorded in
instrument log books?
Are acceptability requirements established
for instrumentation?
Have instrument calibration schedules been
established?
Yes





No





Comment





FEEDBACK AND CORRECTIVE ACTION
Item
Does the QA program clearly state who is
responsible for taking corrective action
when analytical problems are encountered?
Are corrective action follow-up procedures
described?
Are corrective actions documented?
Yes



No



Comment



                                   98

-------
                         APPENDIX D (Continued)




                SAMPLING AND SAMPLE HANDLING PROCEDURES
CONFIGURATION CONTROL
Item
Are procedures described for configuration con-
trol of monitoring systems such as air monitors,
water quality monitors, flow monitors?
If an air or water pollution monitor is relo-
cated, would this be recorded in project files?
Yes


No


Comment


SYSTEMS RELIABILITY
Item
Are procedures described for maintaining the
reliability of data generating equipment and
instrumentation?
Do QA procedures require the documentation of
system reliability (e.g., percent uptime,
percent reliable data, etc.)?
Are reliability data used in revising
maintenance and/or instrument replacement
schedules?
Yes



No



Comment



                                   99

-------
                         APPENDIX D (Continued)




                            QUALITY CONTROL




QUALITY CONTROL PROCEDURES
Item
Is duplicate analysis required on a minimum
of 5 to 15 percent of all samples?
Have acceptance criteria been established for
the precision of duplicate analysis results?
Are spiked sample analyses required on a
routine basis (e.g., 5 percent of samples) to
determine analytical recoveries?
have acceptance criteria been established for
spiked sample results?
Are quality control charts used to monitor
analyst performance on routine analyses?
Are reagent blank analyses run with each
set of samples?
Are split sample analyses used as part of
the quality control program?
Are a minimum of three and preferably more
standards required for standard curves?
Do routine procedures require that standard
curves bracket sample concentrations?
Yes









No









Comment









                                   100

-------
                         APPENDIX D (Continued)




CONTROL CHECKS AND INTERNAL AUDITS
Item
Does the QA Supervisor or other supervisory
personnel audit laboratory procedures on a
routine, periodic basis?
If keypunching is required, have procedures
been established to identify keypunch errors?
Does the laboratory routinely use available EPA
reference test samples and/or standard reference
materials available from other sources (e.g.,
NBS) to evaluate analytical performance?
Are reference test samples used to evaluate
analytical performance at least semi -annually?
Are the results of intralaboratory performance
tests maintained by the QA Supervisor and used
to improve analytical performance?
Are reference test samples routinely prepared
and submitted to the analyst as "blind"
samples?
Does the laboratory participate in inter-
laboratory performance tests?
Are field sampling procedures audited at
periodic intervals by the QA Supervisor or
other supervisory personnel?
Yes








No








Comment








                                   101

-------
                         APPENDIX D (Continued)




                             DATA HANDLING




DATA HANDLING, REPORTING, AND RECORDKEEPING
Item
Are procedures described for the proper
handling, routing, and reviewing of field
and laboratory data?
Is each sample assigned a unique identification
number upon collection?
Are laboratory notebooks used to record vital
operational information?
Are laboratory bench data reported in a
clear, logical format which permits review
and confirmation of calculations?
Are procedures described for the accurate and
complete labeling of strip charts?
Are field notebooks required for all field
data and observations?
Is the general content of field notebooks
specified?
Are field notebook entries reviewed by
supervisory personnel?
Are field notebooks signed by the individual
recording the information and by reviewers?
Are sample chain-of-custody procedures
described?
Are QA reports routinely prepared for
management review?
Are all data and records retained a minimum
of three years?
Yes












No












Comment












                                   102

-------
                         APPENDIX D (Continued)
DATA VALIDATION
Item
Does the QA program require routine cross-
checking of manual data calculations?
Are supervisory personnel required to .review
all laboratory data?
Are acceptance criteria established for
laboratory data?
Are data collected by on-line or in situ systems
routinely checked for consistency and accuracy?
Are computer programs validated before use?
Are data validation activities recorded for
future reference?
Does the QA program describe the handling
of invalid data?
Yes







No







Comment







                                   103

-------
                            APPENDIX D  (Continued)
Overall, does the QA program described in the document  reviewed meet  the QA
program requirements of lERL-Ci?  	 Yes   	 No
Describe where the QA program departs  from lERL-Ci requirements  and  explain
what corrective action, if any, is necessary.
                                      104

-------
                                 APPENDIX E

                    PROJECT QUALITY ASSURANCE CHECKLIST
This checklist has been prepared to assist Project Officers in reviewing
project work plans or study plans with respect to sampling and analytical
quality assurance.  The checklist is not comprehensive; it does not address
all factors that would affect the quality of the sampling and analytical
data on all projects.  It does, however, represent the type of detail that
should be included in a well-written Project Work Plan.  The use of this
checklist in reviewing project work plans should allow the Project Officer
to judge the adequacy of proposed quality assurance procedures.

This checklist applies to a contractor's specific work plan as applied to a
project or task.  It should be used to insure that the contractor has
prepared a work plan which includes sufficient quality assurance to complete
the project in a satisfactory manner.

Appendix E differs from Appendix D in that it applies specifically to a
project or task, whereas the checklist in Appendix D applies to the
contractor's general Quality Assurance Program.

This checklist should be completed by the Project Officer.
                                     105

-------
                           APPENDIX E (Continued)

                    PROJECT QUALITY ASSURANCE CHECKLIST

                            GENERAL INFORMATION
Laboratory:
Street Address:
Mailing Address (If Different):
City  	   State  	  Zip

Laboratory Telephone No. :   Area Code 	  No. 	

Laboratory Director:  	

Project Manager:  	
Contract/Grant Number:

Contract/Grant Title:

Project Officer:  	
Review Conducted By:

Agency and Address:
Telephone No. :   Area Code 	  No.
                                   Reviewer Signature               Date


QA manual returned to contractor/grantee 	.
                                            Date
                                     106

-------
                         APPENDIX E (Continued)
PROJECT OBJECTIVES
Item
Is a clear statement made of the objectives of
the project and the use of the sampling and
analysis data?
Has a statement been made, or can the level of
importance to be attached to the QA considera-
tions be derived from stated project objectives?
Yes


No


Comment
«i

PROJECT STAFFING
Item
Has a project QA Supervisor been assigned
to the project team?
Is the project organization structure
appropriate to accomplish the QA objectives
of the project?
Do personnel assigned to this project have
the appropriate educational background to
successfully accomplish project objectives?
If any special training or experience is
required, is it represented on the project
staff?
Will the training of personnel be required
specifically for this project? If so, is it
covered in the project plan?
Is there adequate staffing to accomplish the
planned work in a high-quality manner within
the project schedule?
Yes






No






Comment






                                  107

-------
                         APPENDIX E (Continued)





FACILITIES, EQUIPMENT, AND INSTRUMENTATION
Item
Is appropriate and adequate sampling
equipment available?
Will appropriate sample containers be used
for the parameters measured?
If in situ, on-line, or monitoring instrumenta-
tion's to be used, is it clearly specified as
to make, model, and performance specifications?
Are the performance specifications of all on-
line or in situ instrumentation adequate to
meet project reliability and data quality
requirements?
Has a plan been made to optimize system relia-
bility by requiring periodic performance checks,
calibration, and preventive maintenance?
Are procedures described for documenting and
controlling the configuration of all systems?
Is laboratory instrumentation and equipment
suitable to meet the data quality needs of
the project?
Yes







No







Comment







SAMPLING PLAN AND METHODS
Item
Are sampling procedures described in detail?
Will standard sampling methods be used where
available?
Are precautions described to avoid sample
contamination?
Yes



No



Comment



                                   108

-------
                         APPENDIX E (Continued)




SAMPLING PLAN AND METHODS (continued)
Item
Are site or system diagrams included which show
clearly and precisely the location of sample
collection?
Will any samples be taken in duplicate in
order to define sampling variability?
Are background samples to be taken?
Is sampling frequency adequate and appropriate
to the purposes of the project?
Is the sampling program assigned to assure all
samples are representative of the source?
Will field notebooks be used to record
important observations and on-site data?
Will field data or system data be verified
by supervisory personnel on a routine basis?
Do sampling plans allow for delivery of the
samples to the laboratory in time to meet
maximum holding time limitations?
Will the samples be preserved? If so, will
EPA-accepted preservation methods be used?
If any new sampling methods are to be used,
will they be adequately tested before use?
Yes










No










Comment










                                   109

-------
                         APPENDIX E (Continued)




ANALYTICAL PLAN AND METHODS
Item
Will standard analytical (EPA-approved)
procedures be used where appropriate and
available?
Does the project plan include a copy of all
non-standard analytical procedures?
If any new analytical procedures are to be used,
will they be adequately tested before use?
Will use of the analytical methods specified
result in data of adequate detection limit,
accuracy, and precision to meet the require-
ments of the project?
Will duplicate analyses be conducted on at
least 10 percent of the samples?
Will spike sample analyses be conducted on
at least 5 percent of the samples?
Will reagent blank samples be run?
Will split sample analysis be conducted?
Will any field spiked samples be processed?
Will instruments and measurement systems be
calibrated with adequate frequency (at least
daily)?
Will calibration materials that are traceable
to NBS standards be used where available?
Yes











No











Comment











                                   110

-------
                         APPENDIX E (Continued)
DATA MANAGEMENT
Item
Will data be validated before entering
into automated data systems?
Will automated data handling programs or
computer models be adequately documented
and verified before use?
Will mathematical and computer models be
verified by actual data?
Is the statistical treatment of the data
described and does it meet 'project
requirements?
Will a project QA report be prepared to
summarize all quality control data?
Yes





No





Comment





PROJECT SCHEDULE
Item
Does the project plan show adequate time to
accomplish the sampling program, and does it
allow for uncontrollable delays, such as bad
weather?
Will interim sampling and analysis program
results be reported to the Project Officer
for review and comment?
Does the project schedule allow sufficient time
between sample collection and reporting of the
data to apply adequate analytical quality
control, including supervisory review?
Yes



No



Comment



                                   111

-------
                           APPENDIX E (Continued)
Overall, does the Project Work Plan meet the QA requirements of lERL-Ci?
       Yes          No
Describe changes or improvements that should be incorporated into a revised
work plan?
                                      112

-------
                                 APPENDIX F

                   QUALITY ASSURANCE PRE-AUDIT WORKSHEET
The purpose of the Quality Assurance Pre-Audit Worksheet is  to  provide  the
auditing agency with information with which to familiarize itself with  the
project.  It should be sent to the contractor, completed by  the contractor,
and returned before any audit trip is scheduled.

This form is to be used when the decision has been made to perform  a  project
audit.

This checklist should be completed by the contractor's project  personnel.
                                INSTRUCTIONS

1.  It is not necessary to type your replies.  Completion of  the  form  in  ink
    will suffice.  Alternatively, appropriate sections of the project  plan
    may be attached where appropriate.

2.  You are not limited to yes and no answers.  Feel  free to  elaborate at
    any point in the form.

3.  Careful attention should be given to proper completion of this  form
    since the information supplied will have a direct bearing on  conclusions
    drawn and recommendations made concerning the evaluation  of your
    laboratory.

4.  Answer only those questions contained in the worksheet which  are project
    specific.  Indicate non-applicable questions.
                                     113

-------
                           APPENDIX F (Continued)

                    QUALITY ASSURANCE PRE-AUDIT WORKSHEET


                                                  Date 	

Laboratory:  	 	
Street Address:
Mailing Address (If Different):
City 	  State 	  Zip

Laboratory Telephone No. :   Area Code 	  No. 	

Laboratory Director:  	
Quality Assurance Supervisor:
Pre-Audit Worksheet Completed By (if more than one, please indicate which
sections):

                  Name                               Title
No. of Contract or Grant for Which Audit Is Being Performed:

Title of Contract:
Audit to be Conducted By:

Agency and Address:
                        -Do Not Write Below This Line-
Telephone No. :  Area Code 	  No.

Quality Assurance Audit Scheduled On:  	
                                     114

-------
                           APPENDIX F (continued)
A.  ORGANIZATION AND PERSONNEL

A.I.  Please use a simple block diagram to illustrate the structure of your
      organization and how the laboratory functions within it.  Identify key
      management personnel.
                            ORGANIZATIONAL CHART
                                     115

-------
                           APPENDIX F  (continued)
A.2.  Please use a simple block diagram to illustrate  the organization of  the
      project sampling and sample analysis functions.  Identify:  key project
      management personnel, the quality assurance supervisor,  the person
      directly responsible for sampling, and the person directly responsible
      for analyses.
                        PROJECT ORGANIZATIONAL CHART
                                      116

-------
A.3.  PROJECT PERSONNEL




Laboratory
                                           APPENDIX F (continued)
Contract No.
Date

Position















Name















Academic
Training















Years
Experience















Special Training















Responsibilit ies
On This Project
















-------
                           APPENDIX F (continued)
A.4.  Describe or illustrate the planned schedule  for completion  of  the
      project.  Show the schedule  for all sampling activites  and  the  period
      of time scheduled for sample analysis.
                              PROJECT SCHEDULE
                                      118

-------
                         APPENDIX F (continued)
B.  FACILITIES
Instrumentation:  List major instrumentation (including field instrumen-
tation and equipment) required for the performance of this project.
Prepare separate lists for radiation, chemistry, air, biology, micro-
biology, etc.
        Item
Manufacturer
Model
Age
Condition
                                   119

-------
                                                 APPENDIX F  (continued)
     0.   ANALYTICAL METHODOLOGY

     List all  laboratory methods  used  in  the  performance  of  this  project.
     biology,  radiation, air,  microbiology,  etc.
Use separate pages for chemistry,
     Scientific  Area
fO
O

Parameter













Name or
Description of Method













Reference
(Cite Page and Year)













Latest Reference Sample Check
(Cite Agency of Origin and Date)














-------
                           APPENDIX F (continued)
D.  SAMPLING AND SAMPLE HANDLING PROCEDURES

D.I.  Describe the sampling plan and draw a diagram indicating sampling
      locations, how to locate the sampling points and other pertinent
      information.  Please indicate the scientific area (biology, radiation,
      chemistry, air, etc.) and use separate sheets if addressing more than
      one field.
Scientific Area
                                     121

-------
                           APPENDIX F (continued)


D.2.  Describe the information obtained in field notebooks or worksheets,
D.3.  Describe sample chain-of-custody procedures employed by your
      laboratory.
                                     122

-------
                                            APPENDIX F (continued)
D.4.   List sample types collected,  methods of preservation, etc.   Please use separate pages for chemistry,
      biology, radiation, air, microbiology,  etc.
                                 SAMPLE COLLECTING,  HANDLING, AND PRESERVATION
Scientific Area

Parameter













Container Used













Preservative Used













Normal Maximum
Holding Time













Method of Transport














-------
                           APPENDIX F (continued)









E.  QUALITY CONTROL




E.I.  Are analytical quality control records available for review?
E.2.  Are duplicate sample analyses conducted?




      If so, at what frequency?  	
E.3.  Are spiked (recovery) sample analyses conducted?




      If so, at what frequency?  	
E.4.  Are performance test samples analyzed?




      If so, how often?                 	
E.5.  Are quality control charts maintained?
E.6.  Does the laboratory have a Quality Assurance Manual?




      If so, is it available for review?  	
E.7.  Does each analyst have access to approved and documented analytical




      procedures?  	




E.8.  Has a quality assurance plan been prepared for the sampling required by




      this project?	




E.9.  How are quality control data used?  	
                                     124

-------
                           APPENDIX F (continued)









F.  DATA HANDLING




F.I.  How long are laboratory records and field data retained after




      completion of a project?  	
F.2.  Have procedures been established for cross-checking laboratory




      calculations?  	




F.3.  Have procedures been established for cross-checking reported data?
F.4.  Are laboratory data reviewed by supervisory personnel?
                                      125

-------
                                 APPENDIX G

                  INSTRUMENTATION, EQUIPMENT, AND PERSONNEL
                      SKILL RATING FOR SPECIFIC METHODS
     The arbitrary rating numbers used in  this appendix  for  the degree  of
skills required are:

     Rating 1—A semi-skilled sub-professional with  limited  background

     Rating 2—An experienced aide  (sub-professional) with a background  in
          general laboratory techniques and some knowledge of chemistry, or  a
          professional with modest  training or experience

     Rating 3—Requires a good background  in analytical  techniques

     Rating 4—Requires an individual with experience on complex
          instrumentation, some degree of  specialization, and the ability to
          interpret results

     Specific analytical methods are presented as  follows.

                                                                     Page

WATER

     GENERAL ANALYTICAL METHODS	127
     TRACE METALS	130
     OTHER REFERENCE METHODS FOR METALS	131
     NUTRIENTS, ANIONS, AND ORGANICS	133
     OTHER PARAMETERS	139

AMBIENT AIR	139

SOURCE AIR	141

BIOLOGY	142

RADIATION	145
                                     126

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER

GENERAL ANALYTICAL METHODS

1.   Alkalinity

    (a)  Electrometric Titration, Manual:  pH meter, Type I or II
         as defined in ASTM D1293 	 1

    (b)  Electrometric Titration, Automated:  An automatic
         titrimeter meeting the pH meter specifications in (a) 	 1

    (c)  Automated, Methyl Orange:  AutoAnalyzer with appropriate
         analytical manifold and 550-nm filters	 2, 3

2.   Biochemical Oxygen Demand (BOD), 5-day, 20°C

    (a)  Modified Winkler with Full-Bottle:  BOD incubation bottles;
         BOD incubator	 2 , 3

    (b)  Probe Method:  No specific probe is recommended as superior
         in the 1974 EPA Methods Manual	 2, 3

3.   Chemical Oxygen Demand (COD)

    Reflux Apparatus	 2 , 3

4.   Residue, Total

    Gravimetric, dried at 103-105°C:  Blender (if samples contain oil
    or grease); Porcelain, vycor, or platinum evaporating dishes;
    Muffle furnace, 550°C; Steam bath or 98°C oven; Drying oven,
    103-105°C; Dessicators; Analytical balance, 200-g capacity,
    weighting to 0.1 mg	 1, 2

5.   Residue, Total Filterable

    Glass Fiber Filtration, dried at 180°C:  Glass fiber filter discs
    (0.45-um glass fiber filter); Filter holder,  membrane filter
    funnel, or Gooch crucibles and adapter; Suction flask; Porcelain,
    vycor, or platinum evaporating dishes; Muffle -furnace, 550°C;
    Steam bath; Drying oven, 180°C; Dessicators;  Analytical balance,
    200-g capacity, weighting to 0.1 mg	 1, 2
                                     127

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No,
WATER (Continued)

6.  Residue, Total Non-Filterable

    Glass Fiber Filtration, Dried at 103-105°C:  Same as (5), except
    drying oven is at 103-105°C and steam bath, muffle furnace, and
    evaporating dishes are not required	 1, 2

7.  Residue, Total Volatile

    Gravimetric, Dried at 550°C:  Same as (5)	 1, 2

8.  Ammonia (as N)

    (a)  Distillation and Titration:  All glass distillation
         apparatus (Kjeldahl); Standard titration apparatus	 2, 3

    (b)  Distillation and Nesslerization:  All-glass distillation
         apparatus (Kjeldahl); Nessler tubes, 50 ml, matched set,
         APHA standard; Spectrophotometer or filter photometer for
         use at 425 nm with light path _>. 1 cm	 2, 3

    (c)  Distillation and Ammonia Electrode:  All-glass distillation
         apparatus (Kjeldahl); Electrometer (pH meter) with expanded
         mV scale or specific ion meter; Ammonia-selective electrode;
         Magnetic stirrer, thermally-insulated, and Teflon-coated
         stirring bar	 2, 3

    (d)  Automated Colorimetric Phenate Method:  AutoAnalyzer with
         appropriate analytical manifold and 630-660 nm filter	 2, 3

9.  Total Kjeldahl Nitrogen (as N)

    (a)  Digestion, Distillation, and Titration:  Same as 8(a) with
         suction takeoff to remove SO^ fumes during digestion	 2, 3

    (b)  Digestion, Distillation, and Nesslerization:  Same as 8(b)
         with suction takeoff to remove SO^ during digestion	 2, 3

    (c)  Digestion, Distillation, and Ammonia Electrode:   Same as
         8(c) with suction takeoff to remove 803 fumes during
         digestion	 2, 3

    (d)  Automated Phenate Method:  AutoAnalyzer with appropriate
         analytical manifold and 630-nm filter	 2, 3
                                     128

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

     (e)  Automated Selenium Method:  AutoAnalyzer with appropriate
          analytical manifold and 630- or 650-nm filters	 2, 3

10.  Nitrate (as N)

     (a)  Cadmium Reduction Method  (Nitrate-Nitrite):  Glass fiber
          or membrane filters and associated apparatus; Copper/
          cadmium reduction column; Spectrophotometer or filter
          photometer for use at 540 nm with light path > 1 cm	 2, 3

     (b)  Automated Cadmium Reduction Method (Nitrate-Nitrite):
          Glass fiber or membrane filters and associated apparatus;
          Copper/cadmium reduction column; AutoAnalyzer with
          appropriate analytical manifold and 540-nm filter	 2, 3

     (c)  Brucine Method:  Spectrophotometer or filter photometer
          for use at 410 nm; Water bath at 100°C (Temperature
          control is critical:  all sample tubes must be held at the
          same temperature, and temperature must not drop
          significantly when tubes are immersed in bath); Water bath
          at 10-15°C; Neoprene-coated wire rack for holding sample
          tubes in baths; Glass sample tubes (40-50 ml)	 2, 3

11.  Phosphorus, Total as Ortho (as P)

     (a)  Single Reagent (Ascorbic Acid Reduction Method):
          Spectrophotometer or filter photometer for use at 650 nm
          (less sensitive) or 880 nm; Acid-washed,  detergent-free
          glassware; Hotplate or autoclave (for persulfate
          digestion)	 2, 3

     (b)  Automated Colorimetric Ascorbic Acid Reduction Method:
          Acid-washed, detergent-free glassware; Hotplate or
          autoclave (for persulfate digestion); AutoAnalyzer with
          appropriate analytical manifold and 650-660 nm or 880-nm
          filter	 2, 3

12.  Acidity

     (a)  Hydrogen Peroxide Digestion and Electrometric Titration:
          pH meter, Type I or II as defined in ASTM D1293	 1, 2

     (b)  Hydrogen Peroxide Digestion and Phenolphthalein End-Point
          Titration:  No special equipment other than standard
          laboratory glassware	 1, 2
                                     129

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No,
WATER (Continued)

13.  Organic Carbon, Total (TOC)

     Combustion and infrared method (CC^) or  flame ionization
     method (CH^); Blender; Apparatus for total and dissolved
     organic carbon	. ... 2, 3

14.  Hardness, Total

     (a)  EDTA Titration:  No special equipment other than standard
          laboratory glassware	 1

     (b)  Automated Colorimeter:  AutoAnalyzer with appropriate
          analytical manifold and 520-nm filter	 2, 3

15.  Nitrate (as N)

     (a)  Manual Colorimeter Diazotization:   Spectrophotoraeter for
          use at 540 nm with cells > 1 cm; Nessler tubes or
          volumetric flasks, 50 ml	 2

     (b)  Automated Colorimetric Diazotization:  Glass  fiber or
          membrane filters and associated apparatus; AutoAnalyzer
          with appropriate analytical manifold and 540-nm filter	 2, 3

TRACE METALS

EPA specifies atomic absorption as at least one of the  reference
methods for many metals.  The required equipment in each case will
include:  (1) an atomic absorption spectrophotometer, (2) the hollow
cathode (or electrode-less discharge) lamp for each metal, and (3) the
fuels and other apparatus specified below.  Design features of some
common atomic absorption spectrophotometers (as of March 1979) are
discussed in the EPA Handbook for Analytical  Quality Control in
Water and Wastewater Laboratories.  If extraction procedures are  to
be used, special reagents are required, but no special  equipment
other than standard laboratory glassware.

16.  Metal by Atomic Absorption	 2, 3
                                     130

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No,
WATER (Continued)

OTHER REFERENCE METHODS FOR METALS

17.  Aluminum

     Eriochrome Cyanine R Colorimetric Method:  Spectrophotometer
     for use at 535 nm, or Filter photometer with 525-535 nm
     filters (green), or Nessler tubes, 50 ml	 2, 3

18.  Arsenic

     Gaseous Hydride—Silver Diethyldithiocarbamate Colorimetric
     Method:  Arsine generator and absorption tube; Spectrophoto-
     meter for use at 535 nm, or Filter photometer with 530-540 nm
     filter (green)	 2, 3

19.  Beryllium

     Aluminon Method:  Spectrophotometer or filter photometer for
     use at 515 nm with 5-cm cells	 2, 3

20.  Boron

     Curcumin Method:  Spectrophotometer or filter photometer for
     use at 540 nm with cells > 1 cm; Vycor or platinum evaporating
     dishes, 100-150 ml; Water bath, 55 _+ 2°C; Ion exchange column,
     50 cm x 1.3 cm (diameter)	 2, 3

21.  Cadmium

     Dithizone Colorimetric Method:   Spectrophotometer or filter
     photometer for use at 515 nm	 2, 3

22.  Calcium

     EDTA Titration:  No special equipment	 1

23.  Chromium VI

     Diphenylcarbazide Colorimetric:  Membrane or sintered glass
     filter; Spectrophotometer or filter photometer for use at
     540 nm with cells > 1 cm	 2, 3
                                     131

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

24.  Chromium, Total

     Oxidation and Diphenylcarbazide Colorimetric:  Membrane
     or sintered glass filter; Spectrophotometer or filter
     photometer for use at 540 nm with cells _>. 1 cra	 2, 3

25.  Copper

     Neocuproine Colorimetric:  Spectrophotometer for use at
     457 nm with cells ^ 1 cm, or filter photometer with narrow-band
     violet filter (max.  transmittance at 450-460 nm) and cells
     2^ 1 cm, or Nessler tubes, 50 ml	 2, 3

26.  Iron

     0-Phenanthroline Colorimetric:   Spectrophotometer or filter
     photometer for use at 510 nm with cells > 1 cm,  or Nessler
     tubes, 100ml	7	 2, 3

27.  Lead

     Dithizone Colorimetric:  Spectrophotometer or filter photometer
     for use at 520 nm with cells > 1 cm; pH meter	 2, 3

28.  Magnesium

     Gravimetric:  No special equipment	 2

29.  Mercury

     Manual Cold Vapor Technique (Water or Sediment):  Commercially
     available mercury analyzer employing this technique, or atomic
     absorption Spectrophotometer with open sample presentation area
     for mounting 10-cm absorption cell; Mercury hollow cathode lamp;
     Recorder, multi-range, variable speed, compatible with UV
     detection system; Absorption cell, 10 cm, quartz end windows,
     vapor inlet and outlet ports; Air pump, peristaltic, 1 liter/
     minimum; Flowmeter;  Aeration tubing and drying tube (or
     incandescent lamp to warm cell); Autoclave (optional,  for
     digestion procedure)	 2, 3

30.  Nickel

     Heptoxime Colorimetric Method:   Spectrophotometer or filter
     photometer for use at 445 nm with cells 2. 1 cm	 2, 3
                                     132

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

31.  Potassium

     (a)  Colorimetric:  Spectrophotometer for use at 425 nm with
          cells _> 1 era, or Filter photometer with violet filter
          (maximum transmittance near 425 mm) and > 1 cm cells,
          or Nessler tubes, 100 ml; Centrifuge and 25-ml centrifuge
          "tubes	2, 3

     (b)  Flame Photometric:  Flame photometer, direct-reading or
          internal-standard, and associated equipment for measure-
          ment at 768 nm	 2, 3

32.  Sodium

     Flame Photometric:  Flame photometer, direct-reading or
     internal-standard, and associated equipment for measurement
     at 589 nm; For low-solids water, air filter and blower for
     burner housing, oxyhydrogen flame, and polyethylene or
     Teflon cups, bottles, etc	 2, 3

33.  Vanadium

     Colorimetric (Catalysis of Gallic Acid Oxidation):  Spectro-
     photometer or filter photometer for use at 415 nm with 1-5 cm
     cells; Water bath, 25^0.5°C	 2, 3

34.  Zinc

     Dithizone Colorimetric Method:  Spectrophotoraeter or filter
     photometer for use at 535 or 620 nm with 2-cm cells, or
     Nessler tubes, matched; pH meter	 2, 3

NUTRIENTS, ANIONS, AND ORGANICS

35.  Organic Nitrogen (as N)

     Kjeldahl Nitrogen Minus Ammonia Nitrogen:  See (8) and (9)
     above	 2 , 3

36.  Sulfate (as 804)

     (a)  Gravimetric:  Analytical balance, weighing to 0.1 mg;
          Steam bath; Drying oven, 180°C; Muffle furnace, 800°C;
          Appropriate filters or crucibles	 2
                                     133

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

     (b)  Turbidimetric:  Nepheloraeter, or Spectrophotometer or
          filter photometer for use at 420 nm with 4-5 cm cells;
          Magnetic stirrer with timer or stopwatch	 2

     (c)  Automated Colorimetric Barium Chloroanilate:  Auto-
          Analyzer with appropriate analytical manifold and
          520-nm filter; Magnetic stirrer	 2, 3

37.  Sulfide (as S)

     Titrimetric Iodine:  No special equipment other than standard
     laboratory glassware	 2

38.  Sulfite (as S03)

     Titrimetric lodide-Iodate:  No special equipment other than
     standard laboratory glassware	 2

39.  Bromide

     Titrimetric lodide-Iodate:  No special equipment other than
     standard laboratory glassware	 2

40.  Chloride

     (a)  Silver Nitrate:  No special equipment other than standard
          laboratory glassware	 1

     (b)  Mercuric Nitrate:  No special equipment other than
          standard laboratory glassware	 1

     (c)  Automated Colorimetric Ferricyanide:  AutoAnalyzer with
          appropriate analytical manifold and 480-nm filter	 2, 3

41.  Cyanide, Total

     (a)  Distillation and Silver Nitrate Titration:  Cyanide
          distillation apparatus; Koch microburet, 5 ml	 2, 3

     (b)  Distillation and Pyridine-Pyrazolone (or Pyridine-
          Barbituric Acid) Colorimetric:   Cyanide distillation
          apparatus; Spectrophotometer or filter photometer for
          use at 578 or 620 nm with > 1-cra cells	 2, 3
                                     134

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

42.  Fluoride

     (a)  Distillation—SPADNS:  Bellack distillation apparatus;
          Spectrophotometer for use at 570 nra with > 1-cra cells,
          or Filter photometer with green-yellow filter (max.
          transmittance 550-580 nm) and > 1-cm cells	 2, 3

     (b)  Automated Complexone Method:  AutoAnalyzer with
          appropriate analytical manifold and 650-nm filter	 2, 3

     (c)  Fluoride Electrode:   Electrometer; Fluoride ion activity
          electrode; Reference electrode, single junction, sleeve-
          type; Magnetic mixer	 2

43.  Chlorine, Total Residual

     (a)  Starch-Iodide Titration:  No special equipment other than
          standard laboratory glassware	 2

     (b)  Amperometric Titration:  Amperoraetric end-point detection
          apparatus consisting of noble metal electrode, salt bridge,
          and silver-silver chloride reference electrode cell unit
          connected to microammeter with appropriate electrical
          accessories; Agitator	 2

44.  Oil and Grease

     (a)  Gravimetric:  Separatory funnels or soxhlet apparatus;
          Vacuum	 2 , 3

     (b)  Infrared:  Separatory funnels; Infrared Spectrophotometer,
          double beam, with 1-, 5-, and 10-cra cells	 2, 3

45.  Phenols

     (a)  Colorimetric (4-AAP Method with Distillation):  Phenols
          distillation apparatus; Spectrophotometer or  filter
          photometer for use at 460 nm (following chloroform
          extraction) or 510 nm and 1- to 10-cm cells;  pH meter	 2, 3

     (b)  Automated 4-AAP Method:  AutoAnalyzer with appropriate
          analytical manifold and 505- or 520-nm filter	 2, 3
                                     135

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

46.  Surfactants

     Methylene Blue Colorimetric:  Spectrophotometer or  filter
     photometer for use at 625 nm with > 1-cm cells	 2

47.  Algicides

     Gas Chromatography:  There is no reference procedure  for
     algicides as a class, and, therefore, detailed equipment
     requirements cannot be specified.  For general discussions of
     gas chromatography and its application in environmental
     monitoring, see the EPA Training Manual for Pesticide Residue
     Analysis in Water and the EPA Methods Manual for Analysis of
     Pesticide Residues in Human and Environmental Samples	 3, 4

48.  Benzidine

     Diazotization and Colorimetric:  Spectrophotometer, scanning,
     510-370 nm; Cells, 1- to 5-cm pathlength, 20-ml max. volume	   3

49.  Chlorinated Organic Compounds (Except Pesticides)

     Gas Chromatography:  There is no reference procedure  for
     chlorinated organic compounds as a class, and, therefore,
     detailed equipment requirements cannot be specified.  Gas
     chromatography with electron capture, microcoulometry, or
     electrolytic conductivity detection may be appropriate for
     individual compounds or groups of compounds.  For general
     discussions of gas chromatography and its application in
     environmental monitoring, see the EPA Training Manual for
     Pesticide Residue Analysis in Water and the EPA Methods Manual
     for Analysis of Pesticide Residues in Human Environmental
     Samples	 3 , 4

50.  Pesticides

     There is no single reference procedure for pesticides as a
     class.  However, specific reference procedures for  several
     sub-classes are available from EMSL, U.S. EPA, Cincinnati,
     Ohio.  To be qualified in this parameter, the  laboratory should
     be equipped to analyze for all specified sub-classes.  The
     analysis of pesticides at the levels normally  found in waste-
     water and other environmental sources requires special expertise
     and experience, in addition to up-to-date, well-maintained,
                                     136

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No,
WATER (Continued)

     calibrated instrumentation and apparatus.  The equipment
     lists below are based on the EMSL methods; for further
     information on the equipment and methodology of pesticide
     analysis, see the EPA Training Manual for Pesticide Residue
     Analysis in Water and the EPA Methods Manual for Analysis of
     Pesticide Residues in Human and Environmental Samples

     (a)  Organochlorine Pesticides:  Gas chromatograph with
          (1) Glass-lined injection port, (2) One or more of the
          following detectors:  electron capture, radioactive
          (H3 or Ni^3), microcoulometric titration;
          electrolytic conductivity, (3) Recorder, potentio-
          metric, 10" strip chart, and (4) Appropriate Pyrex
          gas chromatographic columns; Snyder columns, 3-ball
          (macro) and 2-ball  (micro), and other K-D glassware;
          Appropriate columns for liquid-solid partition
          chromatography; Blender; and Special materials, such
          as PR Grade Florisil and pesticide standards	  3, 4

     (b)  Organophosphorus Pesticides:  Gas chromatograph with
          (1) Glass-lined injection port, (2) One or more of
          the following detectors:  flame photometric, 526-nm
          phosphorus filter; electron capture, radioactive
          (H3 or Ni63), (3) Recorder, potentiometric, 10"
          strip chart, and (4) Appropriate Pyrex gas chromato-
          graphic columns; Snyder columns, 3-ball (macro) and
          2-ball (micro), and Other K-D glassware; Appropriate
          columns for liquid-solid partition chroraatography;
          Blender; Special materials, such as PR Grand Florisil
          Woelm neutral alumina, and pesticide standards	  3, 4

     (c)  Polychlorinated Biphenyls (PCBs):  Gas chromatograph
          with (1) Glass-lined injection port, (2) One or more
          of the following detectors:  electron capture,
          radioactive (H3 or Ni"3), microcoulometric
          titration, electrolytic conductivity, (3) Recorder,
          potentiometric, 10" strip chart, and (4) Appropriate
          Pyrex gas chromatographic columns;  Snyder column,
          3-ball (macro); Appropriate columns for liquid-solid
          partition chromatography; Low-pressure regulator
          (0-5 psig) with low-flow needle valve; Blender;
          Special materials, such as PR Grade Florisil,  high-
          quality silica gel, and Aroclor (PCB) standards	  3, 4
                                     137

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

     (d)  Triazine Pesticides:  Gas chromatograph with (1) Glass-
          lined injection port, (2) Electrolytic conductivity
          detector, (3) Recorder,  potentiometric, 10" strip
          chart, and (4) Appropriate Pyrex gas chromatographic
          column; Snyder columns,  3-ball (macro) and 2-ball
          (micro), and Other K-D glassware; Appropriate
          columns for liquid-solid partition chromatography;
          Blender; Special material, such as PR Grade Florisil
          and pesticide standards	 3, 4

     (e)  0-Aryl Carbamate Pesticides:   Thin layer chromatography
          plates, 200 x 200 mm, coated with Silica Gel G, 0.25 mm;
          Associated TLC apparatus, including spotting template,
          developing chamber, and  sprayer (20 ml)	 3, 4

51.  Organics

     Gas chromatography/mass spectrometry	 4

52.  Organics

     High-pressure liquid chromatography	 4

53.  Specific Conductance (mho/cm @ 25°C)

     Wheatstone bridge:  Commercial conductivity meter, or
     apparatus consisting of (1) Wheatstone bridge (reading to
     1% accuracy or better), (2) Appropriate source of electrical
     current, (3) Specific conductance cell, (4) Water bath, 25CC,
     with racks	 1

54.  Turbidity (Jackson Units)

     Turbidimeter Method:  Nephelometric turbidimeter	 1

55.  Streptococci Bacteria, Fecal  (Number/100 ml)

     (a)  Membrane Filter:  Autoclave (to 121°C); Filter membranes;
          Petri culture dishes; Incubator, 35 + 0.5°C, ca. 90%
          relative humidity; Low-power (10-15xT, binocular,
          wide-field, dissecting microscope and light source	 2, 3

     (b)  MPN:  Autoclave (to 121°C); Inoculation tubes; Incubator,
          35 + 0.5°C	 2, 3
                                     138

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
WATER (Continued)

     (c)  Plate Count:  Autoclave (to 121°C); Petri culture dishes;
          Incubator, 35 + 0.5°C; Microscope and light source, or
          colony counter; Petri culture dishes; incubator,
          35 _+ 0.5°C; Microscope and light source, or colony
          counter	 2, 3

56.  Coliforra Bacteria, Fecal (Number/100 ml)

     (a)  MPN:  Autoclave (to 121°C); Inoculation tubes;  Incubator,
          35 _+ 0.5°C; Water bath, 44.5jf 0.2°C	 2, 3

     (b)  Membrane Filter:  Autoclave (to 121°C); Filter membranes;
          Petri culture dishes; Water bath, 44.5 _+ 0.2°C; Low-power
          (10-15X), binocular, wide-field, dissecting microscope
          and light source	 2, 3

57.  Coliforra Bacteria, Total (Number/100 ml)

     (a)  MPN:  Same as 56 (a)	 2, 3

     (b)  Membrane Filter:  Same as 56 (b)	 2, 3

OTHER PARAMETERS

58.  Temperature:  Good quality mercury-filled or dial-type
     centrigrade thermometer, or a thermistor	 1

59.  pH:  pH meter (electrometer using either glass electrode and
     reference, such as saturated calomel, or a combination glass
     and reference electrode)	 1

AMBIENT AIR

60.  Sulfur Dioxide (ug/m-3 or ppm)

     (a)  Pararosaniline Method:  Absorber; Pump; Air flowmeter
          or critical orifice; Spectrophotoraeter for use at 548 nm,
          band width < 15 nm, with 1-cm cells	 2, 3

     (b)  Automated Pararosaniline:  Autoanalyzer with appropriate
          manifold and 548-nm filter	 2, 3

     (c)  Continuous Analyzer:  EPA-designated equivalent method
          instrumentation	 2 , 3
                                     139

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No,
AMBIENT AIR (continued)


61.  Suspended Particulates (ug/m^)

     (a)  High-Volume Sampler:  High-volume sampler; Shelter  for
          sampler; Flow measurement equipment, including  (1)  Rota-
          meter,  (2) Orifice calibration unit, (3) Differential
          manometer, (4) Positive displacement meter; Barometer;
          Environment for conditioning filters; Analytical balance:
          chamber to hold unfolded 8" x 10" filters, sensitivity
          =0.1 mg; Glass fiber filters; Acceptable alternative
          equipment for flow measurement (3-6):  Exhaust  orifice
          meter,  interfaced with a circular chart recorder	 1, 2

     (b)  Continuous High-Volume:   EPA-designated equivalent
          instrumentation	 2 , 3

62.  Carbon Monoxide (ug/m^ or ppm)

     Non-Dispersive Infrared Spectrometry:  Carbon monoxide
     analyzer; Pump, flow control valve, and  flowraeter; In-line
     filter for particles (2-10 urn); Moisture control (refriger-
     ation unit or drying tube)	 2, 3

63.  Photochemical Oxidant (63) (ug/m-^ or ppm)

     Chemiluminescence, Continuous:  Commercial photochemical
     oxidant (63) analyzer, or apparatus consisting of
     (1) Detector cell, (2) Flowmeters (air and ethylene),
     (3) Air inlet filter (Teflon, 5 m), (4) Photomultiplier  tube,
     (5) High voltage power supply, (6) Direct current amplifier,
     (7) Recorder, (8) Ozone source (low pressure Hg lamp/quartz
     tube) and dilution system; Apparatus for calibration
     (KI -> I2 Spectrophotoraetric Method)	 2, 3

64.  Total Hydrocarbons (Corrected for Methane) GC - FID

     Method:  Commercially-available THC, CH^, and CO Analyzer;
     Pump, flow control valves, automatic switching valves, and
     flowmeter; In-line filter (3-5 um); Stripper or precolumn;
     Oven (for column and catalytic converter)	 2, 3
                                     140

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
AMBIENT AIR (continued)


65.  Nitrogen Dioxide (ug/nH or ppm)

     (a)  Arsenite 24-Hour Sampling Method:  Sampling train
          (bubbler, trap, membrane filter, 27-gauge hypodermic
          needle, air pump, calibration equipment); Standard
          glassware (volumetrics, pipets, graduated cylinders,
          etc.); Spectrophotoraeter or colorimeter for use at
          540 nm	 2, 3

     (b)  Continuous Chemiluminescent Method:  Commercial chemi-
          luminescent analyzer [generally including particulate
          filter, thermal converter, (N02	NO), ozone generator,
          reaction chamber, optical filter, photomultiplier tube,
          and vacuum pump; instrument will be specified as EPA
          equivalent method]; Calibration apparatus (Gas-Phase
          Titration Method)  [generally including air flow con-
          troller, air flowmeters, pressure regulator for NO
          cylinder, NO flowmeters, capillary restriction, ozone
          generator, reaction chamber and mixing bulb, sample
          manifold, NO detector, iodometric calibration apparatus]... 2, 3

     (c)  Griess-Saltzman Colorimetric, Continuous:  Sampling
          train; Colorimeter for use at 550 nm	 2, 3

SOURCE AIR

66.  Stack Gas Velocity (EPA Method 2):  Pitot tube	 2

67.  Dry Molecular Weight of Gas (EPA Method 3)

     (a)  Or sat	 2

     (b)  Gas chromatograph with thermal conductivity detector	 2

68.  Stack Gas Moisture (EPA Method 4):  Midget impingers;
     Sample metering pump	 2

69.  Particulates (EPA Method 5):  Heated filter holder;
     Impingers; Sample metering pump; Analytical balance;
     Heated probe	 2, 3

70.  Sulfur Dioxide (EPA Method 6):  Impingers; Sample
     metering pump; Burettes	 2, 3
                                     141

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
SOURCE AIR (Continued)

71.  Nitrogen Oxide (EPA Method 7):  Round-bottom flasks; Vacuum
     pump; Hot plate;  Spectrophotometer	  2, 3

72.  Sulfuric Acid Mist (EPA Method 8):  Source sampling train;
     Burettes	  2, 3

73.  Visible Emissions (EPA Method 9):  Stopwatch	  2

74.  Carbon Monoxide (EPA Method 10):  Non-dispersive infrared
     analyzer	  2, 3

75.  Hydrogen Sulfide  (EPA Method 11):  Same as 70	  2, 3

76.  Fluoride

     (a)  EPA Method 13A:   Spectrophotometer	  2, 3

     (b)  EPA Method 13B:   Fluoride-specific ion electrodes;
          Specific ion meter	  2, 3

77.  Sulfur Compounds  (EPA Method 15):  Gas chromatograph with
     flame photometric detector; All Teflon dilution system; Mass
     flowmeter; Heated sampling line; Permeation system	  3, 4

78.  Sulfur Compounds  (EPA Method 16):  Two gas chromatographs
     with flame photometric detectors; All Teflon dilution system;
     Mass flowmeter; Heated sampling line; Permeation system	  3, 4

79.  Particulates (EPA Method 17):  In-stack filter holder;
     Impingers; Sample metering pump; Analytical balance	  2, 3

BIOLOGY

80.  Wildlife and Terrestrial Methods

     (a)  Cover and Habitat Assessment:  Ocular instrumentation	  3, 4

     (b)  Mapping and Vegetation Analysis:  Planimeter;
          Statistical instrumentation	  2, 3, 4

     (c)  Photo Interpretation:  Stereoscope	  3, 4

     (d)  Biomass Determination:  General laboratory
          instrumentation	  1, 2, 3
                                     142

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
BIOLOGY (Continued)

     (e)  Species Composition/Density:   Statistical
          instrumentation	  2, 3, 4

     (f)  Ecosystem Analysis:   Statistical instrumentation;
          Graphical instrumentation	  3, 4

     (g)  Community Analysis:   Same as  (f)	  3, 4

     (h)  Systems Analysis:  Computer	  4

     (i)  Inventory of Mammals:   Ocular instrumentation;
          Mechanical traps	  1,2,3,4

     (j)  Inventory of Birds:   Same as  (i)	  2, 3, 4

     (k)  Inventory of Reptiles  and Amphibians:   Same as (i)	  2, 3, 4

     (1)  Inventory of Terrestrial Insects:   Same as (i)	  2, 3, 4

81.   Bioassay

     (a)  Culture Maintenance:   General laboratory equipment	  2

     (b)  Routine Static Assays:   General laboratory equipment	  2

     (c)  Chronic Assays:  Proportional diluters; Metering
          pumps ; Photometers	  3

     (d)  Non-Routine Assays:   Dependent on type of assay	  3, 4

     (e)  Physiological/Biochemical:  Multi-channel recorder;
          Photometer	  3, 4

82.   Fish:  Age and Growth

     (a)  Collection:  Electrofisher; Nets;  Seines; Trawls;
          Traps; Toxicants	  1, 2

     (b)  Scale Analysis:  Scale press; Scale reader	  2

83.   Fish:  Population Dynamics

     (a)  Collection:  Same as  82(a)	  1, 2
                                     143

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
BIOLOGY (Continued)

     (b)  Mark and Recapture Methods:  Method-specific materials	  1, 2,  3

     (c)  Predictive Models:  Statistical instrumentation	  3, 4

84.  Fish:  Feeding Habits

     (a)  Collection:  Same as 73(a)	  1, 2

     (b)  Stomach Content Analysis	  2, 3

85.  Phytoplankton

     (a)  Collection:  Plankton net	  1, 2,  3

     (b)  Analysis:  Microscope with Sedwick-Rafter Cell	  3, 4

86.  Zooplankton

     (a)  Collection:  Zooplankton net	  1, 2,  3

     (b)  Microscope

87.  Periphyton

     (a)  Collection:  Glass slides attached to floats or other
          sampling apparatus	  1, 2,  3

     (b)  Analysis:  Microscope	  3, 4

88.  Aquatic and Terrestrial

     (a)  Specific Macrophyte Collection:  Macrophyte standard
          sampler	  2, 3

     (b)  Macrophyte Taxonomy:  Ocular instrumentation	  2, 3,  4

89.  Chlorophyll

     (a)  Spectrophotometric:  Variable wavelength spectro-
          photometer	  2 , 3

     (b)  Fluorometric:  Fluorometer	  2, 3
                                     144

-------
                           APPENDIX G (continued)

                                                                    Skill
                                                                  Rating No.
RADIATION

The analysis of radiological parameters requires special expertise
and experience, in addition to up-to-date, we11-maintained,
calibrated instrumentation and apparatus.

90.  Alpha, Total (pCi/liter):  Windowless Gas-Flow Proportional
     Counter and associated equipment, or Thin Window Gas-Flow
     Proportional Counter and associated equipment, or Alpha
     Scintillation Counter and associated equipment, or Alpha
     Spectrometer System (Surface Barrier Type) and associated
     equipment	 2, 3, 4

91.  Alpha Counting Error (pCi/liter):  Same as 90	 2, 3, 4

92.  Beta, Total (pCi/liter):   Windowless  Gas-Flow Proportional
     Counter and associated equipment, or  Thin Window Gas-Flow
     Proportional Counter and associated equipment, or Beta
     Scintillation Counter and associated equipment, or Liquid
     Scintillation Counter and associated equipment	 2, 3, 4

93.  Beta Counting Error (pCi/liter):  Same as 92	 2, 3, 4

94.  Radium, Total (pCi/liter):   Windowless gas-flow proportional
     counter and associated equipment, or  Thin window gas-flow
     proportional counter and associated equipment, or Alpha
     scintillation counter and associated  equipment, or Alpha
     spectrometer (surface barrier type) system and associated
     equipment, or Radon gas counting system and associated
     equipment	 2, 3, 4
                                     145

-------
                               APPENDIX H




       SAMPLE  PRESERVATION  METHODS  AND  RECOMMENDED  HOLDING  TIMES







                   TABLE H-l.   WATER PARAMETERS(l»2^
Measurement
Acidity
Alkalinity
Arsenic
Biological Oxygen
Demand (BOD)
Bromide
Chemical Oxygen
Demand (COD)
Chloride
Chlorine Req.
Color
Cyanides

Dissolved Oxygen
Probe
Winkler
Fluoride
Hardness
Iodide
MBAS
Metals
Dissolved

Suspended
Total
Mercury
Dissolved




Vol. Req.
(ml)
100
100
100
1,000
100

50
50
50
50
500


300
300
300
100
100
250

200


100

100




Container
P, G<3)
P, G
P, G
P, G
P, G

P, G
P, G
P, G
P, G
P, G


G only
G only
P, G
P, G
P, G
P, G

P, G




P, G




Preservative
Cool, 4°C
Cool, 4°C
HN03 to pH<2
Cool, 4°C
Cool, 4°C

H2S04 to PH<2
None Req .
Cool, 4°C
Cool, 4°C
Cool, 4°C
NaOH to pH 12

Det . on site
Fix on site
Cool, 4°C
Cool, 4°C
Cool, 4°C
Cool, 4°C

Filter on site
HN03 to pH<2
Filter on site
HN03 to pH<2

Filter
HN03 to pH<2



Holding
Time^7'
24 Hours
24 Hours
6 Months
6 Hours (*)
24 Hours

7 Days
7 Days
24 Hours
24 Hours
24 Hours


No Holding
No Holding
7 Days
7 Days
24 Hours
24 Hours

6 Months

6 Months
6 Months

38 Days
(Glass)
13 Days
(Hard
Plastic)
(continued)
                                  146

-------
                         TABLE H-l  (continued)
Vol. Req.
Measurement (ml)
Total




Nitrogen
Ammonia

Kjeldahl

Nitrate

Nitrite
NTA
Oil & Grease

Organic Carbon

pH

Phenolics


Phosphorus
Orthophosphate,
Dissolved
Hydrolyzable
Total
Total, Dissolved

Residue
Filterable
Non-Filterable
Total
Volatile
Settleable Matter
Selenium
Silica
Specific
Conductance
100





400

500

100

50
50
1,000

25

25

500


50

50
50
50


100
100
100
100
1,000
50
50
100

Container
P, G





Pf*
> «

P, G

P, G

P, G
P, G
G only

P, G

P, G

G only


P, G

P, G
P, G
P, G


P, G
P, G
P, G
P, G
P, G
P, G
P only
P, G

Preservative
HN03 to pH<2





Cool, 4°C
H2S04 to PH<2
Cool, 4°C
H2S04 to pH<2
Cool, 4°C
H2S04 to pH<2
Cool, 4°C
Cool, 4°C
Cool, 4°C
H2S04 to pH<2
Cool, 4°C
H2S04 to pH<2
Cool, 4°C
Det. on site
Cool, 4°C
H3P04 to pH<4
1.0 g CuS04/l
Filter on site
Cool, 4°C
Cool, 4°C
H2S04 to pH<2
Cool, 4eC
Filter on site
Cool, 4°C

Cool, 4°C
Cool, 4°C
Cool, 4°C
Cool, 4°C
None Req .
HN03 to pH<2
Cool, 4°C
Cool, 4°C

Holding
Time^7'
38 Days
(Glass)
13 Days
(Hard
Plastic)

24 Hours (5)

24 Hours ^5)

24 Hours (5)

24 Hours (5)
24 Hours
24 Hours

24 Hours

6 Hours(4)

24 Hours


24 Hours (5)

24 Hours (5)
24 Hours (5)
24 Hours (5)


7 Days
7 Days
7 Days
7 Days
24 Hours
6 Months
7 Days
24 Hours ^6^

(continued)
                                  147

-------
                         TABLE H-l (continued)
Measurement
Sulfate
Sulfide
Vol. Req.
(ml)
50
50
Container Preservative
P,
P,
G
G
Cool
2 ml
, 4
°C
zinc
Holding
Time (7?
7
24
Days
Hours
acetate
Sulfite
Temperature
Threshold Odor
Turbidity
50
1,000
200
100
P,
P,
G
P,
G
G
only
G
Cool
Det.
Cool
Cool
, 4
on
, 4
, 4
°C
site
°C
°C
24
No
24
7
Hours
Holding
Hours
Days
1.  More specific instructions for preservation and sampling  are  found
    with each procedure as detailed in this manual.  A general discus-
    sion on sampling water and industrial wastewater may be found  in
    ASTM, Part 31, p. 68-78 (1978).

2.  U.S. Environmental Protection Agency Office of Research and Develop-
    ment.  Environmental Monitoring and Support Laboratory.   1979.
    Methods for Chemical Analyses of Water and Wastes.  Cincinnati,
    Ohio.  EPA-625/6-74-003.

3.  Plastic or Glass.

4.  If samples cannot be returned to the laboratory in less than  6 hours
    and holding time exceeds this limit, the  final reported data  should
    indicate the actual holding time.

5.  Mercuric chloride may be used as an alternate preservative at  a
    concentration of 40 mg/1, especially if a longer holding  time  is
    required.  However, the use of mercuric chloride is discouraged
    whenever possible.
                                                     •*»
6.  If the sample is stabilized by cooling, it should be warmed to 25°C
    for reading, or temperature correction should be made  and results
    reported at 25°C.

7.  It has been shown that samples properly preserved may  be  held  for
    extended periods beyond the recommended holding time.
                                  148

-------
                   TABLE H-2.  AMBIENT AIR SAMPLES
                            Recommended
Parameter                   holding time            Preservation method
Particulate Filters         Indefinite              Store in controlled
                                                    atmosphere of <50%
                                                    relative humidity

Sulfur Dioxide              30 days, if             Store at <4CC after
(Pararosaniline Method)     properly stored         collection, during
                                                    transport, and
                                                    before analysis

Nitrogen Oxides             6 weeks                 Samples are stable
(Sodium-Arsenite Method)                            for 6 weeks at room
                                                    temperature

Fluoride                    None                    Collect and store
                                                    in plastic
                                                    containers
                                  149

-------
                                 APPENDIX  I

                      QUALITY ASSURANCE AUDIT CHECKLIST

This form has-been prepared to be used in  the on-site  audit  of  projects.
Personnel performing the audit should list any  questions  which  were  not  clear
from the information provided by the Quality Assurance Pre-Audit  worksheet  in
this checklist and obtain answers during the actual  site  visit.

This form is to be used in the actual performance  of an on-site project
audit.

This checklist is to be completed by the personnel involved  in  the on-site
audit.
                                      150

-------
                         APPENDIX I (Continued)




                   QUALITY ASSURANCE AUDIT CHECKLIST
Laboratory:
Street Address:
Mailing Address (If Different):
City 	  State	Zip




Laboratory Telephone No. :   Area Code 	  No. 	




Laboratory Director:
Quality Assurance Supervisor:
Personnel Contacted During Audit:




                  Name                               Title
Contract Number:




Contract Title:




Project Officer:
Audit Conducted By:




Agency and Address:
Telephone No.:  Area Code 	  No.







Date Audit Performed:
                                   151

-------
                         APPENDIX I  (continued)
A.  ORGANIZATION AND PERSONNEL

A.I.  Review the Pre-Audit Worksheet  and list questions from the Organ-
      ization and Personnel section of the Pre-Audit Worksheet to be
      discussed during the QA audit.
Al.
Q2.
A2.
Q3.
A3.
Q4.
A4.
Q5.
A5.
                                  152

-------
                          APPENDIX  I  (continued)
A.2.  Organization and  Personnel Checklist

Do personnel assigned to this project have the
appropriate educational background to success-
fully accomplish the objectives of the program?
Do personnel assigned to this project have the
appropriate level and type of experience to
successfully accomplish the objectives of this
program?
Is project organization appropriate to
accomplish the objectives of this' program?
Is the project adequately staffed to meet
project commitments in a timely manner?
Are project reporting relationships clear?
If any special training or experience is
required, is it represented on the project staff?
Does the laboratory have a Quality Assurance
Supervisor who reports to senior management
levels?
Was the Project Manager available during the
QA audit?
Was the Quality Assurance Supervisor available
during the QA audit?
Does the project schedule show adequate time to
accomplish the sampling program and does it allow
for uncontrollable delays, such as bad weather?
Does the project schedule allow sufficient time
between sample collection and reporting of the
data to apply adequate analytical quality control,
including supervisory review of the data?
Yes











No











Comment











                                     153

-------
                           APPENDIX  I  (continued)

A. 3.  Does the project organization  plan  and  schedule  give  adequate  attention
      and time to the sampling and analysis effort?
      Comments on project organization  and  schedule:
A.4.  Are the personnel assigned to this project generally  qualified  to
      accomplish the objectives of the program?  	
      Comments on personnel:
                                     154

-------
                           APPENDIX I (continued)
B.  FACILITIES
When touring the facilities, give special attention to:  (a) the overall
appearance of organization and neatness, (b) the proper maintenance of
facilities and instrumentation, (c) the general adequacy of the facilities
to accomplish the required work, and (d) sampling equipment required  for  the
project.

B.I.  General Facilities Checklist

Does the laboratory appear to have adequate
workspace (120 sq. feet, 6 linear feet of
unencumbered bench space per analyst)?
Are voltage control devices used on major
instrumentation (e.g., GC/MS, spectropho-
tometers)?
Does the laboratory have a source of distilled/
demineralized water?
Is the conductivity of distilled/demineralized
water routinely checked and recorded?
Is the analytical balance located away from draft
and areas subject to rapid temperature changes?
Has the balance been calibrated within one year?
Are exhaust hoods provided to allow
organized work with volatile materials?
Is the laboratory maintained in a clean and
organized manner?
Are safe and contamination-free work areas
provided for the handling of toxic or radio-
active materials?
Are the radioactive and/or toxic chemical
handling areas either a stainless steel bench
or an impervious material covered with absorbent
material?
Yes










No










Comment










                                     155

-------
APPENDIX I (continued)

Are adequate facilities provided for storage of
samples, including cold storage?
Are chemical waste disposal facilities adequate?
Are contamination-free areas provided for trace
level analytical work?
Can the laboratory supervisor document that trace
metals-free water is available for preparation of
standards and blanks?
Is organic-free water available for preparation of
standards and blanks?
If biotesting is to be conducted, are adequate
environment-controlled facilities available (e.g.,
light, temperature control)?
Is the required field instrumentation and sampling
equipment properly maintained?
Is adequate safety equipment (fire extinguishers,
showers, eyewash stations) located throughout the
laboratory?
If bacteriological analyses are to be conducted,
is an aseptic work area provided?
Are bacteriological incubators maintained at the
proper temperature (35 + 0.5°C for total coliform
and fecal streptococcus^ 44. 5 +• 0.2°C for fecal
coliform)?
Are boats, motors, vehicles, and other mobile
facilities available as required?
Yes











No











Comment











          156

-------
                           APPENDIX I (continued)
B.2.  Instruments.   List the major laboratory and in situ analytical
      instruments that will be used.  Complete an instrument evaluation  form
      on each one.
               Instrument
Analysis
                                      157

-------
                          APPENDIX I  (continued)




                         INSTRUMENT EVALUATION
          Instrument:
Instrument Mfg.




Model:
Year of Acquisition:
Condition:
Calibration Frequency:
Service Maintenance Frequency:




Other Pertinent Information:

Are Manufacturer's operating manuals readily
available to the operator?
Is there a calibration protocol available to the
operator?
Are calibrations kept in a permanent record?
Is a permanent service record maintained?
Has the instrument been modified in any way?
yes





no





                                         SATISFACTORY?
Comments:
                                  158

-------
          Instrument:
                          APPENDIX I (continued)




                         INSTRUMENT EVALUATION
Instrument Mfg.




Model:
Year of Acquisition:
Condition:
Calibration Frequency:
Service Maintenance Frequency:




Other Pertinent Information:

Are Manufacturer's operating manuals readily
available to the operator?
Is there a calibration protocol available to the
operator?
Are calibrations kept in a permanent record?
Is a permanent service record maintained?
Has the instrument been modified in any way?
yes





no





                                         SATISFACTORY?
Comments:
                                  159

-------
                         APPENDIX I (continued)
                         INSTRUMENT EVALUATION
          Instrument:
Instrument Mfg.
Model:
Year of Acquisition:
Condition:
Calibration Frequency:
Service Maintenance Frequency:
Other Pertinent Information:

Are Manufacturer's operating manuals readily
available to the operator?
Is there a calibration protocol available to the
operator?
Are calibrations kept in a permanent record?
Is a permanent service record maintained?
Has the instrument been modified in any way?
yes





no





                                         SATISFACTORY?
Comments:
                                  160

-------
                        APPENDIX I (continued)
C.  ANALYTICAL METHODOLOGY

C.I.  Review the Pre-Audit Worksheet and list items from the Analytical
      Methodology section of the Pre-Audit Worksheet to be discussed
      during the QA audit.
Al
Q2.
A2.
Q3.
A3.
Q4.
A4.
Q5.
A5.
                                 161

-------
                          APPENDIX  I  (continued)
C.2.  Conduct discussions with two or more individuals who have
      analytical responsibilities in connection with the project.  The
      following points should be addressed to determine each individ-
      ual's awareness and application of QA/QC procedures:

      1.  Specific project responsibilities,
      2.  Level of knowledge of the analytical methods used,
      3.  Awareness of and adherence to the laboratory's QC procedures,
          and
      4.  Appearance and accuracy of the work records.
   Analyst
                       Name                      Responsibility

   Comments:
   Analvst
                       Name                      Responsibility

   Comments:
                                  162

-------
Comments:
Comments:
                      APPENDIX I (continued)
Analyst
                    Name
Responsibility
Analyst
                    Name
Responsibility
                                163

-------
                           APPENDIX I (continued)
C.3.   Analytical  Methodology Checklist.
Item
Are standard methods (e.g., EPA, ASTM, Standard
Methods for the Examination of Water and
Wastewater) used when available?
Have standard methods been altered in any way?
If so, is it justified?
Are written analytical procedures provided to
the analyst?
Are reagent grade or higher purity chemicals
used to prepare standards?
Are samples analyzed within the linear range
of the method in all cases?
Does the standard curve bracket the concen-
tration of the samples on each sample run?
Are fresh analytical standards prepared at
the required frequency?
Are standards run periodically during a
long sample run?
Are reference standards properly labeled with
concentrations, date of preparation, and the
identity of the person preparing the sample?
Do the analysts record bench data in a neat
and accurate manner?
Is the appropriate instrumentation used in
accordance with standard procedures?
Are methods used which are appropriate to the;
sample matrix (e.g., saline waters, wastewaters)?
Yes












No












Comment












                                     164

-------
                         APPENDIX I  (continued)
Item
Are analytical detection limits adequate for
the purposes of the project?
Are the analytical procedures used adequately
documented? For example, if a standard method
is not available, is a written procedure
incorporated into the project plan?
Are all strip charts properly labeled with
instrument conditions, date, and sample
numbers?
Are samples properly handled (e.g., organized,
chilled as necessary, appropriate containers)
before, during, and after analysis?
Yes




No




Comment




C.4.  Are the analytical methods used satisfactory to  accomplish  the  objectives
      of the program? 	  Are laboratory practices  acceptable?  	
   Comments on analytical methods and practices:
                                   165

-------
                        APPENDIX I (continued)
 D.   SAMPLING AND SAMPLE HANDLING

 D.I.   Review the Pre-Audit  Worksheet  and  list  the  items  from  the  Sampling
       and  Sample Handling section  of  the  Pre-Audit  Worksheet  to be
       discussed.
_
 Al.
 Q2.
 A2.
 03.
 A3.
 QA.
 Q5.
 A5.
                                  166

-------
                         APPENDIX I (continued)
D.2.  Conduct discussions with two or more individuals who have sampling
      responsibilities in connection with the project.  The following
      points should be addressed to determine each individual's aware-
      ness and application of appropriate sampling procedures:

      1.  Specific project responsibility,
      2.  Level of knowledge of acceptable sampling procedures,
      3.  Adherence to the project sampling plan, and
      4.  Neatness and accuracy of field records.
   Field
   Technician
                         Name                      Responsibility

   Comments:
   Field
   Technician
                         Name                      Responsibility

   Comments:
                                 167

-------
                          APPENDIX I  (continued)
D.3.   Sampling  Equipment and Procedures Checklist
Item
Are the sampling procedures specifically
defined in the project QA plan or other
referenced document?
Is the sampling program well organized?
Have appropriate techniques been used in
selecting sampling sites?
Are proper containers used for sample
collection, transport, and storage?
Are sample containers properly prepared before
sample collection to avoid sample contamination?
Containers for organics should be solvent rinsed;
for trace metals, acid rinsed.
Are the proper preservatives used in the samples
for each parameter? (See Appendix H)
Are permanent labels affixed to sample containers?
Do the sample labels contain adequate information
(date, time, sample location, samples) and a
unique sample identification number?
Are proper techniques used to collect representa-
tive samples while avoiding sample contamination?
Are duplicate samples collected? What frequency?
Are samples shipped promptly to the laboratory in
order to meet recommended holding time deadlines?
(See Appendix H)
Are chain-of-custody records available for inspec-
tion? Are they neat and understandable? Have the
required custody signatures been obtained?
Yes












No












Comment












                                     168

-------
                        APPENDIX I  (continued)

D.4.  Field Notebooks.  Review one or more field notebooks and determine if
      the following information is recorded.
Item
Is a permanent bound notebook used to
record all field data and observations?
Is the notebook reasonably neat and
organized, considering the use under
adverse field conditions?
Are field instrument and in situ instru-
ment calibration data recorded daily?
Are sample location, time, and number
accurately and completely recorded?
Are in situ data neatly recorded in an
understandable manner?
Are ambient data (i.e., weather)
recorded when appropriate?
Are the necessary engineering data (e.g.,
flow, operating conditions) recorded?
Have supervisory personnel reviewed the field
notebook and so indicated by their signature?
Yes








No








Comment








D.5.  Is the sampling program adequate to accomplish the objectives of the
      project?  	
   Comments on the sampling program and sample collection:
                                 169

-------
                         APPENDIX  I (continued)
 E.   QUALITY CONTROL


 E.I.   Review the  Pre-AudiL  Worksheet  and  list  items  from  the Quality
       Control  section  of the  Pre-Audit Worksheet  to  be discussed.


~oT
 Al.
 Q2.
 A2.
 Q3.
 A3.
 Q4.
 A4.
 Q5.
 A.5
                                   170

-------
                          APPENDIX I  (continued)




E.2.  Quality Control  Manual Checklist
Item
Does the laboratory maintain a
Quality Control Manual?
Does the manual address the important
elements of a QC program, including the
fol lowing :
a. Personnel?
b. Facilities and equipment?
c. Configuration control of instruments?
d. Documentation of procedures?
e. Procurement and inventory practices?
f. Preventive maintenance?
g. Reliability of data?
h. Data validation?
i. Feedback and corrective action?
j. Instrument calibration?
k. Pecord keeping?
1 . Internal audits?
Does the QC Manual specify the frequency of
duplication and spiked sample analysis?
Is at least 10 percent sample duplication
required?
Are QC responsibilities and reporting
relationships clearly defined?
Yes

















No

















Comment

















                                   171

-------
                       APPENDIX I  (continued)
E.3.   Quality  Control Procedures Checklist
Item
Select a representative number of
analyses from the project list and review
historical quality control data for these
parameters. Are QC records adequate for
the purposes of the project?
Are reference standards analyzed with
each set of samples?
Have standard curves been adequately
documented?
Are laboratory standards traceable to
the National Bureau of Standards, where
appropriate?
Have standards been analyzed every
20 or fewer samples to verify that the
analytical method is in control?
Have the prescribed number (QC Manual or
Project Plan) of duplicate and spiked
samples been analyzed?
Do duplicate data fall within
acceptable limits of precision?
Are recoveries, calculated from spiked
sample data, acceptable?
Are quality control charts maintained
for each routine analysis?
Do QC records show corrective action
when analytical results fail to meet
QC criteria?
Do supervisory personnel review the
data and QC results?
Yes











No











Comment











                                    172

-------
                        APPENDIX I (continued)

E.4.  Are quality control procedures and  records generally adequate  to
accomplish the objectives of the project?   	
   Comments on quality control procedures  and  records:
                                  173

-------
                       APPENDIX I (continued)
F.  DATA HANDLING


F.I.  Review the Pre-Audit Worksheet  and list items from the Data

      Handling section of the Pre-Audit Worksheet to be discussed.


Ql.
Al.
Q2.
A2.
Q3.
A3.
04,
A4.
Q5,
A5.
                                 174

-------
                           APPENDIX I (continued)




F.2.  Data Handling Checklist
Item
Ask for a demonstration of data handling
procedures from initial sample check-in
to reporting of the final data. Are
these procedures clear and adequate to
avoid data errors?
Are data calculations checked by a
second person?
Are data calculations documented?
Do records indicate corrective action
that has been taken on rejected data?
Are limits of detection determined and
reported properly?
Are results which are below the analytical
detection limit reported as such?
Are the data reported to a justifiable
number of significant figures?
Are all data and records retained at least
3 years beyond completion of the project?
Are quality control data (e.g., standard
curve, results of duplication and spikes)
accessible for all analytical results?
Are data reported in the appropriate units
(e.g., ppm, mg/1, dry weight, metric
measure)?
Yes










No










Comment










                                     175

-------
                       APPENDIX I  (continued)

F.3.  Are data handling procedures adequate  to  accomplish  the  objectives
      of the project and to trace the accompanying quality control
      results?
   Comments on data handling:
                                 176

-------
                           APPENDIX I (continued)
G.   SUMMARY




G.I.  Summary Checklist
Item
Do responses to audit questions indicate that
project and supervisory personnel are aware
of QA and its application to the project?
Do project and supervisory personnel place
positive emphasis on QA/QC?
Have responses with respect to QA/QC aspects
of the project been open and direct?
Has a cooperative attitude been displayed
by all project and supervisory personnel?
Are the personnel assigned to the
project qualified?
Does the organization place the proper
emphasis on quality assurance?
Have any QA/QC deficiencies been
discussed before leaving?
Is the overall quality assurance
adequate to accomplish the objectives
of the project?
Are any corrective actions required?
If so, list the necessary actions below.
Yes









No









Comment









                                     177

-------
G.2.  Summary Comments and Corrective Actions
                                    178

-------
                                   GLOSSARY
agency:  the United States Environmental Protection  Agency (EPA).

analytical or reagent blank:  a  blank  used  as  a  baseline  for the analytical
     portion of a method.  For example,  a blank  consisting of a sample from a
     batch of absorbing  solution  used  for normal  samples,  but processed
     through the analytical  system  only, and used  to adjust  or correct
     routine analytical  results.

blank or sample blank:   a sample  of  a  carrying agent (gas, liquid,  or solid)
     that is normally used to selectively capture  a  material of interest  and
     that is subjected to the usual  analytical or  measurement process to
     establish a zero baseline or background value,  which  is used  to adjust
     or correct routine  analytical  results.

confidence interval:  a  value interval  that has  a  designated probability  (the
     confidence coefficient) of  including some defined  parameter of the
     populat ion.

confidence limits:  the  outer boundaries of a  confidence  interval.

contract:  the legal  instrument  reflecting  a relationship  between  the Federal
     Government and a State  or local government  or other  recipient:
     (1) whenever the principal  purpose  of  the instrument  is the acquisition,
     by purchase, lease, or  barter,  of  property  or services  for the direct
     benefit or use of the Federal Government; or  (2) whenever an  executive
     agency determines in a  specific instance  that the  use of a type of
     procurement contract is appropriate.

cooperative agreement:   the  legal instrument reflecting the  relationship
     between the Federal Government  and  a State  or local  government or other
     recipient whenever:  (1) the principal purpose  of  the relationship is
     the transfer of money,  property,  services,  or anything  of value to the
     State or local government or other  recipient  to accomplish a  public
     purpose of support  or stimulation  authorized  by Federal statute, rather
     than acquisition, by purchase,  lease,  or  barter, of  property  or services
     for the direct benefit  or use  of  the Federal  Government; and  (2) sub-
     stantial involvement is anticipated between  the executive agency acting
     for the Federal Government  and  the  State  or  local  government  or other
     recipient during performance of the contemplated activity.

data validation:  a systematic effort  to review  data to identify any outliers
     or errors and thereby cause  deletion or flagging of  suspect values to


                                     179

-------
     assure the validity of the data  to  the user.  This  "screening"  process
     may be done by manual and/or computer methods,  and  it may  utilize  any
     consistent technique such as sample  limits  to screen out impossible
     values or complicated acceptable relationships  of the data with  other
     data.

extramural review:  technical and scientific review  of a research  or
     demonstration proposal by a qualified individual not an employee of  the
     Environmental Protection Agency, such as  an employee of industry or  an
     academic institution.

grant:   the legal instrument reflecting  the relationship between  the  Federal
     Government and a State or local  government  or other recipient in which:
     (1) the principal purpose of the relationship is the transfer of money,
     property, services, or anything  of  value  to the State or local  govern-
     ment or other recipient in order to  accomplish  a public purpose  of
     support or stimulation authorized by Federal statute, rather  than
     acquisition, by purchase, lease, or  barter, of  property or services  for
     the direct benefit or use of the Federal  Government; and (2)  no  sub-
     stantial involvement is anticipated  between the executive  agency,  acting
     for the Federal Government, and  the  State or local  government or other
     recipient during performance of  the  contemplated activity.

grantee:  any individual, agency, or  entity that has been awarded  a  grant
     pursuant to grant regulations or has received a cooperative  agreement.

in-house project:  a project carried  out  by EPA  staff in EPA facilities.

intramural review:  technical and scientific review  of a research  or
     demonstration proposal by a qualified employee  of the Environmental
     Protection Agency.

measures of dispersion or variability:  measures of  the  differences,  scatter,
     or variability of values of a set of numbers.  Measures of the  disper-
     sion or variability are the range,  the standard deviation, the  variance,
     and the coefficient of variation.

performance audit:  planned independent  (duplicate) sample checks  of  actual
     output made on a random basis to arrive at  a quantitative  measure  of the
     quality of the output.  These independent checks are made  by  an  auditor
     subsequent to the routine checks by  a field technician or  laboratory
     analyst.

performance test sample:  a sample or sample concentrate (to be diluted to a
     specified volume before analysis) of known  (to EPA  only) true value
     which has been statistically established  by interlaboratory  tests.
     These samples are commonly provided  to laboratories to test  analytical
     performance.  Analytical results are reported to EPA for evaluation.

pre-application:  a preliminary proposal  outlining the intent of  a proposed
     project.  Letter format is normally  used, in which  case the  program


                                      180

-------
     office to which the pre-application is referred responds  directly  to  the
     submitter to encourage or discourage  followup.

pre-award survey:  on-site inspection, review, and discussions with  a
     prospective grantee or prospective contractor at his/her  facilities.
     Discussions would normally include, but not be limited  to,  the  proposed
     project plan, personnel, procedures,  schedule, and  facilities.  Normally
     conducted after receipt of a "best and final" offer, but  prior  to  final
     selection of a contractor.

proficiency testing:  special series of planned  tests to determine  the
     ability of  field technicians or laboratory  analysts who normally perform
     routine analyses.  The results may be used  for comparison against
     established criteria, or for relative comparisons among the  data from a
     group of technicians or analysts.

program:  the technical office or staff that has responsibility  for  a part of
     the Agency's operations.   For R&D grants,  the "programs" are  the  Office
     of Research and Development, the Office of  Air Quality  Planning and
     Standards,  the Office of Solid Waste  Management Programs, and  the  Office
     of Mobile Sources Air Pollution Control.

project officer:  the EPA official designated  in the grant or  contract  agree-
     ment as the Agency's principal contact with the grantee on  a particular
     grant.  This person is the individual responsible for project  monitoring
     and for recommendations on or approval of proposed  project  changes.

quality:  the totality of feature and characteristics of a product  or service
     that bears  on its ability to satisfy  a given  purpose.   For  pollution
     measurement systems, the product is pollution measurement data, and  the
     characteristics of major importance are accuracy, precision, and
     completeness.  For monitoring systems, "completeness,"  or the  amount  of
     valid measurements obtained relative  to the amount  expected  to  have  been
     obtained, is usually a very important measure of quality.   The  relative
     importance  of accuracy, precision, and completeness depends  upon
     particular  purpose of the user.

quality assurance:  actions taken by the Laboratory (lERL-Ci)  line  organi-
    zation under  the specific auspices of  the Office of  the  Director, to
    assure that  quality control policies and procedures  are  being properly
    implemented  and appropriate levels of  accuracy, reliability,  and com-
    parability are being achieved in the sampling  and analysis activities
    (including data reduction and handling) of the Laboratory  to  fulfill  the
    Laboratory's  assigned mission.

quality assurance manual:  an orderly assembly of  management policies,
     objectives, principles, and general procedures by which an  agency  or
     laboratory  outlines how it intends to produce quality data.

quality assurance plan:  an orderly assembly of  detailed and specific
     procedures  by which an agency or laboratory delineates  how  it  produces
     quality data for a specific project or measurement  method.   A  given

                                     181

-------
     agency or laboratory would have only one quality  assurance  manual,  but
     would have a quality assurance plan for each of its  projects  or  programs
     (group of projects using the same measurement methods;  for  example,  a
     laboratory service group might develop a plan by  analytical  instrument
     since the service is provided to a number of projects).

quality control:  actions taken by the Laboratory (lERL-Ci)  organization (on
     in-house projects) and by contractors/grantees (on extramural  projects)
     in day-to-day activities to achieve desired accuracy, reliability,  and
     comparability in the results obtained  from sampling  and  analysis
     activities.  Review by contractors/grantees of their  overall  quality
     control activities is "quality assurance" to them, but  "quality  control"
     from the Laboratory's viewpoint.

range:   the difference between the maximum  and minimum values  of  a  set of
     values.  When the number of values is  small (i.e., 12 or  less),  the
     range is a relatively sensitive (efficient) measure  of  variability.

reliability (general):  the ability of an item or system  to  perform a
     required function under stated conditions for a stated  period  of  time.

reliability (specific):  the probability that an item will perform  a  required
     function under stated conditions for a stated period  of  time.

representative sample:  a sample taken to represent a  lot  or  population  as
     accurately and precisely as possible.  A representative  sample may  be
     either a completely random sample or a stratified sample, depending upon
     the objective of the sampling and the  conceptual  population  for  a given
     situation.

spiked sample:  a normal sample of material (gas, solid,  or  liquid) to which
     is added a known amount of some substance of interest.   The  extent  of
     the spiking is unknown to those analyzing the sample.   Spiked  samples
     are used to check on the performance of a routine analysis  or  the
     recovery efficiency of a method.

standard deviation:  the square root of the variance of a  set  of  values:


                    Ui - X)2
          s =
                     n - 1

     If the values represent a sample from a  larger population:

                 N
                    (X£ - u)2
                i=l	
          s =
                       N
                                      182

-------
     where u is the true arithmetic mean  of  the  population.   The property of
          the standard deviation  that makes  it most  practically meaningful is
          that it is  in the  same  units  as  the values  of  the  set, and univer-
          sal statistical tables  for  the  normal  (and  other)  distributions are
          expressed as a function  of  the  standard  deviation.   Mathematically,
          the tables  could just as easily  be expressed  as  a  function of the
          variance.

standard reference material  (SRM):  a material produced  in quantity, of which
     certain properties have been  certified  by the National  Bureau of
     Standards (NBS)  or other  agencies  to  the extent  possible to satisfy its
     intended use.  The material  should be  in a  matrix  similar to actual
     samples to be measured  by a  measurement system  or  be  used directly in
     preparing such a matrix.  Intended uses include:   (1) standardization of
     solutions, (2) calibration of equipment, and  (3) monitoring the accuracy
     and precision of measurement  systems.

standard reference sample (SRS):   a carefully prepared  material produced from
     or compared against an  SRM (or other  equally  well  characterized
     material) such that there is  little  loss of accuracy.  The sample should
     have a matrix similar to  actual  samples used  in  the measurement system.
      These samples are intended  for  use  primarily as reference standards to:
     (1) determine the precision  and  accuracy of measurement  systems,
     (2) evaluate calibration  standards,  and (3) evaluate  quality control
     reference samples.  They may  be  used  "as is"  or  as  a  component of a
     calibration or quality  control measurement  system.  Examples:  an
     NBS-certified sulfur dioxide  permeation device  is  an  SRM.  When used in
     conjunction with an air dilution device, the  resulting  gas becomes an
     SRS.  An NBS-certified  nitric oxide  gas is  an SRM.  When diluted with
     air, the resulting gas  is an  SRS.

standardization:   a physical or mathematical adjustment  or correction of a
     measurement system to make the measurements conform to  predetermined
     values.  The adjustments or  corrections are usually based on a
     single-point calibration  level.

     calibration standard:   a  standard  prepared  by the  analyst for the pur-
          pose of calibrating an  instrument.  Laboratory control standards
          are prepared independently  from  calibration standards for most
          methods.

     detection limit:  that  number obtained  by adding two  standard deviations
          to the average value obtained for  a series  of  reagent blanks that
          are analyzed over  a  long time period (several  weeks or months).

     duplicate analyses:  the collection  of  two  samples  from the same
          fieldsite which are analyzed  at  different  times  but usually on the
          same day.

     laboratory control standard:  a  standard of known  concentration prepared
          by the analyst.
                                      183

-------
     reference standard:  a solution obtained  from  an  outside  source having  a
          known value and analyzed as a blind  sample.

     relative percent error for duplicate analyses:  the difference between
          the measured concentration for the duplicate  pair  times  100  and
          divided by the average of the concentration.

     relative percent error for laboratory control  standards:   the difference
          between the measured value and the theoretically correct value
          times 100 and divided by the correct value.

     relative percent error of a reference sample analysis:  the difference
          between the correct and measured values times 100  and divided by
          the correct concentration.

Standards Based Upon Usage

     calibration standard:  a standard used to quantitate the  relationship
          between the output of a sensor and a property to be  measured.
          Calibration standards should be traceable  to  standard reference
          materials or primary standards.

     quality control reference sample (or working standard):   a material used
          to assess the performance of a measurement or portions thereof.  It
          is intended primarily for routine intralaboratory  use in maintain-
          ing control of accuracy and would be prepared from or traceable  to
          a calibration standard.

Standards Depending Upon "Purity" or Established Physical or Chemical
Constants

     primary standard:  a material having a known property that is stable,
          that can be accurately measured or derived from established
          physical or chemical constants, and  that  is  readily  reproducible.

     secondary standard:  a material having a  property  that  is  calibrated
          against a primary standard.

standards in naturally-occurring matrix:  standards  relating to the pollutant
     measurement portions of air pollution measurement  systems  may be  cate-
     gorized according to matrix, purity, or use.   Standards in a  naturally-
     occurring matrix include Standard Reference Materials and  Standard
     Reference Samples.

statistical control chart (also Shewhart control chart):  a  graphical  chart
     with statistical control limits and plotted values (usually in chrono-
     logical order) of some measured parameter for  a series  of  samples.  Use
     of the charts provides a visual display of  the  pattern  of  the data,
     enabling the early detection of time trends and shifts  in  level.  For
     maximum usefulness in control, such charts  should  be plotted  in a timely
     manner, i.e., as soon as the data are available.
                                     184

-------
system audit:  a systematic on-site qualitative review of facilities,  equip-
     ment, training, procedures, recordkeeping, validation,  and reporting
     aspects of total (quality assurance) system to arrive at a measure of
     the capability and ability of the system.   Even though  each element of
     the system audit is qualitative in nature, the evaluation of each ele-
     ment and the total may be quantified and scored on some subjective
     basis.

Test Variability

     accuracy:  the degree of agreement of a measurement (or an average of
          measurements of the same thing), X, with an accepted reference or
          true value, T, usually expressed as the difference between the two
          values, X-T, or the difference as a percentage of  the reference or
          true value, 100(X-T)/T, and sometimes expressed as a ratio,  X/T.

     bias:  a systematic (consistent) error in test results.  Bias can exist
          between test results and the true value (absolute  bias, or lack of
          accuracy), or between results from different sources (relative
          bias).  For example, if different laboratories analyze a homo-
          geneous and stable blind sample, the relative biases among the
          laboratories would be measured by the differences  existing among
          the results from the different laboratories.  However, if the true
          value of the blind sample were known, the absolute bias or lack of
          accuracy from the true value would be known for each laboratory.

     precision:  a measure of mutual agreement among individual measurements
          of the same property, usually under prescribed similar conditions.
          Precision is most desirably expressed in terms of  the standard
          deviation but can be expressed in terms of the variance, range, or
          other statistics.  Various measures of precision exist depending
          upon the "prescribed similar conditions."

     replicates:  repeated but independent determinations of the same  sample,
          by the same analyst, at essentially the same time  and same condi-
          tions.  Care should be exercised in considering replicates of a
          portion of an analysis and replicates of a complete analysis.  For
          example, duplicate titrations of the same digestion are not  valid
          replicate analyses, although they may be valid replicate titra-
          tions.  Replicates may be performed to any degree, e.g., dupli-
          cates, triplicates, etc.

     reproducibility:  the precision, usually expressed as a standard  devia-
          tion, measuring the variability among results of measurements of
          the same sample at different laboratories.

variance:  mathematically, for a sample, the sum of squares  of the differ-
     ences between the individual values of a set and the arithmetic mean of
     the set, divided by one less than the number of values.
                                      185
                                                    EPA-RTF LIBRARY

-------
                                   TECHNICAL REPORT DATA
                            (Please read Instructions on the reverse before completing)
 1. REPORT NO.
  EPA-600/9-79-046
                                                           3. RECIPIENT'S ACCESSION-NO.
4. TITLE ANDSUBTITLE
  Quality  Assurance Guidelines  for
  lERL-Ci  Project Officers
                                                           5. REPORT DATE
                                                            December, 1979 issuing date
             6. PERFORMING ORGANIZATION CODE
7. AUTHOR(S)

  Charles L.  Stratton and John D.  Bonds
                                                           8. PERFORMING ORGANIZATION REPORT NO.
9. PERFORMING ORGANIZATION NAME AND ADDRESS
  Environmental Science and Engineering,  Inc.
  P. 0. Box 13^
  Gainesville,  Florida  326oh
             10. PROGRAM ELEMENT NO.

              C2HN1E	
             11. CONTRACT/GRANT NO.
                                                             68-03-2656
 12. SPONSORING AGENCY NAME AND ADDRESS
  Industrial  Environmental Research  Laboratory
  Office of Research and Development
  U. S. Environmental Protection Agency
  Cincinnati,  Ohio  U5268
             13. TYPE OF REPORT AND PERIOD COVERED
             Final Report  9/78-12/79
             14. SPONSORING AGENCY CODE

                    EPA/600/12
15. SUPPLEMENTARY NOTES
16. ABSTRACT
  This document  provides guidelines  to Industrial Environmental Research Laboratory-
  Cincinnati  (lERL-Ci) Project Officers for (l)  incorporating quality assurance  (QA)
  criteria  in contract procurement and grant awards,  (2)  monitoring quality assurance
  of extramural  projects, and (3)  conducting QA audits for  projects involving sampling
  and analysis activities.  The Project Officer's responsibilities are described  for
  the initiation,  monitoring, and satisfactory conclusion of contracts, research  and
  demonstration  grants, and cooperative agreements of the type normally funded by IERL-
  Guidance  is provided to assure QA  is adequately addressed  during project conception
  and solicitation and that prospective grantees are informed of QA requirements.  A
  technical evaluation system is presented for the evaluation of the QA aspects of pro-
  posals and  grant applications.  The  basic elements of an acceptable QA program  and of
  a project QA plan are described, and the Project Officer's role in QA monitoring is
  discussed.   A  procedure is described for conducting QA audits of active projects.
  Checklists  are included to assist  the Project Officer.
 7.
                                KEY WORDS AND DOCUMENT ANALYSIS
                  DESCRIPTORS
                                              b.IDENTIFIERS/OPEN ENDED TERMS
                          c.  COSATI Field/Group
  Quality Assurance
  Sampling  and  Analysis
  Contract  Procurement
  QA
  Quality Control
18. DISTRIBUTION STATEMENT

     Release  to Public
i9. SECURITY CLASS (This Report)
   Unclassified	
21. NO. OF PAGES
    196
                                              20. SECURITY CLASS (Thispage)
                                                 Unclassified
                                                                         22. PRICE
EPA Form 2220-1 (9-73)
                                            18 b
                                                                > U.S. GOVERNMENT PRINTING OFFICE: 1980 -657-146/3599

-------