information h aur
USEPA QUALITY ASSURANCE AUDITOR  IS
SCHEDULED FOR A VISIT:  WHAT CAN  I EXPECT
(U.S.) ENVIRONMENTAL PROTECTION AGENCY
RESEARCH TRIANGLE PARK, NC
1995
 U.S. DEPARTMENT OF COMMERCE
 National Technical Information Service

-------
BIBLIOGRAPHIC INFORMATION
P896-139134
Report Nos : EPA/600/A-95/146
Title : USEPA Quality Assurance Auditor Is Scheduled for a Visit: What Can I Expect.
Date : 1995
Authors . L. 0. Childers.
Performing Organization : Environmental Protection Agency. Research Triangle Park. NC.
National Exposure Research Lab.
Sponsoring Organization : *Air and Waste Management Association. Pittsburgh. PA.
NTIS Field/Group Codes : 68A (Air Pollution & Control). 99A (Analytical Chemistry). 70F
( Public Administration & Government)
Price : PC A03/MF AOl
Availability : Available from the National Technical Information Service. Springfield.
VA. 1b1
Number of Pages : 16p
Keywords : *Air pollution monitoring. *Audjts *Data collection, Remote sensing.
Standards compliance. Laboratory tests. Fourier transformation. Infrared spectroscopy.
Requirements. Forms(Paper). Administrative procedures. Quality assurance. Reprints.
Abstract : Environmental studies involving data collection activities conducted by or
br tne United States Environmental Protection A9ency (USEPA) are required to undergo
a review of their data collection activities. This review is usually in the form of an
independent quality assurance (QA) audit. Those studies involving open-path Fourier
transform infrared (OP/FuR) monitoring re also expected to meet this requirement. To
help evaluate the effectiveness of QA pr rams designed for OP/FTIR studies, an audit
questionnaire was developed. This questionnaire provides a meaningful tool in
assessing the study’s compliance with Agency and laboratory QA requirements and
currently accepted quality control (OC) procedures. This paper introduces this audit
questionnaire and discusses the basis for its development.

-------
                                                                   EPA/600/A-95/146
      The USEPA QA Auditor is Scheduled for a Visit.  What Can I Expect?
                                    Lconor Ortiz Childcrs
                         United States Environmental Protection Agency
                                79 Alexander Drive, (MD-77B)
                               Research Triangle Park, NC 27711
ABSTRACT
       Environmental studies involving data collection activities conducted by or for the United
States Environmental Protection Agency (USEPA) zre required to undergo a review of their data
collection activities.  This review is usually in the form of an independent quality assurance (QA)
audit.  Those studies involving open-path Fourier transform infrared (OP/FTIR) monitoring are also
expected to meet this requirement.  This technology,- however,  presents special challenges for both
the instrument operator and the QA auditor.  First, the instrumentation and data-analysis methods
are evolving.  Secondly,  research conducted in the area of QA  has not yet produced agreed-upon
procedures.  Finally, the use of this instrumentation for USEPA-sponsored studies is fairly new.  As
a result, many QA auditors will have limited experience in conducting audits on OP/FTIR systems.
       Regardless of the technology used to collect environmental measurements, the QA auditor
will expect the field team to develop an adequate QA program  before beginning any data-collection
activities. To help evaluate  the effectiveness of QA programs designed for OP/FTIR studies, an
audit questionnaire was developed.  This  questionnaire provides a meaningful tool in assessing the
study's compliance with Agency and laboratory QA requirements and currently accepted quality
control (QC) procedures.  This paper introduces this audit questionnaire and discusses the basis for
its development.

INTRODUCTION
       Each year, the USIIPA spends  millions of dollars on environmental studies.  If a particular
study is conducted in an efficient manner there should be little  to no difference between what was
expected and what was delivered.  If the  amount, quality, or type of data collected was inadequate,
this indicates a performance gap.1  A study that experiences these problems does not usually meet
the needs, expectations, or goals of the end user. Such performance gaps between the end user and
the data collector lead to waste and rework.
       At one time rework accounted  for 20-40% of the budgetary  needs at the USEPA.2 To help
prevent the problems of waste and rework, the USEPA developed Agency-wide QA requirements
for all environmental studies involving data collection activities. The primary goal of these QA
requirements is to produce data of a known quality that satisfies the objectives defined by the data
user.  The role of the QA auditor is to evaluate the effectiveness of the QA program developed for a
particular study to determine if the data collected will meet this primary  goal.  However, to conduct
an audit there must first exist assessment criteria against which the  study can be evaluated. For the
field of OP/FTIR monitoring, this presents some challenges. First, the instrumentation and  data-
analysis methods are evolving.  Secondly, research conducted in the area of QA has not yet
produced agreed-upon procedures. In addition to these issues, many QA auditors presently have
limited experience in conducting audits on OP/FTIR systems, given that the use of this instrumenta-
tion for USEPA-sponsored studies is fairly new.
       In spite of these challenges, an audit was  conducted on a field study using OP/FTTR systems.
Audit procedures and criteria were developed using various reference materials available to the QA
auditor. The Agency-wide QA requirements, the laboratory's  Quality Management Plan, and
routine auditing procedures  were used as general guidance in conducting the audit and in developing
the audit aupstionnaire.  The more technical aspects of the audit questionnaire were based on the

-------
QA/QC research conducted in support of the laboratory and the study documents written by’the field
teams, such as Work Plans, QA Project Plans, and Standard Operating Procedures (SOPs).
DISCUSSION
There are various organizational units within the USEPA that develop QAJQC policies,
procedures, methods, and materials. The QA auditor uses this information to develop the proce-
dures and materials used in the audit. The information used to establish the framework for the
OP/FuR audit and questionnaire are discussed in the sections below.
Agency-wide OA Requirements
The policies and requirements that serve as the basis for much of the Agency’s QA activities
are developed by the Quality Assurance Division (QAD), formerly known as the Quality Assurance
Management Staff. QAD is under the newly formed National Center for Environmental Research
and Quality Assurance of the Office of Research and Development.
The policies established by EPA Order 5360. r 3 and the guidance developed by QAD set forth
those QA activities which are required and those which are recommended for a QA Program. The
required activities include the development of Quality Management Plans and QA Project Plans and
the assessment of studies through QA audits and reviews. The recommended activities include the
development of Data Quality Objectives (DQOs) and SOPs. QAD defines these QA activities and
other key terms in specific ways and provides useful requirement- and guidance documents as
reference materials for the auditor and auditee. 48 A brief summary of the material used for the
OP/FuR audit and questionnaire is given in Table 1.
Ouality Management Plan of the Laboratory
Although the OP/FFIR study that was audited was not sponsored by the Atmospheric
Research and Exposure Assessment Laboratory (AREAL) in Research Triangle Park, NC, the QA
audit team came from this laboratory. Therefore, the Quality Management Plan of AREAL was
used to govern parts of the audit. Recently, AREAL was reorganized and now forms part of the
National Exposure Research Laboratory (NERL). Until the NERL Quality Management Plan is
completed, the one for AREAL will continue to be implemented.
AREAL’s Quality Management Plan incorporates the Agency-wide requirements and further
defines these requirements as they apply to the project needs of the laboratory. According to this
plan: I) the required document of preference is the Quality Systems Implementation Plan (QSIP),
which combines the QA Project Plan, Work Plan, and DQOs into one document, as opposed to the
QA Project Plan; 2) the use of the DQO process is required unless the project meets a specified
ct iteria or it receives an exemption by senior management; and 3) the development of SOPs is
required of all routine and repetitive activities of a project, even those projects categorized as
Research and Development.
QAIOC Research Conducted by the Laboratory
The Methods Branch and Quality Assurance Branch of NERL conduct and direct various
QA/QC research and development activities. General QA Handbooks, developed by the Quality
Assurance Branch, are applicable to many types of environmental studies. ” Various documents
and articles directly related to OP/FTIR systems have been produced in support of the Methods
Branch) 2 ’ 4 Although others have addressed QA/QC issues related to OP/FTIR monitoring,’ 5 ’ 6 the
research conducted for the Methods Branch has produced generally-applicable QA/QC procedures,
which formed the technical basis for the audit questionnaire.

-------
THE QA AUDIT AND QUESTIONNAIRE
The audit conducted on the OPIFTIR field study was done in a manner most often used by
the staff auditors of the Quality Assurance Branch of NERL. This general audit protocol is
normally followed:
1. Gather and review the necessary study documents from the field team, such as Work
Plans, QA Project Plans, QSIPs, SOPs, DQOs, etc.
2. Gather arid review other documeuts pertinent to the study being audited.
3. Conduct pre-audic meetings or briefings.
4. Develop audit materials, such as questionnaires, surveys, checklists, test samples, etc.
based on information gathered in Steps #1, #2, and //3.
5. Conduct the.audit. This includes on-site observations, interviews, and tests.
6. Conduct post-audit meetings or briefings.
7. Compile the audit fmdings.
8. Write the draft audit report and distrib ute it for internal review and clearance.
9. Distribute the draft audit report to the field team for review and comments.
10. Compile the responses to the draft audit report.
ii. Write the final audit report and distribute it for internal review and clearance.
12. Distribute the final audit report to the field team and other appropriate individuals.
As Step #4 indicates, many items contribute to the development of the audit materials. The
study documents prepared by the field team are essential for this purpose. Because of this, it is
important that: 1) they are written according to Agency and laboratory guidelines; 2) they contain
only those activities intended for implementation, and 3) they are made available to the auditor in a
timely manner. Before the study documents are submitted to the QA auditor, they should be
reviewed by someone who is familiar with the study method and the appropriate level of QA/QC
needed to assure the quality of the data. For example, the review of written plans for OPIFrIR
systems should be made by someone familiar with this technology. Under extraordinary circum-
stances, it may be necessary to revise the study documents once in the field. If this occurs, the
approved revision should be provided to the auditor so that they can make the appropriate changes
to their audit materials.
The audit questionnaire developed for the OPIFTIR study was generally based on the
Agency-wide QA requirements and the AREAL Quality Management Plan. It was technically based
on the QAIQC research conducted on behalf of the Methods Branch and Quality Assurance Branch
and the Work Plans, QA Project Plans, and SOPs written by the field teams. Table 2 shows this
audit questionnaire in its entirety except for those portions only relevant to the field teams that were
audited. The questionnaire is divided into these major sections:
I. General Information: This section gathers some basic information about the field
study, auditor(s), and field team. This section asks for names, addresses, and years
of experience of the field operator(s).
II. OP/FuR Instrument: This section gathers information about the type of instrument
used in the field study. This section asks for the manufacturer, model, type, age, and
owner of the instniment.
Ill. Study Documentation: This section determines if the documentation requirements
were met. The section asks about the study’s classification, QA Project Plan, SOPs,
log book, and control charts. In the future, this section may be expanded to include
inquiries about other documents, such as DQOs and QSIPs.

-------
IV. Pre-Field Activities: This section surveys the measurements recorded and procedures
used by the operator prior to employing the instrument in the field to: 1) determine
compliance with the prescribed procedta:es’ 2 ’ 4 and 2) determine if the operator estab-
lished a pre-fleld database against which the measurements collected in the field may
be compared. This section poses questions about detector saturation, return-beam
intensities, stray light, black body radiation, electronic noise, RMS noise, and
resolution.
V. Study Activities: This section surveys the QC and data acquisition procedures em-
ployed in the field to determine compliance with the prescribed procedures. ’ 2 14 The
first part of this section inq&res about QC activities involving electronic noise, RMS
noise, return-beam intensities at specific wavenumbers, stray light, black body
radiation, detector saturation, wavenumber shifts, selected resolution, precision,
accuracy, and the use of a QA cell. In the other two parts, questions are posed about
the end use of the data, how the detecior was cooled, the operating path length for the
field study, if the beam was attenuated, the spectral resolution and measurement time
used, and how the background and water vapor spectra were obtained.
VI. Data Analysis: This section surveys the methods used to analyze the field spectra.
This section inquires about the form in which the data were collected (interferogram
files or single-beam spectra), the target compounds of the field study, the wave-
number regions, apodization, zero filling, reference spectra, and how the stray light
and black body measurements will ‘ndled in the data analysis.
This audit questionnaire was tested in the field and proved to be a useful tool in the
evaluation process. It provided a standard approach in assessing each field team’s performance an’i
compliance with specific audit criterIa. As the technology of OP/PTIR moves forward through
continued research and field use, this questionnaire will be refined to reflect the advancements.
CONCLUSIONS
The quality of environmental databases is important to the USEPA. The Agen:y has
developed extensive QA requirements and guidance to ensure that the data collected are of a known
quality; of the amount, quality, and type needed to meet pre-defmed objectives; and adequate to
support the decisions that the Agency will make based on the data. Those performing work for the
USEPA, including contractors, must design and implement an adequate QA Program that is in
compliance with the Agency-wide QA requirements, the sponsoring- or auditing laboratory’s Quality
Management Plan, and generally accepted QA/QC procedures contained in the scientific literature.
The audit questionnaire introduced in this paper indicates some of the QA/QC issues that
should be considered for a QA program for those employing remote sensing techniques, such as
OP/FTIR monitoring. Until other QA/QC procedures are developed Agency-wide or within this
particular research field, this type of questionnaire will be used as the primary assessment tool for
studies audited by NERL.
ACKNOWLEDGEMENTS
The author wishes to thank the following individuals for providing some of the background
information used in this paper: Bill Mitchell, David Hinton, Linda Porter, Ronald Patterson,
Johnnie Pearson, Nancy Weitworth, Gary Johnson, and Tom Dickson.
REFER!2NCES
1. Organizational Dynamics, the; The Quality Course For EPA, 1991; pp 22-29.

-------
2. Wentworth, N. and Johnson, 0.; U.S. Environmental Protection Agency, Washington, D.C.,
course prcsentation, 1992.
3. U.S. Environmental Protection Ager’y; Policy and Program Requirements to hnpleinent the
Mandatory Quality Assurance Program, EPA Order 5360.1, 1984.
4. U.S. Environmental Protection Agency; EPA Requirements for Quality Managetnent Plans,
Doc. No. EPA QA/R-2, 1994.
5. U.S. Environmental Protection Agency; EPA Requirements for Quality Assurance Project
Plans for Environmental Data Operations, Doc. No. EPA QA/R-5, 1994.
6. U. S. Environmental Protection Agency; Orientation to Quality Assurance Management,
Course Notebook, 1995.
7. U.S. Environmental Protection Agency; Guidance for the Data Quality Objectives Process,
Doc. No. EPA QA/G-4, 1994.
8. U.S. Environmental Protection Agency; Guidance on Standard Operating Procedures for
Environmental Data Operations, Doe. No. EPA QA/G-6, in preparation.
9. U.S. Environmental Protection Agency; Quality Assurance Handbook for Air Pollution
Measurement Systems, Vol. 1, A Field Guide to Environmental Quality Assurance, Doe. No.
EPAJ600IR-94/038a, 1994.
10. U.S. Environmental Protection Agency; Quality Assurance Handbook for Air Pollution
Measurement Systems, Vol. II, Ambient Air Specific Methods (interim Edition), Doe. No.
EPA/600/R-94/038b, 1994.
11. U.S. Environmental Protection Agency; Quality Assurance Handbook for Air Pollution
Measurement Systeris. Vol. 11<, Meteorological Measurements (as revised August, 1989),
Doe. No. EPA/6001R-941038d, 1994.
12. Russwunn, G.M. and Childers, J.W.; Fr-JR Open-Path Monitoring Guidance Document, SP-
4420-95-04; ManTech Environmental Technology, mc: Research Triangle Park, 1995.
13. Russwurm, G.M., Childers, J.W., and Jr., Thompson, E.L.; Effects of stray light in FTIR
instruments on open path measurements, TM in Proceedings of Optical Sensing for Environmen-
tal and Process Monitoring, VIP-37; Air & Waste Management Association: Pittsburgh,
1995; pp 339-346.
14. Childers, J.W., Russwurm, G.M., and Jr., Thompson, E.L.; TM Quality assurance consider-
ations in a long-term FT-IR monitoring program, • in Proceedings of Optical Sensing for
Environmental and Process Monitoring, VIP-37; Air & Waste Management Association:
Pittsburgh, 1995; pp 389-395.
15. Kagann, R.H., Jolley, J.G., Shoop, D.S., et al.; •Vali jon of open-path FTIR data at
treatment, storage, and disposal facilities, u in Proceedings of Optical Sensing for Environ-
mental Monitoring, SP-89; Air & Waste Management Association: Pittsburgh, 1993; pp 437-
442.

-------
16. Kricks, R.J., Scotto, R.L., Pritchett, T.H. et al.; “Quality assurance issues concerning the
cperation of open-path FuR spectrometers,” in Proceedings of Optical Remote Sensing.
Applications to Environmental and Industrial Safety Problems, SP-81; Air & Waste Manage-
ment Association: Pittsburgh, 1992; pp 224-231.
Copies of the Agency’s requirement- and guidance documents can be obtained from U.S. EPA, ORD Publications Office, Center for
Environmental Research Information (CERI), 26 W. Martin Luther King Drive, Cincinnati, OH, 45268, phone: 513-569-7562.
Table 1. Terminology and information relevant to Agency-wide QA Activities.
• Quality Management Plan: “A formal document that describes the quality system in terms of
the organizational structure, functional responsibilities of management and staff, lines of
authority, and required interfaces for those planiiing, implementing, and assessing all
activities conducted.” The laboratory Quality Management Plan is prepared and approved by
the senior management of the laboratory and then reviewed and approved by QArJ. Once
approved, it is valid for up to five years but must undergo an internal review on an annual
basis. The ten elements required in a Quality Management Plan, formerly called a QA
Program Plan, are discussed in a new document, EPA QA/R-2. 4
• QA Project Plan: “A formal document describing in comprehensive detail the necessary QA,
QC and other technical activities that must be implemented to ensure that the results of the
work performed will satisfy the stated perfonnance criteria.” The QA Project Plan serves as
the blueprint for conducting the study and collecting and assessing the data. Its development
is the responsibility of USEPA personnel, but the actual preparation may be assigned to
groups outside the Agency, such as contractors, an assistance agreement holder, or interagen-
cy agreement holder. Prior to the start of any data collection, the QA Project Plan must be
reviewed and approved by the USEPA Project Manager and QA Manager. The new
ãocument EPA QA/R-5 discusses the elements that must be addressed in a QA Project Plan. 5
There are now twenty-six required elements instead of the previous sixteen. Appendix A in
EPA QAIR-5 lists the former sixteen elements and their new counterparts.
• Data Quality Objectives (DQOs): “Qualitative and quantitative statements derived from the
DQO Process that clarify study tecimical and quality objectives, define the appropriate type of
data, and specify tolerable levels of potential decision errors that will be used as the basis for
establishing the quality and quantity of data needed to support decisions.” The development
of DQOs is strongly recommended but is not mandatory. If deemed necessary, a USEPA
organization may add them as a special requirement to the QA Project Plan. The new
guidance document for DQOs, EPA QA/G-4, 7 supersedes all previous guidance.
• Standard Operating Procedures (SOPs): “A written document that details the method for an
operation, analysis, or action with thoroughly prescribed techniques and steps, and that is
officially approved as the method for performing certain routine or repetitive tasks.” The
development of SOPs is strongly recommended but is not mandatory. If deemed necessary, a
USEPA organization may add them as a special requirement to the QA Project Plan. SOPs
must be reviewed by individuals with appropriate training and experience and approved by at
least the immediate USEPA supervisor.’ The new guidance for SOPs will be contair d in
EPA QA/G-6.’ 4

-------
Table 2; Audit questionnaire for OP1FTJ.R Systems.
I. General Information
OPIFFIR Audit Questionnaire
Date(s) of Field Audit:
Title of Field Study:
Name(s) of Auditor(s)
Business Address:
Name(s) of Field Operator(s):
Years of Experience working with OP/FTIR Instruments:
Business Address:
II. OP/FflR Instrument
Manufacturer:
Model:
Bistatic Monostatic Age of OPIFFIR Instrument:
Owner of OP/FflR Instrument:
III. Study Documentation
If a question requires a YES/NO answer place an “X” in the appropriate box.
YES NO
1. How would you classif ’ this field study?
data can stand alone, legally defensible, for compliance purposes
data will be used for rule making or policy making
exploratory/preliminary work, data will be used tc design future
purely a research and development project
2. Do you have a Quality Assurance Project Plan for this study?
studies
f
.
I
tf yes, to whom was it submitted?
By whom was it reviewed?
By whom was it approved?
If no, why not?
3. Do you have Standard Operating Procedures for this instrument 9
4. Do you keep a log book for this instrument 9
If yes, what items do you record in this log book?
5. Do you use control charts for this instrument?
I
If yes, what measurements do you record on these control charts?

-------
OPIFTIR Audit Questionnaire
IV. Pre-Fieki Activities - procedures performed prior to going to the field
If a question requires a YES/NO answer place an AX” in the appropriate box. NO
1. In general, how do you determine if the detector is saturated?
2. If you see evidence of saturation what remedial action do you take?
3. Did you determine the path length at which the detector is no longer
saturated (the minimum working distance of the instrument)’)
If yes, what is thispath length?
What procedure did you use?
4. Did you determine the path length at which t.he return-beam intensity
levels off or reaches the noise level (the maximum working distance of
the instrument) 9
If yes, what is this path length?
What procedure did you use?
5. Were the path lengths used for this study between the minimum and
maximum working distances of your instrument?
6. Did you determine the return-beam intensities at specific wavenumbers as
a function of path length?
If yes, what procedure and wavenumbers did you use?
7. Did you determine the return-beam intensities at these wavenumbers?
987 cm’ If yes, at what path length(s)?
2500 cm’ If yes, at what path length(s)?
4400 cm’ If yes, at what path length(s)? ______
8. Did you measure for stray light?
If yes, how did you measure the stray light? ___________
9. Do you know what percentage of the return signal is due to stray light?.
If yes, what is this percentage?
10. Do you routinely subtract the signal due to stray light from the single-
beam spectra 7
11. Did you measure for black body radiation?
If yes, how did you measure the black body radiation?
12. Do you know what percentage of the total signal is due to black body
radiation 9
If yes, what is this percentage? ______ ____
13. Do you routinely subtract the signal due to black body radiation from the
single-beam spectra 7

-------
OP/FflR Audit Questionnaire
If a question requires a YES/NO answer place an X” in the appropriate box.
14. Did you measure the electronic noise of the instrument 9
If yes, how did you measure the electronic noise?
15. Did you measure the random or RMS noise 9
If yes, over what wavenuinber range was this measurement taken?
How did you measure the RMS noise?
16. Did you determine •if the instrument is operating at the selected
resolution 9
If yes, how was this determined?
17. Describe any other pre-field activities that you performed on the instrument.
I
I
V. Study Activities - procedures performed while in the field
A. Quality Control Procedures
1. Did you measure any of the following parameters for QC purposes:
a. Electronic Noise’ I
If yes, how often?
What procedure did you use?
b. Random or RMS Noise 9
If yes, how often?
What procedure did you use?
c. Return-Beam Intensities at Specific Wavenumbers?
If yes, how often?
What procedure did you use?
d. Stray Light”
If yes, how often?
What procedure did you use?
e. Black Body Radiation 9
If yes, how often?
What procedure did you use?
2. Did you check for detector saturation at the operating path length?
If yes, how often?
What procedure did you use?
3. Did you check for possible wavenumber shifts?
If yes, how often?
What procedure did you use?
I
I
I
I
I
I
I
I
I
I
YES INO

-------
OP/Jciix Audit Questionnaire _____ ____
If a question requires a YES/NO answer place an “X” in the appropriate box. YES NO
4. Did you determine if the instrument was operating at the selected
resolution?
If yes, how often?
What procedure did you use?
5. How did you determine the precision of the instrument?
_____ I used the ambient level of N 2 0.
_____ I used the ambient level of CH 4 .
_____ I used an internal QA cell and known concentrations of gases.
Other: ________________
How reliable, repeatable, and reproducible is the procedure that was used?
If you used a QA cell, what target gases and concentrations did you use? What was the
maximum absorbance value of the target gases?
6. How did you determine the accuracy of the instrument?
_____ I used the ambient level of N 2 0.
_____ I used the ambient level of CH 4 .
_____ I used an internal QA cell and known concentrations of gases.
_____ I used an external QA cell and known concentrations of gases.
____ Other: _______________
How reliable, repeatable, and reproducible is the procedure that you used?
If you used a QA cell, what target gases and concentrations did you use? What was thern
maximum absorbance value of the targc gases?
7. What level of precision and accuracy did you obtain with the instrument?
Precision:
Accuracy:
8. What level of precision and accuracy is needed for your end-use?
Precision:
Accuracy:
9. Describe any other QA/QC checks that were perfonned on this instrument? S ie how
often these were performed and give a brief description of the procedure(s) used.
B. Acquisition of Field Spectra
1. What type of data were you responsible for collecting during this study? What will be
its end-use?
2. Was your detector cooled with liquid nitrogen? I I
If yes, how often did you have to refill the Dewar?
If no, how was the detector cooled?

-------
OP/F IR Audit Questionnaire
If a question requires a YES/NO answer place an M X” in the appropriate box.
3. What was the operating path length used in this study?
How was the path length measured?
4. Did you use screens or other means to attenuate the beam while
collecting data?
If yes, of what material are they made?
5. What spectral resolution did you use?
How did you determine the resolution to use?
6. What measurement time did you use?
C. Acquisition of Background Spectra and Water Vapor Spectra
1. Did you take background spectra at the field site 9
If yes, how often?
If no, what background spectra do you plan to use?
2. What method(s) did you use to obtain the background spectra (Ia)?
______ synthetic
_____ upwind
_____ short path
______ averaged
____ Other: _______________
. I
I
What are the limitations of the method(s) used, if any? ____________
3. When acquiring the ambient blackbody background spectra did you see
the presence of emission bands for atmospheric gases? ___________
If yes, how will these be handled when analyzing your data?
4. What method did you use to obtain the water vapor spectra?
_____ obtained from long path ZR cell in the Iaboratoiy
_____ generated from one of the field spectra by subtracting out other ambient gases
_____ used a commercial library
_____ used the HI-TRAN datat.se to calculate spectrum
____ Other: _______________
5. How did you handle the changing concentration of water vapor? ___________
6. Did you record the partial pressure (in torr) of the water vapor? I
If yes, how often?

-------
OP/FuR Audit Questionnaire
VI. Data Analysis
question requires a YES/NO an.nver place an “X” in the appropriate box. I YES I NO
When collecting data did you store the interferogram fiics or the single-
beam spectra?
If you saved the interferogram files, will they also be saved after
conversion to single-beam spectra’
List all the target compounds for this field study.
What techniques will you use to determine the concentration of the target compounds?
_____ comparison method
_____ scaled subtraction technique
_____ multicomponent analysis
____ Other:
Over what wavenumber regions will you analyze for the target compounds?
Note: This questionnaire was compressed to meet the symposium’s paper-length requirement. Normally, sufficient space is
provided to record all answers.
If a
1.
2.
3.
4.
Why did you choose these wavenumber regions to analyze?
5. What kind of apodization will you use?
6. Will you use any zero filling 9 I
7. How will the signal due to stray light be handled in the data analysis?
8. How will the signal due to black body radiation be handled in the data analysis?
9. What is the source of your reference spectra?
If you develop your own, enclose a write-up of your procedure.

-------
DISCLAIMER:
The information in this document has been funded by the
Environmental Protection Agency. It has been subjected to the
Agency’s peer and administrative review, and it has been approved
for publication as an EPA document. Mention of trade names or
commercial products does not conL;titute endorsement or
recommendation for use.

-------