EPA-600/2-76-083 March 1976 Environmental Protection Technology Series DEVELOPMENT AND TRIAL FIELD APPLICATION OF A QUALITY ASSURANCE PROGRAM FOR DEMONSTRATION PROJECTS Industrial Environmental Research Laboratory Office of Research and Development U.S. Environmental Protection Agency Research Triangle Park, North Carolina 27711 ------- RESEARCH REPORTING SERIES Research reports of the Office of Research and Development, U.S. Environmental Protection Agency, have been grouped into five series These five broad categories were established to facilitate further development and application of environmental technology. Elimination of traditional grouping was consciously planned to foster technology transfer and a maximum interface in related fields. The five series are: 1. Environmental Health Effects Research 2. Environmental Protection Technology 3. Ecological Research 4. Environmental Mqnitorihg 5,: ,-,Soe'i6ecjonomic Environmental Studies This report-,has. p&ek -assignee! to the ENVIRONMENTAL PROTECTION TECHNOLOGY series. This'series describes research performed to develop and demonstrate instrumentation, equipment, and methodology to repair or prevent environmental degradation from point and non-point sources of pollution. This work provides the new or improved technology required for the control and treatment of pollution sources to meet environmental quality standards EPA REVIEW NOTICE This report has been reviewed by the U. S. Environmental Protection Agency, and approved for publication. Approval does not signify that the contents necessarily reflect the views and policy of the Agency, nor does mention of trade names or commercial products constitute endorsement or recommendation for use. This document is available to the public through the National Technical Informa- tion Service, Springfield. Virginia 22161. ------- EPA-600/2-76-083 March 1976 DEVELOPMENT AND TRIAL FIELD APPLICATION OF A QUALITY ASSURANCE PROGRAM FOR DEMONSTRATION PROJECTS by James Buchanan Research Triangle Institute P.O. Box 12194 Research Triangle Park, NC 27709 Contract No. 68-02-1398, Task 20 ROAP No. ABA-011 Program Element No. EHB-557 EPA Task Officer: L. D. Johnson Industrial Environmental Research Laboratory Office of Energy, Minerals, and Industry Research Triangle Park, NC 27711 Prepared for U.S. ENVIRONMENTAL PROTECTION AGENCY Office of Research and Development Washington, DC 20460 ------- ACKNOWLEDGMENTS The work on this project was performed by the Systems and Measurements Division of the Research Triangle Institute. Mr. Frank Smith, Supervisor, Quality Assurance Section, served as the project leader. Dr. James Buchanan of the Quality Assurance Section was responsible for the coordination of the program. Institute staff members Dr. D. E. Wagoner and Mr. Larry Hackworth, analytical chemists, Mr. Leon Bissette, an electrical engineer, and Dr. Buchanan, a physical chemist, were major contributors to the program. Project officer for the Environmental Protection Agency was Dr. L. D. Johnson of the Process Measurements Branch of the Industrial Environmental Research Laboratory. The Research Triangle Institute acknowledges the cooperation and assistance of the project officer and Dr. R. Statnick of the Process Measure- ments Branch. The Institute also appreciates the assistance and guidance provided by Mr. John Williams, the EPA project officer for the Shawnee wet limestone scrubber demonstration. Finally, gratitude is extended to Mr. Joe Barkley and Mr. Ken Metcalf of TVA and to Mr. Dewey Burbank of Bechtel Corporation for their cooperation at the Shawnee test site. iii ------- TABLE OF CONTENTS SECTION 1.0 INTRODUCTION ] 2,0 MAJOR COMPONENTS OF A QUALITY CONTROL PROGRAM 3 2-1 QUALITY ASSURANCE ASPECTS OF THE RFP 3 2-2 EVALUATION OF QUALITY CONTROL IN THE PROPOSAL 3 2-3 EVALUATION OF QUALITY CONTROL IN THE WORK PLAN 4 2-4 MANAGEMENT COMMITMENT TO QUALITY CONTROL 4 2-5 QUALITY CONTROL IN THE ORGANIZATIONAL STRUCTURE 4 2-6 ASSESSMENT OF QUALITY CONTROL REQUIREMENTS 5 2.7 SPECIFIC AREAS OF CONCERN FOR DEMONSTRATION PROJECT QUALITY CONTROL PROGRAMS 5 2-7.1 FACILITIES AND EQUIPMENT 6 2-7.2 CONFIGURATION CONTROL 6 2.7.3 PERSONNEL TRAINING 6 2.7.4 DOCUMENTATION CONTROL 7 2.7.5 CONTROL CHARTS 7 2-7.6 IN-PROCESS QUALITY CONTROL 7 2-7.7 PROCUREMENT AND INVENTORY PROCEDURES 9 2-7.8 PREVENTIVE MAINTENANCE 9 2-7.9 RELIABILITY 9 2.7.10 DATA VALIDATION 9 2-7.11 FEEDBACK AND CORRECTIVE ACTION 10 2-7.12 CALIBRATION PROCEDURES 10 lv ------- TABLE OF CONTENTS (CON,) SECTION PAGE 3,0 GUIDELINES FOR DEMONSTRATION PROJECT QUALITY ASSURANCE PROGRAMS ii 3-1 GENERAL STATEMENTS H 3-2 THE ON-SITE QUALITATIVE SYSTEMS REVIEW H 3-3 THE PERFORMANCE AUDIT 12 3-4 MATERIAL BALANCES 12 3-5 ASSESSMENT OF DATA QUALITY 12 3-6 ASSESSMENT AND MODIFICATION OF THE ONGOING QUALITY ASSURANCE PROGRAM 14 4,0 A SHORT-TERM QUALITY ASSURANCE PROGRAM IMPLEMENTED AT THE SHAWNEE SCRUBBER FACILITY 17 4.1 THE CONTROL LABORATORY 17 4.1-1 MEASUREMENT OF pH 17 4.1-2 SLURRY ANALYSIS 19 4.1.3 OVERALL LABORATORY EVALUATION 24 4-2 GAS STREAM SAMPLING 24 4.2.1 PARTICULATE MASS LOADING 24 4.2.1.1 PITOT TUBE COMPARISON 25 4.2.1.2 TEMPERATURE MEASUREMENT 25 4.2.1.3 MOISTURE MEASUREMENT 25 4.2.1.4 VOLUME MEASUREMENT 26 4.2.2 SULFUR DIOXIDE CONCENTRATION DETERMINATIONS 26 4-3 PROCESS INSTRUMENTATION 27 4-4 RECOMMENDATIONS 29 v ------- TABLE OF CONTENTS (CON,) SECTION 5,0 EVALUATION OF THE SHORT-TERM QUALITY ASSURANCE PROGRAM AT SHAWNEE 31 5-1 QUALITATIVE SYSTEMS REVIEW 31 5-2 QUANTITATIVE PERFORMANCE AUDIT 31 5-2.1 SCHEDULING 31 5.2.2 EQUIPMENT AND INSTRUMENTATION 32 5-2.3 PERSONNEL SELECTION 33 APPENDIX A QUALITY AUDIT CHECKLIST FOR DEMONSTRATION PROJECTS 36 APPENDIX B STANDARD TECHNIQUES USED IN QUANTITATIVE PERFORMANCE AUDITS 61 APPENDIX C COMPARISON ON ANALYSIS OF LIMESTONE SLURRY 65 ------- LIST OF TABLES TABLE NO. PAGE 1 COMPARISONS OF pH 18 2 MEAN VALUES FOR SLURRY ANALYSES, BY LABORATORY 21 3 COMPARISON OF SHAWNEE RESULTS WITH MEAN VALUE OF COOPERATING LABORATORIES 23 4 COMPARISON OF S0¥ DETERMINATIONS 28 LIST OF FIGURES FIGURE NO. 1 STANDARD QUALITY CONTROL CHART, Vii ------- 1.0 INTRODUCTION The major objective of this project was to develop a general quality assurance (QA) program for EPA demonstration projects, using the wet limestone scrubber facility at the Shawnee steam plant, Paducah, Kentucky, as an example project. A second objective was to field test the QA program at the Shawnee facility and carry out whatever modifications were necessary in light of that field trial. Two concurrent final reports deal separately with these objectives and should be consulted for further treatment of areas of interest mentioned here. These reports are: 1. Guidelines for Demonstration Project Quality Assurance Programs, EPA-600/2-76-081. 2. A Quality Assurance Program for the Environmental Protection Agency Wet Limestone Scrubber Demonstration Project, Shawnee jiteam-Electric Plant, Paducah, Kentucky, EPA-600/2-76-080. Reference will occasionally be made to these reports. For conciseness, reports 1. and 2. will be referred to as the guidelines report and the Shawnee report, respectively. Organization of this report is as follows: section 2.0 discusses signifi- cant areas for quality control (QC) from the RFP to the day-to-day QC program; section 3.0 outlines a quality assurance program for demonstration projects; section 4.0 treats the specific project studied, the wet limestone scrubber at the Shawnee steam plant; and section 5.0 is an evaluation of the QA program implemented at Shawnee. The major areas investigated at Shawnee were the analytical (control) laboratory, the gas stream measurement systems, and the process monitoring and control instrumentation. To facilitate the reading of this report, two terms should be carefully defined. These terms are quality control and quality assurance. Concise definitions are given herewith. Quality control: the overall system of activities the purpose of which is to provide a quality of product or service that meets the needs of users. Quality assurance: A system of activities the purpose of which is to provide assurance that the overall quality control job is being done effectively ------- Recommendation of a quality control system as such does not fall within the scope of this project. It is important, however, to be aware of the elements of such a system if one is to act as a monitor of QC by means of QA work. For this reason, the first major area addressed by this report is the QC system itself. ------- 2.0 MAJOR COMPONENTS OF A QUALITY CONTROL PROGRAM 2.1 Quality Assurance Aspects of the RFP The design of the RFP is predicated on stating as clearly as possible what the objectives of the project are; e.g., to design, construct, and maintain a given control system, systematically examining the interaction of appropriate system parameters. The quality of the data obtained from the project will depend upon numerous factors—instrumentation, personnel, sampling technique, sampling size, statistical expertise. It is therefore critical that the RFP be as explicit as possible* in delineating two things—what quality data are expected, and how that quality is to be insured. Since most RFP's are limited in length, it would usually be inappropriate to include more than a brief (one- or two-paragraph) statement of QC require- ments. Nevertheless, it is most important that the bid solicitation be as explicit as possible concerning QC. 2.2 Evaluation of Quality Control in the Proposal The proposal should contain a statement as to the precise position the bidder's company takes regarding quality control programs. This should include past projects and the quality control program effectiveness in that project. In particular, there should be a clear and explicit response to the QC require- ments stated in the RFP. This response must be compared directly, item-by-item, with other proposals submitted against the RFP. The evaluation should result in a determination of a "figure of merit" for the bidder's quality control organization and the competence of the staff. If a contractor has a good proposal but is unclear on some phases of data quality, it would seem worthwhile to have him clarify his proposal by asking him to answer specific questions. If the answers to these questions are still vague, it is a good indication that the quality for these phases of the project may be questionable if this contractor carries out the project. 'it is understood that, because of the nature of the proposed work, it may not be possible to specify either the expected data quality or the way in which data quality is to be insured. ------- 2.3 Evaluation of Quality Control in the Work Plan The work plan should be a detailed accounting of the actual steps to be taken to complete the work delineated in the proposal and should be in direct accord with the requirements of the RFP and other agreements with the project officer. Particular attention should be placed on mutually agreed upon critical areas in order to realize the collection of data having acceptable precision, accuracy, representativeness, and completeness. In cases where the submitted proposal has been accepted but lacks the completeness required by the project officer, finalized negotiations to remove the problem areas should be directly addressed in the work plan showing the details of the work to be done. The work plan must be submitted to the project ofJ'icer before any work is begun by the contractor. The plan can be accepted in draft form, which will allow for minor changes prior to the final plan's acceptance and approval. 2.4 Management Commitment to Quality Control No quality control program, regardless of the amount of planning or level i of effort expended, will be effective without the explicitly visible support of top management. The support should be expressed initially as the project gets underway and periodically throughout the duration of the program. The support of top management then filters down through middle and lower management to the operators, resulting in a program where QC is practiced on a day-to-day basis, rather than being an additional program or nuisance. Quality control must be built-in, functional area within the total program, and this is not possible without continuing obvious management support. 2.5 Quality Control in the Organization Structure Support for quality control is most visible when the organizational struc- ture has provision for personnel whose authority and responsibilities lie in the area; i.e., a quality control coordinator (QCC) and/or any other staff appropriate to the program. The QCC is responsible for the organization's entire QC program, and this person's judgment determines the effectiveness of the program. The basic function of the QCC should be the fulfillment of the QC objective of management in the most efficient and economical manner ------- commensurate with insuring continuing completeness, accuracy, and precision of the data produced. The responsibilities and authority of the QCC are detailed in the Quality Assurance Handbook for Air Pollution Measurement Systems, Vol. 1, Principles, EPA-600/9-76-005.* The QCC should have, within the main organizational structure, a subordi- nate organization for QC activities (auditing, calibration, quality control). He should have authority for assignment of QC duties and for coordination of the entire program, and must not be directly subordinate to operational person- nel in the project. 2.6 Assessment of Quality Control Requirements The establishment of a QC program for a demonstration project requires first of all the setting, in as quantitative a manner as possible, of project objectives. The desired precision and accuracy of each measurement should be specified, as well as the technical means of attaining this degree of data quality; i.e., the tasks to be performed. Once this is done, it is efficient to group the tasks organizationally and assign responsibility for the QC function to each task group. It is inevitable that problems are incurred in each step of the planning and establishment of the QC program. Some of these cannot be resolved until the program enters the functional stage. What is important initially is that these problems be identified and clearly stated, so that they can be resolved as quickly as possible once the program gets underway. 2.7 Specific Areas of Concern for Demonstration Project Quality Control Programs A quality control program for a demonstration project serves to: 1. Evaluate the overall adequacy of the project insofar as data quality is concerned; 2. Identify potential as well as existing problems in the data-producing system, from measurement to data reduction; Some of the general discussion of QC programs in this report has been taken from this document. ------- 3. Stimulate research into and discussion of alternative methods for obtaining data of the required quality. It is advisable to delineate a number of important aspects of the project which have direct bearing on data quality, and to discuss each of these in some detail. 2.7.1. Facilities and Equipment An obvious beginning point in the assessment of an ongoing program is a general survey of the facilities and equipment available for day-to-day opera- tion of the project. Are they adequate for the job at hand? Do standards exist for evaluation of facilities, equipment, and materials? The laboratories, data processing sections, and other operation areas should be neat and orderly, within common-sense limits imposed by the nature of the facility. A neat, well-organized laboratory area serves to inspire neatness and organization among the laboratory workers. Good laboratory maintenance, particularly for certain types of instru- mentation, requires complete manuals, kept in a convenient place so that they are readily available to appropriate personnel. Responsibility for keeping up with all necessary manuals should be given to an individual, with the under- standing that he must devise a system (checkin-checkout) for quick location of each document. 2.7.2 Configuration Control The documentation of design changes in the system must be carried out unfailingly. Procedures for such documentation should be written, and be accessible to any individual responsible for configuration control. 2.7.3 Personnel Training It is highly desirable that there be a programmed training system for new employees. This system should include motivation toward producing data of acceptable quality standards. This is to be preferred to on-the-job training, which may be excellent or slipshod, depending upon a number of circumstances. ------- A thorough personnel training program should focus particular attention on those people whose work directly affects data quality (calibration personnel, bench chemists, etc.). These people must be cognizant of the quality standards fixed for the project and the reasons for those standards. They must be made aware of the various ways of achieving and maintaining quality data. As these people progress to higher degrees of proficiency, their accomplishments should be reviewed and then documented. 2.7.4 Documentation Control Procedures for making revisions to technical documents must be clearly written out, with the lines of authority indicated. The revisions themselves should be written and distributed to all affected parties, thus insuring that the change will be implemented and will become permanent. 2.7.5 Control Charts Control charts are essential as a routine day-to-day check on the consis- tency or "sameness" of the data precision. A control chart should be kept for each measurement that directly affects the quality of the data. Typically, control charts are maintained for duplicate analyses, percent isokinetic sampling rate, calibration constants, and the like. An example control chart is given as figure 1. The symbol a (sigma) represents a difference, d, of one standard deviation unit in two duplicate measurements, one of which is taken as a standard, or audit value. Two cr is taken as a warning limit and 3a as a control limit. 2.7.6 In-Process Quality Control During routine operation, critical measurement methods should be checked for conformance to standard operating conditions (flow rates, reasonableness of data being produced, and the like). The capability of each method to pro- duce data within specification limits should be ascertained by means of appro- priate control charts. When a discrepancy appears in a measurement method, it should be analyzed and corrected as soon as possible. ------- 30 2a -0 -20 -30 CHECK NO. ACTION LIMIT • XL WARNING LIMIT -CL WARNING LIMIT ACTION LIMIT •— LCL 8 10 DATE/TIME OPERATOR PROBLEM AND CORRECTIVE ACTION Figure 1, Standard quality control chart. ------- 2.7.7 Procurement and Inventory Procedures There should be well-defined and documented purchasing guidelines for all equipment and reagents having an effect on data quality. Performance specifi- cations should be documented for all items of equipment having an effect on data quality. In the case of incoming equipment, there should be an established and documented inspection procedure to determine if procurements meet the quality assurance and acceptance requirements. The results of this inspection procedure should be documented. Once an item has been received and accepted, it should be documented in a receiving record log giving a description of the material, the data of the receipt, results of the acceptance test, and the signature of the responsible individual. It is then placed in inventory, which should be maintained on a first-in, first-out basis. 2.7.8 Preventive Maintenance It is most desirable that preventive maintenance procedures be clearly defined and written for each measurement system and its support equipment. When maintenance activity is necessary, it should be documented on standard forms maintained in log books. A history of the maintenance record of each system serves to throw light on the adequacy of its maintenance schedule and parts inventory. 2.7.9 Reliability The reliability of each component of a measurement system relates directly to the probability of obtaining valid data from that system. It follows that procedures for reliability data collection, processing, and reporting should be clearly defined and in written form for each system component. Reliability data should be recorded on standard forms and kept in a log book. If this procedure is followed, the data can be utilized in revising maintenance and/or replacement schedules. 2.7.10 Data Validation Data validation procedures, defined ideally as a set of computerized and manual checks applied at various appropriate levels of the measurement process, ------- should be clearly defined, in written form, for all measurement systems. Cri- teria for data validation must be documented. The required data validation activities (flow-rate checks, analytical precision, etc.) must be recorded on standard form in a log book. Any demonstration project should, on a random but regular basis, have quality audits performed by in-house personnel. These audits must be independ- ent of normal project operations, preferably performed by the QCC or appointees of the QCC. The audits should be both qualitative and quantitative (i.e., they should include both system reviews and independent measurement checks). For the system review, a checklist is desirable to serve as a guide for the reviewer. Such a checklist is included as appendix A of this report. The quantitative aspect of the audit will vary depending on the nature of the project. Some guidelines for quantitative audits are given in appendix B. 2.7.11 Feedback and Corrective Action Closely tied to the detection of invalid data is the problem of establish- ment of a closed loop mechanism for problem detection, reporting, and correction. Here it is important that the problems are reported to those personnel who can take appropriate action. A feedback and corrective action mechanism should be written out, with individuals assigned specific areas of responsibility. 2.7.12 Calibration Procedures Calibration procedures are the crux of any attempt to produce quality data from a measurement system. For this reason it is extremely important that the procedures be technically sound and consistent with whatever data quality re- quirements exist for that system. Calibration standards must be specified for all systems and measurement devices, with written procedures for assuring, on a continuing basis, traceability to primary standards. Since calibration personnel change from time to time, the procedures must be, in each instance clearly written in step-by-step fashion. Frequency of calibration should be set and documented, subject to rescheduling as the data are reviewed. Full documentation of each calibration and a complete history of calibrations per- formed on each system are absolutely essential. This permits a systematic review of each system reliability. 10 ------- 3,0 GUIDELINES FOR DEMONSTRATION PROJECT QUALITY ASSURANCE PROGRAMS 3.1 General Statements The objective of quality assurance is to independently assess the quality control program of the project. This assessment should normally take two major forms: (1) a qualitative audit (systems review), and (2) a quantitative per- formance audit. These are discussed in detail below, as sections 3.2 and 3.3 respectively. The frequency of a qualitative and/or a performance audit obviously should be dictated by the specific project. It is recommended that a minimum frequency be once each calendar year. The initial systems review and performance audit should take place within the first quarter of the first project year. Subse- quent scheduling should be dependent on the requirements of management and the apparent quality of the day-to-day data being obtained. More frequent auditing may be necessary in the initial stages of the project. 3.2 The Qualitative Audit The objective of the qualitative audit is to assess and document facilities; equipment; systems; recordkeeping; data validation; operation, maintenance, and calibration procedures; and reporting aspects of the total quality control pro- gram for demonstration projects. The review should accomplish the following: 1. Identify existing system documentation—i.e., maintenance manuals, organizational structure, operating procedures, etc; 2. Evaluate the adequacy of the procedures as documented; 3. Evaluate the degree of use of and adherence to the documented procedures in day-to-day operations based on observed conditions (auditor) and a review of applicable records on file. To aid the auditor in performing the review, a checklist is included as appendix A. This checklist will allow for systematic appraisal of the areas mentioned above. 11 ------- 3.3 The Performance Audit In addition to a thorough on-site qualitative audit, quantitative perform- ance audits should be periodically undertaken at each demonstration project. The objective of these audits is to evaluate the quality of project data by independent measurement techniques. It is convenient to classify the major measurement methods into three areas: physical measurements, gas stream measurements, and liquid stream measures (the latter including analysis of any suspended solids). Appendix B lists in matrix form a number of standard tech- niques for auditing in the three major areas just mentioned. Table 1 of appendix B is a compilation of commonly measured physical properties, with a selection of possible measurement, calibration, and audit techniques. Table 2, concentrating on analysis of gas effluent streams, lists the material to be analysed and measurement, calibration, and audit techniques for that material. Finally, table 3 very briefly and generally deals with measurement methods appropriate to liquids and solids. The specific techniques vary widely from project to project, but the audit technique generally involves use of control (reference) samples of known composition and/or splitting a sample among several laboratories for independent analyses. 3.4 Material Balances Material balances serve as a gross indication of the quality of the total measurement system complex of the project. The extent of closure will be directly related to the precision and bias of each measurement taken. In general, both physical measurements of flow rates, temperatures, pressured (and so on), and chemical analysis of material composition will bear on the degree of closure attained. The achievable extent of closure must be estimated for each project and used as a target figure. The frequency with which material balances are run is related to how successful one is in attaining the estimated closure over a significant period of time. 3.5 Assessment of Data Quality Standard methods exist for estimation of the precision and accuracy f measurement data. Efficient usage of the audit data requires that a ratio 12 ------- be followed which gives the best possible estimates of precision and accuracy within the limits imposed by timing, sample size, etc. For a given measurement, the difference between the field (or plant) and the audited results, is used to calculate a mean and standard deviation as follows: n d =Y^ d,/n , E n E - d)2/(n - where d is an estimate of the bias in the measurements (i.e., relative to the audited value). Assuming the audited data to be unbiased, the existence of a bias in the field data can be checked by the appropriate t-test, i.e., sd/n If t is significantly large, say greater than the tabulated value of t with n - 1 degrees of freedom, which is exceeded by change only 5 percent of the time, then the bias is considered to be real and some check should be made for a possible cause of the bias. If t is not significantly large, then the bias should be considered zero, and the accuracy of the data is acceptable. The standard deviation, s,, is a function of both the standard deviation d of the field measurements and of the audit measurements. Assuming the audit values to be much more accurate than the field measurements, then s is an estimate of a{x}, by using the statistical test procedure 13 ------- . 7 -T S {X} o where X /f is the value of a random variable having the chi-square distribution with f = n - 1 degrees of freedom. If X /f is larger than the tabulated value exceeded only 5 percent of the time, then it would be concluded that the test procedure is yielding more variable results due to faulty equipment or opera- tional procedure. The measured values should be reported along with the estimated biases, standard deviations, the number of audits, n, and the total number of field tests, N, sampled (n < N). Estimates (such as s, and d) which are significantly — a different from the assumed population parameters should be identified on the data sheet. 2 The t-test and X -test described above are used to check on the biases and standard deviations separately. Other statistical techniques exist which may apply to specific projects (or to highly specialized areas of a given project). It is usually worthwhile to acquire the services of a statistical consultant in order to more effectively treat the available data. 3.6 Assessment and Modification of the Ongoing Data Quality Program The guidelines put forth in the preceding sections serve as a basis for development of a data quality program specific to the needs of a particular project. A program should not be attempted without a thorough study of the entire facility, supplemented by at least one site visit. It is to be desired that provision for QC be made from the project's inception. The EPA project officer's responsibility is then to see that a program of adequate QC practices is incorporated into the day-to-day project operations, along with periodic QA audits conducted by outside organizations. Implementation of the program, at whatever point in the lifetime of the project, exposes weaknesses of approach and problems that were not anticipated in the planning stages. Certainly it is necessary to maintain maximum flexi- bility of approach as the interface with project realities is made. The 14 ------- Shawnee report documents several problems of implementation and suggests pro- cedures for avoiding those problems in future efforts. They are also discussed in the final section of this report. Generally, one should expect that a degree of modification would be re- quired in the areas listed below: 1. audit instrumentation and general equipment requirements, 2. sampling frequencies, and 3. audit personnel requirements. Experience must always be the final judge of the effectiveness of a data quality program, as one monitors data quality on a continuing basis. 15 ------- 4.0 A SHORT-TERM QUALITY ASSURANCE PROGRAM IMPLEMENTED AT THE SHAWNEE SCRUBBER FACILITY A quality assurance program was implemented at the wet scrubber facility, Shawnee steam Plant, Paducah, Kentucky. This program was carried out by per- sonnel of the Research Triangle Institute, Research Triangle Park, North Carolina. The program consisted of two site visits of approximately 1 week each, occurring October 28-31 and November 17-21, 1975. The first visit was largely occupied with the qualitative audit, using as a guide the checklist provided as appendix A. An evaluation was made of equipment needs for the second visit, when the quantitative performance audit was conducted. This section presents the results of both the qualitative review and the performance audit, with recommendations for a QA program at the facility. 4.1 The Control Laboratory 4.1.1 Measurement of pH One of the major elements for control of the scrubber is pH measurement. For this reason a significant part of the RTI effort was spent in observing and verifying the TVA techniques for pH determination at scrubber inlet and outlet. A portable pH system is used several times a day by TVA operators to check the control room readings obtained from inline pH sensors. These devices are permanently situated in pots through which slurry continuously circulates when the scrubber is operating. Four such sensors monitor the inlet and outlet pH for the TCA and Venturi scrubbers. A series of direct comparison pH measurements were made on November 19-20, 1975. RTI and TVA personnel made simultaneous measurements of slurry pH. The operators were instructed to carry out their measurements routinely, from standardization to cleanup. Also during this period the inline probes were removed from their pots and immersed in pH 5 and 6 buffers. Both RTI and TVA long-lead probes were put into these same buffers, after independent standard- ization. The values obtained in the buffers and in slurry are summarized in table 1. Measurements to the nearest 0.01 pH unit were made using the RTI Acumet pH meter. Readings of such precision were not possible with the TVA 17 ------- Table I. Comparisons* of pH TEST POINT 18161 18252 2816^ 2S254 nH RTI (portable) 5.25 5.13 c** 5.05s 5.24 5.28 5.34 5.06 4.92 4.83 4.77 5.035 6!025 5.035 6.016 nH TVA' (portable) 5.3 5.2 4.9 5.2 5.2 5.3 5.1 4.9 4.9 4.8 ...-,-_- 6.0 5.0 6.0 TVA (inline) 5.16 5.29 5.04 5.34 5.38 5.34 5.25 5.17 5.02 5.08 4.82 5.77 4.99 6.12 Temperature (°Centrigrade) 54 54 16 54 50 51 50 53 54 50 21 21 21 21 Date (mo/day/yr) 11/19/75 11/19/75 11/20/75 11/20/75 11/20/75 11/20/75 11/20/75 11/19/75 11/19/75 11/20/75 11/19/75 11/19/75 11/19/75 11/19/75 Tine (hours) 11:15 15:30 08:45 09:10 09:45 11:15 15:15 11:30 15:30 15:15 10:00 10:30 70:50 11:00 I-1 00 1 Venturi effluent hold tank. 2 Venturi outlet. 3 TCA effluent hold tank. 4 TCA outlet. Unless otherwise noted, measurements are on slurry in inline probe pots. ** Superscript number indicates measurement of a buffer solution of pH 5 or 6. ------- Orion pH meter, where estimates of 0.1 pH unit were made. Actual scale mark- ings on the TVA meter were at 0.2-pH unit intervals. Two observations are in order, after study of table 1: 1. The RTI and TVA portable (long-lead) pH systems agreed within 0.1 unit or better in every comparison made. This verifies the accuracy of the TVA portable system, since the RTI system was standardized against certified buffer. Operator reading errors are probably the largest error source. 2. The inline system readings differed from RTI readings from 0 to 0.3 pH unit, with the mean difference being 0.14 pH unit over 14 readings. Certain- ly it would be unwise to dwell on the significance of the statistics of such a brief study. One point can be made, however, with respect to the confidence placed in the inline pH readings, as follows: it appears unlikely that, using the present system, pH measurements on the slurry can be made to better than 0.1 pH unit. There are a number of factors which militate against greater accuracy, the major one probably being the nature of the slurry itself. This viscous, highly abrasive suspension tends to clog lines, coat out on probe surfaces, etc., making reproducible measurements quite difficult. The non- equilibrium mixture of reactive chemicals has a pH that will change on removal from the scrubber proper; i.e., as it flows into the pots within which measure- ments are made. Another factor is the difficulty of standardization of inline probes. The present system calls for probe removal, cleaning, and standardization roughly each 2 days. The accuracy of the pH reading is surely dependent on the condition of the probe surface, and restandardizing is ideally done shortly before each measurement. 4.1.2 Slurry Analysis As a check on the reliability of the chemical analysis phase of the scrubber operation, a series of slurry samples was collected* and sent to several other laboratories for independent analysis of both the liquid phase and suspended solids. *A11 samples were taken from the venturi effluent hold tank. 19 ------- The laboratories originally selected for participation in this phase of the audit were two TVA laboratories (Chattanooga and Muscle Shoals), E/ETB (EpA-RTP), and RTI. The E/ETB laboratory later declined to participate in the project. Each laboratory was given five l-£ samples of slurry, which were taken concurrently with control laboratory samples. After filtering and drying the solid, it was analyzed for calcium, magnesium, and total sulfur. The filtrate was analyzed for calcium, magnesium, sodium, potassium, and chloride. Complete results of the analyses are given as appendix C. These results are given in matrix form, both by laboratory and by element. This report will present a limited statistical analysis of the data and will comment on the techniques used by each laboratory. The five slurry samples were taken over a 36-hour period, all from the effluent hold tank. There was little apparent change in the slurry composition over this period of time. Taking a simple numerical average of the five analy- ses for each element yields a number which itself has little significance, since it represents the combined effect of analytical uncertainty and slurry composi- tion change over 36 hours. The rationale for obtaining such an average is that, if the analytical technique exhibits a bias, this will result in a number that is correspondingly biased. Thus a comparison can be made among the participating laboratories and the various analytical techniques. Table 2 is a matrix of these averages, with the analytical technique used by the laboratory given underneath the number. A few observations are in order: 1. Analysis results for calcium in the solid were extremely close among the Shawnee, Chattanooga, and Muscle Shoals laboratories, but RTI obtained a considerably lower value by AA. The Shawnee XRF standard value for calcium was established by sending por- tions of the standard to various TVA laboratories, including the Chattanooga and Muscle Shoals facilities. Also, not that these two laboratories both used the same technique, EDTA titration. Further work would be required to determine which technique, EDTA or AA, is inherently more accurate. 2. Results for magnesium show the expected large variation for an element present in low concentration. 20 ------- Table 2. Mean values for slurry analyses, by laboratory ^XLABORATORY ELEMENT ^s^^ Ca (CaO) wt % in solid Mg (MgO) wt % in solid Total Sulfur (S03) wt % in solid Ca, ppm in liquid Mg, ppm in liquid Na, ppm in liquid K, ppm in liquid Cl, ppm in liquid SHAWNEE 23.19 m 0.29 Wf 31.99 m 1929 AA 697 AA 71 AA 128 AA 3580 Pot. Titration CHATTANOOGA 22.93 EDTA 0.60 EDTA 30.00 BaSO, ppt. 1700 EDTA 772 EDTA 65 FE 108 FE 3601 Volhard MUSCLE SHOALS 23.16 EDTA 0.25 AA 30.74 BaS04 ppt. 1759 EDTA 736 AA 39 FE 56* FE 3680 \gNO, titration RTI 19.72 AA 0.42 AA 33.01 BaS04 ppt. 1800 AA 932 AA 74 AA 106 AA 3780 Volhard XRF = X-ray fluorescence EDTA = Ethylenediaminetetracetic acid titration BaSCL ppt. Precipitation as barium sulfate AA = Atomic absorbance FE = Flame emission * Value discarded, second RTI value (different RTI laboratory) of 111 ppm used for statistical purposes. ------- 3. Total sulfur determinations were consistent between Chattanooga and Muscle Shoals; RTI's value was high and Shawnee's came in between, with no large discrepancies. 4. Results in the liquid phase were relatively consistent, although the Muscle Shoals laboratory obtained extremely low numbers for sodium and potassium. Chattanooga, which also used flame emission, got results consistent with the AA determinations of Shawnee and RTI. 5. Chloride determinations showed good consistency across all laboratories, although an interesting sidelight is that some preliminary determinations at RTI, using a chloride-sensing electrode, gave results that were high by roughly 100 percent. This result was duplicated by a second, non-RTI laboratory, indicating that the electrical environment of the slurry liquid phase was unsuitable for chloride determination by ion-selective electrode. Table 3 singles out the Shawnee control laboratory results for comparison with the mean results from the other three laboratories. Close agreement exists for the most critical elements—calcium (both liquid and solid), total sulfur (solid), and chloride (liquid). Magnesium in the solid showed the greatest variation among the comparison laboratories (43 percent) and between Shawnee and the mean of the comparison laboratories (31 percent). Across the board, these results indicate the Shawnee control laboratory is performing routine analyses at about the +20 percent level, referenced to comparable laboratories. In particular, it appears to be obtaining rather good accuracy (+5 to 10 per- cent) in its calcium, sulfur, and chloride analyses. These results are con- sistent with material balance closures of 5 to 15 percent that have been repeatedly obtained by Bechtel Corporation. It should be stressed that the percentages quoted above are simple esti- mates based upon direct comparisons between the Shawnee results and the mean of the cooperating laboratories' results. They are not confidence levels. It would be appropriate to carry out detailed statistical work only if consider- ably more data were obtained. 22 ------- Table 3. Comparison of Shawnee results with mean value of cooperating laboratories ELEMENT Ca (CaO) wt * in solid Mg (MgO) wt % in solid Total Sulfur (SO,) wt % in solid Ca, ppm in liquid Mg, ppm in liquid Na, ppm in liquid K., ppm in liquid Cl , ppm in liquid SHAWNEE RESULTS 23.19 0.29 31.99 1929 697 71 128 3580 MEAN OF CHATTANOOGA, MUSCLE SHOALS & RTI RESULTS 21.94 0.42 31.25 1753 813 59 108* 3687 STANDARD DEVIATION ABOUT MEAN 1.92 0.18 1.57 50 104 18.2 . 2.5 90 COEFFICIENT OF VARIATION ABOUT MEAN (*) 8.8 43 5.0 2.9 12.8 31 2.3 2.4 (SHAWNEE/MEAN) x 100 -100 (*) + 5.7 - 31 + 2.4 + 10.0 - 14.3 + 19.7 + 18.5 - 2.9 U) * Low Muscle Shoals results thrown out, second RTI result (from another RTI laboratory) of 111 ppm used in averaging. ------- A final point is that the Shawnee ionic imbalances for solid and liquid were typically 2 to 3 percent and 10 to 15 percent respectively, both biased negatively, during the sampling period. The close solid ionic balance indi- cates either accurate analyses or balancing positive and negative ion analytical errors. Results of this study appear to validate the accuracy of the analyti- cal procedures used on the slurry solid. The negative ionic imbalance in the liquid phase analyses cannot be rationalized by the data obtained from this comparison study. 4.1.3 Overall Laboratory Evaluation The control laboratory operation appears to be adequate for the routine analytical work it performs. It has no formal quality control program, but bad data may be flagged by either TVA or Bechtel personnel. Acceptance limits on data are not formalized, but "reasonableness" is the experience-based criterion. There are problems associated with the lack of operator training programs, incentives for superior performance and the like, but so long as the laboratory operations remain strictly routine these problems are not likely to seriously hamper the program. Equipment and instrumentation is appropriate for the type of work done, and it is maintained on a regular basis (largely by service contracts). 4.2 Gas Stream Sampling 4.2.1 Particulate Mass Loading Side-by-side duplicate runs were not attempted. The entire sampling procedure was observed, with critical techniques checked repeatedly, during the site visits. Overall performance was evaluated using a checklist. On a scale of 1 to 5, ranging from unacceptable to excellent, the Shawnee particulate loading technique was rated 3 (acceptable). A major problem appeared to be the failure of Shawnee personnel to carry out adequate leak-checking of the sampling train. Specific comments are given below. 24 ------- 4,2.1.1 Pitot tube comparison. A comparison of TVA and RTI pitot tubes was performed at the Venturi inlet only. A check of the outlet tube was not carried out due to the outlet tube misalignment (> 30°) along its roll axis. Side-by-side measurements were performed. Based upon comparison with the RTI (NBS calibrated) pitot tube, the TVA tube Cp factor was 0.879. The assumed value was 0.850. The difference was considered to be negligible. 4.2.1.2 Temperature measurement. A system capable of measuring the stack gas temperature to within 1.5 per- cent of the minimum absolute stack temperature is required. The temperature- measuring system (inlet sampler) was checked versus a calibrated thermocouple and was found to be within 1 percent. 4.2.1.3 Moisture measurement. The impinger section of the EPA sampling train is intended to collect moisture from the sample gases for determination of moisture content. The last impinger contains silica gel to adsorb the water vapor not condensed in the first two impingers. The moisture content of the sample gas leaving the silica gel impinger increases as the exit gas temperature rises. Also, the exit gas moisture content will increase as the sample train vacuum increases at any one sample temperature. Moisture not collected by the condensation system is incorrectly measured as dry gas by the dry test meter and the error is carried through the isokinetic and grain loading calculations. However, if the exit gas temperature is held below 25° C and the rain vacuum is held below 380 mm of Hg, the resulting error in the sample volume will be less than 2 percent. A single RTI reading of exit gas temperature was 22° C. There was evidence of significant moisture accumulation in the silica gel, indicating the presence of some water vapor in the total gas volume measured. This does not likely introduce a large error into the technique, although it would be advisable to make quantitative or semiquantitative checks on the actual water volume collected versus water content of the stack gas. This is not presently being done at Shawnee. 25 ------- 4.2.1.4 Volume measurement. The sampling train was checked for accuracy of volumetric measurement with a calibrated dry test meter (1 cf/revolution) which had been previously cali- brated versus a 1-cf wet test meter. The RTI meter was connected directly to the Shawnee probe tip, so that the actual volume intake at the probe was measured. RTI volume was 15.8 percent lower than TVA volume, indicating a rather large positive bias in the TVA measurement. Critical examination of the TVA sampling system led to the conclusion that the bias could be attributed to leaks in the system (broken or cracked polycarbonate impinger tubes, loose probe tip, etc.). Leakage rate was estimated to be 0.36 cfm at 380 mm of Hg vacuum.* Inaccuracies in volume measurements appear directly in the concentra- tion and particulate mass emission rate determinations. A probe tip diameter check was made with a micrometer. The range of the diameter measurements was 0.7 mm, indicating a severely out-of-round nozzle which should be repaired or replaced. The estimated nozzle area was calculated 2 to be roughly 20 percent lower than the assumed area (0.583 cm calculated, 2 0.7125 cm assumed). An error in the nozzle diameter is quadrupled in the process of determining isokinetic sampling rates and is doubled in the percent of isokinetic sampling calculation. The percent isokinetic, as calculated with respect to the above errors in volume measurement and nozzle diameter could result in either a positive or negative bias, depending upon which factor predominates. 4.2.2 Sulfur Dioxide Concentration Determinations Sulfur dioxide concentrations at the wet limestone scrubber facility are determined by means of du Pont Model 400 photometric analyzers. The analyzers continuously monitor inlet and outlet gas streams of the venturi and TCA units. The RTI audit team collected a total of 23 gas samples, all collected at the venturi inlet. Thirteen of these samples were analyzed by a modified barium chloranilate (colorimetrie) method, the remaining 10 by sodium hydroxide TVA leak-checking was not observed by the audit team. A thorough leak-check would surely have detected such a significant leak-rate. 26 ------- titration. Results are given In table 4. The average bias of the photometric method with respect to the wet chemical methods was +6.9 percent, with a stand- ard deviation of 8.7 percent. These results indicate that the du Pont analyzer at the venturi inlet is yielding data of high quality. At the 95 percent con- fidence level, an individual photometric determination should have a precision of +18 percent of the mean concentration, biased 7 percent high on the average.* Due to the time limitation of the audit team, it was not possible to run checks on the other three analyzers. 4.3 Process Instrumentation A series of calibration checks on the electronic instrumentation was scheduled. Some checks and observations were not carried out because of TVA personnel work schedules, but enough was accomplished for a judgment to be made as to the quality of the instrumentation facilities. Three types of sensors (temperature, differential pressure, and flow rate) are the primary sources of measurement information being recorded and used for the scrubber's mechanical operation control, and four readout devices are employed for visual display of the output signals. The methods of test and calibration are simple, using rudimentary sources of stimuli for sensor exam- ination. Straightforward electrical current measuring instruments are used to monitor currents produced by the transmitters. As performed, the tests are sufficient to maintain the quality of measurement to the degree established by the manufacturers in their design specifications. The performance of equipment over time can best be judged by a review of accurate records which clearly show a life history of each item having a functional part in the operation of a system. The Shawnee facility was judged deficient in recordkeeping for its instrumentation. In spite of this, it was felt that the electronic devices used for physical measurements were being maintained sufficiently to provide pressure, level, and flow information to a +2 percent tolerance of desired nominal values, and temperature information to a +10 percent tolerance of desired information (temperature sensors can be This assumes no bias in the wet methods. 27 ------- Table 4. Comparison of SO determinations X Sampling Train Mo.l Date 11/18/75 11/18/75 11/1.9/75 11/19/75 11/19/75 11/19/75 11/19/75 11/19/75 11/19/75 11/19/75 11/19/75 00 11/19/75 11/19/75 11/19/75 11/20/75 11/20/75 11/20/75 11/20/75 11/20/75 11/20/75 11/20/75 11/20/75 11/20/75 Sample Number 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 Sample Time 17:15-17:25 17:35-17:45 10:15-10:23 10:42-10:50 11:07-11:21 12:08-12:17 12:43-12:49 12:53-13:02 13:39-13:47 13:50-13:59 14:22-14:31 14:32-14:41 15:06-15:15 15:17-15:27 09:16-09:27 09:30-09:40 09:53-10:04 10:05-10:15 10:22-10:33 10:41-10:52 10:58-11:09 11:11-11:22 11:37-11:48 SO by Barium Cliloranllate (ppm) 4187 2736 Sample Voided Sample Voided 2261 2172 2309 1958 1582 1827 1750 1724 1690 S0x Contained in Isopropranol Scrubber (ppm) 543 650 530 563 530 551 523 643 610 541 637 Total 4730 3386 2791 2735 2839 2509 2105 2470 2360 2265 2327 SO by* NaOH Tit rat ion 2562 2252 3053 2832 2812 2757 2435 2385 2459 2294 S02 by TVA (DuPont Analy 2750 2750 3625 3585 3366 3167 3051 3046 3016 2985 2998 2947 2893 2909 2506 2509 2505 2508 2570 2600 2606 2648 2719 TVA-RTI RTI X 10° zer) >7.3 +9.0 - 23.3 +5.9 -0.1 +9.1 +6.5 +9.1 +6.6 +3.8 +4.9 + 15.9 + 19.0 +3.0 +1.4 +5.2 +8.9 +5.7 + 15.1 + 15.4 + 16.8 SO, analyzer accepted on a total acid determination (TVA analysis) Average Bias: +6.9% Std. Dev.: 8.69% 95% Confidence Interval + 18.08% ------- calibrated to a +2 percent tolerance of a known temperature—the inaccuracies are estimated to be high because of the lack of knowledge of the thermodynamics of the stack gases being measured). 4.4 Recommendations Specification of a quality control program for the Shawnee scrubber project does not fall within the scope of this report. The recommendations made in the following paragraphs apply to the implementation of a (qualitative and perform- ance audit) QA program.* This type of program normally should be carried out by an organization which has no special interest in the data; i.e., no self- interest to protect and no preconceptions as to the quality of the information forthcoming. On the other hand, the organization should be reputable and well qualified to carry out the type of auditing program desired. In the case of EPA demonstration projects, EPA may wish to contract a third party to handle the audit program, or it may handle the program by means of its own QA staff. In either case, it is quite important that the auditing be done competently and objectively. It is recommended that the wet limestone scrubber operation located at the Shawnee steam-electric plant be externally audited twice each calendar year. Timing of the audit program, which normally should take 1 work week, should be coordinated among the auditing team, EPA, TVA, and the Bechtel Corporation. Some advance notice is necessary in order to insure cooperation of operational personnel. It is not recommended that the audits be scheduled on a regular basis, since by definition an audit is conducted without extensive "preparation" at the project being audited. Advance notice to EPA and Bechtel supervisory staff should be at least 2 weeks, so that the audit team can be apprised of special test and analysis schedules which may alter its audit procedure or cause postponement of the audit itself. Advance notice to TVA staff (senior chemist, instrumentation foreman) should be at least 1 week. *It is important that the Shawnee project continue to develop its own QC program internally. 29 ------- It is recommended that the audit team concentrate its efforts in the following major areas: 1. Verification of pH measurements at inlet and outlet, on both TCA and venturi scrubbers. An accurate pH meter brought in by the audit team, with appropriate buffer solutions, should be used. Measurement of pH is critical to efficient process control at this facility. 2. Independent chemical analysis of slurry samples by several laboratories. A continuing audit program can aid in establishing acceptance limits and method biases. 3. Verification of particulate mass loading and sulfur dioxide measurement systems. If possible, side-by—side operation of TVA and audit team sampling trains should be carried out, with independent analyses of the collected samples. A wet chemical technique such as total acid titration should be used to check the SCL analyzer response. If a duplicate sampling train cannot be used by the audit team, then critical measurement parameters should be identified and checked. For stack sampling procedure, this includes (at a minimum): a. Sample volume measurement check by means of a calibrated wet test or dry gas meter;* b. Pitot tube (C_ factor) check by means of an NBS calibrated pi tot tube; c. Thermometer and thermocouple checks with a calibrated temperature measurement system; d. Stack gas moisture content check by means of an absorbing impinger train. 4. Electronic checks on process instrumentation and physical measurement techniques. For suggestions as to techniques available, see "Process Stream Volumetric Flow Measurement and Gas Sample Extraction Methodology," by Brooks and Williams. This manual (TRW Document No. 24916-6028-RU-OO) was prepared under EPA Contract No. 68-02-1412, for the Process Measurements Branch of IEBL. 30 ------- 5.0 EVALUATION OF THE SHORT-TERM QUALITY ASSURANCE PROGRAM AT SHAWNEE The objective of this project was to devise, implement, and modify a general quality assurance program for IERL demonstration projects. It is ap- propriate then to evaluate the procedures which RTI used in its qualitative and performance audit at the Shawnee scrubber facility. 5.1 Qualitative Audit The audit checklist (appendix A) is a valuable device for assuring a balanced review of each major area. For the control systems laboratory opera- tions, the checklist was quite useful. It is recommended that either this checklist or a similar one be used each time a review is done. One note of caution is in order. No checklist can substitute for experience and common sense. Each facility will require a somewhat different approach by a QA team. The checklist alone will not suffice. It will be particularly appropriate with certain phases of a project, but in other areas a good part of the questionnaire will not be applicable. 5.2 Quantitative Performance Audit The quantitive performance audit at Shawnee emphasized several potential problem areas of which an audit team should be cognizant. These are discussed in the subsections below. 5.2.1 Scheduling If samples are to be taken for round robin purposes, it is important that the laboratory being audited perform complete analyses on each sample to be distributed to outside laboratories. This requires close supervision by the member of the auditing team responsible for sample collection. At Shawnee, the sampling schedule (at the time of the audit) called for analysis of the solid at certain times of the day, and both solid and liquid at other times. Because of the preoccupation of the senior chemist with routine operations, this schedule was not made clear and several samples (which required complete analyses for comparison purposes) were collected at times calling for solid 31 ------- analysis only. When this was discovered, it was necessary to ship some surplus samples (which RTI was fortunate to have kept) back to the Shawnee laboratory for liquid analysis. Had the surplus samples not been available, it would have been necessary to collect a second series of samples at another time. This example serves to point out the need for complete understanding, on the part of both senior laboratory personnel and audit team personnel, of the requirements of the audit. A related scheduling problem occurred when the Shawnee instrumentation foreman was unable to provide RTI with the assistance needed to complete cali- bration and testing of the facilities' electronic process monitoring equipment. It is of overriding importance that efficient scheduling of audit activities occur, since the total time allowed is normally a few days to 1 week. Un- scheduled operational problems will occur which could not have been anticipated, and these must be handled as expeditiously as possible. 5.2.2 Equipment and Instrumentation It should be unnecessary to emphasize the desirability of having reliable, high-quality audit equipment and instrumentation. Where feasible, the audit team should have duplicate (or at least equivalent) backup items of that equip- ment which is crucial to the audit. For example, if pH measurements are to be made, two probes should be on hand. If possible, two meters should also be available. During the Shawnee audit, a measurement problem developed due to a shielded probe lead making electrical contact with the meter chassis. This problem had not been discovered earlier because the meter had previously been used only in nongrounded environments (such as glass beakers). The metal slurry pots were directly grounded to the scrubber framework and a relatively large extraneous signal was being read by the meter. This signal, by its magni- tude, swamped the circuit and pegged the meter dial each time a pH reading was attempted. Fortunately, the problem was diagnosed as a missing insulating washer, and a washer was fabricated in time to take a series of readings. It would have been much more desirable to have had a second meter immediately available, since pH comparisons were certainly an important aspect of this work. Spare glassware (and other fragile items) is obviously desirable, since in an unfamiliar environment laboratory workers are more prone to make mistakes resulting in breakage. 32 ------- 5,2.3 Personnel Selection An audit team must be competent and versatile. The RTI team assigned to the Shawnee project consisted of a physical chemist, an analytical chemist, and an electrical engineer. Each man knew his responsibility. An audit visit, because of its short duration, does not allow for inefficient use of personnel time or audit equipment. The services of the electrical engineer were useful when the pH meter malfunction was discovered, pointing up the advantages of a diversity of technical talent. A team composed entirely of chemists might not have been able to repair the meter in time to make the required number of measurements. 33 ------- APPENDIX A QUALITATIVE AUDIT CHECKLIST FOR DEMONSTRATION PROJECTS 35 ------- APPENDIX A QUALITATIVE AUDIT CHECKLIST FOR DEMONSTRATION PROJECTS This checklist is designed to: 1. Identify existing system documentation; i.e., maintenance manuals, organizational structure, operating procedures, etc. 2. Evaluate the adequacy of the procedures as documented. 3. Evaluate the degree of use of and adherence to the documented procedures in day-to-day operations based on observed conditions (auditor) and a review of applicable records on file. The checklist gives three descriptions to each facet of a quality control system. In all cases the "5" choice is the most desirable and effective mode of operation; "3" is marginal and tolerable; "1" is definitely unacceptable and ineffective as a mode of operation. It is not always possible to describe accurately all options with only three choices. Therefore, a "2" or "4" rating may be selected if the evaluator feels that an in-between score is more descriptive of the actual situation. After all the applicable questions are answered, an average is computed to give an overall indication of the quality system effectiveness. Generally, a rating of 3.8 or better is considered acceptable. A rating between 2.5 and 3.8 indicates a need for improvement but no imminent threat to project performance as it stands. For the control laboratory, the results are as follows: 1. Of 82 check questions, 65 were answered on site; 2. Average score was 3.0 (5.0 maximum), indicating a satisfactory but not outstanding program as presently operated; 3. The control laboratory was judged weak in its quality control organization, procurement, and inventory procedures, and in its personnel training policy; 4. Strong points were its day-to-day "in-process" quality assurance, its calibration procedures, and its facilities and equipment. The completed questionnaire, with indicated judgments in specific areas, is given herewith. These judgments are for the control laboratory operation only. 36 ------- A.I QUALITY ORGANIZATION SCORE (1.1) Overall responsibility for quality assurance (or quality control) for the organization is: (a) Assigned to one individual by title (e.g., Quality Control Coordinator). 5 (b) Assigned to a specific group within the organization. 3 (c) Not specifically assigned but left to the discre- tion of the various operational, analytical, inspec- tion, and testing personnel. 1 (1.2) The Quality Control Coordinator is located in the organization such that: (a) He has direct access to the top management level for the total operation, independent of others in- volved in operational activities. 5 (b) He performs as a peer with others involved in operational activities, with access to top manage- ment through the normal chain of command. 3 (c) His primary responsibility is in operational activities, with quality assurance as an extra or part-time effort. 1 (1.3) Data reports on quality are distributed by the Quality Control Coordinator to: (a) All levels of management.* 5 (b) One level of management only. 3 (c) The quality control group only. 1 (1.4) Data Quality Reports contain: (a) Information on operational trends, required actions, and danger spots. 5 (b) Information on suspected data/analyses and their causes. 3 (c) Percent of valid data per month. 1 *Management at appropriate levels in all applicable organizations such as subcontractors, prime contractor, EPA. 37 ------- A.2 THE QUALITY SYSTEM SCORE (2.1) The quality control system is: (a) Formalized and documented by a set of procedures which clearly describe the activities necessary and sufficient to achieve desired quality objec- tives, from procurement through to reporting data to the EPA/RTP. 5 (b) Contained in methods procedures or is implicit in those procedures. Experience with the materials, product, and equipment is needed for continuity of control. 3 (c) Undefined in any procedures and is left to the cur- rent managers or supervisors to determine as the situation dictates. 1 (2.2) Support for quality goals and results is indicated by: (a) A clear statement of quality objectives by the top executive, with continuing visible evidence of its sincerity, to all levels of the organization. 5 (b) Periodic meetings among operations personnel and the individual(s) responsible for quality assurance, on quality objectives and progress toward their achieve- ment . 3 (c) A "one-shot" statement of the desire for product quality by the top executive, after which the quality assurance staff is on its own. 1 (2.3) Accountability for quality is: (a) Clearly defined for all sections and operators/ analysts where their actions have an impact on quality. (b) Vested with the Quality Control Coordinator who must use whatever means possible to achieve quality goals. (c) Not defined. 38 ------- A.2 THE QUALITY SYSTEM (continued) SCORE (2.4) The acceptance criteria for the level of quality of the demonstration projects routine performance are: (a) Clearly defined in writing for all characteris- tics . 5 (b) Defined in writing for some characteristics and some are dependent on experience, memory and/or verbal communication. 3 (c) Only defined by experience and verbal communica- tion. 1 (2.5) Acceptance criteria for the level of quality of the project's routine performance are determined by: (a) Monitoring the performance in a structured pro- gram of inter- and intralaboratory evaluations. 5 (b) Scientific determination of what is technically feasible. 3 (c) Laboratory determination of what can be done using currently available equipment, techniques, and manpower. 1 (2.6) Decisions on acceptability of questionable results are made by: (a) A review group consisting of the chief chemist or engineer, quality control, and others who can render expert judgment. 5 (b) An informal assessment by quality control. 3 (c) The operator/chemist. 1 39 ------- A.2 THE QUALITY SYSTEM (continued) SCORE (2.7) The quality control coordinator has the authority to: (a) Affect the quality of analytical results by in- serting controls to assure that the methods meet the requirements for precision, accuracy, sensi- tivity, and specificity. 5 (b) Reject suspected results and stop any method that projects high levels of discrepancies. 3 (c) Submit suspected results to management for a decision on disposition. 1 A.3 IN-PROCESS QUALITY ASSURANCE (3.1) Measurement methods are checked: (a) During operation for conformance to operating conditions and to specifications, e.g., flow rates, reasonableness of data, etc. 5 (b) During calibration to determine acceptability of the results. 3 (c) Only when malfunctions are reported. 1 (3.2) The capability of the method to produce within specification limit is: (a) Known through method capability analysis (X-R Charts) to be able to produce consistently acceptable results. (b) Assumed to be able to produce a reasonably acceptable result. (c) Unknown. (3.3) Method determination discrepancies are: (a) Analyzed immediately to seek out the causes and apply corrective action. 5 (b) Checked out when time permits. 3 (c) Not detectable with present controls and procedures. 1 40 ------- A.3 IN-PROCESS QUALITY ASSURANCE (continued) SCORE (3.4) The operating conditions (e.g., flow rate, range, temperature, etc.) of the methods are: (a) Clearly defined in writing in the method for each significant variable. 5 (b) Controlled by supervision based on general guide- lines . 3 (c) Left up to the operator/analyst. 1 (3.5) Auxiliary measuring, gaging, and analytical instruments are: (a) Maintained operative, accurate, and precise by regular checks and calibrations against stable standards which are traceable to the U.S. Bureau of Standards. 5 (b) Periodically checked against a zero point or other reference and examined for evidence of physical damage, wear or inadequate maintenance. 3 (c) Checked only when they stop working or when ex- cessive defects are experienced which can be traced to inadequate instrumentation. 1 A.4 CONFIGURATION CONTROL (4.1) Procedures for documenting, for the record, any design change in the system are: (a) Written down and readily accessible to those individuals responsible for configuration con- trol. 5 (b) Written down but not in detail. 3 (c) Not documented. 1 41 ------- A.4 CONFIGURATION CONTROL (continued) SCORE (4.2) Engineering schematics are: (a) Maintained current on the system and subsystem levels. 5 (b) Maintained current on certain subsystems only. 3 (c) Not maintained current. 1 (4.3) All computer programs are: (a) Documented and flow charted. 5 (b) Flow charted. 3 (c) Summarized. 1 (4.4) Procedures for transmitting significant design changes in hardware and/or software to the EPA project officer are: (a) Documented in detail sufficient for implementation. 5 (b) Documented too briefly for implementation. 3 (c) Not documented. 1 A.5 DOCUMENTATION CONTROL (5.1) Procedures for making revisions to technical documents are: (a) Clearly spelled out in written form with the line of authority indicated and available to all involved personnel. 5 (b) Recorded but not readily available to all personnel. 3 (c) Left to the discretion of present supervisors/mana- gers. 1 42 ------- A.5 DOCUMENTATION CONTROL (continued) SCORE (5.2) In revising technical documents, the revisions are: (a) Clearly spelled out in written form and distrib- uted to all parties affected, on a controlled basis which assures that the change will be implemented and permanent. 5 (b) Communicated through memoranda to key people who are responsible for effecting the change through whatever method they choose. 3 (c) Communicated verbally to operating personnel who then depend on experience to maintain continuity of the change. 1 (5.3) Changes to technical documents pertaining to opera- tional activities are: (a) Analyzed to make sure that any harmful side effects are known and controlled prior to revision effectiv- ity. 5 (b) Installed on a trial or gradual basis, monitoring the product to see if the revision has a net bene- ficial effect. 3 (c) Installed immediately with action for correcting side effects taken if they show up in the final results. 1 (5.4) Revisions to technical documents are: (a) Recorded as to date, serial number, etc. when the revision becomes effective. 5 (b) Recorded as to the date the revision was made on written specifications. 3 (c) Not recorded with any degree of precision. 1 43 ------- A.5 DOCUMENTATION CONTROL (continued) SCORE (5.5) Procedures for making revisions to computer software programs are: (a) Clearly spelled out in written form with the line of authority indicated. 5 (b) Not recorded but changes must be approved by the present supervisor/manager. 3 (c) Not recorded and left to the discretion of the programmer. 1 (5.6) In revising software program documentation, the re- visions are: (a) Clearly spelled out in written form, with reasons for the change and the authority for making the change distributed to all parties affected by the change. 5 (b) Incorporated by the programmer and communicated through memoranda to key people. 3 (c) Incorporated by the programmer at his will. 1 (5.7) Changes to software program documentation are: (a) Analyzed to make sure that any harmful side effects are known and controlled prior to revision effectivity. 5 (b) Incorporated on a trial basis, monitoring the results to see if the revision has a net bene- ficial effect. 3 (c) Incorporated immediately with action for detecting and correcting side effects taken as necessary. 1 44 ------- A.5 DOCUMENTATION CONTROL (continued) SCORE (5.8) Revisions to software program documentation are: (a) Recorded as to date, program name or number, etc., when the revision becomes effective. 5 (b) Recorded as to the date the revision was made. 3 (c) Not recorded with any degree of precision. 1 A.6 PREVENTIVE MAINTENANCE (6.1) Preventive maintenance procedures are: (a) Clearly defined and written for all measurement systems and support equipment. 5 (b) Clearly defined and written for most of the measure- ment systems and support equipment. 3 (c) Defined and written for only a small fraction of the total number of systems. 1 (6.2) Preventive maintenance activities are documented: (a) On standard forms in station log books. 5 (b) Operator/analyst summary in log book. 3 (c) As operator/analyst notes. 1 (6.3) Preventive maintenance procedures as written appear adequate to insure proper equipment operation for: (a) All measurement systems and support equipment. 5 (b) Most of the measurement systems and support equip- o ment. J (c) Less than half of the measurement systems and sup- port equipment. *• 45 ------- A.6 PREVENTIVE MAINTENANCE SCORE (6.4) A review of the preventive maintenance records indicates that: (a) Preventive maintenance procedures have been carried out on schedule and completely documented. 5 (b) The procedures were carried out on schedule but not completely documented. 3 (c) The procedures were not carried out on schedule all the time and not always documented. 1 (6.5) Preventive maintenance records (histories) are: (a) Utilized in revising maintenance schedules, de- veloping an optimum parts/reagents inventory and development of scheduled replacements to minimize wear-out failures. 5 (b) Utilized when specific questions arise and for estimating future work loads. 3 (c) Utilized only when unusual problems occur. 1 A.7 DATA VALIDATION PROCEDURES (7.1) Data validation procedures are: (a) Clearly defined in writing for all measurement systems. 5 (b) Defined in writing for some measurement systems, some dependent on experience, memory, and/or verbal communication. 3 (c) Only defined by experience and verbal communica- tion. 1 46 ------- A.7 DATA VALIDATION PROCEDURES (continued) SCORE (7.2) Data validation procedures are: (a) A coordinated combination of computerized and manual checks applied at different levels in the measurement process. 5 (b) Applied with a degree of completeness at no more than two levels of the measurement process. 3 (c) Applied at only one level of the measurement pro- cess (7.3) Data validation criteria are documented and include: (a) Limits on: (1) operational parameters such as flow rates; (2) calibration data, (3) special checks unique to each measurement; e.g., succes- sive values/averages ; (4) statistical tests; e.g., outliers; (5) manual checks such as hand calcula- tions . (b) Limits on the above type checks for most of the measurement systems. (c) Limits on some of the above type checks for only the high-priority measurements. (7.4) Acceptable limits as set are reasonable and adequate to insure the detection of invalid data with a high probability for: (a) All measurement systems. 5 (b) At least 3/4 of the measurement systems. 3 (c) No more than 1/2 of the measurement systems. 1 47 ------- A.7 DATA VALIDATION PROCEDURES (continued) SCORE (7.5) Data validation activities are: (a) Recorded on standard forms at all levels of the measurement process. 5 (b) Recorded in the operator's/analyst's log book. 3 (c) Not recorded in any prescribed manner. 1 (7.6) Examination of data validation records indicates that: (a) Data validation activities have been carried out as specified and completely documented. 5 (b) Data validation activities appear to have been performed but not completely documented. 3 (c) Data validation activities, if performed, are not formally documented. 1 (7.7) Data validation summaries are: (a) Prepared at each level or critical point in the measurement process and forwarded to the next level with the applicable block of data. 5 (b) Prepared by and retained at each level. 3 (c) Not prepared at each level nor communicated between levels. 1 (7.8) Procedures for deleting invalidated data are: (a) Clearly defined in writing for all levels of the meas- urement process, and invalid data are automatically deleted when one of the computerized validation cri- teria is exceeded. 5 (b) Programmed for automatic deletion when computerized validation criteria are exceeded but procedures not defined when manual checks detect invalid data. 3 (c) Not defined for all levels of the measurement pro- cess. 1 48 ------- A.7 DATA VALIDATION PROCEDURES (continued) SCORE (7.9) Quality audits (i.e., both on-site system reviews and/or quantitative performance audits) independent of the normal operations are: (a) Performed on a random but regular basis to ensure and quantify data quality. 5 (b) Performed whenever a suspicion arises that there are areas of ineffective performance. 3 (c) Never performed. 1 A.8 PROCUREMENT AND INVENTORY PROCEDURES (8.1) Purchasing guidelines are established and documented for: (a) All equipment and reagents having an effect on data quality. 5 (b) Major items of equipment and critical reagents. 3 (c) A very few items of equipment and reagents. 1 (8.2) Performance specifications are: (a) Documented for all items of equipment which have an effect on data quality. 5 (b) Documented for the most critical items only. 3 (c) Taken from the presently used items of equipment. 1 (8.3) Reagents and chemicals (critical items) are: (a) Procured from suppliers who must submit samples for test and approval prior to initial shipment. 5 (b) Procured from suppliers who certify they can meet all applicable specifications. 3 (c) Procured from suppliers on the basis of price and delivery only. l 49 ------- A.8 PROCUREMENT AND INVENTORY PROCEDURES (continued) SCORE (8.4) Acceptance testing for incoming equipment is: (a) An established and documented inspection procedure to determine if procurements meet the quality assurance and acceptance requirements. Results are document- ed. 5 (b) A series of undocumented performance tests performed by the operator before using the equipment. 3 (c) The receiving document is signed by the responsible individual indicating either acceptance or rejection. 1 (8.5) Reagents and chemicals are: (a) Checked 100% against specification, quantity, and for certification where required and accepted only if they conform to all specifications. 5 (b) Spot-checked for proper quantity and for shipping damage. 3 (c) Released to analyst by the receiving clerk without being checked as above. 1 (8.6) Information on discrepant purchased materials is: (a) Transmitted to the supplier with a request for corrective action. 5 (b) Filed for future use. 3 (c) Not maintained. j_ (8.7) Discrepant purchased materials are: ( a) Submitted to a review by Quality Control and Chief Chemist for disposition. 5 (b) Submitted to Service Section for determination on acceptability. 3 (c) Used because of scheduling requirements. 1 50 ------- A.8 PROCUREMENT AND INVENTORY PROCEDURES (continued) SCORE (8.8) Inventories are maintained on: (a) Flrst-in, first-out basis. 5 (b) Random selection in stock room. 3 (c) Last-in, first-out basis. 1 (8.9) Receiving of materials is: (a) Documented in a receiving record log, giving a description of the material, the date of receipt, results of acceptance test, and the signature of the responsible individual. 5 (b) Documented in a receiving record log with material title, receipt date, and initials of the individual logging the material in. 3 (c ) Documented by filing a signed copy of the requisi- tion. 1 (8.10) Inventories are: (a) Identified as to type, age, and acceptance status. 5 (b) Identified as to material only. 3 (c) Not identified in writing. 1 (8.11) Reagents and chemicals which have limited shelf life are: (a) Identified as to shelf life expiration data and systematically issued from stock only if they are still within that date. 5 (b) Issued on a first-in, first-out basis, expecting that there is enough safety factor so that the expiration date is rarely exceeded. 3 (c) Issued at random from stock. 1 51 ------- A.9 PERSONNEL TRAINING PROCEDURES SCORE (9.1) Training of new employees is accomplished by: (a) A programmed system of training where elements of training, including quality standards, are included in a training checklist. The employee's work is immediately rechecked by supervisors for errors or defects and the information is fed back instanta- neously for corrective action. 5 (b) On-the-job training by the supervisor who gives an overview of quality standards. Details of quality standards are learned as normal results are fed back to the chemist. . 3 (c) On-the-job learning with training on the rudi- ments of the job by senior coworkers. 1 (9.2) When key personnel changes occur: (a) Specialized knowledge and skills are retained in the form of documented methods and descriptions. (b) Replacement people can acquire the knowledge of their predecessors from coworkers, supervisors, and detailed study of the specifications and memoranda. (c) Knowledge is lost and must be regained through long experience or trial-and-error. (9.3) The people who have an impact on quality, e.g., cali- bration personnel, maintenance personnel, bench chemists, supervisors, etc., are: (a) Trained in the reasons for and the benefits of standards of quality and the methods by which high quality can be achieved. (b) Told about quality only when their work falls below acceptable levels. (c) Are reprimanded when quality deficiencies are directly traceable to their work. 52 ------- A.9 PERSONNEL TRAINING PROCEDURES (continued) SCORE (9.4) The employee's history of training accomplishments is maintained through: (a) A written record maintained and periodically reviewed by the supervisor. 5 (b) A written record maintained by the employee. 3 (c) The memory of the supervisor/employee. 1 (9.5) Employee proficiency is evaluated on a continuing basis by: (a) Periodic testing in some planned manner with the results of such tests recorded. 5 (b) Testing when felt necessary by the supervisor. 3 (c) Observation of performance by the supervisor. 1 (9.6) Results of employee proficiency tests are: (a) Used by management to establish the need for and type of special training. 5 (b) Used by the employee for self-evaluation of needs. 3 (c) Used mostly during salary reviews. 1 A.10 FEEDBACK AND CORRECTIVE ACTION (10.1) A and corrective action mechanism to assure that problems are reported to those who can correct them and that a closed loop mechanism is established to assure that appropriate corrective actions have been taken is: » (a) Clearly defined in writing with individuals assigned specific areas of responsibility. 5 (b) Written in general terms with no assignment of responsibilities. 3 (c) Not formalized but left to the present supervisors/ managers. 1 53 ------- A.10 FEEDBACK AND CORRECTIVE ACTION (continued) SCORE (10.2) Feedback and corrective action activities are: (a) Documented on standard forms. 5 (b) Documented in the station log book. 3 (c) Documented in the operator's/analyst's notebook. 1 (10.3) A review of corrective action records indicates that: (a) Corrective actions were systematic, timely, and fully documented. 5 (b) Corrective actions were not always systematic, timely, or fully documented. 3 (c) A closed loop mechanism did not exist. 1 (10.4) Periodic summary reports on the status of corrective action are distributed by the responsible individual to: (a) All levels of management. 5 (b) One level of management only. 3 (c) The group generating the report only. 1 (10.5) The reports include: (a) A listing of major problems for the reporting period; names of persons responsible for correc- tive actions; criticality of problems; due dates; present status; trend of quality performance (i.e., response time, etc.); listing of items still open from previous reports. 5 (b) Most of the above items. ' 3 (c) Present status of problems and corrective actions. 1 54 ------- A.ll CALIBRATION PROCEDURES SCORE (11.1) Calibration procedures are: (a) Clearly defined and written out in step-by-step fashion for each measurement system and support device. 5 (b) Defined and summarized for each system and device. 3 (c) Defined but operational procedures developed by the individual. 1 (11.2) Calibration procedures as written are: (a) Judged to be technically sound and consistent with data quality requirements. 5 (b) Technically sound but lacking in detail. 3 (c) Technically questionable and lacking in detail. 1 (11.3) Calibration standards are: (a) Specified for all systems and measurement devices with written procedures for assuring, on a con- tinuing basis, traceability to primary standards. 5 (b) Specified for all major systems with written procedures for assuring traceability to pri- mary standards. 3 (c) Specified for all major systems but no procedures for assuring traceability to primary standards. 1 (11.4) Calibration standards and traceability procedures as specified and written are: (a) Judged to be technically sound and consistent with data quality requirements. 5 (b) Standards are satisfactory but traceability is not verified frequently enough. 3 (c) Standards are questionable. 1 55 ------- A.11 CALIBRATION PROCEDURES (continued) SCORE (11.5) Frequency of calibration is: (a) Established and documented for each measurement system and support measurement device. 5 (b) Established and documented for each major meas- urement system. 3 (c) Established and documented for only certain measurement systems. 1 (11.6) A review of calibration data indicates that the frequency of calibration as implemented: (a) Is adequate and consistent with data quality requirements. 5 (b) Results in limits being exceeded a small frac- tion of the time. 3 (c) Results in limits being exceeded frequently. 1 (11.7) A review of calibration history indicates that: ( a) Calibration schedules are adhered to and results fully documented. 5 (b) Schedules are adhered to most of the time. 3 (c) Schedules are frequently not adhered to. 1 (11.8) A review of calibration history and data validation records indicates that: (a) Data are always invalidated and deleted when calibration criteria are exceeded. (b) Data are not always invalidated and/or deleted when criteria are exceeded. (c) Data are frequently not invalidated and/or deleted when criteria are exceeded. 56 ------- A.11 CALIBRATION PROCEDURES (continued) SCORE (11.9) Acceptability requirements for calibration results are: (a) Defined for each system and/or device requiring calibration including elapsed time since the last calibration as well as maximum allowable change from the previous calibration. 5 (b) Defined for all major measurement systems. 3 (c) Defined for some major measurements systems only. 1 (11.10) Acceptability requirements for calibration results as written are: (a) Adequate and consistent with data quality require- ments : 5 (b) Adequate but others should be added. 3 (c) Inadequate to ensure data of acceptable quality. 1 (11.11) Calibration records (histories) are: (a) Utilized in revising calibration schedules (i.e., frequency). 5 (b) Utilized when specific questions arise and re- viewed periodically for trends, completeness, etc. 3 (c) Utilized only when unusual problems occur. 1 A. 12 FACILITIES/EQUIPMENT (12.1) Facilities/Equipment are: (a) Adequate to obtain acceptable results. 5 (b) Adequate to obtain acceptable results most of the time. 3 (c) Additional facilities and space are needed. 1 57 ------- A.12 FACILITIES/EQUIPMENT (continued) SCORE (12.2) Facilities, equipment, and materials are: (a) As specified in appropriate documentation and/or standards. 5 (b) Generally as specified in appropriate standards. 3 (c) Frequently different from specifications. 1 (12.3) Housekeeping reflects an orderly, neat, and effective attitude of attention to detail in: (a) All of the facilities. 5 (b) Most of the facilities. 3 (c) Some of the facilities. 1 (12.4) Maintenance Manuals are: (a) Complete and readily accessible to maintenance personnel for all systems, components, and devices. 5 (b) Complete and readily accessible to maintenance personnel for all major systems, components, and devices. 3 (c) Complete and accessible for only a few of the systems. 1 A.13 RELIABILITY (13.1) Procedures for reliability data collection, processing, and reporting are: (a) Clearly defined and written for all system com- ponents . (b) Clearly defined and written for major components of the system. (c) Not defined. 58 ------- A.13 RELIABILITY (continued) SCORE (13.2) Reliability data are: (a) Recorded on standard forms. 5 (b) Recorded as operator/analyst notes. 3 (c) Not recorded. 1 (13.3) Reliability data are: (a) Utilized in revising maintenance and/or replace- ment schedules. 5 (b) Utilized to determine optimum parts inventory. 3 (c) Not utilized in any organized fashion. 1 59 ------- APPENDIX B STANDARD TECHNIQUES USED IN QUANTITATIVE PERFORMANCE AUDITS 61 ------- Ambient air techniques' to Pollutant EPA Bias (absolute, method or percent of or mean con centra- number tion) SO 6 0 ^ NO, N00 Chemilumi- 0 2 NO nescent X Photochem- Chemilumi- -35 to -15% ° ical oxi- nescent from 0.05 to dants 0.50 ppm CO NDIR +2.5 Precision coefficient Within- laboratory 5-13 yg/m3, from x = 0-1000 yg/m3 7-8% at 100 yg/m (0.05 ppm) 0.0033 + ° 0.0255 x (0-0.5 ppm) 3 0.6 mg/m (absolute, or of variation) Between- laboratory 10-25 yg/m3 from x = . 0-1000 yg/m _ 0.0008 + 0.0355 x (0-0.5 ppm) -0.0051 + 0.0690 x (0.15-0.5 ppm) 0.8 - 1.6 mg/m3 (non- linear varia- tion) over 0-60 mg/m3 Comments Lower limit of detec- tion is 25 yg/m3. Flow rate changes, sampling train leakage are prim- ary error sources. Lower limit of detec- tion is 10 yg/m3 (0.005 ppm) . Errors are asso- ciated with calibration and instrument drift (from zero and span settings) . Lower detection limit is 0.0065 ppm Lower detection limit is 0,3 mg/m3. Interference of water vapor is signifi- cant. Reference EPA-R4- 73-028d d EPA-R4- 73-028c EPA-R4- 73-028a ------- Ambient air techniques (con.) ON U> Pollutant Particulates EPA method or number High- Volume Bias (absolute, or percent of mean concentra- tion) No information Precision (absolute, or coefficient of variation) Within- Between- laboratory laboratory 3% 3.7% Comments Minimum detectable limit is 3 mg. Shorter samp- Reference EPA-R4- 73-0 2 8b NO Arsenite -3% (50-300 yg/m3) 8 yg/m (50-300 lag/ m3) 11 yg/m (50-300 yg/ m3) ling periods give less precise results, biased high. A tentative method. Lower detectable limit is 9 yg/m3. EPA-R4- 73-280o . This table is a summary of information contained in the cited references, all of which are quality assurance guideline manuals published by EPA. Collaborative test results are cited, if available, in the manuals . x = pollutant concentration. °EPA-650/4-75/016. Guidelines for Development of a Quality Assurance Program for the Continuous Measurement of Nitrogen Dioxide in the Ambient Air (Chemilumines cent) , Smith & Nelson, Research Triangle Institute, Research Triangle Park, N.C. 27709. ------- Source Sampling Techniques' Pollutant EPA me t hod or numb er S02 6 so and 8 S03/H2S04 NO 7 X CO 10 Particulates 5 Visible 9 emissions Be 104 Bias (absolute, or percent of mean concentra- tion) 0 -2% (analysis only) -2% (analysis only) 0 +7 ppm No information +1.4% opacity -20%, average Precision coefficient Within- laboratory 3.9 o 0.1 g/ni 60% 7% 13 ppm 10-30% 2% opacity 44% (absolute, or of variation) Between- laboratory 5.5 •3 0.11 g/m 65% 10% 25 ppm 20-40% 2.5% 58% Comments Major error source is dif- ficulty of obtaining repro- ducible titration end- points . Minimum detect- able limit is 3 ppm. Same analysis technique as Method 6 above. Grab sample; largest error source is failure to re- calibrate spectrophoto- meter. Analyzer drift and C02 interference are largest problems . Minimum detect- able limit is 20 ppm. Numerous small error sources associated with stack sampling. Good results depend to a great extent on the effec- tive training of observers. Reference EPA-650/ 14-74- 005-e EPA-650/ 14-74- 005-g EPA-650/ 14-74- 005-f EPA-650/ 14-74- 005-h EPA-650/ 14-74- 00 5- d EPA-650/ 14-74- 005-i EPA-650/ 14-74- 005-k aThis table is a summary of information contained in the cited references, all of which are quality assurance guidelines manuals. ------- APPENDIX C COMPARISON OF ANALYSES OF LIMESTONE SLURRY 65 ------- APPENDIX C COMPARISON OF ANALYSIS OF LIMESTONE SLURRY Cooperating laboratories were: 1. TVA Power Service Center Laboratory, Chattanooga, Tennessee - Mr. John Rose, contact 2. TVA Power Service Center Laboratory, Muscle Shoals, Alabama - Dr. Guerry McClellon, contact 3. Research Triangle Institute, Research Triangle Park, North Carolina - Dr. D. E. Wagoner, contact Results from RTI laboratories are presented in two sections. One set of data was obtained on slurry which was filtered at the Shawnee Laboratory. The second set of data results from analysis of samples filtered in the RTI laboratory. The first eight matrixes present results for each element: calcium, mag- nesium, and total sulfur in the solid; and calcium, magnesium, sodium, potas- sium, and chloride in the liquid. The next five matrixes give results of all analyses for each laboratory, with Shawnee results listed first. The last four matrixes break down the total sulfur and calcium analyses into results by a standard "wet" technique, by X-ray fluorescence using Shawnee standard values, and by X-ray fluorescence using RTI-derived standard values on the Shawnee standard material. 66 ------- Table 1. Analysis for calcium in slurry solid Sample: Ca (as CaO) Wt % Laboratory RTI (Shawnee filtered) RTI (RTI filtered) Chattanooga (TVA) Muscle Shoals (TVA) ------- Table 2. Analysis for magnesium in slurry solid Sample: Mg (as MgO) Wt % Laboratory 11/18/75 1100 1500 2300 11/19/75 1100 2300 Shawnee 0.30 0.29 0.28 0.27 0.29 oo RTI (Shawnee filtered) 0.43 0.48 0.39 0.36 0.42 RTI (RTI filtered) 0.61 0.52 0.34 0.39 0.35 Chattanooga (TVA) 0.65 0.56 0.56 0.61 0.61 Muscle Shoals (TVA) 0.25 0.25 0.24 0.25 0.24 ------- Table 3. Analysis for total sulfur in slurry solid ^^" ^*»w Sample: ^XJS (as SO-) v^ Laboratory ^"V^ 1 Shawnee 1 RTI 1 (Shawnee filtered) 1 RTI (RTI filtered) 1 Chattanoog i (TVA) 1 Muscle Shoals (TVA) 11/18/75 11/19/75 1100 1500 2300 1100 2300 34.78 36.70 28.60 30.5 31.3 34.16 32.00 27.00 30.7 30.9 31.22 33.03 31.13 30.0 31.4 28.17 30.53 28.78 29.2 29.7 31.64 32.80 32.03 29.6 30.4 VO ------- Table 4. A-.?.]ysis for calcium in slurry filtrate Sample: Ca (ppm) Laboratory 11/18/75 1100 1500 2300 11/19/75 11002300 Shawnee 1720 1710 1810 2090 2315 RTI (Shawnee filtered) 1775 1825 1810 1708 1885 RTI (RTI filtered) 1700 1810 1730 1720 1825 Chattanooga (TVA) 1756 1740 1676 1596 1732 Muscle Shoals (TVA) 1787 1787 1716 1787 1716 ------- Table 5. Analysis for magnesium in slurry filtrate Sample: Mg (ppm) Laboratory 11/18/75 1100 1500 2300 11/19/75 1100 2300 Shawnee 733 699 662 691 698 RTI (Shawnee filtered) 785 945 813 1000 1115 RTI (RTI filtered) 730 805 805 795 770 Chattanooga (TVA) Muscle Shoals (TVA) 768 734 763 724 724 724 780 816 724 784 ------- Table 6. Analysis for sodium in slurry filtrate Sample: Na Laboratory 11/18/75 1100 1500 2300 11/19/75 1100 2300 Shawnee 71 57 70 73 82 ho RTI (Shawnee filtered) 71 69 79 75 77 RTI (RTI filtered) 153 161 180 176 145 Chattanooga (TVA) 66 62 64 66 69 Muscle Shoals (TVA) 41 37 37 41 41 ------- Table 7. Analysis for potassium in slurry filtrate Sample: K Laboratory 11/18/75 1100 1500 2300 11/19/75 1100 2300 Shawnee 118 121 123 126 153 OJ RTI (Shawnee filtered) 103 101 105 108 111 RTI (RTI filtered) 116 102 114 118 116 Chattanooga (TVA) 107 96 107 116 116 Muscle Shoals (TVA) 58 50 58 58 ------- Table 8. Analysis for chloride in slurry filtrate Sample: Cl Laboratory 11/18/75 1100 1500 2300 11/19/75 1100 2300 Shawnee 3651 3580 3545 3545 3580 RTI (Shawnee filtered) 3697 3700 3855 3660 3987 RTI (RTI filtered) 3621 3638 3754 3519 3566 Chattanooga (TVA) 3692 3543 3571 3571 3628 Muscle Shoals (TVA) 3800 3700 3600 3600 3700 ------- Table 9. Laboratoty: Shawnee (TVA) Ul SOLID (Wt.%) Ca (CaO) Mg (MgO) TS (S03) 11/18/75 11/19/75 1100 1500 2300 1100 2300 24.73 0.30 34.78 19.03 0.48 32.00 LIQUID (ppm) Ca Mg Na K 1720 733 71 118 1710 699 57 121 22.78 0.28 31.22 21.05 0,27 28.17 22.87 .29 31.64 1810 662 70 123 2090 691 73 126 7^1 S 698 82 153 Cl 3651 3580 3545 3545 3580 ------- Table 10. Laboratory: RTI (Shawnee filtered) SOLID (Wt. %) Ca (CaO) Mg (MgO) TS (S03) 11/18/75 11/19/75 1100 1500 2300 1100 2300 22.12 0.43 36.7 19.03 0.48 32.00 19.10 0.39 33.03 18.41 0.36 30.53 19.46 0.42 32.80 LIQUID (ppm) Ca Mg Na K 1775 785 71 103 1825 945 69 101 1810 913 79 105 1708 1000 75 108 1885 1115 77 111 Cl 3697 3700 3855 3660 3987 ------- Table 11. Laboratory: RTI SOLID (Wt. %) Ca (CaO) Mg (MgO) TS (S03) 11/18/75 11/19/75 1100 1500 2300 1100 2300 17.99 0.61 28.6 19.96 0.52 27.0 18.63 0.34 31.13 18.13 0.39 28.78 19.32 0.35 32.03 LIQUID (ppm) Ca Mg Na K 1700 730 153 116 1810 805 101 102 1730 805 180 114 1720 795 176 118 1825 770 145 116 Cl 3621 3638 3754 3519' 3566 ------- Table 12. Laboratory: Muscle shoals (TVA) oo SOLID (Wt. %) Ca (CaO) Mg (MgO) TS (S03) 11/18/75 11/19/75 1100 1500 2300 1100 2300 23.6 0.25 31.3 23.1 0.25 30.9 23.6 0.24 31.4 22.6 0.25 29.7 22.9 0.24 36.4 LIQUID (ppm) Ca Mg Na K 1787 724 41 58 1787 724 37 54 Cl 3800 3700 1716 724 37 50 3600 1787 724 41 58 3600 1716 784 41 58 3700 ------- Table 13. Laboratory: Chattanooga (TVA) VO SOLID (Wt. %) Ca (CaO) Mg (MgO) TS (S03) 11/18/75 11/19/75 1100 1500 2300 1100 2300 23.45 0.65 30.5 23.24 0.56 30.7 23.02 0.56 30.0 22.46 0.61 29.2 22.46 0.61 29.6 LIQUID (ppm) Ca Mg Na K 1756 768 66 107 1740 734 62 96 1676 763 64 107 1596 780 66 116 1732 816 67 116 Cl 3692 3543 3571 3571 3628 ------- Table 14. Total sulfur determinations, Shawnee filtered samples analyzed at RTI wt % (as SO ) 11/18/75 1100 1500 2300 11/19/75 1100 2300 By BaCl2 precipitation 36.7 32.00 33.03 30.53 32.80 By X-Ray fluorescence oo o Using Shawnee X-Ray standard number for TS Using RTI determined number for TS 22.08* 21.25 23.15 27.75 22.95 Shawnee X-Ray Standard (as so3)* Shawnee given RTI determined 28.45 28.38 * Shawnee and RTI TS determinations on the XRF standard were virtually identical, so the Shawnee TS value only was used in calculating wt % TS in each sample. ------- Table 15. Total sulfur determinations, RTI filtered and analyzed samples BaCl2 precipitation X-Ray fluorescence Using Shawnee X-Ray standard number for TS Using RTI-determined number for TS 11/18/75 11/19/75 1100 1500 2300 1100 2300 28.6 27.00 31.13 28.78 32.03 JO. 28 29.8 34. 4b oo ------- Table 16. Calcium determinations, Shawnee filtered samples analyzed at RTI oo to wt % (as CaO) By AA 11/18/75 11/19/75 1100 1500 2300 1100 2300 22.12 19.03 19.60 18.41 19.46 By X-Ray fluorescence Using Shawnee X-Ray standard number for CaO Using RTI determined number for CaO Shawnee X-Ray Standard (as CaO) Shawnee given RTI determined 23.13 20.09 25.41 22.06 22.25 19.32 25.76 22.37 22.89 19.88 25.41 22.06 ------- Table 17. Calcium determinations, RTI filtered and analyzed samples oo u> wt % (as CaO) •••••^•^••••••••••^•^^•••••••••••^^••••••••••^^•^^••^^ By AA By X-Ray fluorescence Using Shawnee X-Ray standard number for CaO Using RTI-determined number for CaO 11/18/75 11/19/75 1100 mmmmmmmmmmmm*mm*m*mmmmimi^m 17.99 1500 B^|MMa|BHBaH^IM^l^MH^^l^HHIHHI^ 17.96 2300 ••^•••••••••••••••••••••MMMMBM 18.63 1100 BBMMHHBI^V^BHB.HHIHMMM| 18.13 2300 •••^^^•••^^^••••••••••••••i 19.32 28.35 24.61 25.52 22.16 26.17 22.72 27.68 24.04 26.32 22.85 ------- TECHNICAL Rf PORT DATA (Please read Instructions on the reverse before completing) 1. REPORT NO. EPA-600/2-76-083 2. 3. RECIPIENT'S ACCESSION>NO. 4. TITLE AND SUBTITLE Development and Trial Field Application of a Quality Assurance Program for Demonstration Projects 5. REPORT DATE March 1976 6. PERFORMING ORGANIZATION CODE I. AUTHOR(S) 8. PERFORMING ORGANIZATION REPORT NO. James Buchanan 9. PERFORMING OROANIZATION NAME AND ADDRESS Research Triangle Institute P.O. Box 12194 Research Triangle Park, NC 27709 10. PROGRAM ELEMENT NO. EHB-557; ROAP ABA-011 11. CONTRACT/GRANT NO. 68-02-1398, Task 20 12. SPONSORING AGENCY NAME AND ADDRESS EPA, Office of Research and Development Industrial Environmental Research Laboratory Research Triangle Park, NC 27711 13. TYPE OF REPORT AND PERIOD COVERED Task Final: 7-12/75 14. SPONSORING AGENCY CODE EPA-ORD 15. SUPPLEMENTARY NOTES Project officer for this report is L.D. Johnson, Mail Drop 62, Ext 2557. 16. ABSTRACT Tne repOr£ outlines results of a project: to develop a set of quality assu- rance guidelines for EPA demonstration projects; to implement a short-term quality assurance program at the EPA wet limestone scrubber facility at the Shawnee steam/ electric plant; and to modify the guidelines in light of the Shawnee operating exper- ience. The set of quality assurance guidelines and detailed results of the Shawnee program are included in two other reports prepared during the project. 7. KEY WORDS AND DOCUMENT ANALYSIS DESCRIPTORS Air Pollution Quality Assurance Duality Control crubbers limestone 'lue Gases Industrial Processes Instruments Sulfur Dioxide Dust Sampling Weight Measurement b.lDENTIFIERS/OPEN ENDED TERMS Air Pollution Control Stationary Sources Field Application Control Laboratory Particulate c. COSATI Field/Group I3B~ 3H,14D 14B 07B 07A 11G I08G 21B 8. DISTRIBUTION STATEMENT Unlimited 19. SECURITY CLASS (ThisReport) Unclassified 21. NO. OF PAGES 84 20. SECURITY CLASS (Thispage) Unclassified 22. PRICE EPA Form 2220-1 (9-73) ------- |