United States Environmental Protection Agency Environmental Monitoring Systems Laboratory Us Vegas NV 89193-3478 Research and Development EPA/800/S4-89/031 Nov. 1989 Project Summary Direct/Delayed Response Project: Quality Assurance Plan for Preparation and Analysis of Soils from the Mid-Appalachian Region of the United States M. L. Papp, R. D. Van Remortel, C. J. Palmer, G. E. Byers, B. A. Schumacher, R. L. Slagie, J. E. Teberg, and M. J. Miah The quality assurance (QA) policy of the U.S. Environmental Protection Agency (EPA) requires every monitoring and measurement project to have a written and approved quality assurance project plan. The purpose of this quality assurance plan Is to specify the policies, organization, objectives, and the quality evaluation and quality control activities needed to achieve the data quality requirements of the Direct/Delayed Response Project (ODRP), Mid-Appalachian Soil Survey (MASS). These specifications are used to assess and control measurement errors that may enter the system at various phases of the project, e.g., during soil sampling, preparation, and analysis. This Project Summary was developed by EPA's Environmental Monitoring Systems Laboratory, Las Vegas, NV, to announce key findings of the research project that Is fully documented In a separate report of the same title (see Protect Report ordering Information at back). Project Description The DDRP is conducted as part of the interagency, federally mandated National Acid Precipitation Assessment Program which addresses the concern over potential acidification of surface waters by atmospheric deposition within the United States. The overall purpose of the project is to characterize geographic regions of the United States by predicting the long-term response of watersheds and surface waters to acidic deposition. The EPA is assessing the role that atmospheric deposition of sulfur plays in controlling long-term acidification of surface waters. Recent trend analyses have indicated that the rate of sulfur deposition is either unchanging or slowly declining in the Northeastern United States, but is increasing in the Southeastern United States. If a "direct" response exists between sulfur deposition and surface water alkalinity, then the extent of current effects on surface water probably would not change much at current levels of deposition, and conditions would improve as the levels of deposition decline. If surface water chemistry changes in a "delayed" manner, e.g., due to chemical changes in the watershed, then future changes in water chemistry (even with level or declining rates of deposition) become difficult to predict. This range of potential effects has clear and significant implications to public policy decisions on sulfur emissions control strategies. The ODRP focuses on regions of the United States that have been identified as ------- potentially sensitive to surface water acidification. The MASS is the third of three DDRP regional surveys and includes portions of Pennsylvania, Virginia, and West Virginia. Surface water acidification in the Mid-Appalachian region was studied during the Eastern Lakes Survey in 1984 and in the National Stream Survey, Phase I, in 1985. Information gained during these and other surveys has been used to optimize the quality assurance program in the MASS. Field and laboratory data collected in the DDRP are included in the system description modeling and are then used to assess the key processes that regulate the dynamics of base cation supply and sulfur retention within watersheds. Integrated watershed data are used to calibrate three dynamic simulation models that predict future regional responses to acidic deposition. Uncertainty and error propagation estimates are an important part of all levels of analysis. The DDRP staff at the EPA Environmental Research Laboratory in Corvallis, Oregon, is involved in all aspects of the MASS. Responsibilities of this laboratory include the development of an experimental design, QA oversight for soil mapping and sampling, and data analysis and interpretation. The DDRP staff at the EPA Environmental Monitoring Systems Laboratory in Las Vegas, Nevada, is involved in matters relating to QA implementation, logistics, and analytical services. The responsibilities of this laboratory include the development and implementation of operational procedures and quality evaluation/control criteria for soil sample preparation and analysis, and the development of data verification procedures for field and laboratory data. The use of interagency agreements by EPA has allowed the DDRP to engage support from several groups that have expertise in areas of importance to the project. The DDRP staff at Oak Ridge National Laboratory is involved in structuring and managing large data bases to satisfy DDRP data analysis requirements. The soil survey staff of the United States Department of Agriculture, Soil Conservation Service is involved in DDRP soil mapping and sampling. Laboratory analysis support is solicited through competitive bid by independent contract laboratories. Field and Laboratory Operations Field mapping and sampling operations in the MASS are conducted by Soil Conservation Service soil scientists under the supervision of EPA representatives. Approximately five sampling crews, each composed of three to four qualified persons, are selected. These crews are responsible for the description of soil and site characteristics, sample collection, and shipment of samples to the preparation laboratory. Members of each crew are soil scientists that are experienced with the universally recognized National Cooperative Soil Survey characterization procedures. The crew leader is responsible for selecting sampling sites and documenting all field data. The sampling protocols specify the procedures for ensuring the integrity of the samples. Specific instructions on excavating and draining (if necessary) soil pits are given. Crews are provided with standard containers for sample collection. Samples are kept as cool as possible in the field and during transport to the preparation laboratory. In order to maintain their integrity and to reduce the possibility of biological degradation, samples are stored at a temperature of 4°C within 24 hours of sampling. It is anticipated that 150 pedons are to be sampled during the MASS, with each pedon yielding an average of six routine samples from its component horizons. To allow the determination of bulk density, natural soil clods are collected in triplicate, where possible from each soil horizon. Where clods cannot be collected, known volumes of soil are collected in duplicate. Using a volume replacement method, the volume of soil is estimated by replacing a small excavated soil cavity with a known volume of foam beads. If this method does not yield satisfactory samples, a volume filling method is attempted by filling a container of known volume with soil and attempting to pack it to the same density normally encountered in the horizon being sampled. The preparation laboratory is the intermediate link between the sampling crews and the analytical laboratories. The laboratory is located at the EPA Environmental Monitoring Systems Laboratory in Las Vegas, Nevada, although previous DDRP surveys utilized the services of multiple university laboratories for sample preparation. It was thought that consolidating t preparation activities at a single labc tory might result in better samp handling, improved sample integrity, a higher data quality. The laboratory serv as a central location for soil process! and for introducing double-blii measurement quality samples into t sample flow. The soil samples collect by the sampling crews are sent \ overnight courier to the preparati laboratory, where the laboratory st processes the samples and prepar homogeneous, anonymous subsampl that are shipped to the analytic laboratories. To accomplish these tas successfully, the preparation laboratc must uniformly track, process, and stc all samples. After a bulk soil sample is air dried, t soil is carefully disaggregated a sieved. The less than 2-mm soil retained in a labeled plastic bottle a placed in cold storage when not undi going processing. The rock fragments < retained for gravimetric analysis. In ore to obtain representative volumes of soil is necessary to prepare homogeneo subsamples from the less than 2-mm s fraction using a riffle splitter. T subsamples are placed in labeled plas bottles and are organized in batches tl are shipped to separate contra analytical laboratories for general a elemental analyses. Each batch norms contains 40 samples but may have many as 42 samples. With the except! of a quality control audit sample, samples are randomly placed within batch and cannot be distinguished by t recipient laboratory. The unused pprtio of the bulk soil samples are archived cold storage. All raw data obtained are recorded or series of preparation laboratory raw di forms and are immediately entered int< personal computer to ensure prop tracking, processing, and evaluation the samples. The computer is direc linked to the QA staff and allows real-tir data evaluation and tracking of samp by both parties. The precision a accuracy criteria for measurement qua samples are checked while the samp are being processed, permitting 1 prompt identification of any discrc ancies. During shipment, the samples receiv by the analytical laboratories c undergo segregation both by particle s and by density; therefore, each sam must be re-homogenized by thorou mixing prior to the removal of aliquots ------- analysis. The raw data generated during sample analysis are entered on a personal computer using a specially designed entry and verification system. Data reports are then generated to assist in the evaluation of data quality. Before a batch of analytical data is accepted all completeness and quality control requirements must be met. In addition, the following documents must be updated constantly at the analytical laboratory and must be available, upon request, to the analysts and the supervisor involved in the project: standard operating procedures, laboratory QA plan, list of in-house samples, instrument performance data, control charts for all check samples and detection limit check samples, and quality control data reports. Frequent communications, i.e., two or three contacts each week, are maintained with each analytical laboratory in order to obtain current sample status and to discuss any problems that may occur during analyses. These discussions are to be recorded in a logbook by the QA staff and the laboratory staff. Preliminary and final data are available for access via electronic transfer. The preliminary data are reviewed for anomalies and if a problem is identified, the laboratory is iQtified. Corrective action or reanalysis •nay be suggested. Legal chain-of-custody procedures are not required for this study; however, sample custody must be documented. The sampling crew leader is responsible for all samples in the field as well as the sample shipments. As soil shipments are received in the preparation laboratory, they are immediately logged in and checked against the packing slip and the pedon description forms for proper labeling and quantity. If any discrep- ancies are found, the soil sampling task leader and the field crew leader are notified. Following the receipt of samples, the appropriate laboratory managers are responsible for sample integrity until the samples are archived. Quality Assurance Program The data collection criteria provide a balance between constraints of time and cost and the quality of data necessary to achieve the DORP research objectives. The MASS QA Plan is designed to accomplish the following general objectives: • establish the QA criteria used to control md assess data collected in the survey, • provide standardized sampling, preparation, and analytical methods and procedures, • utilize assessment samples and procedures to verify the quality of the data, • perform field and laboratory on-site audits to ensure that all activities are properly performed and that discrepancies are identified and resolved, • evaluate the data and document the results in QA reports to EPA manage- ment. The raw soil characterization data for the DDRP surveys are collected during four major operational phases consisting of mapping, sampling, preparation, and analysis. A certain amount of data measurement uncertainty is expected to enter the system at each phase. Grouping of the data, e.g., by sampling class/horizon configurations, also increases uncertainty. The sampling population itself is a source of confounded uncertainty that is extremely difficult to quantify. Generally, the data quality objectives for the MASS encompass the overall allowable uncertainty from sample measurement and from the sampling population that the data users are willing to accept in the analytical results. Because of the many possible confounded sources of uncertainty and the data collection focus of this QA plan, overall data quality objectives for the survey are not described herein. Rather, the plan focuses on the definition, implementation, and assessment of measurement quality objectives (MQOs) that are specified for the entire sample preparation and analysis phases of data collection and portions of the field sampling phase. The MQOs are more or less specific goals defined by the data users that clearly describe the data quality that is sought for each of the measurement phases. The MQOs are defined according to the following six attributes: • Detectability—the lowest concentration of an analyte that a specific analytical procedure can reliably detect, • Precision—the level of agreement among multiple measurements of the same characteristic, • Accuracy—the difference between an observed value and the true value of the parameter being estimated, • Representativeness—the degree to which the data collected accurately represent the population of interest, • Completeness—the quantity of data that is successfully collected with respect to the amount intended in the experimental design, • Comparability—the similarity of data from different sources included within individual or multiple data sets; the similarity of analytical methods and data from related projects across regions of concern. An attempt has been made to define each quality attribute individually for each measurement phase of the program (sampling, preparation, analysis) or collectively for overall measurement system uncertainty on a parameter-by- parameter basis. The term "uncertainty" is used as a generic term to describe the sum of all sources of error associated with a given portion of the measurement system. In order to reduce measurement uncertainty in the final data values, error occurring in each phase must be identified and rectified during that phase of the survey. Measurement uncertainty for the MASS is controlled through well- defined protocols, system audits, and the introduction of measurement quality samples at each measurement phase. Initial MQOs were established on the basis of the requirements of EPA data users and on the selection of appropriate methods to obtain the data. The MQOs were reviewed by persons familiar with analytical methods and soil characterization techniques, including soil chemists and laboratory personnel. Modifications to the MQOs and protocols were implemented based on information gained from the DORP Northeastern region and Southern Blue Ridge Province surveys, from peer review comments, or according to a particular analytical procedure or instrument. Measurement quality samples are placed at a rate of about one sample in four in each batch of analytical samples to determine measurement uncertainty. The measurement quality samples are used during various phases of the survey and allow the QA staff to control and assess the quality of the data collection process. Included in each batch of samples are: audit samples that have known ranges of analyte concentration ------- (detectability, precision, and accuracy), duplicates of routine samples introduced at each measurement phase (precision and representativeness), and an audit sample for use in laboratory quality control (accuracy). There are separate analytical within-batch precision and accuracy MQOs for organic and mineral soils because of differences in methodology and the expected wide variability in analyte concentrations. For this reason, mineral and organic samples are organized and analyzed in separate batches (see below) The quality evaluation samples are those samples which are known to the QA staff but are either blind or double blind to the sampling crews, preparation laboratory, or analytical laboratory. A blind sample has a concentration range that is unknown to the analysts, whereas a double blind sample cannot be distinguished from routine sample and has a concentration range that is unknown to the analyst. These samples provide an independent check on the analytical process and can be used to evaluate whether the MQOs have been met for any given run or batch, or for all batches, i.e., overall measurement uncertainty. Important characteristics of the audit samples include their similarity to routine samples in matrix type and concentration level, their homogeneity and stability, and their defined accuracy windows. Every quality evaluation sample has a specific purpose in the data assessment scheme. In order to produce data of consistently high quality, the contract laboratories are required to analyze certain types of quality control samples that are known to the laboratory staff and that can be used by the analysts to identify and control analytical measurement uncertainty. Each of these samples has certain specifications that must be met before data for the parameter or batch are accepted. The control samples are non- blind samples procured under contract to assist the laboratories in meeting laboratory MQOs and include soil samples, e.g., analytical duplicates, and non-soil samples, e.g., reagent blanks. The samples allow the laboratory manager and the QA staff to assess whether the physical and chemical analysis is under control. Quality Assurance Implementation The quality assurance program is implemented through training, on-site systems audits, independent assess- ments, and other procedures used to control and assure the quality of the data being collected. Verification of these data is accomplished through a series of computer and manual checks of data quality. Corrective action for errors made at the preparation and analytical laboratories is accomplished primarily through the application of a QA reanalysis template for each analysis of interest. The templates provide a precision and accuracy checklist of the measurement quality samples for each batch and i useful in deciding whether to reanalyze particular parameter. The analytical data verification is multi-faceted, computerized approach provide a concise and consists assessment of the data. The over process if highlighted by the Laboratc Entry and Verification Information Sysfa (LEVIS). The LEVIS programs a implemented on personal computers a facilitate the data entry and qual control sample evaluation at the analytii laboratories as well as the evaluation laboratory performance by the QA sti The system is a menu-driven prodi that is designed as a two-pha operation, where phase one is analytical laboratory system and pha two is a quality assessment system. T LEVIS initiative was pursued becau previous surveys had shown that mam verification of the data was both lab and time-intensive, and allowed only limited amount of real-time correcti action on the part of the laboratories. An internal consistency program used to generate routine data outliers each sample batch. Analytical data frc each parameter are correlated agaii corresponding data from all otr analytical parameters measured in t MASS. For each parameter, t parameter pair with the strongest lin< relationship is identified and evaluate Soil chemistry relationships are anotl tool used to examine the interr consistency of the routine sample data is expected that approximately Status and Assessment of Measurement Quality Samples Status of Sample* Sample Type Low-range field audit" Field audit pair Field duplicate Low-range lab audit" Lab audit pair* Preparation duplicate Quality control audit Manager's sample if Per Batch 1 2 1 1 2 2 1 - QA Staff K K - K K - K K Sampling Crew B B B -- - - -- - Preparation Laboratory B B B - - B -- B Analytical Laboratory DB DB DB DB DB DB B - Assessment Purpose" System D.A P,A P - - - - -- Sampling D.A P.A P.R - -- -- - -- Preparation D.A P.A P - - P.R •- A Analyst A P.A P D P.A P A -- • K = known concentration, B = blind, DB = double blind. *> D - detectabilHylcontamination, P - precision, A * accuracy, R - representativeness. * Not placed in organic soil batches. " Triplicate in organic soil batches. ------- percent of the data will not comply with the relationships; these anomalous data are examined by a staff soil chemist who either qualifies them or assigns appro- priate flags signifying the discrepancy. Data Quality Assessment and Reporting The assessment of detectability is accomplished on a parameter basis at four levels: (1) compliance with contract- required detection limits; (2) calculation of actual instrument detection limits; (3) calculation of estimated system detection limits; and (4) identification of routine samples having concentrations below the system detection limits. The results can be grouped in tabular form to allow comparisons among the values for any parameter of interest. A statistical evaluation procedure that has been developed by the QA staff and data users is applied to the data in order to assess precision as a function of confounded data collection uncertainty. An additive step-function model is used, where an observed value of any soil characteristic is considered as the sum of the "true" or accepted value and an error term. Precision is evaluated for each variance segment of the range of concentration for a given analyte. The accuracy windows for the laboratory audit samples are based on previous interlaboratory analyses of the same sample material by the same protocols. The objective of setting windows is to specify a range of acceptable single values based on a mean and standard deviation computed from a number of previously observed values. The prediction intervals used for the accuracy windows are generally determined with confidence intervals constructed around the mean of these values, using appropriate weighting factors. The sampling aspect of repre- sentativeness is assessed by comparing the individual site and pedon classifications with the component sampling classes to which the soils are assigned. Representativeness of the measurement quality samples is assessed by comparing the concentration ranges of data from the duplicate samples to the overall concentration range of the routine sample data. Representativeness of the analytical samples is identified by assessing the homogenization and subsampling procedures at the preparation and analytical laboratories using precision estimates form the duplicate samples. Sampling completeness is assessed by comparing the actual number of soil pedons and associated horizons sampled to that number specified in the MASS design. Completeness of the sample collection, preparation, and analysis is calculated using data from the verified data base, while completeness of the data form a data user's perspective, i.e., amount of usable data, is determined using the validated data base. Following completion of the MASS, a comparison is made across the Northeastern, Southern Blue Ridge, and Mid-Appalachian regions that focuses on method differences, audit sample results, laboratory effects, and other QA features of the surveys. Comparison of the DORP data bases to other similar data bases may also be undertaken. Summary statistics are used to collate individual values into groups that enable the data users to discern trends of interest among the surveys. Task leaders for the various stages of the MASS provide a written summary of operations to the project managers on a quarterly basis. These reports describe the kinds of data collected as well as summarize the QA activities associated with the data. The summary of QA activities includes the following: • Overview of QA activities, • List of changes to the QA program, • Results of system and performance audits, • Assessment of data quality based on the verified data bases, • Documentation of unfavorable incidents and corrective actions, • Distribution of updated control charts, • Results of special studies. Reports relating to data quality assessment are also prepared by QA staff. These reports include interpretation of performance audits and replicate sample data, estimates of measurement error, and identification of any major discrepancies found during the assessment. In addition to this QA Plan, the QA staff produces laboratory operations manuals for use in the preparation and analysis of soil samples. Upon completion of the data verification activities, summary QA reports of the sample preparation and sample analysis data are produced and distributed to all cooperating DDRP staff and data users. Data Management System The purpose of data base management for the MASS is to facilitate the collection, entry, review, modification, and distri- bution of all data associated with the survey. Additional tasks include the data analysis for report generation, statistical analysis, verification, and data file security. The MASS final data base contains three progressive versions of the data gathered: (1) raw data base. (2) verified data base, and (3) validated data base. ------- M. L Papp, R. D. Van Remortel, C. J. Palmer, G. E. flyers, fl. A. Schumacher, R. L Slagle, J. E. Teberg, and M. J. Mian are with Lockheed Engineering and Sciences Company, Las Vegas, NV 89119. L J. Blume is the EPA Project Officers (see below). The complete report, entitled "Direct/Delayed Response Project: Quality Assurance Plan tor Preparation and Analysis of Soils from the Mid-Appalachian Region of the United States,'' (Order No. PB 90-116 971/AS; Cost $31.00, subject to change) will be available only from: National Technical Information Service 5285 Port Royal Road Springfield, VA 22161 Telephone: 703-487-4650 The EPA Project Officer can be contacted at. Environmental Monitoring Systems Laboratory U.S. Environmental Protection Agency Las Vegas, NV 89193-3478 United States Environmental Protection Agency Center for Environmental Research Information Cincinnati OH 45268 Official Business Penalty for Private Use $300 EPA/600/S4-89/031 I&ICUOI ACE.CI CHICAGO ------- |