United States Environmental Protection Agency Office of Research and Development Washington, DC 20460 EPA/600/R-04/012 May 2004 &EPA Innovative Technology Verification Report Field Measurement Technology for Mercury in Soil and Sediment Milestone Inc.'s Direct Mercury Analyzer (DMA)-80 ------- EPA/600/R-04/012 May 2004 Innovative Technology Verification Report Milestone Inc.'s Direct Mercury Analyzer (DMA)-80 Prepared by Science Applications International Corporation Idaho Falls, ID Contract No. 68-C-00-179 Dr. Stephen Billets Characterization and Monitoring Branch Environmental Sciences Division Las Vegas, Nevada 89193-3478 National Exposure Research Laboratory Office of Research and Development. U.S. Environmental Protection Agency ------- Notice The U.S. Environmental Protection Agency through its Office of Research and Development funded and managed the research described here under contract to Science Applications International Corporation. It has been subjected to the Agency's peer and administrative review and has been approved for publication as an EPA document. Mention of trade names or commercial products does not constitute endorsement or recommendation for use. ------- UNITED STATES ENVIRONMENTAL PROTECTION AGENCY Office of Research and Development Washington. DC 20460 MEASUREMENT AND MONITORING TECHNOLOGY PROGRAM VERIFICATION STATEMENT TECHNOLOGY TYPE: Field Measurement Device APPLICATION: Measurement for Mercury TECHNOLOGY NAME: Milestone Inc.'s Direct Mercury Analyzer (DMA)-SO COMPANY: Milestone Inc. ADDRESS: 160B Shelton Rd. Monroe, CT 06468 WEB SITE: http://www.milestonesci.com TELEPHONE: (203) 261-6175 VERIFICATION PROGRAM DESCRIPTION The U.S. Environmental Protection Agency (EPA) created the Superfund Innovative Technology Evaluation (SITE) and Measurement and Monitoring Technology (MMT) Programs to facilitate deployment of innovative technologies through performance verification and information dissemination. The goal of these programs is to further environmental protection by substantially accelerating the acceptance and use of improved and cost-effective technologies. These programs assist and inform those involved in design, distribution, permitting, and purchase of environmental technologies. This document summarizes results of a demonstration of the Direct Mercury Analyzer (DMA)-SO developed by Milestone Inc. PROGRAM OPERATION Under the SITE and MMT Programs, with the full participation of the technology developers, the EPA evaluates and documents the performance of innovative technologies by developing demonstration plans, conducting field tests, collecting and analyzing demonstration data, and preparing reports. The technologies are evaluated under rigorous quality assurance(QA) protocols to produce well-documented data of known quality. The EPA National Exposure Research Laboratory, which demonstrates field sampling, monitoring, and measurement technologies,selected Science Applications International Corporation as the verification organization to assist in field testing five field measurement devices for mercury in soil and sediment. This demonstration was funded by the SITE Program. DEMONSTRATION DESCRIPTION In May 2003, the EPA conducted a field demonstration of the DMA-80 and four other field measurement devices for mercury in soil and sediment. This verification statement focuses on the DMA-80; a similar statement has been prepared for each of the other four devices. The performance of the DMA-80 was compared to that of an off-site laboratory using the reference method, "Test Methods for Evaluating Solid Waste" (SW-846) Method 7471B (modified). To verify a wide range of performance attributes, the demonstration had both primary and secondary objectives. The primary objectives were: (1) Determining the instrument sensitivity with respect to the Method Detection Limit (MDL) and Practical Quantitation Limit (PQL); ------- (2) Determining the analytical accuracy associated with the field measurement technologies; (3) Evaluating the precision of the field measurement technologies; (4) Measuring the amount of time required for mobilization and setup, initial calibration, daily calibration, sample analysis, and demobilization; and (5) Estimating the costs associated with mercury measurements for the following four categories: capital, labor, supplies, and investigation-derived waste (IDW). Secondary objectives for the demonstration included: (1) Documenting the ease of use, as well as the skills and training required to properly operate the device; (2) Documenting potential health and safety concerns associated with operating the device; (3) Documenting the portability of trie device; (4) Evaluating the device durability based on its materials of construction and engineering design; and (5) Documenting the availability of the device and associated spare parts. The DMA-80 analyzed 59 field soil samples, 13 field sediment samples, 42 spiked field samples, and 59 performance evaluation (PE) standard reference material (SRM) samples in the demonstration. The field samples were collected in four areas contaminated with mercury, the spiked samples were from these same locations, and the PE samples were obtained from a commercial provider. Collectively, the field and PE samples provided the different matrix types and the different concentrations of mercury needed to perform a comprehensive evaluation of the DMA-80. A complete description of the demonstration and a summary of the results are available in the Innovative Technology Verification Report: "Field Measurement Technology for Mercury in Soil and Sediment—Milestone Inc.'s Direct Mercury Analyzer (DMA)-80"(EPA/600/R-04/012). TECHNOLOGY DESCRIPTION The DMA-80 is an atomic adsorption spectrophotometer based on mercury vaporization, amalgamation, desorption, and analysis of samples using an adsorbance spectrophotometer. Mercury samples are heated to 750° to 800°C, causing organic materials to be decomposed and mercury to be vaporized in a carriergas of oxygen. The oxygen flow carries the vaporized mercury to the amalgamator, where it deposits on gold-covered molecular sieves. Potential interferents are carried out of the system with the continuous gas stream. The mercury deposits are then desorbed as the amalgamator is heated; vaporized mercury is transported to the spectrophotometer for analysis. The spectrophotometer uses a mercury vapor lamp as its light source. Light from the lamp is directed through an excitation filter before it irradiates the vaporized mercury contained in a quartz cuvette. The detector utilizes two sequential cuvettes: one for low concentration samples and the other for high concentration samples. Light which is not absorbed by the mercury vapors, then passes through an emission filter before being measured by the detector. Results are transmitted to the system controller, where concentrations are calculated based on sample mass and the detector response relative to a calibration curve. During the demonstration, no extraction or sample digestion was required. Individual samples were mixed manually using a stainless steel spatula. (Note that samples were already considered to be homogeneous based upon the standard operating procedure used by SAIC to homogenize and aliquot all samples.) This same spatula was used to transfer the sample to a nickel weigh boat designed to fit the auto sampler. The sample was then weighed on a digital balance and placed on the 40-slot, auto sampler tray. The sample weight was automatically relayed to the DMA-80 controller; sequential sample numbers were automatically entered by the software in the data table in the location corresponding to the auto sampler location (1 -40). Site-specific sample identification numbers were entered manually. The sample was analyzed, and the device displayed the mercury concentration in parts per million, which is equivalent to a soil concentration in milligrams per kilogram. IV ------- ACTION LIMITS Action limits and concentrations of interest vary, and are project specific. There are, however, action limits which can be considered as potential reference points. The EPA Region IX Preliminary Remedial Goals for mercury are 23 mg/kg in residential soil and 310 mg/kg in industrial soil. VERIFICATION OF PERFORMANCE To ensure data usability, data quality indicators for accuracy, precision, representativeness, completeness, comparability, and sensitivity were assessed for the reference method based on project-specific QA objectives. Key demonstration findings are summarized below for the primary objectives. Sensitivity: The two primary sensitivity evaluations performed for this demonstration were the MDL and PQL. Both will vary dependent upon whether the matrix is a soil, waste, or aqueous solution. Only soils/sediments were tested during this demonstration, and therefore, MDL calculations and PQL determinations for this evaluation are limited to those matrices. By definition, values measured below the PQL should not be considered accurate or precise and those below the MDL are not distinguishable from background noise. Method Detection Limit - The evaluation of an MDL requires seven different measurements of a low concentration standard or sample. Following the procedures established in the 40 Code of Federal Regulations (CFR) Part 136, the MDL is estimated between 0.049 and 0.068 mg/kg. The equivalent calculated MDL for the referee laboratory is 0.0026 mg/kg. The calculated MDL is only intended as a statistical estimation and not a true test of instrument sensitivity. Practical Quantitation Limit - The PQL for this instrument is approximately 0.082 mg/kg (the concentration of a SRM used during the demonstration) for soil and sediment materials. It is possible that the PQL may be as low as the MDL but there were no SRMs tested at this lower concentration. The referee laboratory PQL confirmed during the demonstration is 0.005 mg/kg, with a %D < 10%. Accuracy: The results from the DMA-80 were compared to the 95% prediction interval for the SRM materials and to the referee laboratory results (Method 7471B). DMA-80 results were within SRM 95% prediction intervals 93% of the time, which suggests significant equivalence to certified standards. The number of Milestone average values less than 30% different from the referee laboratory results or SRM reference values; however, was 16 of 30 different sample lots. Only 2 of 30 Milestone average results have relative percent differences greater than 100% for this same group of samples. However, when making the comparison between Milestone and ALSI data, and taking into account the possible bias associated with both sets of data, this comparison may be within reasonable expectations for considering these two separate analyses to be equivalent. With the exception of a slight low bias for the referee laboratory and a slight high bias for the DMA-80 (similar to biases observed during other inter-laboratory studies), the data sets for the DMA-80 compared to the referee laboratory were considered to be similar and within expected statistical variation. Precision: The precision of the Milestone field instrument is very comparable to the referee laboratory precision, and within expected precision variation for soil and sediment matrices. The overall average relative standard deviation (RSD) was 23.7% for the referee laboratory and 19.4% for Milestone. Both the laboratory and Milestone precision results are within the predicted 25% RSD objective for precision expected from both analytical and sampling variance. Measurement Time: From the time of sample receipt, Milestone required 22 hours and 10 minutes to prepare a draft data package containing mercury results for 173 samples. One technician performed all setup, calibration checks, sample preparation-and analysis, and equipment demobilization. Individual analyses took 5 minutes each (from the time the sample was injected until results were displayed), but the total time per analysis averaged nearly 8 minutes when all field activities and data package preparation were included in the calculation. Measurement Costs: The cost per analysis based upon 173 samples, when renting the DMA-80, is $35.90 per sample. The cost per analysis for the 173 samples, excluding rental fee, is $18.55 per sample. Based on the 3-day field demonstration, the total cost for equipment rental and necessary supplies is estimated at $6,210. The cost breakout by category is: capital costs, 48.3%; supplies, 9.5%; support equipment, 4.5%; labor, 14.5%; and IDW, 23.2%. ------- Key demonstration findings are summarized below for the secondary objectives. Ease of Use: Based on observations made during the demonstration, the DMA-80 is easy to operate, requiring one field technician with a basic knowledge of chemistry acquired on the job or in a university and training on the DMA-80: A 1-day training course on instrument operation is offered at additional cost; this training would likely be necessary for most device operators who have no previous laboratory experience. Potential Health and Safety Concerns: No significant health and safety concerns were noted during the demonstration. The only potential health and safety concerns identified were the generation of mercury vapors and the use of oxygen as the carrier gas. The vendor recommends and can provide a mercury filter; oxygen can be safely handled using standard laboratory procedures. Portability: The DMA-80 was not easily portable (by hand) due to its size (80 cm by 42 cm by 30 cm high) and weight (56 kg). It was easy to set up and can be taken any place accessible to a small van or SUV. The instrument is better characterized as mobile rather than field portable. It operates on 110 or 220 volt AC current; no battery power supply is available. Durability: The DMA-80 was well designed and constructed for durability. The auto sampler piston required re-alignment once early in the demonstration, an operation normally required after shipment. In two incidents related to piston alignment, one sample was dropped by the weigh boat injector and the auto-sampler tray laterjammed. These problems were easily rectified, requiring less than 5 minutes each to troubleshoot and fix. Availability of the Device: The DMA-80 is readily available for lease, or purchase. DMA-80 rental is available on a limited basis. Spare parts and consumable supplies can be added to the original DMA-80 order or can be received within 24 to 48 hours of order placement. Supplies and standards not provided by Milestone are readily available from laboratory supply firms. PERFORMANCE SUMMARY In summary, during the demonstration, the DMA-80 exhibited the following desirable characteristics of a field mercury measurement device: (1) good accuracy, (2) good precision, (3) high sample throughput, (4) low measurement costs, and (5) ease of use. During the demonstration the DMA-80 was found to have the following limitation: (1) non-portable due to the instrument size and weight. The demonstration findings collectively indicated that the DMA-80 is a reliable field measurement device for mercury in soil and sediment. NOTICE: EPA verifications are based on an evaluation of technology performance under specific, predetermined criteria and appropriate quality assurance procedures. The EPA makes no expressed or implied warranties as to the performance of the technology and does not certify that a technology will always operate as verified. The end user is solely responsible for complying with any and all applicable federal, state, and local requirements. VI ------- Foreword The U.S. Environmental Protection Agency (EPA) is charged by Congress with protecting the nation's natural resources. Under the mandate of national environmental laws, the Agency strives to formulate and implement actions leading to a compatible balance between human activities and the ability of natural systems to support and nurture life. To meet this mandate, the EPA' s Office of Research and Development provides data and scientific support that can be used to solve environmental problems, build the scientific knowledge base needed to manage ecological resources wisely, understand how pollutants affect public health, and prevent or reduce environmental risks. The National Exposure Research Laboratory is the agency's center for investigation of technical and management approaches for identifying and quantifying risks to human health and the environment. Goals of the laboratory's research program are to (1) develop and evaluate methods and technologies for characterizing and monitoring air, soil, and water; (2) support regulatory and policy decisions; and (3) provide the scientific support needed to ensure effective implementation of environmental regulations and strategies. The EPA's Superfund Innovative Technology Evaluation (SITE) Program evaluates technologies designed for characterization and remediation ofcontaminated Superfund and Resource Conservation and Recovery Act (RCRA) sites. The SITE Program was created to provide reliable cost and performance data in order to speed acceptance and use of innovative remediation, characterization, and monitoring technologies by the regulatory and user community. Effective monitoring and measurement technologies are needed to assess the degree of contamination ata site, provide .data that can be used to determine the risk to public health or the environment, and monitor the success or failure of a remediation process. One component of the EPA SITE Program, the Monitoring and Measurement Technology (MMT) Program, demonstrates and evaluates innovative technologies to meet these needs. Candidate technologies can originate within the federal government orthe private sector. Through the SITE Program, developers are given the opportunity to conduct a rigorous demonstration of their technologies under actual field conditions. By completing the demonstration and distributing the results, the agency establishes a baseline foracceptance and use of these technologies. The MMT Program is managed by the Office of Research and Development's Environmental Sciences Division in Las Vegas, NV. Gary Foley, Ph. D. Director National Exposure Research Laboratory Office of Research and Development VII ------- Abstract Milestone's Direct Mercury Analyzer (DMA-80) was demonstrated under the U.S. Environmental Protection Agency Superfund Innovative Technology Evaluation Program in May 2003 at the Oak Ridge National Laboratory (ORNL) in Oak Ridge, TN. The purpose of the demonstration was to collect reliable performance and cost data for the DMA-80 and four other field measurement devices for mercury in soil and sediment. The key objectives of the demonstration were: 1) determine sensitivity of each instrument with respect to a vendor-generated method detection limit (MDL) and practical quantitation limit (PQL); 2) determine analytical accuracy associated with vendor field measurements using field samples and standard reference materials (SRMs); 3) evaluate the precision of vendor field measurements; 4) measure time required to perform mercury measurements; and 5) estimate costs associated with mercury measurements for capital, labor, supplies, and investigation-derived wastes. The demonstration involved analysis of standard reference materials (SRMs), field samples collected from four sites, and spiked field samples for mercury. The performance results for a given field measurement device were compared to those of an off-site laboratory us ing Deference method, "Test Methods for Evaluating Solid Waste" (SW-846) Method 7471B. The sensitivity, accuracy, and precision measurements were successfully completed. The DMA-80 performed well in all these categories. During the demonstration, Milestone required 22 hours and 10 minutes for the analysis of 173 samples. The measurement costs were estimated to be $6,210 for Milestone's DMA-80 rental option or $35.90 per sample; $18.55 per sample excluding rental fees. The, DMA-80 exhibited good ease of use and durability; as well as no major health and safety concerns. However, the device portability is somewhat limited by its size. Despite these limitations, the demonstration findings collectively indicated that the DMA-80 is a reliable field mobile measurement device for mercury in soil. VIII ------- Contents Notice ii Verification Statement iii Foreword ; vii Abstract viii Contents ix Tables xii Figures xiii Abbreviations, Acronyms, and Symbols xiv Acknowledgments xvi Chapter Page 1 Introduction 1 1.1 Description of the SITE Program 1 1.2 Scope of the Demonstration 2 1.2.1 Phase I 2 1.2.2 Phase II :...-. 2 1.3 . Mercury Chemistry and Analysis 3 1.3.1 Mercury Chemistry 3 1.3.2 Mercury Analysis ,. 4 2 Technology Description 6 2.1 Description of Atomic Absorption Spectroscopy 6 2.2 Description of the DMA-80 6 2.3 Developer Contact Information 8 3 Field Sample Collection Locations and Demonstration Site 9 3.1 Carson River 10 3.1.1 Site Description 10 3.1.2 Sample Collection 10 3.2 Y-12 National Security Complex 11 3.2.1 Site Description 11 3.2.2 Sample Collection 11 3.3 Confidential Manufacturing Site 11 3.3.1 Site Description 11 ix ------- Contents (Continued) Chapter • Page 3.3.2 Sample Collection 12 3.4 Puget Sound 12 3.4.1 Site Description 12 3.4.2 Sample Collection : 12 3.5 Demonstration Site 13 3.6 SAIC GeoMechanics Laboratory 14 4 Demonstration Approach 15 4.1 Demonstration Objectives 15 4.2 Demonstration Design 16 4.2.1 Approach for Addressing Primary Objectives 16 4.2.2 Approach for Addressing Secondary Objectives 20 4.3 Sample Preparation and Management 21 4.3.1 Sample Preparation 21 4.3.2 Sample Management 24 4.4 Reference Method Confirmatory Process 25 4.4.1 Reference Method Selection 25 4.4.2 Referee Laboratory Selection 25 4.4.3 Summary of Analytical Methods 27 4.5 Deviations from the Demonstration Plan 28 5 Assessment of Laboratory Quality Control Measurements 29 5.1 Laboratory QA Summ.ary 29 5.2 Data Quality Indicators for Mercury Analysis 29 5.3 Conclusions and Data Quality Limitations 30 5.4 Audit Findings 32 6 Performance of the DMA-80 33 6.1 Primary Objectives 34 6.1.1 Sensitivity , 34 6.1.2 Accuracy 36 6.1.3 Precision 44 6.1.4 Time Required for Mercury Measurement : 47 6.1.5 Cost ." '. ; 48 6.2 Secondary Objectives 49 .6.2.1 Ease of Use 49 6.2.2 Health and Safety Concerns 52 6.2.3 Portability of the Device 53 6.2.4 Instrument Durability , 54 6.2.5 Availability of Vendor Instruments and Supplies 54 7 Econom ic Analysis 56 7.1 Issues and Assumptions 56 7.1.1 . Capital Equipment Cost 56 7.1.2 Cost of Supplies 57 ------- Contents (Continued) Chapter Page 7.1.3 Support Equipment Cost 57 7.1.4 Labor Cost 57 7.1.5 Investigation-Derived Waste Disposal Cost 58 7.1.6 Costs Not Included 58 7.2 DMA-80 Costs 59 7.2.1 Capital Equipment Cost 59 7.2.2 Cost of Supplies 60 7.2.3 Support Equipment Cost 61 7.2.4 Labor Cost , 61 7.2.5 Investigation-Derived Waste Disposal Cost 61. 7.2.6 Summary of DMA-80 Costs 62 7.3 Typical Reference Method Costs ' 63 8 Summary of Demonstration Results 64 8.1 Primary Objectives 64 8.2 Secondary Objectives . . . . • 65 9 Bibliography 68 Appendix A - Milestone Comments 69 Appendix B - Statistical Analysis : 70 XI ------- Tables Table Page 1-1 Physical and Chemical Properties of Mercury 4 1-2 Methods for Mercury Analysis in Solids or Aqueous Soil Extracts 5 3-1 Summary of Site Characteristics 10 4-1 Demonstration Objectives 15 4-2 Summary of Secondary Objective Observations Recorded During the Demonstration 20 4-3 Field Samples Collected from the Four Sites 22 4-4 Analytical Methods for Non-Critical Parameters 28 5-1 MS/MSD Summary 30 5-2 LCS Summary ' 30 5-3 Precision Summary v 31 5-4 Low Check Standards 31 6-1 Distribution of Samples Prepared for Milestone and the Referee Laboratory 33 6-2 Milestone SRM Comparison '. 38 6-3 ALSI SRM Com parison 38 6-4 Accuracy Evaluation by Hypothesis Testing " 39 6-5 Number of Samples Within Each %D Range ' 41 6-6 Concentration of Non-Target Analytes 41 6-7 Evaluation of Precision 45 6-8 Time Measurements for Milestone : . . 48 7-1 Capital Cost Summary for the DMA-80 60 7-2 Carrier Gas Cost Summary 60 7-3 Mercury Trap Costs . . . 61 7-4 Labor Costs 61 7-5 IDW Costs 62 7-6 Summary of Rental Costs for the DMA-80 62 7-7 DMA-80 Costs by Category 63 8-1 Distribution of Samples Prepared for Milestone and the Referee Laboratory : 65 8-2 Summary of DMA-80 Results for the Primary Objectives 66 8-3 Summary of DMA-80 Results for the Secondary Objectives 67 B-1 Summary of Unified Hypothesis Test 72 XII ------- Figures Figure Page 2-1 Schematic of DMA-80 7 2-2 Photograph of the DMA-80 during the field demonstration 8 3-1 Tent and field conditions during the demonstration at Oak Ridge, TN 13 3-2 Demonstration site and Building 5507. . 13 4-1 Test sample preparation at the SAIC GeoMechanics Laboratory * 23 6-1 Data plot for low concentration sample results 42 6-2 Data plot for high concentration sample results 43 6-3 Calibration result screen 49 6-4 Calibration curve screen .50 6-5 System control display screen 51 6-6 Sample peak screen 52 7-1 Capital equipment costs 59 xiii ------- Abbreviations, Acronyms, and Symbols % Percent %D Percent difference °C Degrees Celsius ug/kg Microgram per kilogram AAS Atomic absorption spectrometry ALSI Analytical Laboratory Services, Inc. bgs Below ground surface cm Centimeter CFR Code of Federal Regulations Cl Confidence Interval COC Chain of custody DMA-80 Direct Mercury Analyzer DOE Department of Energy EPA United States Environmental Protection Agency g Gram H&S . Health and Safety Hg Mercury HgCI2 Mercury (II) chloride IDL Instrument detection limit IDW Investigation-derived waste ITVR Innovative Technology Verification Report kg Kilogram L Liter mL/min Milliliter per minute LCS Laboratory control sample LEFPC Lower East Fork Poplar Creek m • Meter MDL Method detection limit mg Milligram mg/kg Milligram per kilogram mL Milliliter mm Millimeter MMT Monitoring and Measurement Technology MS/MSD Matrix Spike/ Matrix Spike Duplicate NERL National Exposure Research Laboratory ng Nanogram nm Nanometer . . ORD Office of Research and Development xiv ------- Abbreviations, Acronyms, and Symbols (Continued) ORR Oak Ridge Reservation ORNL Oak Ridge National Laboratory OSW ER Office of Solid Waste and Emergency Response ppb Parts per billion PPE Personal protective equipment ppm Parts per million PQL Practical quantitation limit QA Quality assurance QAPP Quality Assurance Project Plan QC Quality control RPD Relative percent difference RSD Relative standard deviation SAIC Science Applications International Corporation SITE Superfund Innovative Technology Evaluation SOP Standard operating procedure SRM Standard reference material SW-846 Test Methods for Evaluating Splid Waste; Physical/Chemical Methods TOC Total organic carbon TOM Task Order Manager UL Underwriters Laboratory UEFPC Upper East Fork of Poplar Creek Y-12 Y-12 Oak Ridge Security Complex, Oak Ridge, TN xv ------- Acknowledgments The U.S. Environmental Protection Agency (EPA) Superfund Innovative Technology Evaluation wishes to acknowledge the support of the following individuals in performing the demonstration and preparing this document: Elizabeth Phillips of the U.S. Department of Energy Oak Ridge National Laboratory (ORNL); Stephen Childs, Thomas Early, Roger Jenkins, and Monty Ross of the UT-Battelle ORNL; Dale Rector of the Tennessee Department of Environment and Conservation (TDEC) Department of Energy Oversight; Mikhail Mensh of Milestone, Inc.; Leroy Lewis of the Idaho National Engineering and Environmental Laboratory, retired; Ishwar Murarka of the EPA Science Advisory Board, member; Danny Reible of Louisiana State University; Mike Bolen, Joseph Evans, Julia Gartseff, Sara Hartwell, Cathleen Hubbard, Kevin Jago, Andrew Matuson, Allen Motley, John Nicklas, Maurice Owens, Nancy Patti, Fernando Padilla, Mark Pruitt, James Rawe, Herb Skovronek, and Joseph Tillman of Science Applications International Corporation (SAIC); Scott Jacobs and Ann Vega of the EPA National Risk Management Research Laboratory's Land Remediation and Pollution Control Division; and Brian Schumacher of the EPA National Exposure Research Laboratory. This document was QA reviewed by George Brilis of the EPA National Exposure Research Laboratory. xvi ------- Chapter 1 Introduction The U.S. Environmental Protection Agency (EPA) under the Office of Research and Development (ORD), National Exposure Research Laboratory (NERL), conducted a demonstration to evaluate the performance of innovative field measurement devices for their ability to measure mercury concentrations in soils and sediments. This Innovative Technology Verification Report (ITVR) presents demonstration performance results and associated costs of Milestone's Direct Mercury Analyzer (DMA) -80. The vendor-prepared comments regarding the demonstration are presented in Appendix A. The demonstration was conducted as part of the EPA Superfund Innovative Technology Evaluation (SITE) Monitoring and Measurement Technology (MMT) Program. Mercury contaminated soils and sediments, collected from four sites, within the continental U.S., comprised the majority of samples analyzed during the evaluation. Some soil and sediment samples were spiked, with mercury (II) chloride (HgCI2) to provide concentrations not occurring in the field samples. Certified standard reference material (SRM) samples were also used to provide samples with certified mercury concentrations and to increase the matrix variety. The demonstration was conducted at the Department of Energy (DOE) Oak Ridge National Laboratory (ORNL) in Oak Ridge, TN during the week of May 5, 2003. The purpose of the demonstration was to obtain reliable performance and cost data for field measurement devices in order to 1) provide potential users with a better. understanding of the devices' performance and operating costs under well-defined field conditions and 2) provide the instrument vendors with documented results thatcan assist them in promoting acceptance and use of their devices. The results obtained using the five field mercury measurement devices were compared to the mercury results obtained for identical sample sets (samples, spiked samples, and SRMs) analyzed ata referee laboratory. The referee laboratory, which was selected prior to the demonstration, used a well-established EPA reference method. 1.1 Description of the SITE Program Performance verification of innovative environmental technologies is an integral part of the regulatory and research mission of the EPA. The SITE Program was established by EPA's Office of Solid Waste and Emergency Response (OSWER) and ORD under the Superfund Amendments and Reauthorization Act of 1986. The overall goal of the SITE Program is to conduct performance verification studies and to promote the. acceptance of innovative technologies that may be used to achieve long-term protection of human health and the environment. Theprogram isdesigned tomeetthree main objectives: 1) identify and remove obstacles to the development and commercial use of innovative technologies; 2) demonstrate promising innovative technologies and gather reliable performance and cost information to support site characterization and cleanup activities; and 3) develop procedures and policies that encourage the use of innovative technologies at Superfund sites, as well as at other waste sites or commercial facilities. The SITE Program includes the following elements: The MMT Program evaluates innovative technologies that sample, detect, monitor, or measure hazardous and toxic substances in soil, water, and sediment samples. These technologies are expected to provide better, faster, or more cost-effective methods for ------- producing real-time data during site characterization and remediation studies than conventional technologies. The Remediation Technology Program conducts demonstrations of innovative treatment technologies to provide reliable performance, cost, and applicability data for site cleanups. The Technology Transfer Program provides and disseminates technical information in the form of updates, brochures, and other publications that promote the SITE Program and participating technologies. The Technology Transfer Program also offers technical assistance, training, and workshops in the support of the technologies. A significant number of these activities are performed by EPA's Technology Innovation Office. The Field Analysis of Mercury in Soils and Sediments demonstration was performed under the MMT Program. The MMT Program provides developers of innovative hazardous waste sampling, detection, monitoring, and measurement devices with an opportunity to demonstrate the performance of their devices under actual field conditions. The main objectives of the MMT Program are as follows: • Test and verify the performance of innovative field sampling and analytical technologies that enhance sampling, monitoring, and site characterization capabilities. Identify performance attributes of innovative technologies that address field sampling, monitoring, and characterization problems in a cost-effective and efficient manner. Prepare protocols, guidelines, methods, and other technical publications that enhance acceptance of these technologies for routine use: The MMT Program is administered by the Environmental Sciences Division of the NERL in Las Vegas, NV. The NERL is the EPA center for investigation of technical and management approaches for identifying and quantifying risks to human health and the environment. The NERL mission components include 1) developing and evaluating methods and technologies for sampling, monitoring, and characterizing water, air, soil, and sediment; 2) supporting regulatory and policy decisions; and 3) providing technical support to ensure the effective implementation of environmental regulations and strategies. 1.2 Scope of the Demonstration The demonstration project consisted of two separate phases: Phase I involved obtaining information on prospective vendors having viable mercury detection instrumentation. Phase II consisted of field and planning activities leading up to and including the demonstration activities. The following subsections provide detail on both of these project phases. 1.2.1 Phase I Phase I was initiated by making contact with knowledgeable sources on the subject of "mercury in soil" •detection devices. Contacts included individuals within EPA, Science Applications International Corporation (SAIC), and industrywhere measurement of mercury in soil was known to be conducted. Industry contacts included laboratories and private developers of mercury detection instrumentation. In addition, the EPA Task Order Manager (TOM) provided contacts for "industry players" who had participated in previous MMT demonstrations. SAIC also investigated university and other research-type contacts for knowledgeable sources within the subject area. These contacts led to additional knowledgeable sources on the subject, which in turn led to various Internet searches. The Internet searches were very successful in finding additional companies involved with mercury detection devices. All in all, these research activities generated an original list of approximately 30 companies potentially involved in the measurement of mercury in soils. The list included both international and U.S. companies. Each of these companies was contacted by phone or email to acquire further information. The contacts resulted in 10 companies that appeared to have viable technologies. Due to instrument design (i.e., the instrument's ability to measure mercury in soils and sediments), business strategies, and stage of technology development, only 5 of those 10 vendors participated in the field demonstration portion of phase II. 12.2 Phase II Phase II of the demonstration project involved strategic planning, field-related activities for the demonstration, data analysis, data interpretation, and preparation of the ITVRs. Phase II included pre-demonstration and demonstration activities, as described in the following subsections. ------- 1.2.2.1 Pre-Demonstration Activities The pre-demonstration activities were completed in the fall 2002. There were six objectives for the pre-demonstration: Establish concentration ranges for testing.vendors' analytical equipment during the demonstration. Collect soil and sediment field samples to be used in the demonstration. Evaluate sample homogenization procedures. Determine mercury concentrations in homogenized soils and sediments. Selecta reference method and qualify potential referee laboratories for the demonstration. Provide soil and sediment samples to the vendors for self-evaluation of their instruments, as a precursor to the demonstration. As an integral part of meeting these objectives, a pre- demonstration sampling event was conducted in September 2002 to collect field samples of soils and sediments containing different levels of mercury. The field samples were obtained from the following locations: Carson River Mercury site - near Dayton, NV Y-12 National Security Complex - Oak Ridge, TN A confidential manufacturing facility - eastern U.S. Puget Sound - Bellingham Bay, WA Immediately after collecting field sample material from the sites noted above, the general mercury concentrations in the soils and sediments were confirmed by quick turnaround laboratory analysis of field-collected subsamples using method SW-7471B. The field sample materials were then shipped to a soil preparation laboratory for homogenization. Additional pre-demonstration activities are detailed in Chapter 4. 1.2.2.2 Demonstration Activities Specific objectives, for this SITE demonstration were developed and defined in a Field Demonstration and Quality Assurance Project Plan (QAPP) (EPA Report # EPA/600/R-03/053). The Field Demonstration QAPP is available through the EPA ORD web site (http://www.epa.gov/ORD/SITE) or from the EPA Project Manager. The demonstration objectives were subdivided into two categories: primary and secondary. Primary objectives are goals of the demonstration study that need to be achieved for technology verification. The measurements used to achieve primary objectives are referred to as critical. These measurements typically produce quantitative results that can be verified using inferential and descriptive statistics. Secondary objectives are additional goals of the demonstration study developed for acquiring other information of interest about the technology that is not directly related to verifying the primary objectives. The measurements required forachieving secondary objectives are considered to be noncritical. Therefore, the analysis of secondary objectives is typically more qualitative in nature and often uses observations and sometimes descriptive statistics. The field portion of the demonstration involved evaluating the capabilities of five mercury-analyzing instruments to measure mercury, concentrations in soil and sediment. During the demonstration, each instrument vendor received three types of samples 1) homogenized field samples referred to as "field samples", 2) certified SRMs, and 3) spiked field samples (spikes). Spikes were prepared by adding known quantities of HgCI2 to field samples. Together, the field samples, SRMs, and spikes are referred to as "demonstration samples" for the purpose of this ITVR. All demonstration samples were independently analyzed by a carefully selected referee laboratory. The experimental design for the demonstration is detailed in Chapter 4. 1.3 Mercury Chemistry and Analysis 1.3.1 Mercury Chemistry Elemental mercury is the only metal that occurs as a liquid at ambient temperatures. Mercury naturally occurs, primarily within the ore, cinnabar, as m ercury sulfide (HgS). Mercury easily forms amalgams with many other metals, including gold. As a result, mercury has historically been used to recover gold from ores. Mercury is ionically stable; however, it is very volatile for a metal. Table 1-1 lists selected physical and chemical properties of elemental mercury. . ------- Table 1-1. Physical and Chemical Properties of Mercury Properties Data Appearance Hardness Abundance Density @ 25'C Vapor Pressure @ 25 'C Volatilizes @ Solidifies @ Silver-white, mobile, liquid. Liquid 0.5% in Earth's crust 13.53g/mL 0.002 mm 356 °C -39 °C Source: Merck Index, 1983 Historically, mercury releases to the environment included a number of industrial processes such as chloralkali manufacturing, copper and zinc smelting operations, paint application, waste oil combustion, geothermal energy plants, municipal waste incineration, ink manufacturing, chemical manufacturing, paper mills, leather tanning, pharmaceutical production, and textile manufacturing. In addition, industrial and domestic mercury-containing products, such as thermometers, electrical switches, and batteries, are disposed of as solid wastes in landfills (EPA, July 1995). Mercury is also an indigenous compound at many abandoned mining sites and is, of course, found as a natural ore. At mercury-contaminated sites, mercury exists in mercuric form (Hg2+),mercurous form (Hg22*), elemental form (Hg°), and alkylated form (e.g., methyl or ethyl mercury). Hg22+ and Hg2* are the more stable forms under oxidizing conditions. Under mildly reducing conditions, both organically bound mercury and inorganic mercury may be degraded to elemental mercury, which can then be converted readily to methyl or ethyl mercury by biotic and abiotic processes. Methyl and ethyl mercury are the most toxic forms of mercury; the alkylated mercury compounds are volatile and soluble in water. Mercury (II) forms relatively strong complexes with CI" and CO32'. Mercury (II) also forms complexes with inorganic ligands such as fluoride (F~), bromide (Br~), iodide (I"), sulfate (SO42~), sulfide (S2~), and phosphate (PO43') and forms strong complexes with organic ligands, such as sulfhydryl groups, amino acids, and humic and fulvicacids. The insoluble HgS is formed under mildly reducing conditions. 1.3.2 Mercury Analysis There are several laboratory-based, EPA promulgated methods for the analysis of mercury in solid and liquid hazardous waste matrices. In addition, there are several performance-based methods for the determination of various mercury species. Table 1-2 summarizes the commonly used methods for measuring mercury in both solid and liquid matrices, as identified through a review of the EPA Test Method Index and SW-846. A discussion of the choice of reference method is presented in Chapter 4. ------- Table 1-2. Methods for Mercury Analysis in Solids or Aqueous Soil Extracts Method Analytical Type(s) of Technology ' Mercury analyzed Approximate Concentration Range Comments SW-7471B CVAAS SW-7472 ASV SW-7473 TD, amalgamation, andAAS SW-7474 AFS EPA 1631 CVAFS EPA 245.7 CVAFS EPA 6200 FPXRF inorganic mercury 10-2,000 ppb organo-mercury inorganic mercury 0.1-10,000 ppb organo-mercury inorganic mercury 0.2 - 400 ppb organo-mercury inorganic mercury 1 ppb - ppm organo-mercury inorganic mercury 0.5 -100 ppt • organo-mercury inorganic mercury 0.5 - 200 ppt organo-mercury inorganic mercury >30 mg/kg Manual cold vapor technique widely used for total mercury determinations Newer, less widely accepted method Allows for total decomposition analysis Allows for total decomposition analysis; less widely used/reference Requires "trace" analysis procedures; written for aqueous matrices; Appendix A of method written for sediment/soil samples Requires "trace" analysis procedures; written for aqueous matrices; will require dilutions of high-concentration mercury samples Considered a screening protocol AAS = Atomic Absorption Spectrometry AAF = Atomic Fluorescence Spectrometry AFS = Atomic Fluorescence Spectrometry ASV = Anodic Stripping Voltammetry CVAAS = Cold Vapor Atomic Absorption Spectrometry CVAFS = Cold Vapor Atomic Fluorescence Spectrometry FPXRF = Field Portable X-ray Fluorescence EPA = U.S. Environmental Protection Agency mg/kg = milligram per kilogram ppb = parts per billion ppm = parts per million ppt = parts per trillion SW = solid waste TD = thermal decomposition ------- Chapter 2 Technology Description This chapter provides a detailed description of 1) the thermal decomposition method of atomic absorption spectroscopy (AAS), which is the type of technology on which Milestone's instrument is based, and 2) a detailed description of the DMA-80. 2.1 Description of Atomic Absorption Spectroscopy The principle of analysis used by the DMA-80 is AAS preceded by thermal decomposition, catalytic reduction, and amalgamation desorption. AAS is based on the direct relationship between the absorption of light of a specific wavelength by gas-phase atoms of an inorganic analyte, and the concentration of those atoms. Because samples analyzed by AAS are usually liquids or solids, the analyte atoms or ions must be vaporized in a flame or graphite furnace prior to the determination. The vaporized atoms absorb light of an analyte-specific wavelength, and make transitions to higher electronic energy levels. The analyte concentration is directly proportional to the amount of light absorbed. Concentration measurements are usually determined from a working curve after calibrating the instrument with standards of known concentration. In reference to AAS as a general analytical application, thermal decomposition, followed by atomic absorption spectroscopy, is common; however, the mechanism of analyte recovery for the determination step may vary. Examples include cold vapor traps and amalgamation desorption. When operating the DMA-80, a sample of known mass is placed in the drying and decomposition furnace and heated to 750 Celsius (°C).The liquid or solid sample is dried and organic materials are decomposed. AAS utilizes the absorption of light by an element, in this case mercury vapor, as compared to a standard to quantify the mass of analyte present in a sample. The absorption of light is proportional to the concentration of analyte present. The wavelength of the light source is specific to the contaminant of interest. For mercury, the wavelength is 254 nm. 2.2 Description of the DMA-80 The Milestone DMA-80 is an integrated system that utilizes thermal decomposition, catalytic reduction, amalgamation desorption, and AAS to rapidly analyze solid and liquid samples. Applications and Specifications - The Milestone DMA-80 enables analysts to rapidly determine total mercury concentrations in solid and liquid samples without sample pretreatment or digestion. Maximum sample sizes are 500 uL and 500 mg, respectively, for liquid and solid samples. According to Milestone, individual sam pie results are available in approximately 5 minutes and up to 40 samples can be processed, start to finish, in a 4-hour period. Per Milestone, results are reportedly independent of matrix, detection limits range from 0.5 to 600 ng mercury on a mass basis, and reproducibility (measurement error for two or more samples) is less than 1.5 percent. Results from this demonstration are reported in Chapter 6. In areas where mercury contamination in the soil is an existing problem, the background signal may be significantly increased due to airborne dust containing mercury. As.with other AAS technologies, memory effects between analyses may be encountered when analyzing a sample of high mercury content (e.g., 400 ng) prior to analyzing one of low content (e.g., 25 ng). Typically, to minimize memoryeffects, samples are analyzed in batches of low and high concentrations, analyzing those of low ------- concentration first. If this batching process cannot be accomplished, a blank analysis, with an extended decomposition time, maybe required following the analysis of a highly-concentrated sample to limit memory effects. Co-absorbing gases, such as free chlorine and certain organics (as indicated in Methods 7470A and 7471B), should not interfere due to the release of decomposition products by the decomposition furnace, removal of some decomposition products by the decom position catalyst, and the selective entrapment of mercury vapor on the amalgamator. As with other analytical devices, field conditions that may affect accuracy and precision include sample homogeneity, sample handling errors, unpredictable matrix effects, and sample and cell contamination (EPA, 1998). Because no sample digestion or pre-treatment is required, no reagents are utilized. As a result, the only waste materials are residual sample material, excess sample, and decontamination solution. The DMA-80 volatilizes mercury into the oxygen stream flowing through the instrument, which ultimately exhausts to ambient air. The instrument exhaust may be attached to a fume hood with a filter, or a mercury trap may be assembled and attached in the field, based on instructions provided by the vendor. Figure 2-1 presents a schematic diagram of the thermal decomposition, catalytic reduction unit, and amalgamation desorption furnace for the DMA-80. Figure 2-1. Schematic of DMA-80 The Milestone DMA-80 is approximately 56 kg and has a dimension of 80 cm by 42 cm by 30 cm (height). The terminal has a dimension of 33 cm by 27 cm by 26 cm (height) and weighs less than 2 kg. The instrument operates on either a 110V or 230V AC source at 50 to 60 Hz. The unit is equipped with a built-in, 40-position auto sampler for solids and liquids. An optional analytical balance can be provided for automatic sample weight data transfer. (A steady table is needed for accurate weight determination.) Sample weight data transfer can be accomplished from other appropriate balances by utilizing a 9-pin connector. Other required equipment includes a micro spatula, tweezers, and digital pipets with 10-100 pi and 100-1,000 ul ranges. ------- Operation - Liquid or solid samples are introduced into the DMA-80 (Figure 2-2) individually or using the auto sampler. The sample is initially dried in an oxygen stream passing through a quartz tube located inside a controlled heating coil. A separate cylinder supplies oxygen as a carrier gas; the flow rate is approximately 200 mL/minute at 60 psig. The combustion gases are further decomposed on a catalytic column at 750 °C. Mercury vapor is collected on a gold amalgamation trap and subsequently desorbed for quantitation. Mercury content is determined using a single beam spectrophotometer with two sequential, flow- through measurement cells. The light source is a low- pressure mercury vapor lamp. The instrument detector is a silicon UV photo-detector at 253.65 nm, with a 254 nm interference filter having a 9-nm bandwidth. Figure 2-2. Photograph of the DMA-80 during the field demonstration. Each cell has its own calibration curve. The system provides automatic switch-over between the low and high working ranges. The low range is 0-35 ng mercury; the high range is 35-600 ng mercury. Calibration standards are not provided with the instrument; however, the electronic instructions provided by Milestone included embedded links to internet sites where standards can be purchased. Results are displayed using a touch-screen, Pentium- based control terminal equipped with a keyboard and a mouse. The Windows-based system control software provides automatic data storage; edit functions to create, modify, and store commonlyused methods; andoptionsto select single or auto sample. Standard data include absorbance, mercury mass (nanogram), and total mercury concentration (parts per billion). Data can be printed to a standard printer or stored. 2.3 Developer Contact Information Additional information about the DMA-80 can be obtained from the following source: Milestone i60 B Shelton Road Monroe, CT 06468 Telephone:(203)261-6175 Fax: (203)261-6592 Email: techsales@milestonesci.com Internet: www.milestonesci.com ------- Chapter 3 Field Sample Collection Locations and Demonstration Site As previously described in Chapter 1, the demonstration in part tested the ability of all five vendor instruments to measure mercury concentrations in demonstration samples. The demonstration samples consisted of field- collected samples, spiked field samples, and SRMs. The field-collected samples comprised the majority of demonstration samples. This chapter describes the four sites from which the field samples were collected, the demonstration site, and the sample homogenization laboratory. Spiked samples were prepared from these field samples.. Screening of potential mercury-contaminated field sample sites was conducted during Phase I of the project. Four sites were selected for acquiring mercury-contaminated samples that were diverse in appearance, consistency, and mercury concentration. A key criterion was the source of the contamination. These sites included: Carson River Mercury site - near Dayton, NV The Y-12 National Security Complex (Y-12) - Oak Ridge, TN A confidential manufacturing facility -eastern U.S. Puget Sound - Bellingham Bay, WA Site Diversity - Collectively, the four sites provided sampling areas with both soil and sediment, having variable physical consistencies and variable ranges of mercury contamination. Two of the sites (Carson River and Oak Ridge) provided both soil and sediment samples. A third site (a manufacturing facility) provided just soil samples and a fourth site (Puget Sound) provided only sediment samples. Access and Cooperation - Site representatives were instrumental in providing site access, and in some cases, guidance on the best areas to collect samples from relatively high and low mercury concentrations. In addition, representatives from the host demonstration site (ORNL) provided a facility for conducting the demonstration. . At three of the sites, the soil and/or sediment sample was collected, homogenized by hand in the field, and subsampled for quick turnaround analysis. These subsamples were sent to analytical laboratories to determine the general range of mercury concentrations at each of the sites. (The Puget Sound site did not require confirmation of mercury contamination due to recently acquired mercury analytical data from another, ongoing research project.) The field-collected soil and sediment samples from all four sites were then shipped to SAIC's GeoMechanics Laboratory for a more thorough sample homogenization (see Section 4.3.1) and subsampled for redistribution to vendors during the pre-demonstration vendor self-evaluations. All five of the technology vendors performed a self- evaluation on selected samples collected and homogenized during this pre-demonstration phase of the project. For the self-evaluation, the laboratory results and SRM values were supplied to the vendor, allowing the vendor to determine how well it performed the analysis on the field samples. The results were used to gain a preliminary understanding of the field samples collected and to prepare for the demonstration. Table 3-1 summarizes key characteristics of samples collected at each of the four sites. Also included are the sample matrix, sample descriptions, and sample depth intervals. The analytical results presented in Table 3-1 are based on referee laboratory mercury results for the demonstration samples. ------- Table 3-1. Summary of Site Characteristics Site Name Carson River Mercury site Y-1 2 National Security Complex Confidential manufacturing site Puget Sound - Bellingham Bay Sampling Area Carson River Six Mile Canyon Old Hg Recovery Bldg. Poplar Creek Former plant building Sediment layer Underlying Native Material Sample Matrix Sediment Soil Soil Sediment Soil Sediment Sediment Depth water/sediment interface 3 - 8 cm bgs 0 - 1 m bgs 0 - 0.5 m bgs 3.6 -9 m bgs 1.5 -1.8m thick 0.3 m thick Description Sandy silt, with some organic debris present (plant stems and leaves) Silt with sand to sandy silt Silty-clay to sandy-gravel Silt to coarse sandy gravel Silt to sandy silt Clayey-sandy silt with various woody debris Medium-fine silty sands Hg Concentration Range 10ppb- 50 ppm 10 ppb- 1.000 ppm 0.1 - 100 ppm 0.1 - 100 ppm 5- 1,000 ppm 10 - 400 ppm 0.16- 10 ppm bgs = below ground surface. 3.1 Carson River 3.1.1 Site Description The Carson River Mercury site begins near Carson City, NV, and extends, downstream to the Lahontan Valley and the Carson Desert. During the Comstock mining era of the late 1800s, mercury was imported to the area for processing gold and silver ore. Ore mined from the Comstock Lode was transported to mill sites, where it was crushed and mixed with mercury to amalgamate the precious metals. The Nevada mills were located inVirginia City, Silver City, Gold Hill, Dayton, Six Mile Canyon, Gold Canyon, and adjacent to the Carson River between New Empire and Dayton. During the mining era, an estimated 7,500 tons of mercury were discharged into the Carson River drainage, primarily in the form of mercury-contaminated tailings (EPA Region 9, 1994). Mercury contamination is present at Carson Riveras either elemental mercury and/or inorganic mercury sulfides with less than 1%, if any, methylmercury. Mercury contamination exists in soils present at the former gold and silver mining mill sites; waterways adjacent to the mill sites; and sediment, fish, and wildlife over more than a 50-mile length of the Carson River. Mercury is also present in the sediments and adjacent flood plain of the Carson River, and in the sediments of Lahontan Reservoir, Carson Lake, Stillwater Wildlife Refuge, and Indian Lakes. In addition, tailings with elevated mercury levels are still presentat, and around, the historic mill sites, particularly in Six Mile Canyon (EPA, 2002a). 3.1.2 Sample Collection The Carson River Mercury site provided both soil and sediment samples across the range of contaminant concentrations desired for the demonstration. Sixteen near-surface soil samples were collected between 3-8 cm below ground surface (bgs). Two sediment samples were collected at the water-to-sediment interface. All 18 samples were collected on September 23-24, 2002 with a hand shovel. Samples were collected in Six Mile Canyon and along the Carson River. The sampling sites were selected based upon historical data from the site. Specific sampling locations in the Six Mile Canyon were selected based upon local terrain and visible soil conditions (e.g.,'color and particle size). The specific sites were selected to obtain soil samples with as much variety in mercury concentration as possible. These sites included hills, run-off pathways, and dry river bed areas. Sampling locations along the Carson River were selected based upon historical mine locations, localterrain, and river flow. When collecting the soil samples, approximately 3 cm of surface soil was scraped to the side. The sample was then collected with a shovel, screened through a 6.3-millimeter (mm) (0.25-inch) sieve to remove larger material, and collected in4-liter(L)sealable bags identified with a permanent marker. The sediment samples were also collected with a shovel, screened through a 6.3-mm sieve to remove larger material, and collected in 4-L scalable bags identified with a permanent marker. Each of the 4-L scalable bags was placed into a second 4-L 10 ------- sealable bag, and the sample label was placed onto the outside bag. The sediment samples were then placed into ' 10-L buckets, lidded, and identified with a sample label. 3.2 Y-12 National Security Complex 3.2.1 Site Description The Y-12 site is located at the DOE ORNL in Oak Ridge, TN. The Y-12 site is an active manufacturing and developmental engineering facility that occupies approximately 800 acres on the northeast corner of the DOE Oak Ridge Reservation (ORR) adjacent to the city of Oak Ridge, TN. Built in 1943 by the U.S. Army Corps of Engineers as part of the World War II Manhattan Project, the original mission of the installation was development of electromagnetic separation of uranium isotopes and weapon components manufacturing, as part of the national effort to produce the atomic bomb. Between 1950 and 1963, large quantities of elemental mercury were used at Y-12 during lithium isotope separation pilot studies and subsequent production processes in support of thermonuclear weapons programs. Soils at the Y-12 facility are contaminated with mercury in many areas. One of the areas of known high levels of mercury-contaminated soils is in the vicinity of a former mercury use facility (the "Old Mercury Recovery Building" - BuildJng 8110). At this location, mercury-contaminated material and soil were processed in a Nicols-Herschoff roasting furnace to recover mercury. Releases of mercury from this process, and from a building sump used to secure the mercury-contaminated .materials and the recovered mercury, have contaminated the surrounding soils (Rothchild, et al., 1984). Mercury contamination also occurred in the sediments of the East Fork of Poplar Creek (DOE, 1998). The Upper East Fork of Poplar Creek (UEFPC) drains the entire Y-12 complex. Releases of mercury via building drains connected to the storm sewer system, building basement dewatering sump discharges, and spills to soils, all contributed to contamination of UEFPC. Recent investigations showed that bank soils containing mercury along the UEFPC were eroding and contributing to mercury loading. Stabilization of the bank soils along this reach of the creek was recently completed. 3.2.2 Sample Collection Two matrices were sampled at Y-12 in Oak Ridge, TN, creek sediment and soil. A total of 10 sediment samples was collected; one sediment sample was collected from the Lower East Fork of Poplar Creek (LEFPC) and nine sediment samples were collected from the UEFPC. A total of six soil samples was collected from the Building 8110 area. The sampling procedures that were used are summarized below. Creek Sediments - Creek sediments were collected on September 24-25, 2002 from the East Fork of Poplar Creek. Sediment samples were collected from various locations in a downstream to upstream sequence (i.e., the downstream LEFPC sample was collected first and the most upstream point of the UEFPC was sampled last). The sediment samples from Poplar Creek were collected using a commercially available clam-shell sonar dredge attached to a rope. The dredge was slowly lowered to the creek bottom surface, where it was pushed by foot into the sediment. Several drops of the sampler (usually seven or more) were made to collectenough material for screening. On some occasions, a shovel was used to remove overlying "hardpan" gravel to expose finer sediments at depth. One creek sample consisted of creek bank sediments, which was collected using a stainless steel trowel. The collected sediment was then poured onto a 6.3-mm sieve to remove oversize sample material. Sieved samples were then placed in 12-L sealable plastic buckets. The sediment samples in these buckets were homogenized with a plastic ladle and subsamples were collected in 20- mi Mi liter (mL) vials for quick turnaround analyses. Soil - Soil samples were collected from pre-selected boring locations September 25, 2002. All samples were collected in the immediate vicinity of the Building 8110 foundation using a commercially available bucket auger. Oversize material was hand picked from the excavated soil because the soil was too wet to be passed through a sieve. The soil was transferred to an aluminum pan, homogenized by hand, and subsampled to a 20-mL vial. The remaining soil was transferred to 4-L plastic containers. 3.3 Confidential Manufacturing Site 3.3.1 Site Description A confidential manufacturing site, located in the eastern U.S., was selected for participation in this demonstration. The site contains elemental mercury, mercury amalgams, and mercury oxide in shallow sediments (less than 0.3 m deep) and deeper soils (3.65 to 9 m bgs). This site provided soil with concentrations from 5-1,000 mg/kg. The site is the location of three former processes that resulted in mercury contamination. The first process 11 ------- involved amalgamation of zinc with mercury. The second process involved the manufacturing of zinc oxide. The third process involved the reclamation of silver and gold from mercury-bearing materials in a retort furnace. Operations led to the dispersal of elemental mercury, mercury compounds such as chlorides and oxides, and zinc-mercury amalgams. Mercury values have been measured ranging from 0.05 to over 5,000 mg/kg, with average values of approximately 100 mg/kg. 3.3.2 Sample Collection Eleven subsurface soil samples were collected on September 24, 2002. All samples were collected with a Geoprobe® unit using plastic sleeves. All samples were collected at the location of a former facility plant. Drilling locations were determined based on historical data provided by the site operator. The intention was to gather soil samples across a range of concentrations. Because the surface soils were from relatively clean fill, the sampling device was pushed to a depth of 3.65 m using a blank rod. Samples were then collected at pre-selected depths ranging from 3.65 to 9 m bgs. Individual cores were 1-m long. The plastic sleeve for each 1-m core was marked with a permanent marker; the depth interval and the bottom of each core was marked. The filled plastic tubes were transferred to a staging table where appropriate depth intervals were selected formixing. Selected tubes were cut into 0.6-m intervals, which' were emptied into a plastic container for premixing soils. When feasible, soils were initially screened to remove materials larger than 6.3-mm in diameter. In many cases, soils were too wet and clayey to allow screening; in these cases, the soil was broken into pieces by hand and, by using a wooden spatula, oversize materials were manually removed. These soils (screened or hand sorted) were then mixed until the soil appeared visually uniform in color and texture. The mixed soil was then placed into a 4-L sample container for each chosen sample interval. A subsample of the mixed soil was transferred into a 20-mL vial, and it was sent for quick turnaround mercury analysis. This process was repeated for each subsequent sample interval. 3.4 Puget Sound 3.4.1 Site Description The Puget Sound site consists of contaminated offshore sediments. The particular area of the site used for collecting demonstration samples is identified as the Georgia Pacific, Inc. Log Pond. The Log Pond is located within the Whatcom Waterway in Bellingham Bay, WA, a well-established heavy industrial land use area with a maritime shoreline designation. Log Pond sediments measure approximately 1.5 to 1.8-m thick, and contain various contaminants including mercury, phenols, polyaromatic hydrocarbons, polychlorinated biphenyls.and wood debris. Mercury was used as a preservative in the logging industry. The area was capped in late 2000 and early 2001 with an average of 7 feet of clean capping material, as part of a Model Toxics Control Act.interim cleanup action. The total thickness ranges from approximate lyO.15 m along the site perimeterto 3 m within the interior of the project area. The restoration project produced 2.7 acres of shallow sub-tidal and 2.9 acres of low intertidal habitat, all of which had previously exceeded the Sediment Management Standards cleanup criteria (Anchor Environmental, 2001). Mercury concentrations have been measured ranging from 0.16 to 400 mg/kg (dry wt). The majority (98%) of the mercury detected in near-shore ground waters and sediments of the Log Pond is believed to be comprised of complexed divalent (Hg2*) forms such as mercuric sulfide (Bothner, et al., 1980 and Anchor Environmental, 2000). 3.4.2 Sample Collection Science Applications International Corporation (SAIC) is currently performing a SITE remedialtechnology evaluation in the Puget Sound (SAIC, 2002). As part of ongoing work at that site, SAIC collected additional sediment for use during this MMT project. Sediment samples collected on August 20-21, 2002 from the Log Pond in Puget Sound were obtained beneath approximately 3-6 m of water, using a vibra-coring system capable of capturing cores to 0.3 m below the proposed dredging prism. The vibra-corer consisted of a core barrel'attached to a power head. Aluminum core tubes, equipped with a stainless steel "eggshell" core catcher to retain material, were inserted into the core barrel. The vibra-core was lowered into position on the bottom and advanced to the appropriate sampling depth. Once sampling was completed, the vibra-core was retrieved and the core liner removed from the core barrel. The core sample was examined at each end to verify that sufficient sediment was retained for the particular sample. The condition and quantity of material within the core was then inspected to determine acceptability. The following criteria were used to verify whether an acceptable core sample was collected: Target penetration depth (i.e., into native material) was achieved. 12 ------- Sediment recovery of at least 65% of the penetration depth was achieved. Sample appeared undisturbed and intact without any evidence of obstruction/blocking within the core tube or catcher. The percent sediment recovery was determined by dividing the length of material recovered by the depth of core penetration below the mud line. If the sample was deemed acceptable, overlying water was siphoned from the top of the core tube and each end of the tube capped and sealed with duct tape. Following core collection, representative samples were collected from each core section representing a different vertical horizon. Sediment was collected from the center of the core that had not been smeared by, or in contact with, the core tube. The volumes removed were placed in a decontaminated stainless steel bowl or pan and mixed until homogenous in texture and color (approximately 2 minutes). After all sediment for a vertical horizon composite was collected and homogenized, representative aliquots were placed in the appropriate pre-cleaned sample containers. Samples of both the sediment and the underlying native material were collected in a similar manner. Distinct layers of sediment and native material were easily recognizable within each core. 3.5 Demonstration Site The demonstration was conducted in a natural environment, outdoors, in Oak Ridge, TN. The area was a grass covered hill with some parking areas, all of which were surrounded by trees. Building 5507, in the center of the demonstration area, provided facilities for lunch, break, and sample storage for the project and personnel. Most of the demonstration was performed during rainfall events ranging from steady to torrential. Severe puddling of rain occurred to the extent that boards needed to be placed under chairs to prevent them from sinking into the ground. Even when it was not raining, the relative humidity was high, ranging from 70.6 to 98.3 percent. Between two and four of the tent sides were used to keep rainfall from damaging the instruments. The temperature in the afternoons ranged from 65-70 degrees Fahrenheit, and the wind speed was less than 10 mph. The latitude is 36°N, the longitude 35°W, and the elevation 275 m. (Figure 3-1 is a photograph of the site during the demonstration and Figure 3-2 is a photograph of the location.) Figure 3-1. Tent and field conditions during the demonstration at Oak Ridge, TN. Figure 3-2. Demonstration site and Building 5507. 13 ------- 3.6 SAIC GeoMechanics Laboratory Sample homogenization was completed at the SAIC GeoMechanics Laboratory in Las Vegas, NV. This facility is an industrial-type building with separate facilities for personnel offices and material handling. The primary function of the laboratory is for rock mechanics studies. The laboratory has rock mechanics equipment, including sieves, rock crushers, and sample splitters. The personnel associated with this laboratory are experienced in the areas of sample preparation and sample homogenization. In addition to the sample homogenization equipment, the laboratory contains several benches, tables, and open space. Mercury air monitoring equipment was used during the sample preparation activities for personnel safety. 14 ------- Chapter 4 Demonstration Approach This chapter describes the demonstration approach that was used for evaluating the field mercury measurement technologies at ORNL in May 2003. It presents the objectives, design, sample preparation and management procedures, and the reference method confirmatory process used for the demonstration. 4.1 Demonstration Objectives The primary goal of the SITE MMT Program is to develop reliable performance and cost data on innovative, field-ready measurement technologies. A SITE demonstration must provide detailed and reliable performance and cost data in order that potential technology users have adequate information needed to make sound judgements regarding an innovative technology's applicability to a specific site, and to be able to compare the technology to conventional technologies. Table 4-1 summarizes the project objectives for this demonstration. In accordance with QAPP Requirements for Applied Research Projects (EPA,1998), the technical project objectives for the demonstration were categorized as primary and secondary. Table 4-1. Demonstration Objectives Objective Description Method of Evaluation Primary Objectives Primary Objective # 1 Primary Objective # 2. Primary Objective # 3 Primary Objective # 4 Primary Objective # 5 Determine sensitivity of each instrument with respect to vendor-generated MDL and PQL Determine potential analytical accuracy associated with vendor field measurements. Evaluate the precision of vendor field measurements. Measure time required to perform five functions related to mercury measurements: 1) mobilization and setup, 2) initial calibration, 3) daily calibration, 4) sample analysis, and 5) demobilization. Estimate costs associated with mercury measurements for the following four categories: 1) capital. 2) labor. 3) supplies, and 4) investigation-derived wastes. Independent laboratory confirmation of SRMs, field samples, and spiked field samples. Documentation during demonstration; vendor- provided information. Secondary Objectives Secondary Objective # 1 Secondary Objective # 2 Secondary Objective # 3 Secondary Objective # 4 Secondary Objective # 5 Document ease of use, skills, and training required to operate the device properly. Document potential H&S concerns associated with operating the device. Document portability of the device. Evaluate durability of device based on materials of construction and engineering design. Document the availability of the device and its spare parts. Documentation of observations during demonstration; vendor- provided information. Post-demonstration investigation. . 15 ------- Criticaldata support primary objectives and noncritical data support secondary objectives. With the exception of the cost information, primary objectives required the use of quantitative results to draw conclusions regarding technology performance. Secondary objectives pertained to information that was useful and did not necessarily require the use of quantitative results to draw conclusions regarding technology performance. 4.2 Demonstration Design 4.2.1 Approach for Objectives Addressing Primary The purpose of this demonstration was to evaluate the performance of the vendor's instrumentation against a standard laboratory procedure. In addition, an overall average relative .standard deviation (RSD) was calculated for all measurements made by the vendor and the referee laboratory. RSD comparisons used descriptive statistics, not inferential statistics, between the vendor and laboratory results. Other statistical comparisons (both inferential and • descriptive) for sensitivity, precision, and accuracy were used, depending upon actual demonstration results. The approach for addressing each of the primary objectives is discussed in the following subsections. A detailed explanation of the precise statistical determination .used for evaluating primary objectives No. 1 through No.. 3 is presented in Chapter 6. 4.2.1.1 Primary Objective #1: Sensitivity Sensitivity is the ability of a method or instrument to discriminate between small differences in analyte concentration (EPA, 2002b). It can be discussed in terms of an instrument detection limit (IDL), a method detection limit (MOL), and as a practical quantitation limit (PQL). MDL isnota measure of .sensitivity in thesame respect as an IDL or PQL. It is a measure of precision at a predetermined, usually low, concentration. The IDL pertains to the ability of the instrument to determine with confidence the difference between a sample that contains the analyte of interest at a low concentration and a sample that does not contain that analyte. The IDL is generally considered to be the minimum true concentration of an analyte producing a non-zero signal that can be distinguished from the signals generated when no concentration of the analyte is present and with an adequate degree of certainty. The IDL is not rigidly defined in terms of matrix, method, laboratory, or analyst variability, and it is not usually associated with a statistical level of confidence. IDLs are, thus, usually lower than MDLs and rarely serve a purpose in terms of project objectives (EPA, 2002b). The PQL defines a specific concentration with an associated level of accuracy. The MDL defines a lower limit at which a method measurement can be distinguished from background noise. The PQL is a more meaningful estimate of sensitivity. The MDL and PQL were chosen as the two distinct parameters for evaluating sensitivity. The approach for addressing each of these indicator parameters is discussed separately in the following paragraphs. MDL MDL is the estimated measure of sensitivity as defined in 40 Code of Federal Regulations (CFR) Part 136. The purpose of the MDL measurement is to estimate the concentration at which an individual field instrument is able to detect a minimum concentration that is statistically different from instrument background or noise. Guidance for the definition of the MDL is provided in EPA G-5i (EPA, 2002b). The determination of an MDL usually requires seven different measurements of a low concentration standard or sample. Following procedures established in 40 CFR Part 136 for water matrices, the demonstration MDL definition is as follows: where: t, (n-1.0.99) 99th percentile of the t-distribution with n-1 degrees of freedom num ber of m eas urem ents standard deviation of replicate measurements PQL The PQL is another important measure of sensitivity. The •PQL is defined in EPA G-5i as the lowest level an instrument is capable of producing a result that has significance in terms of precision and bias. (Bias is the difference between the measured value and the true value.) It is generally considered the lowest standard on the instrument calibration curve. It is often 5-10 times higher than the MDL, depending upon the analyte, the instrument being used, and the method for analysis; however, it should not be rigidly defined in this manner. 16 ------- During the demonstration, the PQL was to be defined by the vendor's reported calibration or based upon lower concentration samples or SRMs. The evaluation of vendor-reported results for the PQL included a determination of the percent difference (%D) between their calculated value and true value. The true value is considered the value reported by the referee laboratory for field samples or spiked field samples, or, in the case of SRMs, the certified value provided by the supplier. The equation used for the %D calculation is: C calculated where: 'calculated' "true true concentration as determined by the referee laboratory or SRM reference value calculated test sample concentration The PQL and %D were reported for the vendor. The %D for the referee laboratory, at the same concentration, was also reported for purposes of comparison. No statistical comparison was made between these two values; only a descriptive comparison was made for purposes of this evaluation. (The %D requirement forthe referee laboratory was defined as 10% or less. The reference method PQL was approximately 10 ug/kg.) 4.2.1.2 Primary Objective #2: Accuracy Accuracy was calculated by comparing the measured value to a known or true value. For purposes of this demonstration, three separate standards were used to evaluate accuracy. These included: 1) SRMs, 2) field samples collected from four separate mercury- contaminated sites, and 3)spiked field samples. Foursites were to be used for the evaluation; however, the manufacturing site samples proved to be too high in concentration forthe Milestone Field instrument (above 10 mg/kg) and therefore were not analyzed. Samples representing field samples and spiked field samples were prepared at the SAIC GeoMechanics Laboratory. In order to preventcross contamination, SRMs were prepared in a separate location. Each of these standards is discussed separately in the following paragraphs. SRMs The primary standards used to determine accuracy for this demonstration were SRMs. SRMs provided very tight statistical comparisons, although they did not provide all matrices of interest nor all ranges of concentrations. The SRMs were obtained from reputable suppliers, and had reported concentrations at associated 95% confidence intervals(Cls), and 95% prediction intervals. Prediction intervals were used for comparison because they represent a statistically infinite number of analyses, and therefore, would include all possible correct results 95% of the time. All SRMs were analyzed by the referee laboratory and selected SRMs were analyzed by the vendor, based upon instrument capabilities and concentrations of SRMs that could be obtained. Selected SRMs covered an appropriate range for each vendor. Replicate SRMs were also analyzed by the vendor and the laboratory. The purpose for SRM analysis by the referee laboratory was to provide a check on laboratory accuracy. During the pre-demonstration, the referee laboratory was chosen, in part, based upon the analysis of SRMs. This was done to ensure a competent laboratory would be used for the demonstration. Because of the need to provide confidence in laboratory analysis during the demonstration, the referee laboratory analyzed SRMs as an ongoing check for laboratory bias. Evaluation of vendor and laboratory analysis of SRMs was performed as follows. Accuracy was reported for individual sample concentrations of replicate measurements made at the same concentration. Two-tailed 95% CIs were computed according to the following equation: (rv-1.0.975)' 3/,/n where: !(„_,,0.975)= 97.5th percentile of the t-distribution with n-1 degrees of freedom number of measurements standard deviation of replicate measurements The number of vendor-reported SRM results and referee laboratory-reported SRM results that were within the associated 95% prediction interval were evaluated. Prediction intervals were computed in a similar fashion to the Cl, except that the Student's "t" value use "n" equal to infinity and, because they represented "n" approaching infinity, the square root of "n" is dropped from the equation. A final measure of accuracy determined from SRMs is a frequency distribution that shows the percentage of vendor- 17 ------- reported measurements that are within a specified window of the reference value. For example, a distribution within a 30% window of a reported concentration, within a 50% window, and outside a 50% window of a reported concentration. This distribution aspect could be reported as average concentrations of replicate results from the vendor for a particular concentration and matrix compared to the same sample from the laboratory. These are descriptive statistics and are used to better describe comparisons, but they are not intended as inferential tests. Field Samples The second accuracystandard used for this demonstration was actual field samples collected from four separate mercury-contaminated sites. (Only 3 of the 4 sites were used for the Milestone evaluation.) This accuracy determination consisted of a comparison of vendor- reported results for field samples to the referee laboratory results for the same field samples. The field samples were used to ensure that "real-world" samples were tested for each vendor. The field samples consisted of variable mercury concentrations within varying soil and sediment matrices. The referee laboratory results are considered the standard for comparison to each vendor. Vendor sample results for a given field sample were compared to replicates analyzed by the laboratory for the same field sample. (A hypothesis test was used with alpha = 0.01. The null hypothesis was that sample results were similar. Therefore, if the null hypothesis is rejected, then the sample sets are considered different.) Comparisons fora specific matrix or concentration were made in orderto provide additional information on that specific matrix or concentration. Comparison of the vendor values to laboratory values were similar to the comparisons noted previously for SRMs, except that a more definitive or inferential statistical evaluation was used. Alpha = 0.01 was used to help mitigate inter-laboratory variability. Additionally, an aggregate analysis was used to mitigate statistical anomalies (see Section 6.1.2). Spiked Field Samples The third accuracy standard for this demonstration was spiked field samples. These spiked field samples were analyzed by the vendors and by the referee laboratory in replicate in order to provide additional measurement comparisons to a known value. Spikes were prepared to cover additional concentrations not available from SRMs or the samples collected in the field. They were grouped with the field sample comparison noted above. 4.2.1.3 Primary Objective #3: Precision Precision can be defined as the degree of mutual agreement of independent measurements generated through repeated application of a process under specified conditions. Precision is usually thought of as repeatability of a specific measurement, and it is often reported as RSD. The RSD is computed from a specified number of replicates. The more replications of a measurement, the more confidence is associated with a reported RSD. Replication of a measurement may be as few as 3 separate measurements to 30 or more measurements of the same sample, dependent upon the degree of confidence desired in the specified result. The precision of an analytical instrument may vary depending upon the matrix being measured, the concentration of the analyte, and whether the measurement is made for an SRM or a field sample. The experimental design for this demonstration included a mechanism to evaluate the precision of the vendors' technologies. Field samples from the four mercury-contaminated field sites were evaluated by each vendor's analytical instrument. (See previous note concerning Milestone.) During the demonstration, concentrations were predetermined only as low, medium, or high. Ranges of test samples (field samples, SRMs, and spikes) were selected to cover the appropriate analytical ranges of the vendor's instrumentation. It was known prior to the demonstration that not all vendors were capable of measuring similar concentrations (i.e., some instruments were better at measuring low concentrations and others were geared toward higher concentration samples or had other attributes such as cost or ease of use that defined specific attributes of their technology). Because of this fact, not all vendors analyzed the same samples. During the demonstration, the vendor's instrumentation was tested with samples from the four different sites, having different matrices when possible (i.e., depending upon available concentrations) and having different concentrations (high, medium, and low) using a variety of • samples. Sample concentrations for an individual instrument were chosen based upon vendor attributes in terms of expected low, medium, and high concentrations that the particular instrument was capable of measuring. The referee laboratory measured replicates of all samples. The results were used for purposes of precision comparisons to the individual vendor. The RSD for the vendor and the laboratory were calculated individually, using the following equation: 18 ------- %RSD = -x100 X where: s = standard deviation of replicate results x = mean value of replicate results Using descriptive statistics, differences between vendor RSD and referee laboratory RSD were determined. This included RSD comparisons based upon concentration, SRMs, field samples, and different sites. In addition, an overall average RSD was calculated for all measurements made by the vendor and the laboratory. RSD comparisons were based upon descriptive statistical evaluations between the vendor and the laboratory, and results were compared accordingly. 4.2.1.4 Primary Objective #4: Time per Analysis The amount of time required for performing the analysis was measured and reported for five categories: • Mobilization and setup Initial calibration Daily calibration Sample analyses Demobilization Mobilization and setup included the time to unpack and prepare the instrument for operation. Initial calibration included the time to perform the vendor recommended on-site calibrations. Daily calibration included the time to perform the vendor-recommended calibrations on subsequent field days. (Note that this could have been the same as the initial calibration, a reduced calibration, or none.) Sample analyses included the time to prepare, measure, and calculate the results for the demonstration samples, and the necessary quality control (QC) samples performed by the vendor. The time per analysis was determined by dividing the total amount of time required to perform the analyses by the number of samples analyzed (173). In the numerator, sample analysis time included preparation, measurement, and calculation of results for demonstration samples and necessary QC samples performed by the vendor. In the denominator, the total number of analyses included only demonstration samples analyzed by the vendor, not QC analyses nor reanalyses of samples. Downtime that was required or that occurred between sample analyses as a part of operation and handling was considered a part of the sample analysis time. Downtime occurring due to instrument breakage or unexpected maintenance was not counted in the assessment, but it is noted in this final report as an additional time. Any downtime caused by instrument saturation or memory effect was addressed, based upon its frequency and impact on the analysis. Unique time measurements are also addressed in this report (e.g., if soil samples were analyzed directly, and sediment samples required additional time to dry before the analyses started, then a statement was made noting that soil samples were analyzed in X amount of hours, and that sediment samples required drying time before analysis). Recorded times were rounded to the nearest 15-minute interval. The number of vendor personnel used was noted and factored into the time calculations. No comparison on time per analysis is made between the vendor and the referee laboratory. 4.2.1.5 Primary Objective #5: Cost The following four cost categories were considered to estimate costs associated with mercury measurements: Capital costs Labor costs Supply costs Investigation derived waste (IDW) disposal costs Although both vendor and laboratory costs are presented, the calculated costs were not compared with the referee laboratory. A summary of how each cost category was estimated for the measurement device is provided below: The capital cost was estimated based on published price lists for purchasing, renting, or leasing each field measurement device. If the device was purchased, the capital cost estimate did not include salvage value for the device after work was completed. The labor cost was based on the number of people required to analyze samples during the demonstration. The labor rate was based on a standard hourly rate for a technician or other appropriate operator. During the demonstration, the skill level required was confirmed based on vendor input regarding the operation of the device to produce mercury concentration results and observations made in the field. The labor costs were based on: 1) the actual number of hours required to complete all analyses, quality assurance (QA), and reporting; and 2) the assumption that a technician who worked for a portion of a day was paid for an entire 8-hour day. 19 ------- The supply costs were based on any supplies required to analyze the field and SRM samples during the demonstration. Supplies consisted of items not included in the capital category, such as extraction solvent, glassware, pipettes, spatulas, agitators, and similar materials. The type and quantity of all supplies brought to the field and used during the demonstration were noted and documented. Any maintenance and repair costs during the demonstration were documented or provided by the vendor. Equipment.costs were estimated based on this information and standard cost analysis guidelines for the SITE Program. • The IDW disposal costs included decontamination fluids and equipment, mercury-contaminated soil and sediment samples, and used sample residues. Contaminated personal protective equipment (PPE) normally used in the laboratory was placed into a separate container. The disposal costs for the IDW were included in the overall analytical costs for each vendor. After all of the cost categories were estimated, the cost per analysis was calculated. This costvalue was based on the number of analyses performed. As the number of samples analyzed increased, the initial capital costs and certain other costs were distributed across a greater number of samples. Therefore, the per unit cost decreased. For this reason, two costs were reported: 1) the initial capital costs and 2) the operating costs per ana lysis. No comparison to the referee laboratory's method cost was made; however, a generic cost comparison is made. Additionally, when determining laboratory .costs, the associated cost for laboratory audits and data validation should be considered. 4.2.2 Approach for Addressing Secondary Objectives Secondary objectives were evaluated based on observations made during the demonstration. Because of the number of vendors involved, technology observers were required to make simultaneous observations of two vendors each during the demonstration. Four procedures were implemented to ensure that these subjective observations made by the observers were as consistent as possible. First, forms were developed for each of the five secondary objectives. These forms assisted in standardizing the observations. Second, the observers met each day before the evaluations began, at significant break periods, and after each day of work to discuss and compare observations regarding each device. Third, an additional observer was assigned to independently evaluate only the secondary objectives; in order to ensure that a consistent approach was applied in evaluating these objectives. Finally, the SAIC TOM circulated among the evaluation staff during the demonstration to ensure that a consistent approach was being followed by all personnel. Table 4-2 summarizes the aspects observed during the demonstration for each secondary objective. The individual approaches to each of these objectives are detailed further in the following subsections. Table 4-2. Summary of Secondary Objective Observations Recorded During the Demonstration SECONDARY OBJECTIVE General Information Secondary Objective # 1 Ease of Use - Vendor Name - No. of Operators - Observer Name - Operator Names/Titles - Instrument Type - Operator Training - Instrument Name - Training References - Model No. - Instrument Setup Time - Serial No. - Instrument Calibration Time - Sample Preparation Time - Sample Measurement Time H&S PPE Health and Safety Personal Protective Equipment Secondary Objective # 2 H&S Concerns - Instrument Certifications - Electrical Hazards - Chemicals Used - Radiological Sources - Hg Exposure Pathways - Hg Vapor Monitoring - PPE Requirements - Mechanical Hazard - Waste Handling Issues Secondary Objective # 3 Instrument Portability - Instrument Weight - Instrument Dimensions - Power Sources - Packaging - Shipping & Handling Secondary Objective # 4 Instrument DurabiOty - Materials of Construction - Quality of Construction - Max. Operating Temp. - Max Operating Humidity - Downtime - Maintenance Activities - Repairs Conducted 20 ------- 4.2.2.1 Secondary Objective #1: Ease of Use The skills and training required for proper device operation were noted; these included any degrees' or specialized training required by the operators. This information was gathered by interviews (i.e., questioning) of the operators. The number of operators required was also noted. This objective was also evaluated by subjective observations regarding the ease of equipment use and major peripherals required to measure mercury concentrations in soils and sediments. The operating procedure was evaluated to determine if the instrument is easily useable and understandable. 4.2.2.2 Secondary Objective #2: Health and Safety Concerns Health and safety (H&S) concerns associated with device operation were noted during the demonstration. Criteria included hazardous materials used, the frequency and likelihood of potential exposures, and any direct exposures observed .during the demonstration. In addition, any potential for exposure to mercury during sample digestion and analysis was evaluated based upon equipment design. Other H&S concerns, such as basic electrical and mechanical hazards, were also noted. Equipment certifications, such as Underwriters Laboratory (UL), were documented. 4.2.2.3 Secondary Objective #3: Portability of the Device The portability of the device was evaluated by observing transport, measuring setup and tear down time, determining the size and weight of the unit and peripherals, and assessing the ease with which the instrument was repackaged for movement to another location. The use of battery power or the need for an AC outlet was also noted. 4.2.2.4 Secondary Objective #4: Instrument Durability The durability of each device and major peripherals was assessed by noting the quality of materials and construction. All device failures, routine maintenance, repairs, and downtime were documented during the demonstration. No specific tests were performed to evaluate durability; rather, subjective observations were made using a field form as guidance. 4.2.2.5 Secondary Objective #5: Availability of Vendor Instruments and Supplies The availability of each device was evaluated by determining whether additional units and spare parts are readily available from the vendor or retail stores. The vendor's office (or a web page) and/or a retail store was contacted to identify and determine the availability of supplies of the tested measurement device and spare parts. This portion of the evaluation was performed after the field demonstration, in conjunction with the cost estimate. 4.3 Sample Preparation and Management 4.3.1 Sample Preparation 4.3.1.1 Field Samples Field samples were collected during the pre-demonstration portion of the project, with the ultimate goal of producing a set of consistent test soils and sediments to be distributed among all participating vendors and the referee laboratory for analysis.during the demonstration. Samples were collected from the following four sites: Carson River Mercury site (near Dayton, NV) Y-12 National Security Complex (Oak Ridge, TN) Manufacturing facility (eastern U.S.) Puget Sound (Bellingham, WA) The field samples collected during the pre-demonstration sampling events comprised a variety of matrices, ranging from material having a high clay content to material composed mostly of gravelly, coarse sand. The field samples also differed with respect to moisture content; several were collected as wet sediments. Table 4-3 shows the number of distinct field samples that were collected from each of the four, field sites. Prior to the start of the demonstration, the field samples selected for analysis during the demonstration were prepared at the SAIC GeoMechanics Laboratory in Las Vegas, NV. The specific sample homogenization procedure used by this laboratory largely depended on the moisture content and physical consistency of the sample. Two specific sample homogenization procedures were developed and tested by SAIC at the GeoMechanics Laboratory during the pre-demonstration portion of the project. The methods included a non-slurry sample procedure and a slurry sample procedure. A standard operating procedure (SOP) was developed detailing both methods. The procedure was found to be satisfactory based upon the results of replicate samples during the pre-demonstration. This SOP is included as Appendix A of the Field Demonstration Quality Assurance Project Plan (SAIC, August 2003 EPA/600/R-03/053). Figure 4-1 summarizes the homogenization steps of the 21 ------- SOP, beginning with sample mixing. This procedure was used for preparing both pre-demonstration and demonstration samples. Prior to the mixing process (i.e., Step 1 in Figure 4-1), all field samples being processed were visually inspected to ensure that oversized materials were removed and that there were no clumps that would hinderhomogenization. Non-slurry samples were air-dried in accordance with the SOP so that they could be passed Table 4-3. Field Samples Collected from the Four Sites No. of Samples / Matrices multiple times through a riffle splitter. Due to the high moisture content of many of the samples, they were not easily air-dried and could not be passed through a riffle splitter while wet. Samples with very high moisture contents, termed "slurries," were not air-dried, and bypassed the riffle splitting step. The homogenization steps for each type of matrix are briefly summarized as follows. Field Site Carson River Y-12 Manufacturing Site Puget Sound Collected 12 Soil 6 Sediment 10 Sediment 6 Soil 12 Soil 4 Sediment Areas For Collecting Sample Material Tailings Piles (Six Mile Canyon) River Bank Sediments Poplar Creek Sediments Old Mercury Recovery Bldg. Soils Subsurface Soils High-level Mercury (below cap) Low-Level Mercury (native material) Volume Required 4 L each for soil 12 L each for sediment 12 L each for sediment 4 L each for soil 4 L each 12 Leach Preparing Slurry Matrices For slurries (i.e., wet sediments), the mixing steps were sufficiently thorough that the sample containers could be filled directly from the mixing vessel. There were two separate mixing steps of the slurry-type samples. Each slurry was initially mixed mechanically within the sample container (i.e., bucket) in which the sample was shipped to the SAIC GeoMechanics Laboratory. A subsample of this premixed sample was transferred to a second mixing vessel. A mechanical drill equipped with a paint mixing attachment was used to mix the subsample. As shown in Figure 4-1, slurry samples bypassed the sample riffle splitting step. To ensure all sample bottles contained the same material, the entire set of containers to be filled was submerged into the slurry as a group. The filled vials were allowed to settle for a minimum of two days, and the standing water was removed using a Pasteur pipette. The removal of the standing water from the slurry samples was the only change to the homogenization procedure between the pre-demonstration and the demonstration. Preparing "Non-Slurry" Matrices Soils and sediments having no excess moisture were initially mixed (Step 1) and then homogenized in the sample riffle splitter (Step 2). Prior to these steps, the material was air-dried and subsampled to reduce the volume of material to a size that was easier to handle. As shown in Figure 4-1 (Step 1), the non-slurry subsample was manually stirred with a spoon or similar equipment until the material was visually uniform. Immediately following manual mixing, the subsample was mixed and split six times for more complete homogenization (Step 2). After the 6th and final split, the sample material was leveled to form a flattened, elongated rectangle and cut into transverse sections to fill the containers (Steps 3 and 4). After homogenization, 20-ml sample vials were filled and prepared for shipment (Step 5). For the demonstration, the vendor analyzed 173 samples, which included replicates of up to 7 samples per sample lot. The majority of the samples distributed had concentrations within the range ofthe vendor's technology. Some samples had expected concentrations at or below the estimated level of detection for each of the vendor instruments. These samples were designed to evaluate the reported MDL and PQL and also to assess the prevalence of false positives. Field samples distributed to the vendor included sediments and soils collected from all four sites and prepared by both the slurry and dry homogenization procedures. The field samples were segregated into broad sample sets: low, medium, and high mercury concentrations. This gave the vendor the same general understanding of the sample to be analyzed as they would typically have for field application of their instrument. 22 ------- A Test material mixed until visually uniform For non-slurries Mix manually For slurries a) Mix mechanically the entire sample volume b) Subsample slurry, transfer to mixing vessel, and mix mechanically Slurries transferred directly to 20 ml vials (vials submerged into slurry) Npn-slum'es to riffle splitter Combined splits are reintroduced into splitter (6 X) (T . // RIFFLE \V fy SPUTTER V3 . _^,^ ^_-<— ^. •> _£•• ^Ss J^VVX- \l Elongated J rectangular pile ^ / (from 6* solit) 4-r.,_..^ . fj, , . . l4 /! 1 1 1 1 1 II 1 1 1 1 1 II ll\ ,^1 1 1 1 M 1 M 1 1 1 1 1 1 1 N 1 tariSlei CUt 1 TEFLON SURFACE. 1 2S0 mL^als 3 SahmP|e a"q"ots made by transverse cuts J H r~] c Samples shipped @ 4 °C to ** referee lab and Oak Ridge ^ ) (Container numbers will vary) Figure 4-1. Test sample preparation at the SAIC GeoMechanics Laboratory. 23 ------- In addition, selected field samples were spiked with mercury (II) chloride to generate samples with additional concentrations and test the ability of the vendor's instrumentation to measure the additional species of mercury. Specific information regarding the vendor's sample distribution is included in Chapter 6. 4.3.1.2 Standard Reference Materials Certified SRMs were analyzed by both the vendors and the referee laboratory. These samples were homogenized matrices which had a known concentration of mercury. Concentrations were certified values, as provided by the supplier, based on independent confirmation via multiple analyses of multiple lots and/or multiple analyses by different laboratories (i.e., round robin testing). These analytical results were then used to determine "true" values, as well as a statistically derived intervals (a 95% prediction interval) that provided a range within which the true value were expected to fall. The SRMs selected were designed to encompass the same contaminant ranges indicated previously: low-, medium-, and high-level mercury concentrations. In addition, SRMs of varying matrices were included in the demonstration to challenge the vendor technology as well as the referee laboratory. The referee laboratory analyzed all SRMs. SRM samples were intermingled with site field samples and labeled in the same manneras field samples. 4.3.1.3 Spiked Field Samples Spiked field samples were prepared by the SAIC GeoMechanics Laboratory using mercury (II) chloride. Spikes were prepared using field samples from the selected sites. Additional information was gained by preparing spikes at concentrations not previously obtainable. The SAIC GeoMechanics Laboratory's ability to prepare spikes was tested prior to the demonstration and evaluated in order to determine expected variability and accuracy of the spiked sample. Thespiking procedure was evaluated by preparing several different spikes using two different spiking procedures (dry and wet). Based upon replicate analyses results, it was determined thatthe wet, or slurry, procedure was the only effective method of obtaining a homogeneous spiked sample. 4.3.2 Sample Management 4.3.2.1 Sample Volumes, Containers,and Preservation A subset from the pre-demonstration field samples was selected for use in the demonstration based on the sample's mercury concentration range and sample type (i.e., sediment versus soil). The SAIC GeoMechanics Laboratory prepared individual batches of field sample material to fill sample containers for each vendor. Once all containers from a field sample were filled, each container was labeled and cooled to 4 °C. Because mercury analyses were to be performed both by the vendors in the field and by the referee laboratory, adequate sample size was taken into account. Minimum sample size requirements for the vendors varied from 0.1 g or less to 8-10 g. Only the referee laboratory analyzed separate sample aliquots for parameters otherthan mercury. These additional parameters included arsenic, barium, cadmium, chromium, lead, selenium, silver, copper, zinc, oil and grease, and total organic carbon (TOC). Since the mercury method (SW-846 7471B) being used by the referee laboratory requires 1 g for analysis, the sample size sent to all participants was a 20 mL vial (approximately 10 g), which ensured a sufficient volume and mass for analysis by all vendors. 4.3.2.2 Sample Labeling The sample labeling used for the 20 mL vials consisted of an internal code developed by SAIC. This "blind" code was used throughout the entire demonstration. The only individuals who knew the key coding of the homogenized samples to the specific field samples were the SAIC TOM, the SAIC GeoMechanics Laboratory Manager, and the SAIC QA Manager. 4.3.2.3 Sample Record Keeping, Archiving, and Custody Samples were shipped to the laboratory and the demonstration site the week prior to the demonstration. A third set of vials was archived at the SAIC GeoMechanics Laboratory as reserve samples. The sample shipment to Oak Ridge was retained at all times in the custody of SAIC at their Oak Ridge office and until arrival of the demonstration field crew. Sam pies were shipped under chain-of-custody (CoC)and with custody seals on both the coolers and the inner plastic bags. Once the demonstration crew arrived, the coolers were retrieved from the SAIC office. The custody seals on the plastic bags inside the cooler were broken by the vendor upon transfer. Upon arrival at the ORNL site, the vendor set up the instrumentation at the direction and oversight of SAIC. At the start of sample testing, the vendor was provided with a sample set representing field samples collected from a particular field site intermingled with SRM and spiked samples. Due. to variability of vendor instrument 24 ------- measurem ent ranges for m ercury detection, not all vendors received samples from the same field material. All samples were stored in an ice coolerpriorto demonstration startup and were stored in an on-site sample refrigerator during the demonstration. Each sample set was identified and distributed as a set with respect to the site from which it was collected. This was done because, in any field application, the location and general type of the samples would be known. The vendor was responsible for analyzing all samples provided, performing any dilutions or reanalyses as needed, calibrating the instrument if applicable, performing any necessary maintenance, and reporting all results. Any samples that were not analyzed during the day were returned to the vendor for analysis at the beginning of the next day. Once analysis of the samples from the first location were completed by the vendor, SAIC provided a set of samples from the second location. Samples were provided at the time that they were requested by the vendor. Once again, the transfer of samples was documented using a COC form. This process was repeated for samples from each location. SAIC maintained custodyof all remaining sample sets until they were transferred to the vendor. SAIC maintained custody of samples that already had been analyzed and followed the waste handling procedures in Section 4.2.2 of the Field Demonstration QAPPto dispose of these wastes. 4.4 Reference Process Method Confirmatory The referee laboratory analyzed all samples that were analyzed by the vendor technologies in the field. The following subsections provide information on the selection of the reference method, selection of the referee laboratory, and details regarding the performance of the reference method in accordance with EPA protocols. Other parameters that were analyzed by the referee laboratory are also discussed briefly. 4.4.1 Reference Method Selection The selection of SW-846 Method 7471B as the reference method was based on several factors, predicated on information obtained from the technology vendors, as well as the expected contaminant types and soil/sediment mercury concentrations expected in the test matrices. The re are several laboratory-based, promulgated methods for the analysis of total mercury. In addition, there are several performance-based methods for the determination of various mercury species. Based on the vendor technologies, it was determined that a reference method for total mercury would be needed (Table 1-2 summarizes the methods evaluated, as identified through a review of the EPA Test Method Index and SW-846). In selecting which of the potential methods would be suitable as a reference method, consideration was given to the following questions: Was the method widely used and accepted? Was the method an EPA-recommended, or similar regulatory method? The selected reference method should be sufficiently used so that it could be cited as an acceptable method for monitoring and/or permit compliance among regulatory authorities. Did the selected reference method, provide QA/QC criteria that demonstrate acceptable performance characteristics over time? Was the method suitable for the species of mercury that were expected to be encountered? The reference method must be capable of determining, as total mercury, all forms of the chemical contaminant known or likely to be present in the matrices. Would the method achieve the necessary detection limits to evaluate the sensitivity of each vendor technology adequately? • Was the method suitable for the concentration range that was expected in the test matrices? Based on the above considerations, itwas determined that SW-846 Method 7471B (analysis of mercury in solid samples by cold-vapor AAS) would be the best reference method. SW-846 Method 7474, (an atomic fluorescence spectrometry method using method 3052 for microwave digestion of the solid) had also been considered a likely technical candidate; however, because this method was not as widely used or referenced, Method 7471B was considered the better choice. 4.4.2 Referee Laboratory Selection During the planning of the pre-demonstration phase of this project, nine laboratories were sent a statement of work (SOW) for the analysis of mercury to be performed as part of the pre-demonstration. Seven of the nine laboratories responded to the SOW with appropriate bids. Three of the seven laboratories were selected as candidate laboratories based upon technical merit, experience, and pricing. These laboratories received and analyzed blind samples and SRMs during pre-demonstration activities. The referee 25 ------- laboratory to be used for the demonstration was selected from these three candidate laboratories. Final selection of the referee laboratory was based upon: 1) the laboratory's interest in continuing in the demonstration, 2) the laboratory-reported SRM results, 3) the laboratory MDL for the reference method selected, 4) the precision of the laboratory calibration curve, 5) the laboratory's ability to support the demonstration (scheduling conflicts, backup instrumentation, etc.), and 8) cost. One of the three candidate laboratories was eliminated from selection based on a technical consideration. It was determined that one of the laboratories would not be able to meet demonstration quantitation limit requirements. (Its lower calibration standard was approximately 50 ug/kg and the vendor comparison requirements were well below this value.) Two candidates thus remained, including the eventual demonstration laboratory, Analytical Laboratory Services, Inc. (ALSI): Analytical Laboratory Services, Inc. Ray Martrano, Laboratory Manager 34 Dogwood Lane . Middletown, PA 17057 (717)944-5541 In order to make a final decision on selecting a referee laboratory, a preliminary audit was performed by theSAIC QA Manager at the remaining two candidate laboratories. Results of the SRM samples were compared for the two laboratories. Each laboratory analyzed each sam pie (there were two SRMs) in triplicate. Both laboratories were within the 95% prediction interval for each SRM. In addition, the average result from the two SRMs was compared to the 95% CI for the SRM. Calibration curves from each laboratory were reviewed carefully. This included calibration curves generated from previously performed analyses and those generated for other laboratory clients. There were two QC requirements regarding calibration curves; the correlation coefficient had to be 0.995 or greater and the lowest point on the calibration curve had to be within 10% of the predicted value. Both laboratories were able to achieve these two requirements for all curves reviewed and for a lower standard of 10 ug/kg, which was the lower standard required for the demonstration, based upon information received from each of the vendors. In addition, an analysis of seven standards was reviewed for MDLs. Both laboratories were able to achieve an MDL that was below 1 ug/kg. It should be noted that vendor sensitivity claims impacted how low this lower quantitation standard should be. These claims were somewhat vague, and the actual quantitation limit each vendor could achieve was uncertain prior to the demonstration (i.e., some vendors claimed a sensitivity as low as 1 MQ/kg. but it was uncertain at the time if this limit was actually a PQL or a detection limit). Therefore, it was determined that, if necessary, the laboratory actually should be able to achieve even a lower PQ'L than 10 For both laboratories, SOPs based upon SW-846 Method 7471 B, were reviewed. Each SOP followed this reference method. In addition, interferences were discussed because there was some concern that organic interferences may have been present in the samples previously analyzed by the laboratories. Because these same matrices were expected to be part of the demonstration, there was some concern associated with how these interferences would be eliminated. This is discussed at the end of this subsection. Sample throughput was somewhat important because the selected laboratory was to receive all demonstration samples at the same time (i.e., the samples were to be analyzed at the same time in order to eliminate any question of variability associated with loss of contaminant due to holding time). This meant that the laboratory would receive approximately 400 samples for analysis over the period of a few days. It was also desirable for the laboratory to produce a data report within a 21-day turnaround time for purposes of the demonstration. Both laboratories indicated that this was achievable. Instrumentation was reviewed and examined at both laboratories. Each laboratory used a Leeman mercury analyzer for analysis. One of the two laboratories had backup instrumentation in case of problems. Each laboratory indicated that its Leeman mercury analyzer was relatively new and had not been a problem in the past. Previous SITE program experience was another factor considered as partof these pre-audits. This is because the SITE program generally requires a very high level of QC, such that most laboratories are not familiar with the QC required unless they have previously participated in the program. A second aspect of the SITE program is that it generally requires analysis of relatively "dirty" samples and many laboratories are not use to analyzing such "dirty" samples. Both laboratories have been longtime participants in this program. Other QC-related issues examined during the audits included: 1) analyses on other SRM samples not previously examined, 2) laboratory control charts, and 3) precision 26 ------- and accuracy results. Each of these issues was closely examined. Also, because of the desire to increase the representativeness of the samples for the demonstration, each laboratory was asked if sample aliquot sizes could be increased to 1 g (the method requirement noted 0.2 g). Based upon previous results, both laboratories routinely increased sample size to 0.5 g, and each laboratory indicated that increasing the sample size would not be a problem. Besides these QC issues, other less tangible QA elements were examined. This included analyst experience, management "involvement in the demonstration, and internal laboratory QA management. These elements were also factored into the final decision. Selection Summary There were very few factors that separated the quality of these two laboratories. Both were exemplaryin performing mercury analyses. There were, however, some minor differences based upon this evaluation that were noted by the auditor. These were as follows: ALSI had backup instrumentation available. Even though neither laboratory reported any problems with its primary instrument (the Leeman mercury analyzer), ALSI did have a backup instrument in case there were problems with the primary instrument, or in the event that the laboratory needed to perform other mercury analyses during the demonstration time. • As noted, the low standard requirement for the calibration curve was one of the QC requirements specified for this demonstration in order to ensure that a lower quantitation could be achieved. This low standard was 10 ug/kg for both laboratories. ALSI, however, was able to show experience in being able to calibrate much lower than this, using a second calibration curve. In the event that the vendor was able to analyze at concentrations as low as 1 ug/kg with precise and accurate determinations, ALSI was able to perform analyses at lower concentrations as part of the demonstration. ALSI used a second, lower calibration curve for any analyses required below 0.05 mg/kg. Very few vendors were able to analyze samples at concentrations at this low a level. • Management practices and analyst experience were . similar at both laboratories. ALSI had participated in a few more SITE demonstrations than the other laboratory, but this difference was not significant because both laboratories had proven themselves capable of handling the additional QC requirements for the SITE program. In addition, both laboratories had internal QA management procedures to provide the confidence needed to achieve SITE requirements. Interferences for the samples previously analyzed were discussed and data were reviewed. ALSI analyzed two separate runs for each sample. This included analyses with and without stannous chloride. (Stannous chloride is the reagent used to release mercury into the vapor phase for analysis. Sometimes organics can cause interferences in the vapor phase. Therefore, an analysis with no stannous chloride would provide information on organic interferences.) The other laboratory did not routinely perform this analysis. Some samples were thought to contain organic interferences, based on previous sample results. The pre-demonstration results reviewed indicated that no organic interferences were present. Therefore, while this was thought to be a possible discriminator between the two laboratories in terms of analytical method performance, it became mootforthe samples included in this demonstration. The factors above were considered in the final evaluation. Because there were only minor differences in the technical factors, cost of analysis was used as the discriminating factor. (If there had been significant differences in laboratory quality, cost would not have been a factor.) ALSI was significantly lower in cost than the other laboratory. Therefore, ALSI was chosen as the referee laboratory for the demonstration. 4.4.3 Summary of Analytical Methods 4.4.3.1 Summary of Reference Method The critical measurement for this study was the analysis of mercury in soil and sediment samples. Samples analyzed by the laboratory included field samples, spiked field samples, and SRM samples. Detailed laboratory procedures for subsampling, extraction, and analysis were provided in the SOPs included as Appendix B of the Field Demonstration QAPP. These are briefly summarized below. Samples were analyzed for mercury using Method 7471B, a cold-vapor atomic absorption method, based on the absorption of radiation at the 253.7-nm wavelength by mercury vapor. The mercury is reduced to the elemental state and stripped/volatilized from solution in a closed system. The mercury vapor passes through a cell positioned in the light path of an AA spectrophotometer. Absorbance (peak height) is measured as a function of mercury concentration. Potassium permanganate is added to eliminate possible interference from sulfide. As per the 27 ------- method, concentrations as high as 20 mg/kg of sulfide, as sodium sulfide, do not interfere with the recovery of added inorganic mercury in reagent water. Copper has also been reported to interfere; however, the method states that copper concentrations as high as 10 mg/kg had no effect on recovery of mercury from spiked samples. Samples high in chlorides require additional permanganate (as much as 25 ml) because, during the oxidation step, chlorides are converted to free chlorine, which also absorbs radiation at 254 nm. Free chlorine is removed by using an excess (25 mL) of hydroxylamine sulfate reagent. Certain volatile organic materials that absorb at this wavelength may also cause interference. A preliminary analysis without reagents can determine if this type of interference is present. Prior to analysis, the contents of the sample container are stirred, and the sample mixed prior to removing an aliquot for the mercury analysis. An aliquot of soil/sediment (1 g) is placed in the bottom of a biological oxygen demand bottle, with reagent water and aqua regia added. The mixture is heated in a water bath at 95 °C for 2 minutes. The solution is cooled and reagent water and potassium permanganate solution are added to the sample bottle. The bottle contents are thoroughly mixed, and the bottle is placed in the water bath for 30 minutes at 95 °C. After cooling, sodium chloride-hydroxylamine sulfate is added to reduce the excess permanganate. Stannous chloride is then added and the bottle attached to the analyzer; the sample is aerated and the absorbance recorded. An analysis without non-stannous chloride is also included as an interference check when organic contamination is suspected. In the event of positive results of the non- stannous chloride analysis, the laboratory was to report those results to SAIC so that a determination of organic interferences could be made. 4.4.3.2 Summary of Methods Measurements. for Non-Critical A selected set of non-critical parameters was also measured during the demonstration. These parameters were measured to provide a better insightinto the chemical constituency of the field samples, including the presence of potential interferents. The results of the tests for potential interferents were reviewed to determine if a trend was apparent in the event that inaccuracy or low precision was observed. Table 4-4 presents the analytical method reference and method type for these non-critical parameters. Table 4-4. Analytical Methods for Non-Critical Parameters Parameter Arsenic, barium, cadmium, chromium, lead, selenium, silver, copper, zinc Oil and Grease TOO Total Solids Method Reference SW-846 3050/6010 EPA 1664 SW-846 9060 EPA 2540G Method Tvoe Acid digestion, ICP n-Hexane extraction, Gravimetric analysis Carbonaceous analyzer Gravimetric 4.5 Deviations Plan from the Demonstration During the demonstration Milestone determined that their field instrument was not appropriate for analysis of samples with greater than 5 mg/kg mercury. Previously, during pre- demonstration trials Milestone was able to analyze higher concentration sam pies, therefore, samples supplied during the demonstration had concentrations greater than 5 mg/kg. SAIC prepared several sam pies with these higher concentrations, including all samples collected from the Manufacturing Site. During the final day of the demonstration, Milestone was slightly behind in performing sample analysis and therefore a decisbn was made that • because the only samples left to analyze were Manufacturing Site samples and because these samples were all known to be above 5 mg/kg, it was decided by the EPA TOM and Milestone personnel that these samples need not be analyzed. This resulted in the analysis of fewer samples (173), than all other vendors. 28 ------- Chapters Assessment of Laboratory Quality Control Measurements 5.1 Laboratory QA Summary QA may be defined as a system of activities, the purpose of which is to provide assurance that defined standards of quality are met with a stated level of confidence. A QA program is a means of integrating the quality planning, quality assessment, QC, and quality improvement efforts to meet user requirements. The objective of the QA program is to reduce measurement errors to agreed-upon limits, and to produce results of acceptable and known quality. The QAPP specified the necessary guidelines to ensure that the measurement system for laboratory analysis was in control, and provided detailed information on the analytical approach to ensure that data of high quality could be obtained to achieve project objectives. The laboratory analyses were critical to project success, as the laboratory results were used as a standard for comparison to the field method results. The field methods are of unknown quality, and therefore, for comparison purposes the laboratory analysis needed to be a known quantity. The following sections provide information on the use of data quality indicators, and a detailed summary of the QC analyses associated with project objectives. 5.2 Data Quality Indicators for Mercury Analysis To assess the quality of the data generated by the referee laboratory, two important data quality indicators of primary concern are precision and accuracy. Precision can be defined as the degree of mutual agreement of independent measurements generated through repeated application of the process under specified conditions. Accuracy is the degree of agreement of a measured value with the true or expected value. Both accuracy and precision were measured by the analysis of matrix spike/matrix spike duplicates (MS/MSDs). The precision of the spiked duplicates is evaluated by expressing, as a percentage, the difference between results of the sample and sample duplicate results. The relative percent difference (RPD) is calculated as: RPD (Maximum Value - Minimum Value) (Maximum Value + Minimum Value)/2 x100 To determine and evaluate accuracy, known quantities of the target analytes were spiked into selected field samples. All spikes were post-digestion spikes because of the high sample concentrations encountered during the demonstration. Pre-digestion spikes, on high- concentration samples would either have been diluted or would have required additional studies to determine the effect of spiking more analyte and subsequent recovery values. To determine matrix spike recovery, and hence measure accuracy, the following equation was applied: %R = C«-C, -x100 where, Cjj = Analyte concentration in spiked sample Cus = Analyte concentration in unspiked sample C^ = Analyte concentration added to sample Laboratory control samples (LCSs) were used as an additional measure of accuracy in the event of significant 29 ------- matrix interference. To determine the percent recovery of LCS analyses, the equation below was used: „,.-. Measured Concentration _„„ %R = — :—— :—x10Q Theoretical Concentration While several precautions were taken to generate data of known quality through control of the measurement system, the data must also be representative of true conditions and comparable to separate sample aliquots. Representativeness refers to the degree with which analytical results accurately and precisely reflect actual conditions present at the locations chosen for sample collection. Representativeness was evaluated as part of the pre-demonstration and combined with the precision measurement in relation to sample aliquots. Sample aliquoting by the SAIC GeoMechanics Laboratory tested the ability of the procedure' to produce homogeneous, representative,-and comparable samples. All samples were carefully homogenized in order to ensure comparability between the laboratory and the vendor. Therefore, the RSD measurement objective of 25% or less forreplicate sample lot analysis was intended to assess not only precision but representativeness and comparability. Sensitivity was another critical factor assessed for the laboratory method of analysis. This was measured as a practical quantitation limit and was determined by the low standard on the calibration curve. Two separate calibration curves were run by the laboratory when necessary. The higher calibration curve was used for the majority of the samples and had a lowercalibration limit of 25 ug/kg. The lower calibration curve was used when samples were below this lowercalibration standard. The lowercalibration curve had a lower limit standard of 5 ug/kg. The lower limit standard of the calibration curve was run with each sample batch as a check standard and was required to be within 10% of the true value (QAPP QC requirement). This additional check on analytical sensitivity was performed to ensure that this lower limit standard was truly representative of the instrument and method practical quantitation limit. 5.3 Conclusions Limitations and Data Quality Critical sample data and associated QC analyses were reviewed to determine whether the data collected were of adequate quality to provide proper evaluation of the project's technical objectives. The results of this review are summarized below. Accuracy objectives for mercury analysis by Method 7471B were assessed by the evaluation of 23 spiked duplicate pairs, analyzed in accordance with standard procedures in the same manner as the samples. Recovery values for the critical compounds were well within objectives specified in the QAPP, except for two spiked samples summarized in Table 5-1. The results of these samples, however, were only slightly outside specified limits, and given the number of total samples (46 or 23 pairs), this is an insignificant number of results that did notfall within specifications. The MS/MSD results therefore, are supportive of the overall accuracy objectives. Table 5-1. MS/MSD Summary Parameter Value QC Limits Recovery Range Number of Duplicate Pairs Average Percent Recovery No. of Spikes Outside QC Specifications 80%-120% 85.2%-126% 23 108% An additional measure of accuracy was LCSs. These were analyzed with every sample batch (1 in 20 samples) and results are presented in Table 5-2. All results were within specifications, thereby supporting the conclusion that QC assessment met project accuracy objectives. Table 5-2. LCS Summary Parameter QC Limits Recovery Range Number of LCSs Average Percent Recovery No. of LCSs Outside QC Specifications Value 90%- 110% 90% - 100% 24 95.5% 0 Precision was assessed through the analysis of 23 duplicate spike pairs for mercury. Precision specifications were established prior to the demonstration as a RPD less 30 ------- than 20%. All but two sample pairs were within specifications, as noted in Table 5-3. The results of these samples, however, were only slightly outside specified limits, and given the number of total samples (23 pairs), this is an insignificant number of results that did not fall within specifications. Therefore, laboratory analyses met precision specifications. Table 5-3. Precision Summary Parameter Value QC Limits MS/MSD RPD Range Number of Duplicate Pairs Average MS/MSD RPD No. of Pairs Outside QC Specifications RPD<20% 0.0% to 25% 23 5.7% 2 Sensitivity results were within specified project objectives. The sensitivity objective-was evaluated as the PQL, as assessed by the low standard on the calibration curve. For the majority of samples, a calibration curve of 25-500 ug/kg was used. This is because the majority of samples fell within this calibration range (samples often required dilution). There were, however, some samples below this range and a second curve was used. The calibration range for this lower curve was 5-50 ug/kg. In order to ensure that the lower concentration on the calibration curve was a true PQL, the laboratory ran a low check standard (lowest concentration on the calibration curve) with every batch of samples. This standard was required to be within 10%. of the specified value. The results of this low check standard are summarized in Table 5-4. Table 5-4. Low Check Standards Parameter Value QC Limits Recovery Range Number of Check Standards Analyzed Average Recovery Recovery 90% -110% 88.6%-111% 23 96% There were a few occasions where this standard did not meet specifications. The results of these samples, however, were only slightly outside specified limits, and given the number of total samples (23), this is an insignificant number of results that did not fall within specifications. In addition, the laboratory reanalyzed the standard when specifications were not achieved, and the second determination always fell within the required limits. Therefore laboratory objectives for sensitivity were achieved according to QAPP specifications. , As noted previously, comparability and representativeness were assessed through the analysis of replicate samples. Results of these replicates are presented in the discussion on primary project objectives for precision. These results show that data were within project and QA objectives. Completeness objectives were achieved for the project. All samples were analyzed and data were provided for 100% of the samples received by the laboratory. No sample bottles were lost or broken. Other measures of data quality included method blanks, calibration checks, evaluation of linearity of the calibration curve, holding time specifications, and an independent standard verification included with each sample batch. These results were reviewed for every sample batch run by ALSI, and were within specifications. In addition, 10% of the reported results were checked against the raw data. Raw data were reviewed to ensure that sample results were within the calibration range of the instrument, as defined by the calibration curve. A 6-point calibration curve was generated at the start of each sample batch of 20. A few data points were found to be incorrectly reported. Recalculations were performed for these data, and any additional data points that were suspected outliers were checked to ensure correct results were reported. Veryfew calculation or dilution errors were found. All errors were corrected so that the appropriate data were reported. Another, measure of compliance were the non-stannous chloride runs performed by the laboratory for every sample analyzed. This was done to check for organic interference. There were no samples that were found to have any organic interference by this method. Therefore, these results met expected QC specifications and data were not qualified in any fashion. Total solids data were also reviewed to ensure that calculations were performed appropriately and dry weights reported when required. All of these QC checks met 31 ------- QAPP specifications. In summary, all data quality indicators and QC specifications were reviewed and found to be well within project specifications. Therefore, the data are considered suitable for purposes of this evaluation. 5.4 Audit Findings The SAIC SITE QA Manager conducted audits of both field activities and of the subcontracted laboratory as part of the QA measures for this project. The results of these technical system reviews are discussed below. The field audit resulted in no findings or non- conformances. The audit performed at the subcontract laboratory was conducted during the time of project sample analysis. One non-conformance was identified and corrective action was initiated. It was discovered that the laboratory PQL was not meeting specifications due to a reporting error. The analyst was generating the calibration curves as specified above; however, the lower limit on the calibration curve was not being reported. This was immediately rectified and no other findings or non- conformances were identified. 32 ------- Chapter 6 Performance of the DMA-80 Milestone, Inc. analyzed 173 samples from May 5-8,2003 in Oak Ridge, TN. Results for these samples were reported by Milestone, and a statistical evaluation was performed. Additionally, the observations performed by SAIC during the demonstration were reviewed, and the remaining primary and secondary objectives were completed. The results of the primary and secondary objectives, identified in Chapter 1, are discussed in Sections 6.1 and 6.2, respectively. The DMA-80 was used during the pre-demonstration in October, 2002 and during the demonstration by Milestone. Some of the pre-demonstration samples had concentrations above the 5 ppm upper concentration limit identified by Milestone. Results for those samples were reported, and although no statistical evaluation was performed, the results were similarto those reported bythe analytical laboratory. To analyze sam pies above the 5 ppm concentration, a soil to solid dilution using silica gel is prepared. Samples were prepared for Milestone for the demonstration that were above the 5 ppm concentration. After arriving at the demonstration, it was determined by Milestone that the soil to solid dilution was not appropriate for field analyses; therefore, Milestone elected to not analyze the samples from the manufacturing site (with concentrations between approximately 5-1,000 ppm). Subsequently, precision and accuracy were only determined for sam pie concentrations below 5 ppm. Use of Milestone's field instrument for higher sample concentrations would likely add additional unknown variance and therefore, it is not recommended for concentrations above 5 ppm based upon the results of this field study. The distribution of the samples prepared for Milestone Inc. and the referee laboratory, is presented in Table 6-1. Milestone, Inc. received samples at 31 different concentrations for a total of 173 samples. These 173 samples consisted of 20 concentrations in replicates of 7, and 11 concentrations in replicates of 3. Table 6-1. Distribution of Samples Prepared for Milestone and the Referee Laboratory Site Concentration Range soil Sample Type Sediment Spiked Soil SRM Carson River (Subtotal = 75) Puget Sound (Subtotal = 57) Oak Ridge (Subtotal = 41) Subtotal (Total = 173) Low(1-500ppb) Mid (0.5-50 ppm) Hiqh (50->1,000 com) Low (1 ppb - 10 ppm) Hiah (10-500 pom) Low (0.1 -10 ppm) Hiah (10-800 pom) 7 9 0 26 0 17 0 72 10 0 0 0 0 3 0 13 7 14 0 14 0 7 0 42 7 ' 21 0 17 0 14 0 70 33 ------- 6.1 Primary Objectives 6.1.1 Sensitivity Sensitivity objectives are explained in Chapter 4. The two primary sensitivity evaluations performed for this demonstration were the MDL and PQL. Determinations of these two measurements are explained in the paragraphs below, along with a comparison to the referee laboratory. These determinations set the standard for the evaluation of accuracy and precision for the Milestone field instrument. Any sample analyzed by Milestone and subsequently reported below their level of detection was not used as part of any additional evaluations. This was done because of the expectation that values below the lower limit of instrument sensitivity would not reflect the true instrument accuracy and precision. In addition, samples that were reported as greater than 5 mg/kg were not used in the evaluation of primary objectives. Therefore, there were fewer than the 173 samples previously noted in Section 6.0 used in the accuracy and precision evaluation. The sensitivity measurements of MDL and PQL are both dependent upon the matrix and method. Hence, the MDL and PQL will vary, depending upon whether the matrix is a soil, waste, or water. Only soils and sediments were tested during this demonstration and therefore, MDL calculations for this evaluation reflect soil and sediment matrices. PQL determinations are not independent calculations, but are dependent upon results provided by the vendor for the samples tested. Comparison of the MDL and PQL to laboratory sensitivity required that a standard evaluation be performed for all instruments tested during this demonstration. PQL, as previously noted, is defined in EPA G-5i as the lowest level of method and instrument performance with a specified accuracy and precision. This is often defined by the lowest pointon the calibration curve. Our approach was to letthe vendor pro vide the lower limit of quantitation as determined by their particular standard operating procedure, and then test this limit by comparing results of samples analyzed at this low concentration to the referee laboratory results, or comparing the results to a standard reference material, if available. Comparison of these data are, therefore, presented for the lowest concentration sample results, as provided by the vendor. If the vendor provided "non-detect" results, then no formal evaluation of that sample was presented. In addition, the sample(s) was not used in the evaluation of precision and accuracy. Method Detection Limit - The standard procedure for determining MDLs is to analyze a low standard or reference material seven times, calculate the standard deviation and multiply the standard deviation by the V value for seven measurements at the 99th percentile (alpha = 0.01). (This value is 3.143 as determined from a standard statistics table.) This procedure for determination of an MDL is defined in 40 CFR Part 136, and while determinations for MDLs may be defined differently for other instruments, this method was previously noted in the demonstration QAPP and is intended to provide a comparison to other similar MDL evaluations. The purpose is to provide a lower level of detection with a statistical confidence at which the instrument will detect the presence of a substance above its noise level. There is no associated accuracy or precision provided or implied. Several blind standards and field sam pies were provided to Milestone at their estimated lower limit of sensitivity. The Milestone lowerlimit of sensitivity was previously estimated at 0.008 mg/kg. Because there are severaldifferent SRMs and field samples at concentrations close to the MDL, evaluation of the MDL was performed using more than a single concentration. Samples chosen for calculation were based upon: 1) concentration and how close it was to the estimated MDL,.2) number of analyses performed for the same sample (e.g., more than 4), and 3) if non-detects were reported by Milestone for a sample used to calculate the MDL. Then the next highest concentration sample was selected based upon the premise that a non-detect result reported for one of several samples indicates the selected sample is on the "edge" of the instruments detection capability. Milestone ran two separate, blind low standards (SRMs), each, seven times. This included one standard at 0.082 mg/kg and one at 0.62 mg/kg. For testing the method sensitivity claim, another blind SRM was provided to Milestone which was 0.017 mg/kg. This standard was analyzed only three times; therefore, it was not used in the MDL calculation. The two standards used had standard deviations of 0.29908 for the 0.082 mg/kg SRM, and 0.021769 for the 0.62 mg/kg SRM. Multiplying each of these standard deviations by the "t" statistic noted previously, one can calculate MDLs of 0.94 and 0.068 mg/kg, respectively. There was, however, one of the seven results for the 0.082 mg/kg SRM that appears to be outside the expected range of the other six determinations (e.g. almost an orderof magnitude above all other results). If this result is not used, then the recalculated MDL (using 6 values and a T statistic of 3.365) is 0.049 mg/kg. Even when an outlier value was excluded from the calculation of the MDL for the 0.082 mg/kg value, the recalculated MDL is still well above the Milestone claim of 0.008 mg/kg. It 34 ------- would therefore appear that the Milestone claim for sensitivity of 0.008 mg/kg is .not applicable to soil and sediment materials. The objective in estimating an MDL is to run a "low" instrument standard so that a more accurate MDL evaluation can be determined. Discounting the 0.94 mg/kg value calculated for the low standard of 0.082 mg/kg (calculated with an apparent outlier value), the MDL is estimated between 0.049 and 0.068 mg/kg. The equivalent MDL for the referee laboratory based upon analysis of a low standard analyzed seven times is 0.0026 mg/kg. The calculated result is only intended as a statistical estimation, and not a true test of instrument sensitivity. Practical Quantitation Limit - This value is usually calculated by determining a low standard on the instrument calibration curve and it is estimated as the lowest standard at which the instrument will accurately and precisely determine a given concentration within specified QC limits. The PQL is often around 5-10 times the MDL. This PQL estimation, however, is method-and matrix-dependent. In order to determine the PQL, several low standards were provided to Milestone and subsequent %Ds were calculated. Using the MDL calculations presented above, this would translate into a low limit PQL of 0.24 mg/kg. The instrument manufacturer, however, suggests a low calibration of 0.010 mg/kg standard in a 1 g sample (over an order of magnitude below the PQL calculated above) which translates to a 10 ng standard "on column." It would appear, based on the information gained during this demonstration, that this low standard calibration is likely well below instrument capabilities in determining an accurate and precise calculation for a PQL. In fact, a low standard of 0.017 mg/kg was tested during the demonstration, as rioted above. This was run three times by Milestone, and the average value calculated was 0.0089 mg/kg, with a standard deviation of 0.00108 mg/kg. A 95% Clfor this value is, therefore, 0.0077 to 0.010 mg/kg which is outside the range of the reference value determination of 0.017 mg/kg. The %D for the average value reported by Milestone, compared to the reference value of 0.017 mg/kg is 48%. Therefore, it appears that the instrument PQL is above the 0.010 mg/kg value suggested by the manufacturer, and above the 0.017 mg/kg value tested during the demonstration. The PQL may be close to the average MDL determined above, 0.058 mg/kg. The next lowest standard tested during the demonstration was the 0.082 mg/kg SRM. This standard was run seven different times during the demonstration. Seven different blind samples were analyzed by Milestone during several different instrument analytical batches. The average value calculated by Milestone for this standard was 0.206 mg/kg, with a standard deviation of 0.29908 mg/kg. The 95% CIs for this standard is therefore -0.016 to 0.428 mg/kg, and encompasses zero. The %D for this calculated value compared to the reference value is 151%. The reference value falls within the Cl because the standard deviation is -extremely wide (note that the relative standard deviation or coefficient of variation is 145%). Therefore, it would appear that because of this wide CIs range, this may be outside instrument capabilities for a precise evaluation of this low standard. There was, however, one of these seven results that appears to be outside the expected range of the other 6 determinations (see note above). If this result is not used, then the average value is 0.0935 mg/kg, the standard deviation is 0.0134 mg/kg, and a relative standard deviation of 15.7%. The 95% Cl is 0.0828 to 0.1042 mg/kg. This Cl does not quite include the reference value of 0.082 mg/kg, but is a much narrower range. It is close to the given reference value for the SRM, and overlaps the 95% SRM Cl. The %D between the calculated average and reference value for this determination is 13.8%. The laboratory results for this same standard (0.082 mg/kg SRM) estimated an average value of 0.0729 mg/kg, a standard deviation of 0.005 mg/kg, and a relative standard deviation of 6.7%. The %D between the referee laboratory average and the reported standard is 11%. This is given for purposes of comparison to the Milestone result. The next lowest SRM value was 0.62 mg/kg. This was run as seven different blind samples by Milestone. The average value was 0.627 mg/kg, and the standard deviation was 0.0218 mg/kg, with a relative standard deviation of 3.47%. The 95% CIs for this standard is 0.611 to 0.643 mg/kg. The SRM value falls within this CIs. As previously noted, this is a very narrow CIs, suggesting not only an accurate, but also a very precise evaluation. The %D between the calculated average and reference value is 1.1%. The laboratory reported an average value of 0.533 mg/kg and a standard deviation of 0.033 and relative standard deviation of 6.2%. The %D between the calculated average and reference value for the referee laboratory is 14%. It could be suggested that the instrument PQL is above 0.017 mg/kg, perhaps close to 0.082 mg/kg, and below 0.62 mg/kg. Given the information associated with the 35 ------- MDL determination, it would appear that the PQL is likely not below the average MDL, 0.058 mg/kg. Sensitivity Summary The MDL is estimated between 0.049 and 0.068 mg/kg. The equivalent MDL for the referee laboratory based upon analysis of a low standard analyzed seven times is 0.0026 mg/kg. The MDL determination, however, is only a statistical calculation that has been used in the past by EPA and is currently not considered a "true" MDL by SW-846 methodology. SW-846 is suggesting that performance based methods be used and that PQLs be determined using low standard calculations. The low standard calculations suggest that a PQL for the Milestone field instrument is somewhere around 0.082 mg/kg. The referee laboratory PQL confirmed during the demonstration is 0.005 mg/kg. The %D for Milestone fie Id instrumentation at concentrations of 0.082 and 0.62 mg/kg is very comparable to the reference laboratory method suggesting a PQL close to the lower SRM, 0.082 mg/kg for soil and sediment materials. 6.1.2 Accuracy Accuracy is the instrument measurement compared to a standard or true value. For this demonstration, three separate standards were used for determining accuracy. The primary standard is SRMs. The SRMs are traceable to national systems. These were obtained from reputable suppliers with reported concentrations and an associated 95% Cl and 95% prediction interval. The. Cl from the reference material is used as a measure of comparison with the Cl calculated from replicate analyses for the same sample analyzed by the laboratory or vendor. Results are considered comparable if CIs of the SRM overlap with the CIs computed from the replicate analyses by the vendor. While this is not a definitive measure of comparison, it provides some assurance that the two values are equivalent: Prediction intervals are intended as a measure of comparison for a single laboratory or vendor result with the SRM. When computing a prediction interval, the equation assumes an infinite number of analyses, and it is used to compare individual sample results. A 95% prediction interval would, therefore, predict the correct result from a single analysis 95% of the time for an infinite number of samples, if the result is comparable to that of the SRM. It should be noted that the corollary to this statement is that 5% of the time a result will be outside the prediction interval if determined for an infinite number of samples. If several samples are'analyzed, the percentage of results within the prediction interval will be slightly above or below 95%. The more samples analyzed, the more likely the percentage of correct results will be close to 95% if the result for the method being tested is comparable to the SRM. Most SRMs were analyzed in replicates of seven by both the vendor and by the referee laboratory. SRMs for reanalyses were analyzed as replicates of three. In some cases there were apparent outlier results, as noted from the Milestone data. When this occurred, calculations were performed both with and without the outlier data for purposes of comparison. These were statistical outliers, generally an order of magnitude above the other results. There was no other reason to consider these samples as outliers other than statistical anomalies, and therefore, these results were not completely discounted from the analysis. The second accuracy determination used a comparison of vendor results of field samples and SRMs to the referee laboratory results for these same samples. Field samples were used to ensure that "real-world" samples were tested by the vendor. The referee laboratory result is considered as the standard for comparison to the vendor result. This comparison is in the form of a hypothesis test with .alpha = 0.01. (Detailed equations along with additional information about this statistical comparison is included in Appendix B.) It should be noted that there is evidence of a laboratory bias. This bias was determined by comparing average laboratory values to' SRM reference values, and is discussed below. The laboratory bias is low in comparison to the reference value. A bias correction was not made when comparing individual samples (replicate analyses) between the laboratory and vendor; however, setting alpha = 0.01 helps mitigate for this possible bias by widening the range of acceptable results between the two data sets. ' In addition, there appears to be a Milestone bias, and this bias is high in comparison to the average reference value. This will be discussed in more detail later, however, in general the laboratory and Milestone data were within expected ranges, except for two SRMs which were subsequently reanalyzed by the laboratory after the demonstration, as explained below. An aggregate analysis or unified hypothesis test was also performed for all 28 sample lots. (A detailed discussion of this statistical comparison is included in Appendix B.) This analysis provides additional statistical evidence in relation to the accuracy evaluation. A bias term is included in this calculation in order to account for any data bias. 36 ------- The third measure of accuracy is obtained by the analysis of spiked field samples. These were analyzed by the vendors and the laboratory in replicate in order to provide additional measurement comparisons and are treated the same as field samples. Spikes were prepared to cover additional concentrations not available from SRMs or field samples. There is no comparison to the spiked concentration, only a comparison between the vendor and the laboratory reported value. The purpose for SRM analysis by the referee laboratory is to provide a check on laboratory accuracy. During the pre-demonstration, the referee laboratory was chosen, in part, based upon the analysis of SRMs. This was done in order to ensure that a competent laboratory would be used for the demonstration. The pre-demonstration laboratory qualification showed that the laboratory was within prediction intervals for all SRMs analyzed. Because of the need to provide confidence in laboratory analysis during the demonstration, the referee laboratory also analyzed SRMs as an on-going check of laboratory bias. The pre-demonstration laboratory evaluation was conducted to help ensure that laboratory SRM data would fall within expected ranges. It was considered possible, however, that during the demonstration the laboratory may fail to fall within the expected concentration ranges for a particular SRM. This did occur, and laboratory corrective action included a check of the laboratory calibration and calibration criteria for those particular samples. (See Table 6-2 for results.) These QC checks were found to be well within compliance, therefore, the laboratory was asked to recalibrate and rerun the two SRMs, as noted in the table. (SRM values were not provided to the laboratory upon reanalysis, nor was the laboratory told why the samples needed to be reanalyzed.) In particular, the SRM labeled as sample lot 40 had a reference value of 1.12 mg/kg. The laboratory analysis of 7 different blind samples on 7 different runs recorded an average analysis of 0.12 with a relative standard deviation of 30%. This result would suggest that this was not a statistical anomaly, but something else, such as so me type of sample interference or simply a mis-labeled sample lot number. Therefore, it was decided that this sample and one additional SRM (sample lot 44) should be reanalyzed. Results of the reanalysis for the two SRMs showed that the laboratory was well within expected CIs and prediction intervals for both sample lots. Evaluation of vendor and laboratory analysis of SRMs is performed in the following manner. Accuracy was determined by comparing the 95% Cl of the sample analyzed by the vendor and laboratory to the 95% Cl for the SRM. (95% CIs around the true value are provided by the SRM supplier.) This information is provided in Tables 6-2 and 6-3, with notations when the CIs overlap, suggesting comparable results. In addition, the number of SRM results for the vendor's analytical instrumentation and the referee laboratory that are within the associated 95% prediction interval are reported. This is a more definitive evaluation of laboratory and vendor accuracy. The percentage of total results within the prediction interval for the vendor and laboratory are reported in Tables 6-2 and 6-3, respectively. The single most important number from these tables is the percentage of samples within the 95% prediction interval. As noted for the Milestone data, this percentage is 93% with n = 54. (This result is computed after removing statistical outliers.) This suggests that the Milestone data are within expected accuracy accounting for statistical variation. For 8 of the 10 determinations (where 7 samples were analyzed), Milestone average results are above the reference value. This would also suggest that possibly Milestone data are biased high. Six of the eight sample lots overlap with the 95% CIs calculated from the Milestone data, compared to values provided by the supplier of the SRM. This number is also suggestive of a reasonable comparison to the SRM value, accounting for statistical variation. The percentage of samples within the 95% prediction interval for the laboratory data (after taking into account sample reanalysis for apparent anomalies) is 96%. This result also suggests that the ALSI data are within expected accuracy accounting for statistical variation. For 7 of the 10 determinations, ALSI average results are below the reference value. This would also suggest that the ALSI data are potentially biased low. Seven of the ten sample lots overlap with the 95% CIs calculated from the ALSI data, compared to values provided by the supplier of the SRM. This number is also suggestive of a reasonable comparison to the SRM value accounting for statistical variation. 37 ------- Table 6-2. Milestone SRM Comparison Sample No. 37 37 44 36 36 40 38 39 41 43 a b Lot SRM Value/ 95% Cl 0.158/0.132-0.184 0.158/0.132-0.184' 4.7/4.3-5.1 0.082/0.073-0.091 0.082/0.073-0.091' 1.12/1.08-1.17 0.62 / 0.54 - 0.70 1.09/0.94-1.24 2.42/2.16-2.46 3.80/3.50-4.11 Total Samples Total Samples w/o outliers % of samples w/in prediction interval Milestone Avg./ 95% Cl 0.482 / 0.0502 - 0.914 0.307/0.230-0.384 4.40 / 4.23 - 4.57 0.206 / 0 - 0.483 0.0935/0.0781-0.109 1.30/1.18-1.42 0.627 / 0.607 - 0.647 1.09/1.03- 1.15 2.83/1.64-4.02 4.17/3.51 -4.83 Cl Overlap (ves/no) yes no yes yes yes no yes yes yes ves Calculated results w/out suspected outlier value. Prediction interval is estimated based upon n=30. A 95% Cl was provided No. of Samples Analyzed 7 6 7 7 6 7 7 7 7 7 56 54 95% Prediction Interval 0 - 0.357 0 - 0.357 3.0-6.4 0.0579-0.106" 0.0579-0.106" 0.49-1.76 0.200-1.04" 0.303-1.88" 1.30-3.32 2.41 - 5.20 Milestone No. w/in Prediction Interval 5 5 7 6 6 7 7 7 6 7 50 50 93% by the SRM supplier but no prediction interval was given. Table 6-3. ALSI SRM Comparison Sample Lot SRM Value/ 95% (Cl) No. 37 44 44. 36 40 40' 38 39 41 43 0.158/0.132-0.184 0.158/0.132-0.184' 4.7/4.3-5.1 0.082/0.073-0.091 0.082 / 0.073 - 0.091 " 1.12/1.08- 1.17 0.62 / 0.54 - 0.70 1.09/0.94- 1.24 2.42/2.16-2.46 3.80/3.50-4.11 Total Samples % of samples w/in prediction interval ALSI Avg./ 95% Cl 0.139/0.0928-0.185 2.33/1.05-3.61 4.09 / 3.60 - 4.58 0.073 / 0.0684 - 0.0776 0.12/0.087-0.15 1.02/0.464-1.58 0.533 / 0.502 - 0.564 1.24/0.634-1.84 1.79/1.28-2.29 2.76 / 2.51 - 3.01 Cl Overlap (ves/no) yes no yes yes no yes yes yes yes no No. of Samples Analyzed 7 7 3 7 7 3 7 7 7 7 56 95% Prediction Interval 0 - 0.357 0 - 0.357 3.0-6.4 0.0579-0.106" 0.0579-0.106" 0.49-1.76 0.200 - 1.04 ' 0.303-1.88" 1.30-3.32 2.41 - 5.20 ALSI No. w/in Prediction Interval 7 2 3 . 7 0 3 7 6 6 7 42 75% Reanalvsis a b Total Samples • % of samples w/in prediction interval 48 Reanalysis of SRM samples was performed by laboratory based upon QAPP corrective action procedures. Prediction interval is estimated based upon n=30. A 95% Cl was provided by the SRM supplier but no prediction 46 96% interval was given. Hypothesis Testing Sample results from field and spiked field samples for the vendor compared to similar tests by the referee laboratory are used as another accuracy check. Spiked samples were used to cover concentrations not found in the field samples/and they are considered the same as the field samples for purposes of comparison. Because of the limited data available for determining the accuracy of the spiked value, these were not considered the same as reference standards. Therefore, these samples were evaluated in the same fashion as field samples, but they were not compared to individual spiked concentrations. Using a hypothesis test with alpha = 0.01, vendor results for all samples were compared to laboratory results to determirte if sample populations are the same or significantly different. This was performed for each sample lot separately. Alpha was set at 0.01 to help mitigate for inter-laboratory bias as mentioned earlier. This mitigation attempt, however, has limited value because Milestone 38 ------- data are likely biased high, and ALSI data are likely biased low. As a result of this bias, some sample lots would be expected to result in comparisons that would test as significantly different. Because this test does not separate precision from bias, if Milestone's or ALSI's computed standard deviation was large due to a highly variable result (indication of poor precision), the two CIs could overlap, and therefore, the fact that there was no significant difference between the two results may be due to high sample variability. Overall precision, however, as noted from the precision evaluation (section 6.1.3) is within expected ranges for both Milestone and ALSI data. Accordingly, associated RSDs have also been reported in Table 6-4 along with results of the hypothesis testing for each sample lot. Table 6-4. Accuracy Evaluation by Hypothesis Testing Sample Lot No7 Site 03/ Oak Ridge Milestone ALSI 09/ Oak Ridge Milestone ALSI 14/ Oak Ridge Milestone ALSI 371 Oak Ridge SRM Milestone ALSI 44/ Oak Ridge SRM Milestone ALSI 44/ Oak Ridge SRM • Milestone ALSI 02/ Puget Sound Milestone ALSI 05/ Puget Sound Milestone ALSI 081 Puget Sound Milestone ALSI 10/ Puget Sound Milestone ALSI 1 1/ Puget Sound Milestone ALSI 12/ Puget Sound Milestone ALSI 35/ Puget Sound SRM Milestone ALSI 36/ Puget Sound SRM Milestone ALSI 40/ Puget Sound SRM Milestone ALSI Avg. Cone. mg/kg 0.45 0.26 0.70 0.47 5.28 4.75 0.48 0.14 4.40 2.33 4.40 4.09 0.089 0.060 0.28 0.21 0.87 0.55 0.71 0.36 1.35 0.81 1.46 1.08 0.0089 0.0087 0.21 0.073 1.30 0.12 RSD or CV 23.2% 3.8% 39.8% 34.2% 12.6% 27.5% 96.9% 36.4% 4.2% 59.4% 4.2% 4.8% 41.3% 23.6% 15.8% 33.3% 31.6% 13.4% 25.3% 20.5% 19.2% 32.6% 2.6% 2.8% 12.2% 6.3% 144% 6.7% 10.1% 30.0% Number of Measurements 3 3 7 7 3 6 7 7 7 7 7 3 7 7 3 3 3 7 3 3 7 7 3 3 3 7 7 7 7 7 Significantly Different at Alpha = 0.01 no > no no no yes no no no no no yes yes no no yes Relative Percent Difference (Milestone to ALSI) 53.5% 40.5% 10.5% 111% 61.4% 7.3% 38.7% 27.5% 45.2% 64.7% 49.9% 29.9% 1.5% 95.4% 166% 39 ------- Table 6-4. Continued Sample Lot No7 Site MO/ Puget Sound SRM Milestone ALSI 571 Puget Sound Milestone ALSI 01/ Carson River Milestone ALSI 04/ Carson River Milestone ALSI 06/ Carson River Milestone ALSI 15/ Carson River Milestone ALSI 1 61 Carson River Milestone ALSI 18/ Carson River Milestone ALSI 38/ Carson River SRM Milestone ALSI 39/ Carson River SRM Milestone ALSI 41/ Carson River SRM Milestone ALSI 431 Carson River SRM Milestone ALSI 56/ Carson River Milestone ALSI 58/ Carson River Milestone ALSI 59/ Carson River Milestone ALSI a Reanalysis performed CV Coefficient of variance Avg. Cone. mg/kg 1.30 1.02 1.03 0.73 0.24 0.24 0.13 0.11 0.46 0.26 3.75 4.23 8.78 7.14 8.66 10.1 0.63 0.53 1.09 1.24 2.83 1.79 4.17 2.76 0.28 0.23 0.86 0.76 2.02 1.71 RSD or CV 10.1% . 22.0% 20.1% 16.2% 41.6% 37.8% 16.1% 9.1% 15.5% 15.7% 11.1% 24.5% 34.2%' 13.7% 7.8% 8.0% 3.5% 6.2% 6.1% 52.9% 45.5% 30.5% '17.1% 9.6% 27.8% 12.6% 12.3% 8.6% 13.8% 7.9% Number of Measurements 7 3 7 7 7 . 7 7 7 3 7 3 . .7 3 3 3 7 . 7 7 7 7 7 7 7 7 7 7 7 7 7 7 Significantly Different at Alpha = 0.01 no yes no no no no no no yes no no yes no no . no Relative Percent Difference (Milestone to ALSI) 24.1% 33.6% 1.3% 19.7% 55.3% -12.0% 20.6% -15.4% 17.2% -12.9% 45.0% 40.7% 19.9% 13.0% 16.6% due to SRM results outside expected accuracy specifications. Of the 28 sample lots, 7 results are significantly different. This may be a higher number of results outside expected accuracy evaluations for Milestone and ALSI if these two sets of data were equal with alpha = 0.01. (Normally this would mean that 1 in 100 results would be outside expected ranges for both sets of data due to statistical variation.) There were, however, two SRMs analyzed'by the laboratory that appeared in question. Per QAPP specifications, reanalysis for the noted SRMs (sam pie lots 40 and 44) were performed. Upon reanalysis with three replicates, both of the sample lots were within expected accuracy specifications. Hypothesis testing at alpha = 0.01 resulted in no significant difference between ALSI and Milestone for the reanalyzed SRMs. Therefore, the number of sample lots that are significantly different dropped to 5. Most of the relative percent differences are 40 ------- positive (all but three) which indicates that the Milestone result is generally higher than the laboratory result. This is indicative of the previously noted low bias associated with the laboratory data. the second set of three analyses (sample Iot40) produced similar RSD results as had been achieved with 7 replicates, indicating a very precise determination. When accounting for differences between the ALSI and Milestone data, it should be noted that there may be inherent biases in both sets of data. ALSI may be biased low and Milestone may be biased high, therefore comparisons of the two data sets would likely result in an additional number of data sets that were significantly different over and above the 1 in 100 difference as noted previously. Table 6-5. Number of Sample Lots Within Each %D Range <30% >30%. <50% In determining the number of results significantly above or belowthe value reported by the referee laboratory, 16 of 30 Milestone average results were found to have relative percent differences less than 30%. Only 2 of 30 Milestone average results have relative percent differences greater than 100% for this same group of samples (see Table 6-5). The differences are accentuated by the low bias for the laboratory results and the high bias for Milestone results. There appears to be more significant differences in the Puget Sound sample set than any of the other sample lots, which may be due to an interference for these particular samples (see Table 6-6). >50%. <100% >100% Total Positive %D Negative %D Total 13 3 16 7 0 7 5 0 5 Table 6-6. Concentration (in mg/kg) of Non-Target Analytes Lot# Site TOC O&G Aq As Ba 1 Carson River 2 Puget Sound 3 Oak Ridge 4 Carson River 5 Puget Sound 6 Carson River 8 Puget Sound 9 Oak Ridge 10 Puget Sound 1 1 Puget Sound 12 Puget Sound 14 Oak Ridge 15 CarsonRiver 16 Carson River 18 Carson River 35 SRM Canmet SO-3 36 SRM Canmet SO-2 37 SRMCRM-016 38 SRMNWRITH-2 39 SRM NWRI WQB-1 40 SRMCRM020 41 SRM CRM 026 43 SRM CRM 027 44 SRM CRM 021 56 Spiked Lot 1 57 Spiked PS- X1.X4 58 Spiked CR-SO-1 4 59 Soiked CR-SO-1 4 870 3500 2300 2400 3500 7200 8100 3300 4200 3800 3500 7800 2700 2100 1900 NR NR NR NR NR NR NR NR NR . 870 3500 870 870 190 290 530 200 210 200 200 150 130 t30 290 180 70 80 70 NR NR NR NR NR NR NR NR NR 190 290 190 190 <0.5 <0.5 1.8 <0.5 <0.5 <0.5 <0.5 1.9 <0.5 <0.5 <0.5 0.32 3.2 0.5 26 NR NR 0.7 5.8 1 38 0.57 6 - 6.5 <0.5 <0.5 <0.5 <0.5 9 3 4 8 3 4 3 5 3 4 3 2 22 4 17 NR NR 7.8 8.7 23 400 5.4 12 25 9 3 9 9 210 23 ' 150 240 28 32 27 160 24 20 23 41 100 150 46 300 970 79 570 600 25 210 170 590 210 23 210 210 Cd <0.5 <0.5 <0.5 <0.5 <0.5 <0.5 1.0 0.5 <0.5 <0.5 0.8 0.4 <0.5 <0.5 2.0 NR NR 0.47 5.2 2 15 12 12 1.2 <0.5 <0.5 <0.5 <0.5 Cr 19 16 46 17 18 16 17 70 18 18 16 16 13 18 6 26 16 14 120 89 14 27 27 11 19 16 19 19 Cu 13 . 10 20 32 11 9 23 49 8 8 7 9 18 39 62 17 7 16 120 80 730 19 9.9 "4800 13 10 13 13 2 0 2 Pb 3 1 15 12 3 1 99 24 1 1 2 11 18 14 200 14 21 14 190 84 5100 26 52 6500 3 1 3 3 Se <2 <2 <2 <2 <2 <2 2 <2 <2 <2 <2 <2 <2 <2 <2 NR NR 1 0.83 1 6.6 1.9 14 NR <2 <2 <2 <2 Sn <5 <5. <5 <5 <5 <5 <5 <5 <5 <5 <5 <4 <5 <5 <5 NR NR NR NR 3.9 NR NR NR 300 <5 <5 <5 <5 27 3 30 Zn 60 24 55 66 28 24 37 100 24 24 23 74 49 81 390 52 120 70 900 275 3000 140 51 550 60 24 60 60 Ha 0.19 0.04 0.31 0.10 0.16 0.23 0.37 0.66 0.62 0.63 1.1 78 3.3 7.3 9.3 0.02 0.08 0.16 0.62 1.09 1.1 2.4 3.8. 4.7 0.20 0.61 ' 0.74 1.6 CRM = Canadian Reference Material NR = Not Reported by Standard Supplier 41 ------- Upon examination of non-target analyte data (Table 6-6) collected for these samples, no obvious interference was noted. For example, a high organic content may cause interference, but these samples do not necessarily have a higher organic content than other samples tested. In addition, the method 7471B mercury analysis requires that a non-stannous chloride analysis be conducted with each sample analyzed, in order to test for organic interferences. Upon examination of the laboratory data for the sample sets mentioned above, there was no apparent interference noted in the non-stannous chloride analysis. Other interferences caused by additional elements were also not found to be significant. Of course, there could be interferences that were not tested, and therefore, while it may be an interference particular to this sample lot, the exact cause remains unknown. In addition to the statistical summary presented above, data plots (Figures 6-1 and 6-2) are included in order to present a visual interpretation of the accuracy. Two separate plots have been included for the Milestone data. These two plots are divided based upon sample concentration in order to provide a more detailed presentation. Concentrations of samples analyzed by Milestone ranged approximately from 0.01 to 10 mg/kg. The previous statistical summary eliminated some of these data based upon whether concentrations were interpreted to be in the analytical range of the Milestone field instrument. This graphical presentation presents all data points. It shows Milestone data compared to ALSI data plotted against concentration. Sample groups are shown by connecting lines. Breaks between groups indicate a different set of samples at.a different concentration. Sample groups were arranged from lowest to highest concentration. As can be seen by this presentation, samples analyzed by Milestone appear to match well with the ALSI results. There are some outlierdata points, however, most of these data points when averaged with other data points from the same sample group were within approximate averaged concentrations. This is only a visual interpretation and does not provide statistical significance. It does, however, provide a visual interpretation that supports the previous statistical results for accuracy, as presented above. Figure 6-1. Data plot for low concentration sample results. 42 ------- Figure 6-2. Data plot for high concentration sample results. Unified Hypothesis Test SAIC performed a unified hypothesis test analysis to assess the comparability of analytical results provided by Milestone and those provided by ALSI. (See appendix B for a detailed description of this test.) Milestone and ALSI both supplied multiple assays on replicates derived from a total of 28 different sample lots, either field materials or reference materials. (Only samples above the previously estimated Milestone PQL were used.) The Milestone and ALSI data from these assays formed the basis of this assessment. The results of this unified hypothesis test show that the two data sets are similar and therefore Milestone data compared well to the referee laboratory. This confirms previous statistical determinations from above showing very few differences between the two data sets. Milestone analytical results for sample lot 37 were large relative to the concentration provided with the sample reference material, and the Milestone data for sample lot 37 made a substantial contribution to the chi-square statistic. Accordingly, excluding sample lot 37 (n = 27) from the data set resulted in a chi-square statistic of 42.1, which does not exceed the upper 99th percentile of the chi- square distribution with 26 degrees of freedom with value 45.6. So, excluding sample lot 37 data, results from this analysis suggest that the two data sets are the same for the ALSI laboratory and the Milestone field instrument. The null hypothesis tested was that, on average, Milestone and ALSI produce the same results within a given sample lot. Additional information about this statistical' evaluation is included in Appendix B. Accuracy Summary In summary, Milestone data compared to SRM values were within expected accuracy determinations. ALSI data compared to SRM values were also within expected accuracy determinations, after reanalysis of two sample lots. These two comparisons are the best evidence suggesting thatthe Milestone field instrument and the ALSI analysis provide accurate data. The additional comparison of these two data sets (hypothesis test for each sample lot), do not provide evidence contrary to the results of this comparison, butdo not necessarilysupportthis conclusion. 43 ------- The number of Milestone average values less than 30% different from the referee laboratory results or SRM reference values was 16 of 30 different sample lots. However, when making the comparison between Milestone and ALSI data, and taking into account the possible bias associated with both sets of data, the hypothesis test and the %D com parison may be within reasonable expectations for considering these two separate analyses to be equivalent. The unified hypothesis test provides additional evidence - that there is no statistical difference between data sets provided by ALSI and Milestone. Overall, the accuracy evaluations suggest that the Milestone field instrument provides results that are comparable within expected accuracy specifications, and should be considered equivalent. 6.1.3 Precision Precision is usually thought of as repeatability of a specific measurement, and it is often reported as RSD. The RSD is computed from a specified number of replicates. The more replications of a measurement, the higher confidence associated with a reported RSD. Replication of a measurement may be as few as 3 separate measurements to 30 or more measurements of the same sample, depending upon the degree of confidence desired in the specified result. Most samples were analyzed seven times by both Milestone and the referee laboratory. In some cases, samples may have been analyzed as few as three times. This was often the situation when it was believed •thatthe chosen sample, or SRM, was likely to be below the vendor quantitation limit. The precision goal for the referee laboratory, based upon pre-demonstration results is an RSD of 25% or less. A descriptive evaluation for differences between Milestone RSDs and the referee laboratory RSDs was determined. In Table 6-7, the RSD for each separate sample lot is shown for Milestone compared to the referee laboratory. The average RSD was then computed for all measurements made by Milestone, and this value was compared to the average RSD for the laboratory. In addition, the precision of an analytical instrument may vary, depending upon the matrix being measured, the concentration of the analyte, and whether the measurement is made for an SRM or a field sample. To evaluate precision for clearly different matrices, an overall average RSD for the SRMs is calculated and compared to the average RSD for the field samples. This comparison is also included in Table 6-7 and shown for both Milestone and the referee laboratory. The purpose of this evaluation is to determine the field instrument's capability to precisely measure analyte concentrations under real-life conditions. Instrument repeatability was measured using samples from each of three different sites. Within each site, there may be two separate matrices, soil and sediment. Not all sites have both soil and sediment matrices, nor are there necessarily high, medium, and low concentrations for each sample site. Therefore, spiked samples were included to cover additional ranges. (Originally there were 4 different sites chosen for each vendor; however, M ilestone's capability to measure high concentration samples was limited under field conditions. Therefore, because the samples from the manufacturing site were believed to be above Milestone's upperquantitation limit, these samples were notanalyzed.) Table 6-7 shows results from Oak Ridge, Puget Sound, and Carson River. Itwasthoughtthat because these three different field sites represented different matrices, measures of precision may vary from site to site. The average RSD for each site is shown in Table 6-7 and compared between Milestone and the referee laboratory. SRM RSDs are not included in this comparison because SRMs, while grouped with different sites for purposes of ensuring that the samples remained blind during the demonstration, were not actually samples from that site, and were, therefore, compared separately. The RSDs of various concentrations are compared by noting the RSD of the individual sample lots. The ranges of test samples (field, SRMs, and spikes) were selected to cover the appropriate analytical ranges of Milestone's instrumentation. Average referee laboratory values for sample concentrations are included in the table, along with SRM values, when appropriate. These are discussed in detail in the Section 6.1.2 and are included here for purposes of precision comparison. Sample concentrations were separated into approximate ranges: low, medium, and high, as noted in Table 6-7 and Table 6-1. Milestone's field instrument, however, is an atomic absorption instrument and therefore less subject to concentration variations. This means that variations in precision due to varying concentrations is less likely. Because Milestone performed no sample dilution, there are no additional operations that would likely affect precision measurements. 44 ------- Table 6-7. Evaluation of Precision Sample Lot No. Milestone and Lab Avg. Cone, or Reference SRM Value RSD Number of Samples w/in 25% RSD Goal? OAK RIDGE Lot no.. 03 Milestone ALSI Lot no. 09 Milestone ALSI Lot no. 14 Milestone ALSI Lot no.37/ SRM Milestone 'Milestone w/out outlier ALSI Lot no. 441 SRM Milestone ALSI Oak Ridge Avg. RSD Milestone ALSI 0.26 (medium) 0.47 (medium) 4.75.(high) 0.16 (medium) 4.70 (high) 23.2% 3.8% 39.8% 34.2% 12.6% 27.5% 96.9% 23.8% 36.4% 4.2% 59.4% 25.0% 28.9% 3 3 7 7 3 6 7 6 7 7 7 yes ' yes no no yes •no no yes no yes no yes no PUGET SOUND Lot no. 02 Milestone ALSI Lot no. 05 Milestone ALSI Lot no. 08 Milestone ALSI Lot no. 10 Milestone ALSI Lot no. 11 Milestone ALSI Lot no. 12 Milestone ALSI Lot no. 35/ SRM Milestone ALSI Lot no. 36/ SRM Milestone 'Milestone w/out outlier ALSI Lot no. 40/ SRM Milestone ALSI Lot no. 57 Milestone ALSI Puget Sound/ Avg. RSD Milestone ALSI 0.060 (low) 0.21 (medium) 0.36 (medium) 0.55 (medium) 0.81 (medium) 1.08 (high) 0.02 (low) 0.08 (low) 1.12 (high) 0.73 (medium) 41.3% 23.6% 15.8% 33.3% 31.6% 13.4% 25.0% 20.5% 19.2% 32.6% 2.6% 2.8% 12.2% 6.3% 144% 15.7% 6.7% 10.1% 30.0% 20.1% 16.2% 22.3% 20.4% 7 7 3 3 3 7 3 3 7 7 3 3 3 7 7 . 6 7 7 7 7 7 no yes yes no no yes yes yes yes no yes yes yes yes no yes yes . yes no yes yes 45 ------- Table 6-7. Continued Sample Lot No. Milestone and Lab Avg. Cone, or Reference RSD SRM Value Number of Samples w/in 25% RSD Goal? CARSON RIVER Lot no. 01 Milestone ' ALSI Lot no. 04 Milestone ALSI Lot no. 06 Milestone ALSI Lot no. 1 5 Milestone ALSI Lot no. 16 Milestone ALSI Lot no. 1 8 Milestone ALSI Lot no. 38/ SRM Milestone ALSI Lot no. 39/ SRM Milestone ALSI Lot no. 417 SRM Milestone ALSI Lot no. 431 SRM Milestone ALSI Lot no. 56 Milestone ALSI Lot no. 58 Milestone ALSI Lot no. 59 Milestone ALSI Carson River/ Avg. RSD Milestone ALSI 0.24 (medium) 0.11 (medium) 0.26 (medium) 4.23 (high) 7.14 (high) 10.1 (high) 0.62 (medium) 1.09 (high) 2.42 (high) 3.85 (high) 0.23 (medium) 0.76 (medium) 1.71 (high) 41 .6% 37.8% 16.1% 9.1% 15.5% 15.7% 11.1% 24.5% 34.2% 13.7% 7.8% 8.0% 3.5% 6.2% 6.1% 52.9% 45.5% 30.5% 17.1% 9.6% 27.8% 12.6% 12.3% 8.6% 13.8% 7.9% 20.0% 15.3% 7 7 7 7 3 7 3 7 3 . 3 3 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 no no yes yes yes yes yes yes no yes yes yes yes yes yes no no no yes yes no yes yes yes yes yes yes yes SUMMARY STATISTICS Overall Avg. RSD Milestone ALSI Field Samples/ Avg. RSD Milestone ALSI . SRMs/Avg. RSD Milestone ALSI 19.4% 23.7% 22.0% 19.6% 15.3% 26.5% yes yes yes yes yes no 46 ------- Samples below the MDL, as determined in the section discussing sensitivity, were not included in Table 6-7. There appears to be no correlation between concentration (low, medium, or high) and RSD; therefore, no otherformal evaluations of this comparison were performed. The referee laboratory analyzed replicates of all samples analyzed by Milestone. This was used for purposes of precision comparison to Milestone. RSD for the vendor and the laboratory were calculated individually and shown in Table 6-7. Milestone precision is very comparable to the referee laboratory (Table 6-7). The single most important m easure of precision provided in Table 6-7, overall average RSD, is 23.7% forthe referee laboratory compared to the Milestone average RSD of 19.4%. Both of these RSDs are within the predicted 25% RSD objective for precision, expected from both analytical and sampling variance. In addition, field sample precision compared to SRM precision shows that there may be no significant difference between these two sample lots; field sample RSD 19.6% for ALSI and 22.0% for Milestone; SRM RSD 26.5% for ALSIand 15.3% for Milestone. Differences in these overall RSD numbers suggest differences in the two methods and/or instruments but not differences attributable to field . samples or SRMs. This would suggest that not only was there no difference in analysis of these samples, but that the preparation procedure for the field samples (see. Section 4.3.1 for description of field sample homogenization) was very thorough and complete. For purposes of this analysis, spiked samples are considered the same as field samples because these were similar field matrices and the resulting variance was expected to be equal to field samples. The replicate sample RSDs also confirm the pre-demonstration results, showing that sam pie homogenization procedures met their originally stated objectives, and that SRM and field sample variation were not signiffcantly different. There also appears to be no significant site variation between Oak Ridge, Puget Sound, and the Carson River site samples. (See Table 6-7 showing average RSDs for each of these sample lots. These average RSDs are computed using only the results of the field samples and not the SRMs.) In addition, there appears to be no difference in precision for differentconcentrations, as noted in the discussion above. Precision Summary The precision of the Milestone field instrument is very comparable to laboratory precision, and within expected precision variation for soil and sediment matrices. The Milestone field instrument can therefore obtain very precise measurements, equivalentto laboratory variation covering the entire range of the instrument (PQL as determined in Section 6.1.1 and an upper limit set by Milestone of 5 mg/kg) as determined during this demonstration. 6.1.4 Time Required for Mercury Measurement During the demonstration, the time required for mercury measurement activities was measured. Specific activities that were timed included: instrument setup, sample analysis, and instrument disassembly. One field technician performed all operations during the demonstration, with the exception of unloading the DMA-80, which measures 80 by 42 by 30 (H) cm and weighs 56 kg. Setup and disassemble times were measured once. Analytical time was measured each day, beginning when the first blank was started, and continuing until the last blank was completed at the end of the day. Any downtime was noted and then subtracted from the total daily operational time. Finally, the total of the operational time from all four days was divided by the total number of analyses performed. For this calculation, analyses of blanks and calibration standards, and reanalyses of. samples were not included in the total number of samples. Setup time for the DMA-80 consisted of removing the instrument from the shipping container, placement on a level working surface, establishment of all electrical and gas tubing connections, and instrument warmup. The time required to remove the DMA-80 from the shipping container could not be measured, because the device was shipped to the site in the back of a vehicle without any packaging. However, based on information provided by the vendor, it is estimated that two people could remove the device from the corrugated cardboard shipping container in less than 5 minutes. Setup time for other peripheral devices, such as the computer/monitor and analytical balance, was also included in the instrument setup time. These two devices were packaged, along with other supplies, in a large corrugated cardboard box. The balance came in two pieces: the base and the top cover. The balance was set up and leveled in 10 minutes. Setup of the computer/monitor took less than 5 minutes. During the demonstration, the DMA-80 was moved to a table on the first and last days of field activities. The vendor required the assistance of one person to perform this task. It is conceivable that one person could perform this operation, but not all individuals would be able to move 47 ------- the large, heavy instrument without assistance. On the second and third days of the demonstration, the vendor operated out of the back of an SUV, and required no assistance in setting up the DMA-80. It is estimated that this activity took 5 minutes on average. After all devices were set in place, and electrical and gas flow connections were made. The DMA-80 was connected to a power source and to the computer/monitor. The balance was also connected to the power source and the computer/monitor. Gas connections were made from the oxygen cylinder, through a pressure regulator (this part was already completed), and then to the DMA-80. A mercury trap (pre-assembled) was inserted in the vent line, which was then attached to the DMA-80. Overall, the electrical and gas flow connections required approximately 10 minutes. However, if the mercury trap had to be assembled and the gas flow regulator installed, as would be the case for most operations, the total setup time is estimated at 20-30 minutes for the first usage. After that the trap can be used for 3 months without reassembly. After setup was complete, the instrument required approximately 20 minutes to come to operating temperature. It is worth noting that setup of the balance was performed during this time period. Overall, the time required to remove the DMA-80 from its shipping container, set up the device, allow the instrument to reach operating temperature, and set up peripheral devices during, instrument warmup is estimated at approximately 30-40 minutes. Individual sample analysis times were not measured forthe duration of the demonstration. Analysis time was . estimated by recording start and stop times each day and accounting for any instrument downtime due to operator breaks or device failure and maintenance activities. Therefore, the total time for analyses included blanks, calibration standards, and any sample reanalyses; however, the total number of analyses performed includes only demonstration samples (samples, spikes, and SRMs), not vendor blanks, calibration standards, or reanalyses. Table 6-8 presents the time measurements recorded for each of the four days of operation of the DMA-80. Table 6-8. Time Measurements for Milestone Day Day 1 Day 2 Day 3 Day 4 Total Run Time (minutes) 260 590 410 70 1330 Instrument disassembly was measured from the time that sample or blank analyses ended until the instrument was disassembled and placed in the original shipping container. During the demonstration, the balance was disassembled and packaged while the final samples were being analyzed (an advantage of an auto-sampler). This complete process took about 5 minutes. The DMA-80 was not re-packaged because it was not brought to the site in a shipping container. Disassembly of the DMA-80 involved turning off power, disconnecting the power source and interface cables to the computer/monitor, removal of the auto-sampler tray, and disconnecting the oxygen supply. This process required 15 minutes to complete. Packaging would require that the DMA-80 be placed in a custom shipping container with re- enforced corners and buffer spaces. The auto-sampler tray, cables, gas tubing, weigh boats, and any other supplies would need to be packaged also. Finally, the oxygen cylinder would need to be disassembled by closing the main valve, bleeding off any pressure in the line, and removing the plastic tubing and pressure regulator. It is estimated that this complete process would take approximately 30 minutes, not including the time to return the oxygen cylinder to the supplier. Analysis Time Summary In total, Milestone analyzed 173 samples during the demonstration. The turnaround time on individual sample analyses was 5 minutes. However, using the total analytical time reported in Table 6-2 (1330 minutes), 7.7 minutes per analysis is a better approximation of real world operating conditions. It should be noted thatthe number of analyses does not include blanks, standards, and reanalyzed samples. These numbers will vary from site to site depending on project goals (e.g., are "greater than" results acceptable, or m ust all samples be quantified) and sample demands (e.g., high concentration samples or very heterogeneous samples). If project goals require all samples to be quantified, the number of reanalyses and blanks required could be higherand, therefore,the time per analysis could be greater. On the other hand, if sample results can be reported as "greater than" values (as was generally done during the demonstration), then 8 minutes per analysis is a reasonable average time. 6.1.5 Cost Background information, assumptions used in the cost analysis, demonstration results, and a cost estimate are provided in Chapter 7. 48 ------- 6.2 Secondary Objectives This section discusses the performance results for the DMA-80 in terms of secondary objectives described in Section 4.1. These secondary objectives were addressed based on observations of the DMA-80 and information provided by Milestone. 6.2.1 Ease of Use Documents the ease of use, as well as the skills and training required to properly operate the device.' Based on observations made during the demonstration, the DMA-80 is easy to operate, requiring one field technician with a basic knowledge of chemistry acquired on the job or in a university and training on the DMA-80. The vendor provided an SOP, entitled "Getting Started: Calibration and Analysis Procedures," for use with the DMA-80 (see Appendix B). This procedure was evaluated during the demonstration. The procedure was generally easy to understand. SOP Section 1.0, Calibration, could not be evaluated because the vendor performed equipment calibration in the office, prior to shipping the instrument to the field. The vendor performed daily calibration checks as recommended in Section 2.0 of the SOP. Calibration involved weighing a small amount of a standard in a weigh boat, placing it in the auto-sampler, and processing the standard sample through the DMA-80. Figure 6-3 presents an example of typical calibration results. Figure 6-4 shows example 3-point calibration curves for the two cuvettes. The instruction on calibration checks and analyses was clear. Combined with instrument training, this SOP would provide a user with adequate direction on basic use of the DMA-80. Included were instructions on calibrating the instrument and running blanks before processing samples. In addition, Milestone provides a 1-day training course (at the purchaser's cost) and telephone support at no cost to anyone who purchases the DMA-80. Neither of these was evaluated during the demonstration. Current Sample: 1 Editor) System Documentation | Program j Graphic | Calibration | Setup | Nairn: [Test 16/1/2002 Data: [16/01/02iv Bal»nc» S»tup Operator. I Suitrtd mean : 0.00 |ig/kg «d: (UK) pg/kg Sample Sample ID Weight lal Height Hg Rssull Calibration Factor Remarks a 1 Cleaning 1 Blank 1 Blank 1 Blank 1 StdSOng 1 Sid Sing 1 Std 50 ng J Sid 100 ng 0.0001 0-1000 0.1000 0.1WO 0,1000 0.1000 0,1000 0.1000 0.0102 0-0038 0.0023 0.0019 .0.1257 01265 0.1293 0~25U 0.00 am 000 0.00 500 5.00 ,-.500. "moo 0.00 0.00 0.00 000 50.00 5000 50.00 100.00 1 Sid 100 ng 0.1000 0.2B15 10.00 100.00 V Cleaning 0.0001 0.0078 aoo 0:00 1 StdlOOng Sid 200 ng Sid 200 ng [d200ng 0.1000 0.1000 0.1000 0.1000 02518 0.5067 0.4971 10.00 moo 2000 2000 100.00 200.00 200.00 200.00 + A Export Figure 6-3. Calibration result screen. 49 ------- Editor I System I Documentation | Program | Graphic Calibration | Setup | *° Calibration Curve Cuvette 1 Calibration Curve Cuvette 2 500 600 700 r»nnrt L NAME •ding Ha 000 •0:00. :aooi 0.0013' 0.001* 0:0022- i;o:o; 20.00 20.00; 0,23-10- 0.'23S7; 0.4563 0.4654; 0:4679; \Cfrette- |j£ Figure 6-4. Calibration curve screen. Items not covered in the SOP were trouble shooting and maintenance. For example, during the demonstration, the auto sampler jammed and the pneumatic sample insertion arm required realignment twice. Neither of these maintenance items was discussed in the SOP. In addition, there are two crucial operational elements that are not addressed in the SOP. The first is the selection of sample size such that the results will be within the calibration range. Selection of sample size requires an estimate of the expected mercury concentration. This problem is not unique to the DMA-80; any AAS instrument requires an estimate of sample concentration in order to get sample results within a specified calibration, range. Second, no information was provided on how to handle samples that were outside of the calibration range. Procedures implemented during the demonstration included running a blank sample after a sample was above the calibration range (to purge the system of mercury) and reducing sample size on subsequent reanalyses (if quantitative results are reported). These procedures were not described in the SOP; however, the software prompted the analyst to run a clean-out blank. It is not known whether these procedures are covered in the vendor training course. Milestone chose to operate the DMA-80 with one chemist during the demonstration. The chemist held a B.S. degree in chemistry. Milestone claimed that a laboratory or field technician with a high school diploma and basic computer knowledge could operate the equipment after a 1-day training course. Field observations supported this claim. Most operations required either use of a keyboard or mouse with a Microsoft Windows-based system, or alternatively, the use of a touch screen with icons. The prompts and icons were clear and easy to understand. The input screen includes a table which is pre-numbered to correspond to the auto sampler slots. The user enters the sample number in the first column and the sample identification or description in a second column. The operator then performs the sample weighing step and enters the sample weight or the sample weight can be automatically transferred from appropriate balances. A status column shows the sample analysis status. 50 ------- The operator was able to perform sample preparation and analysis on a continuous basis. Sample preparation took approximatelyoneminutepersample. Sample preparation consisted of mixing samples in the original container using a clean stainless steel spatula. A clean weigh boat was placed on the balance (not part of the system, but can be provided), the balance was zeroed, the weigh boat was removed from the balance, and a small amount of sample was placed in the weigh boat. The weigh boat and sample were placed on the balance again. The net weight was displayed on the digital balance and within the input screen for the DMA-80. When the weight stabilized, the operator input the weight by touching a screen icon for the scale. This operation was easy to understand and could be performed by a trained technician. Sample analysis took 7.7 minutes persample, on average. Because sample analysis was automated, sample preparation of additional samples continued during sample analysis of previous samples. Typically, three to four samples were prepared during the time it took to perform an analysis, allowing time for observation of equipment performance. Sample analysis consisted of placing the pre-weighed sample boat in the proper slot on the auto sampler. The slot number corresponded to the number in the input screen. The samples were then automatically advanced as samples were processed. The auto sampler picked up the sample boat and inserted it into the furnace opening. When sample analysis was completed, the sample boat was automatically removed from the furnace and placed back on the auto sampler. The auto sampler then advanced the next sample for analysis. Because this process was automated, it was extremely easy to use. The only potential difficulty was ensuring that the sample boat was placed in the appropriate auto sampler slotso that the results matched with the proper sample number. As with sample preparation, sample analysis was easy to understand and could be performed by a trained technician. As samples were analyzed, vendor-proprietary software screens allowed the userto track the approximate location of the sample mercury in a graphic display of the analyzer furnace, amalgamator, and photo cell (Figure 6-5). Current Sample; 1 Tim*: 00:00:00 STOP Editor System | Documentation | Program j Graphic | Calibration | Setup j DESEES (Heating Program - START |®STOP DMA Status SPECTRO- FURNACE 4 FURNACE 3 FURNACE! FURNACE 1 AUTOSAMPLER ASS CELLS ATOMIZER ' CATAtVST COMBUSTION \ Cylinder. II Sylinderl Detector Shutter Hg Lamp / HEATING CONTROL . j _ OK _____ 3111 Cylinder III Figure 6-5. System control display screen. 51 ------- During the demonstration, samples with concentrations outside of the equipment calibration/operation range were encountered. These samples would result ,in a peak outside of the calibration range. The vendor software flagged these samples (red "X" instead of a green V" for "in range" samples) and prompted the user to run a blank to demonstrate that excess mercury had been purged from the system. The software messages were clear and easy to follow. Another screen presented a graph of the adsorption peak. Figure 6-6 shows a representative peak for a sample that was "in range." (1) Tmax: 7.50 me (1) W0t.S):2.37nc (2) Tmax: 17.50 ssc (2) W(0.5):3.5B n.c Blank-.OJlOOO 01 020304 05 0607 lo^O* 10 11 13 13 14 -15.18 I7|t8 19 20.21 ?2 2324 2S26 2728.2S30 31 333334 35 36 37 », 39,40*1 « 43 Figure 6-6. Sample peak screen. The digital balance was the major peripheral item. The vendor will supply a balance with cables for direct input into thesystem monitor/software, orthe usercan supply his/her. own balance. Though the balance is not part of the required vendor equipment, a balance is a necessary peripheral. Therefore, the balance was evaluated during the demonstration. The reader should note that other brands and models of balances may be used and these may not perform in the same manner as the balance used during the demonstration. The interface of the balance with the monitor/software was seamless. Overall, the balance was easy to use in conjunction with the DMA-80. According to the vendor, the sample weight is currently stored in a second database, requiring re-entry of the data into the main sample database with the corresponding potential for data entry errors. The vendor claims that a new edition of the software will eliminate the need to re-enter sample weights into the sample data base. 6.2.2 Health and Safety Concerns Documents potential health and safety concerns associated with operating the device. No significant health and safety concerns were noted during the demonstration. The only potential health and safety concerns identified were the generation of mercury vapors and the use of oxygen as the carrier gas. The vendor recommends and can provide a mercury filter; oxygen can be safely handled using standard laboratory procedures. 52 ------- Health and safety concerns, including chemical hazards, radiation sources, electrical shock, explosion, and mechanical hazards were evaluated. No chemicals were used in the preparation or processing of samples, except for analytical standards. During this demonstration, the analytical standards were soil SRMs for mercury. These were handled with gloves and the operator wore safety glasses with side shields at all times. Such standard laboratory precautions mitigate the potential for dermal exposure. Similar procedures were also used for soil samples which contained mercury. Because the DMA-80 is designed to thermally convert mercury compounds to mercury vapors as part of the analytical process, and no fume hood was present to exhaust mercury vapors after analysis, inhalation of mercury was a concern. The vendor installed a mercury trap, containing potassium permanganate, in the exhaust line from the DMA-80. Measurements were taken with a Jerome 431-x gold film mercury vaporanalyzer, manufactured by Arizona Instruments Corporation. The instrument has a range of 0.000 to 0.999 mg/m3. In all cases, readings were 0.000 mg/m3 in the breathing zone of the operator. In evaluating electrical shock potential, two factors were evaluated: 1) obvious areas where electrical wires are exposed and 2) safety certifications. No exposed wires were noted during the demonstration. All connections between equipment were made using standard electrical power cords, modem interface lines, and 9-pin cords. Power cords were grounded with ground fault interrupters, and a surge protector was utilized. The DMA-80 was not UL certified, but did have CE certification; no other safety certifications were marked on the transformer. No obvious explosion hazards were noted. The use of oxygen as a carrier gas does present the possibility of explosion in the presence of ignition sources; however, implementation of good laboratory safety practices can mitigate any such hazard. The cylinder needs to be secured both when in use and when not in use. When not in use, the cylinder should be disconnected from the DMA-80 and the cap replaced to prevent damage to the. cylinder valve. The cylinder was clearly marked as oxygen and the appropriate hazard label was present. No serious mechanical hazards were noted during the demonstration. All equipment edges were smooth, minimizing any chance of cuts or scrapes. The hinged lid on the DMA-80 presents the possibility of a pinch hazard, as would any hinged device; however, the lid is not overly heavy, does not need to be routinely opened, and is designed to remain securely in place when the lid is open. 6.2.3 Portability of the Device Documents the portability of the device. The DMA-80 was not easily portable (by hand) due to its size and weight. It was easy to set up and can be taken anywhere that a small van or SUV can go. The instrument is better characterized as mobile rather than field portable. The DMA-80 measured 80 cm (L) by 43 cm (W) by 30 cm (H). The weight was estimated at 45 kg. Also included as a standard feature with the DMA-80 were a controller with monitor and a keyboard; both were light weight and easily portable. The controller measured approximately 38 cm (L) by 23 cm (W) by 22 cm inches high. The one negative aspect of the DMA-80, with respect to portability, was its size and weight. This equipment required the assistance of oneSAIC person to unload from • the transport vehicle to the table used during the demonstration. It should be noted that the DMA-80 can be used out of the back of the vehicle (SUV), and, in fact, was used this way on the second and third days of the demonstration. This device may be better characterized as a "mobile" instrument rather than "field portable". The device is not hand held, and can not be easily moved by hand from one location to another. That said, the DMA-80 can certainly be transported to any place that a small van or SUV can go and would be practical for most field applications. The balance required a flat, stable surface. Because the width of the DMA-80 prevented placement of the balance in the SUV, a table was required. The vendor utilized a marble slab on the table to provide extra stability and simplify leveling the balance. A marble slab is not a part of the standard equipment supplied with the DMA-80. A flat surface area is also required for staging samples while filling weigh boats. The DMA-80 is not equipped with a battery. Operation of the instrument requires a standard electrical source of 110 volts. The vendor asserts that the DMA-80 can be powered by a generator, although this was not evaluated during the demonstration. For the demonstration, the vendor was supplied with a folding table, two chairs, and a tent to provide shelter from inclement weather. In addition, one 1-gallon container 53 ------- each was provided for waste soil and decontamination water utilized to clean weigh boats. A 2-gallon zip-lock bag was furnished fordisposalof used gloves, wipes, and other wasteswhich were contaminated during the demonstration. Finally, a large trash bag was supplied for disposal of non- contaminated wastes. 6.2.4 Instrument Durability Evaluates the durability of the device based on its materials of construction and engineering design. The DM A-80 was well designed and constructed for durability. The outside of the DMA-80 is constructed of sturdy stainless steel. Parts were securely connected with screws and lock washers. The top of the device could be opened to access inner components. The lid was secured with stainless steel hinges. No environmental (e.g., corrosion) or mechanical (e.g., shear stress or impact) tests were performed; however, the outer shell of the instrument appeared to be well-designed and constructed, indicating that the device would likely be durable under field conditions. No evaluation could be made regarding the long-term durability of the furnace, analytical cell, or circuitry. Visual inspection did not indicate that any problems were likely. The vendor offers a standard 1-year warranty a-nd will provide an extended warranty and maintenance plan at the owner's cost. Minor problems were identified during the demonstration with two moving parts on the auto sampler component of the system. During the first day of the demonstration, the vendor adjusted the alignment of the pneumatic arm used to take sample boats from the auto sampler and insert them into the furnace. The vendor explained that this alignment is frequently required after the instrument has been shipped. This alignment took less than 10 minutes to accomplish. During the second day of the demonstration, the 23rd sample was dropped by the auto sampler. The vendor indicated that this was due to minor alignment problems with the pneumatic arm. This resulted in approximately 5 minutes of downtime. Later that same day, the auto sample manifold jammed, causing the loss of one sample in the queue. (It should be noted that there was additional sample available, and a replacement weigh boat was prepared while the instrument ran other samples, resulting in no net downtime due to sample loss.) The auto sampler jam resulted in approximately 5 to 7 minutes of downtime. Pneumatic pressure was released by the operator closing the oxygen tank and bleeding oxygen from the system. The auto sample manifold was then disengaged (it had been bent slightly by the pneumatic pressure when it jammed). The pneumatic arm was re-aligned by loosening set screws, aligning the arm, and resetting the set screws. The auto sampler manifold was reinserted, oxygen pressure reestablished, and the system operation tested with a blank. A new sample was running approximately 7 minutes after the jam originally occurred. Finally, most of the demonstration was performed during rainfall events ranging from steady to torrential. The DMA-80 was located either under a tent (Days 1 and 4) or in the back of the SUV (Days 2 and 3). Even when it was not raining, the relative humidity was high, ranging from 70.6 to 98.3 percent. The high humidity and rainfall had no apparent impact on the reliability of the instrument operation". 6.2.5 A vailabijity of Vendor Instruments and Supplies Documents the availability of the device and spare parts. The DMA-80 is readily available for lease, or purchase. DMA-80 rental is available on a limited basis. Spare parts and consumable supplies can be added to the original DMA-80 order or can be received within 24 to 48 hours of order placement. Supplies and standards not provided by Milestone are readily available from laboratory supply firms. faa^m^Ht^mism^fmti^muutatuteiiiiittm. EPA representatives contacted Milestone regarding the availability of the DMA-80 and supplies. Milestone asserted that95 percent of its current business is purchase or long-term lease arrangement. According to Milestone, such systems are available within 3 to 4 weeks of order placement, but can be expedited with a minimum 2-week turnaround. The DMA-80 also is available for rental on a limited basis (special requests). There is only one unit in the rental pool, so lead time is subject to availability. The instrument comes standard with 40 weigh boats and a complete set of consumable items (catalyst, 54 ------- amalgamator, and o-rings) installed in the instrument so Other supplies and standards, not provided by Milestone, that the instrument is fully operable upon receipt. Spare can be purchased from a laboratory supply firm. Typical consumable items are available as part of a consumables delivery times, per Milestone, for most supplies will range kit or can be ordered individually. These and any other from 1 day (with express delivery) to less than one week. parts are available within 24-48 hours. Cost for capital equipment and supplies are discussed in Chapter 7. 55 ------- Chapter 7 Economic Analysis The purpose of the economic analysis was to estimate the total cost of mercury measurement at a hypothetical site. The cost per analysis was estimated; however, because the cost per analysis would decrease as the number of samples analyzed increased, the total capital cost was also estimated and reported. Because unit analytical costs are dependent upon the total number of analyses, no attempt was made to compare the cost of field analyses with the DMA-80 to the costs associated with the referee laboratory. "Typical" unit cost results, gathered from analytical laboratories, were reported to provide a context in which to review DMA-80 costs. No attempt was made to make a direct comparison between these costs for different methods because of differences in sample throughput, overhead factors, total equipment utilization factors, and other issues that make a head-to-head comparison impractical. This Chapter describes the issues and assumptions involved in the economic analysis, presents the costs associated with field use of the DMA-80, and presents a cost summary for a "typical" laboratory performing sample analyses using the reference method. 7.1 Issues and Assumptions Several factors can affect mercury measurement costs. Wherever possible in this Chapter, these factors are identified in such a way that decision-makers can independently complete a project-specific economic analysis. Milestone offers three options for potential DMA-80 users: 1) purchase of the instrument, 2) monthly rental (with a 3-month minimum), and 3) equipment leasing with an option to purchase at the end of 24 months. Because site and user requirements vary significantly, all three of these options are discussed to provide each user with the information to make a case-by-case decision. A more detailed cost analysis was performed on the equipment rental option for three months or less because this case represents the most frequently encountered field scenario. The results of that cost analysis are provided in section 7.2. 7.1.1 Capital Equipment Cost The DMA-80 (the analytical instrument) comes complete with a 40-position auto sampler; Pentium Controller with keyboard, mouse, and touch screen monitor; Windows™ based software; and a set of stainless steel weigh boats, whether the instrument is purchased, rented, or leased. A portable computer may be substituted for the controller at the user's request. An optional digital balance, with sensitivity to 0.1 mg, is available for purchase from Milestone (no rental or leasing), but not included in the base cost of any of these three options. Alternatively, the user may provide his/her own balance. The vendor claims thatvirtually any balance adaptable to a 9-pin sub-D socket (female) interface cable will com m unicate with the DMA-80 (Milestone, 2003). This claim was not evaluated during the demonstration. A printer can also be purchased as an option from Milestone; no lease agreement or rental is available for the printer. Per the vendor, any Windows compatible printer can be used. The cost quoted by Milestone does not include packaging or freight costs to ship the instrument to the user location. A 1-month, non-refundable deposit is required for rental and lease agreements. The deposit is not applied to payments. A user manual is provided at no cost. An 8-hour training session is available for an additional fee. 56 ------- 7.1.2 Cost of Supplies The cost of supplies was estimated based on the supplies required to analyze demonstration samples, and based on discussions with Milestone. Requirements vary, depending upon whether solid or liquid samples are being analyzed. For purposes of this cost estimate, only supplies required to analyze solid samples are factored into the cost estimate. Supplies required for liquid samples are noted, and approximate prices provided, but those costs are not incorporated into the overall cost estimate because liquid samples were not analyzed during the demonstration. Supplies consisted of consumable items (e.g., standards and compressed oxygen)and non-consumablesthatcould not be returned because they were contaminated or the remainderof a set. Non-consumable supplies consisted of a set of 3 micro-spatulas (for solid samples). Consumable supplies consisted of: Adjustable micro-pipettes (for liquid samples) • Housing and tubing for the mercury trap Calibration standards Compressed oxygen (welding grade) • Potassium permanganate for mercury trap • Glass wool for the mercury trap Silica gel for dilution of high-concentration samples The purchase prices and supply sources were obtained from Milestone. Because theusercannot return unused or remaining portions of supplies, no salvage value was included in the cost of supplies. Personal protective equipment (PPE) supplies were assumed to be part of the overall site investigation or remediation costs; therefore, no PPE costs were included as supplies. During the demonstration, high-concentration samples generally were not quantified; they were usually reported as "greater than" values. In cases where the user wants to quantify high- concentration samples, a dilution material is needed. The vendor recommends silica gel. Even-though silica gel was not used during the demonstration, it could have been used for high-concentration samples. (Milestone made the decision to not use silica gel during the demonstration and therefore this additional variation was not evaluated as part of overall instrument accuracy and precision.) Such samples are likely to be encountered at most other sites. Therefore, the cost was estimated based on the assumption that 25 percent of samples may have to be diluted. 7.1.3 Support Equipment Cost During the demonstration, the DMA-80, controller, and balance were operated using AC power. The costs associated with providing the power supply and electrical energy were not included in the economic analysis; the demonstration site provided AC power at no cost. None of the items mentioned above can operate on DC power, although a portable generator can be used to power the equipment. Because of the large number of samples expected to be analyzed during the demonstration, EPA provided support equipment, including tables and chairs, for the field technician's comfort. In addition, the EPA provided a tent to ensure that there were no delays in the project due to inclement weather. These costs may not be incurred in all cases; however, such equipment is frequently needed in field situations, so these costs were included in the overall cost analysis. 7.14 Labor Cost The labor cost was estimated based on the time required for DMA-80 setup, sample preparation, sample analysis, summary data preparation, and instrument packaging at the end of the day. Setup time covered the time required to take the instrument out of-its packaging, setup all components, and ready the device for operation. However, the DMA-80 was brought to the site in a vehicle and was not in an original shipping container. Therefore, this time was estimated ratherthan measured. Sample preparation involved mixing sam pies with a micro-spatula. Other than the first couple of samples, sample preparation was easily completed while previous samples were being analyzed. Sample analysis was the time required to analyze all samples and submit a data summary. The data summary was strictly a tabulation of results in whatever form the vendor chose to provide. In this case, the vendor transcribed results from the electronicdatabase to the fie Id chain of custody forms (no printer was available in the field). The time required to perform all tasks was rounded to the nearest 5 minutes; however, for the economic analysis, times were rounded to the nearest hour and it was assumed that a field technician who had worked for a fraction of a day would be paid for an entire 8-hour day. Based on this assumption, a daily rate for a field technician was used in the analysis. During the demonstration, EPA representatives evaluated the skill level required for the field technician to analyze and report results for mercury samples. Based on these field observations, a field technician with basic chemistry skills acquired on the job or in a university setting, and a 1-day training course specific to the DMA-80, was considered qualified to operate the instrument. For the economic analysis, an hourly rate of $15 was used for a 57 ------- field technician. A multiplication factor of 2.5 was applied to labor costs to account for overhead costs. Based on this hourly rate and multiplication factor, and an 8-hour day, a daily rate of $300 was used for the economic analysis. Monthly labor rates are based on the assumption of an average of 21 work days per month. This assumes 365 days per year, and non work days totaling 113 days per year (104 weekend days and 9 holidays; vacation days are discounted assuming vacations will be scheduled around short-term work or staff will be rotated during long projects). Therefore, 252 total annual work days are assumed. 7.1.5 Investigation-Derived Waste Disposal Cost Milestone was instructed to segregate its waste into four categories during the demonstration: 1) general trash; 2) lightly contaminated PPE and wipes; 3) contaminated soil (both analyzed and unanalyzed) and other highly contaminated wastes; and 4) wash water used for cleaning micro spatulas and weigh boats. General trash was not included as IDW and is not discussed in this document. Lightly contaminated wastes consisted primarily of used surgical gloves and wipes. " The surgical gloves were discarded for one of three reasons: 1) they posed a risk of cross contamination (noticeably soiled), 2) they posed a potential health and safety risk (holes or tears), or 3) the operator needed to perform other tasks (e.g., using cell phone to provide customer support). The rate of waste generation was in excess of what would be expected in a typical application of this instrument. In addition, the EPA evaluators occasionally contributed used gloves to this waste accumulation point. Wipes were used primarily to clean weigh boats and micro spatulas between samples. In cases where cross contamination is not a major concern (e.g., field screening or all samples are in the same concentration range), lesser amounts of waste would likely be generated. Contaminated soils consisted primarily of soil placed in the weigh boat and then removed because the weight was above the target weight. Soil mass that was analyzed was also placed in this waste container as a precaution. It is expected that such soils would be free of mercury after being heated to high temperatures in the analytical instrument. In some cases, these sample residuals may not need to be handled as hazardous waste. Finally, the vendor generated small amounts of waste water by cleaning weigh boats and micro spatulas. Weigh boats are considered clean after the completion of analyses due to high temperatures in the analytical instrument. Therefore, weigh boats were not washed after analyses were completed; however, during the demonstration, the vendor was required to wash weigh boats for samples that were not analyzed (e.g., part of the sample spilled in placing the weigh boat on the auto sampler. The boats were rinsed with water and dried with a clean wipe to prevent potential cross contamination of low concentration samples. The waste water, contaminated soil, excess sample material, and lightly contaminated gloves and wipes were considered hazardous wastes for purposes of this cost analysis. 7.1.6 Costs Not Included Items for which costs were not included in the economic analysis are discussed in the following subsections, along with the rationale for exclusion of each. Oversight of Sample Analysis Activities. A typical user of the DMA-80 would not be required to pay for customer oversight of sample analysis. EPA representatives observed and documented all activities associated with sample analysis during the demonstration. Costs for this oversight were not included in the economic analysis because they were project specific. For the same reason, costs for EPA oversight of the referee laboratory were also not included'in the analysis. Travel and Per Diem for Field Technician. Field technicians may be available locally. Because the availability of field technicians is primarily a function of the location of the project site, travel and per diem costs for field technicians were not included in the economic analysis. Sample Collection and Management. Costs for sample collection and management activities, including sample homogenization and labeling, are site specific and, therefore, not included in the economic analysis. Furthermore, these activities were not dependent upon the selected reference method or field analytical tool. Likewise, sample shipping, COC activities, preservation of samples, and distribution of samples were specific requirements of this project that applied to all vendor technologies and may vary from site to site. None of these costs was included in the economic analysis. Items Costing Less than $10. The costs of inexpensive items, such as paper towels, were not included in the economic analysis. 58 ------- Documentation Supplies. The costs for digital cameras used to document field activities were not included in project costs. These were considered project-specific costs that would not be needed in all cases. In addition, these items can be used for multiple projects. Similarly, the cost of supplies (logbooks, copies, etc.) used to document field activities was not included in the analysis because they are project specific. Health and Safety Equipment. Costs for rental of the mercury vapor analyzer and the purchase of PPE were considered site specific and, therefore, were not included as costs in the economic analysis. Safety glasses and disposable gloves were required for sample handlers and would likely be required in most cases. However, these costs are not specific to any one vendor or technology. As a result, these costs were not included in the economic analysis. Mobilization and Demobilization. Costs for mobilization and demobilization were considered site specific, and not factored into the economic analysis. Mobilization and demobilization costs actually impact laboratory analysis more than field analysis. When a field economic analysis is performed, it may be possible to perform a single mobilization and demobilization. During cleanup or remediation activities, several mobilizations, demobilizations, and associated downtime costs may be necessary when an off-site laboratory is used because of the wait for analytical results. purchase. Also shown are estimated costs for an optional printer and analytical balance. Figure 7-1 shows the relative costs for the basic capital equipment. These costs reflect the basic DMA-80 system (with standard auto sampler) and the controller/monitor. No options (e.g., balance or printer) and no supply or shipping costs are included. As would be expected, this chart clearly shows that leasing is the most cost-effective option (in terms of capital costs), followed by rental, for short-term projects. As project duration (or use on multiple projects) approaches two years, the purchase option is the most cost-effective. These scenarios cover only capital cost, not the cost of optional or user-supplied equipment, supplies, support equipment, labor, and IDW disposal. 1234- Months Purchase Rental Lease 7.2 DMA-80 Costs This subsection presents information on the individual costs of capital equipment, supplies, support equipment, labor, and IDW disposal for the DMA-80. 7.2.1 Capital Equipment Cost During the demonstration, the DMA-80 was operated for approximately 3 days and was used to analyze 173 samples. Table 7-1 summarizes the DMA-80 capital costs for the three procurement options: rental, lease, and Figure 7-1. Capital equipment costs. The DMA-80 sells for $30,000, including the 40-position auto-sampler, the controller (with mouse, keyboard, display monitor, and software), and related electrical connections. Also included are: • 40 stainless steel weigh boats • Plastic tubing for compressed oxygen connections An instruction manual 59 ------- Table 7-1. Capital Cost Summary for the DMA-80 Item Quantity Unit Cost ($) ' a b c Purchase DMA-80 1 Monthly Rental of DMA-80 " 1 Monthly Lease of DMA-80 " 1 Purchase Balance (Optional) c 1 Purchase Printer (Optional) ° 1 • $30,000 $3,000 $1,450 $1,500 $150 1 -Month $30,000 $9,000 $1,450 $1,500 $150 Ten percent of purchase price with a three month minimum. $1 ,450 per month (24-month lease with $1 buyout). A balance is required, but may be provided by the user. A printer is optional; Total Cost for Selected Project Duration 3-Month $30,000 $9,000 $4,350 $1,500 $150 it may also 6-Month $30,000 $18,000 $8,700 $1,500 $150 be provided by the 12-Month $30,000 $36,000 $17,400 $1,500 $150 user. 24-Month $30,000 $72,000 $34,800 $1,500 $150 These items are considered supplies and are discussed in Subsection 7.2.2. Compressed oxygen is required, but must be obtained from a local supplier, along with the appropriate regulator and cylinder mounting brackets (see Subsection 7.2.2). A balance is also required and can be purchased (no rental or lease) from Milestone for $2950. A balance can be purchased from a laboratory supply company for approximately $1500 to $2,500, depending upon model (wwwl .Fishersci.com, 2003). The lowestcost, $1500, was used in this cost analysis. Alternatively, the user can supply a balance with a 9-pin connector to interface with the DMA-80 (Milestone, 2003). The costs presented in Figure 7-1 do not reflect the cost of the balance because it is optional equipment and can be provided by the user (it may already be owned). A printer. can be purchased for approximately $150; however, as with the balance, no printer costs are included in the cost analysis (or Figure 7-1) because this equipment is optional and may be supplied or already owned by the user. . Balance and printer costs are shown in Table 7-1. 7.2.2 Cost of Supplies Supplies used during the demonstration included solid SRMs, compressed oxygen, micro spatulas, and a mercury trap. NIST soil SRMs sell for $250 each; typically both a high and a low standard will be required for many applications, for a total cost of $500. If sediments are analyzed, a NIST sediment SRM may be obtained for $150. . No costs for a sediment SRM are included in this analysis. These standards have a life-expectancy of one to three years (one year is assumed for this cost analysis). Welding grade compressed oxygen is used as a carrier gas for the DMA-80. It can be obtained from a local source and prices will vary. For this cost analysis, a price of $0.04/L was used. An 80 ft3 (2,265 L) cylinder'will last for approximately 19 days, assuming 10 hours of constant operation of the DMA-80 at a flow rate of 200 mL/min. A regulator is required to reduce the flow rate to 200 mL/min. Purchase of a flow regulator and cylinder brackets is estimated at $200, which is a one-time cost that can be spread over the entire term for longer projects. Alternatively, the cost can be included with the cylinder rental cost as was done for this analysis. Table 7.2 summarizes the costs for the carrier gas, assuming the same number of samples are run per day during each period and a total cost of S0.04/L, which equals $5/day of operation. The rental cost for a mounting bracket and pressure regulator is included in the S0.04/L. Table 7-2. Carrier Gas Cost Summary Months Item Flow Regulator Mounting Bracket Oxygen Total Cost 1 NA NA $105 $105 3 NA NA $315 $315 6 NA NA $630 $630 12 NA NA $1,260 $1,260 24 NA NA $2,520 $2,520 A mercury trap was also required during the demonstration and would likely be needed for most field applications. The trap consisted of a polyethylene drying tube (Fisher Scientific # SN:09-242C or equivalent), glass wool (Fisher Scientific # SN:11-390 or equivalent), and potassium permanganate as the reactive ingredient to "trap" mercury. The polyethylene drying tube costs $20 for a package of 12; it is assumed that this will provide a sufficient supply for 60 ------- up to 2 years. The glass wool costs $35 for 454 g, enough for approximately 100 traps. Therefore, the cost is approximately $0.35 per trap. The glass wool should be changed out every 3 months, so annual costs are $1.40. Potassium permanganate costs $60 for 500 g (enough for 6 traps), or $10 pertrap. Annual costs are $40. Total trap costs are presented in Table 7-3. Table 7-3. Mercury Trap Costs Item Drying Tube Glass Wool KMnO4 Total 1 $20 $35 $10 $65 3 $20 $35 $10 $65 Months 6 $20 $35 $20 $75 12 $20 $35 $40 $95 24 $20 $35 $80 $135 Two to three micro spatulas are normally required to prevent cross contamination and allow time for cleaning. A set of three micro spatulas costs $10, and would be expected to last at least two years. 7.2.3 Support Equipment Cost Milestone was provided with a 10x1 Ofoottentfor protection from inclement weather during the demonstration. It was also provided with one table and two chairs for use during sample preparation and analytical activities. The rental cost for the tent (including detachable sides, ropes, poles, and pegs) was $270 per week. The rental cost for the table and two chairs for one week totaled $6. Total support equipment costs were $276 per week for rental. For longer projects, purchase of support equipmentshould be considered. Two .folding chairs would cost approximately $40. A 10x10 foot tent would cost between $260 and $1,000, depending on the construction materials and the need for sidewalls and other accessories (e.g., sand stakes, counter weights, storage bag, etc.). A cost of $800 was used for this cost analysis. A fold ing table would cost between $80 and $250, depending on the supplier. For purposes of this cost analysis, $160 was used. Total purchase costs'for support equipment are estimated at $1,000. 7.2.4 Labor Cost One field technician was required for 3 days during the demonstration to complete sample analyses and prepare a data summary. Based on a labor rate of $300 per day, total labor cost for application of the DMA-80 was $900 for the 3-day period. Labor costs assume qualified technicians are available locally, and that no hotel or per diem costs are applicable. Table 7-4 summarizes labor costs for various operational periods. The costs presented do not include supervision and quality assurance because these would be associated with use of any analytical instrument. and are a portion of the overhead multiplier built into the labor rate. Table 7-4. Labor Costs Item 1 . Months 6 12 24 Technician $6,300 $18,900 $37,800 $75,600 $151,200 Supervisor NA NA NA NA NA NA NA NA NA NA Quality Control Total $6,300 $18,900 $37,800 $75,600 $151,200 7.2.5 Investigation-Derived Waste Disposal Cost Milestone generated PPE waste, decontaminate solution waste, and excess soil waste. The PPE waste was charged to the overall project due to project constraints. The minimum waste volume is a 5-gallon container. Mobilization and container drop-off fees were $1,040; a 5- gallon soil waste drum was $400, and a 5-gallon liquid waste drum was $400. (These costs were based on a listed waste stream with hazardous waste number U151). The total IDW disposal cost was $1,840. These costs may vary significantly from site to site, depending on whether the waste is classified as hazardous or nonhazardous and whetherexcess sample materialis generated that requires disposal. Table 7-5 presents IDW costs for various operational periods, assuming that waste generation rates were similar to those encountered during the demonstration. 7.2.6 Summary of DMA-80 Costs The total cost for performing mercury analysis is summarized in Table 7-6. This table reflects costs for projects ranging from one to 24 months. The rental option was used for estimating the equipment cost. However, because the minimum rental for the DMA-80 is 3 months, the total cost is inflated by the high capital cost. Additionally, capital costs for rental exceed those for 61 ------- purchase at approximately 10 months, so rental is no longer as cost-effective for projects exceeding this duration. Finally, a lease agreement may be a cost- effective alternative as compared to either rental or purchase for projects lasting less than 21 months. Atthat point, equipment purchase may be more-cost-effective; however, the decision on which purchase option to utilize should be made on a case-by-case, basis. Table 7-5. Item Drop Fee Disposal Total IDW Costs 1 $1.040 $400 $1,440 3 $3,120 $1,200 $4,320 Months 6 $6,240 $2,400 $8,640 12 $12,480 $4,800 $17,280 24 $24,960 $9,600 $34,560 Table 7-6. Summary of Rental Costs for the DMA-80 Item Capital Equipment Monthly Rental of DMA-80 Supplies Micro Spatula (set of 3) ". Solid SRM ° Mercury Trap (all components) Compressed Oxygen d Total Supply Cost Support Equipment " Table (optional) - weekly Chairs (optional) - weekly Tent (for inclement weather only) - weekly Total Support Equipment Cost Labor Field Technician (person day) IDW Drop Fee Disposal Total IDW Costs Total Cost Quantity 1 1 2 1 1 1 2 1 1 NA NA Unit NA set each each L ' each each each hour week Unit Cost ($) $3,000 $10 $250 NA $0 $5 $1 $270 $38 $1,040 $400 a Other than unit costs, all costs are rounded to the nearest $5. b For solid samples and SRMs. c Only for use with solid samples; assumes two SRMs are required 1 -Month $3,000 $10 $500 $65 $105 $680 $20 $10 $800 $830 $6,300 $1,040 $400 $1,440 $12.250 (a low and Total Cost for Selected Project Duration " 3-Month 6-Month 12-Month $9.000 $10 $500 $65 $315 $890 $60 $25 $800 $885 $18,900 $3,120 $1,200 $4,320 $33.995 $18,000 $10 $500 $75 $630 $1,215 $120 $40 $800 $960 $37,800 $6,240 $2,400 $8,640 $66.615 a high standard) with a life e $36,000 $10 $1,000 $95 ' $1,260 $2,365 $160 $40 $800 $1,000 $75,600 $12,480 $4,800 $17,280 $132.245 ixpectancy of 1 24-Month $72,000 $10 $1,500 $135 $2,520 $4,165 $160 $40 $800 $1,000 $151,200 $24,960 $9,600 $34,560 $262.925 year (some standards will have longer shelf lives). Liquid standards are also available and are generally less expensive. d Assumes rental of the cylinder, regulator, and mounting bracket for all time periods, plus the cost of oxygen consumed. e Rental costs were used through the 3-month period for chairs and the 6-month period for the table. Purchase costs were used for longer periods. Purchase costs for the tent were used for all periods. 62 ------- Table 7-7 summarizes costs for the actual demonstration. Note that the 3-month rental cost of the DMA-80 was used for capital costs. Table 7-7. DMA-80 Costs by Category Category Instrument Cost Supplies Support Equipment Labor IDW Disposal Total Category Cost ($) $3,000 $590 $280 $900 $1,440 $6,210 Percentage of Total costs 48.3% 9.5% 4.5% 14.5% 23.2% 100.0% The cost per analyses based upon 173 samples when renting the DMA-80 is $35.90 per sample. The cost per analysis for the 173 samples, excluding instrument cost is $18.55 per sample. 7.3 Typical Reference Method Costs This Section presents costs associated with the reference method used to analyze the demonstration samples for mercury. Costs forother project analyses are not covered. The referee laboratory utilized SW-846 Method 7471B for all soil and sediment samples. The referee laboratory performed 421 analyses over a 21-day time period. A typical mercury analysis cost, along with percent moisture for dry-weight calculation, is approximately $35. This cost covers sample management and preparation, analysis, quality assurance, and preparation of a data package. The total cost for 173 sam pies at $35 would be $6,035. This is based on a standard turnaround time of 21 calendar days. The sample turnaround time from the laboratory can be reduced to 14,7, oreven fewer calendar days, with a cost multiplier between 125% to 300%, depending upon project needs and laboratory availability. This results in a cost range from $6,035 to $18,105. The laboratory cost does not include sample packaging, shipping, or downtime caused to the project while awaiting sample results. 63 ------- Chapter 8 Summary of Demonstration Results As discussed previouslyin this ITVR, the Milestone DMA-80 was evaluated by having the vendor analyze 173 soil and sediment samples. These 173 samples consisted of both medium- and low-concentration field samples from three sites, SRMs, and spiked field samples. Table.8-1 provides a breakdown of the numbers of these samples for each sample type and concentration range or source. Collectively, these samples provided the different matrices, concentrations, and types of mercury needed to perform a comprehensive evaluation of the DMA-80. 8.1 Primary Objectives The primary objectives of the demonstration were centered on evaluation of the field instrument and performance in relation to sensitivity, accuracy, precision, time for analysis, and cost. Each of these objectives was discussed in detail in previous chapters and is summarized in the following paragraphs. The overall demonstration results suggestthat the experimental design was successful for evaluation of the Milestone DMA-80. Quantitative results were reviewed and this instrument was found to be very comparable to standard analyses performed by the laboratory and the collected data provide the evidence to support this statement. The two primary sensitivity evaluations performed for this demonstration were the MDL. and PQL. Following procedures established in 40 CFR Part 136, the MDL for the DMA-80 is likely between 0.049 and 0.068 mg/kg. The equivalent MDL for the referee laboratory is 0.0026 mg/kg. The calculated MDL is only intended as a statistical estimation and not a true test of instrument sensitivity. The PQL for the DMA-80 is likely somewhere around 0.082 mg/kg based upon the analysis of several low concentration standard reference materials. Both the MDL and PQL were determined for soils and sediments and the instrument may be capable of measuring lower concentrations for aqueous samples; however, this is not tested during the demonstration. The referee laboratory PQL, determined as part of the laboratory analysis was 0.005 mg/kg based upon a lower calibration standard. The %D is < 10%. Accuracy was evaluated by comparison to SRMs and comparison to the referee laboratory analysis for field samples. This included spiked field samples forevaluation of additional concentrations not otherwise available. Milestone data compared to SRM values were within expected accuracy determinations. ALSI data compared to SRM values were also within expected accuracy determinations. (DMA-80 results were within SRM 95% prediction intervals 93% of the time, and referee laboratory results were within SRM 95% prediction intervals 96% of the time.) Comparison of the Milestone to the referee laboratory data for all field and spiked samples (including SRMs) based upon hypothesis testing at the alpha = 0.01 level suggest that the two data sets are not dissimilar. Additional aggregate analysis for all collected data also suggests that the two data sets are not dissimilar. The number of Milestone average values less than 30% different from the referee laboratory results or SRM reference values; however, was 16 of 30 different sample lots. Only 2 of 30 Milestone average results have relative percent differences greater than 100% for this same group of samples; however, when making the comparison between Milestone and ALSI data, and taking into account the possible bias associated with both sets of data, this comparison may be within reasonable expectations for considering these two separate analyses to be equivalent. Therefore, it could be concluded that the Milestone DMA- 80 was within the expected accuracy for analysis of mercury in soil comparable to laboratory Method 7471B. 64 ------- Precision was determined by analysis of replicate samples. The single most important measure of precision provided, overall average RSD, is 23.7% for the referee laboratory compared to the Milestone average RSD of 19.4%. Both of these RSDs are within the predicted 25% RSD objective for precision. The precision of the Milestone field instrument is therefore very comparable to laboratory .precision, and within expected precision variation for soil and sediment matrices. Precision was not affected by sample concentration or matrix. Time measurements were based on the length of time the operator spent performing all phases of the analysis, including setup, calibration checks, and sample analysis (including all reanalysis). Milestone analyzed 173 samples in 1,330 minutes over four days, which averaged to 7.7 minutes per sample result. Based on this, an operator could be expected to analyze 62 samples (8 hours x 60 minutes + 7.7 minutes/sample) in an 8-hour day. Cost of the Milestone sample analyses included capital, supplies, labor, support equipment, and waste disposal. The cost per sample was calculated both with and without the cost of the instrument included. This was performed because the first sample requires the instrument purchase, and as the sample number increases, the cost per sample would decrease. A comparison of the field Milestone cost to off-site laboratory cost was not made. To compare the field and laboratory costs correctly, it would be necessary to include the expense to the project while waiting for analyses to return from the laboratory (potentially several mobilizations and demobilizations, stand-by fees, and other aspects associated with field activities). Table 8-2 summarizes the results of the primary objectives.. 8.2 Secondary Objectives Table 8-3 summarizes the results of the secondary objectives. Table 8-1. Distribution of Samples Prepared for Milestone and the Referee Laboratory Sample Type Site Carson River (Subtotal = 75) Puget Sound (Subtotal = 57) Oak Ridge (Subtotal = 41) Subtotal Concentration Range Low(1-500ppb) Mid (0.5-50 ppm) High (50-> 1,000 ppm) Low (1 ppb -10 ppm) High (10-500 ppm) Low (0.1-10 ppm) High (10-800 ppm) • Soil 7 9 0 26 0 17 0 72 Sediment 10 0 0 0 0 3 0" 13 Soiked Soil 7 14 0 14 0 7 0 42 SRM 7 21 0 17 0 14 0 70 65 ------- Table 8-2. Summary of DMA-80 Results for the Primary Objectives Demonstration Objective Instrument Sensitivity Evaluation. Basis Performance Results DMA-80 MDL Method from 40 CFR Part 1 36. Between 0.049 and 0.068 mg/kg Reference Method 0.0026 mg/kg Accuracy Precision POL. Low concentration SRMs or samples. Comparison to SRMs, field, and spiked samples covering the entire range of the instrument calibration. Determined by analysis of'replicate samples at several concentrations. Time per Analysis Timed daily operations for 4 days and divided the total time by the total number of analyses. Cost Costs were provided by Milestone and independent suppliers of support equipment and supplies. Labor costs were estimated based on a salary survey. IDW costs were estimated from the actual costs encountered at the Oak Ridge demonstration. Approximately 0.082 mg/kg 0.005 mg/kg Milestone's DMA-80 is within expected accuracy for laboratory analysis. Milestone's field instrument is very comparable to the referee laboratory analytical method, 7471B. Overall RSD was computed to be 19.4% compared to the referee laboratory RSD of 23.7%. This is a combined measure of precision which includes sampling and aliquoting variations. Milestone's precision is comparable to the laboratory analysis and is not affected by matrix or concentration. One technician performed all setup, calibration checks, sample preparation and analysis, and equipment demobilization. Individual analyses took 5 minutes each, but the total time per analysis averaged approximately 8 minutes per sample. The cost per analyses based upon 173 samples, when renting the DMA-80, is $35.90 per sample. The cost per analyses for the 173 samples, excluding capital cost, is $18.55 per sample. The total cost for equipment rental and necessary supplies during the demonstration is estimated at $6,210. The cost breakout by category is: capital costs, 48.3%; supplies, 9.5%; support equipment, 4.5%; labor, 14.5%; and IDW, 23.2%. 66 ------- Table 8-3. Summary of DMA-80 Results for the Secondary Objectives Demonstration Evaluation Basis Objectives Performance Results Ease of Use Health and Safety Concerns Portability of the Device Instrument Durability Availability of Vendor Instruments and Supplies Field observations during the demonstration. Observation of equipment, operating procedures, and equipment certifications during the demonstration. Review of device- specifications, measurement of key components, and observation of equipment setup and tear down before, during, and after the demonstration. Observation of equipment design and construction, and evaluation of any necessary repairs or instrument downtime during the demonstration. Review of vendor website and telephone calls to the vendor after the demonstration. The DMA-80 is easy to operate, requiring one field technician with a basic knowledge of chemistry acquired on the job or in a university, and training on the DMA-80. No significant health and safety concerns were noted during the demonstration. The only potential health and safety concerns identified were the generation of mercury vapors and the use of oxygen as the carrier gas. The vendor recommends and can provide a mercury filter; oxygen can be safely handled using standard laboratory procedures. The DMA-80 was not easily portable (by hand) due to its size and weight. It was easy to set up and can be taken anywhere that a small van or SUV can go. The instrument is better characterized as mobile rather than field portable. The DMA-80 was well designed and constructed for durability. The DMA-80 is readily available for lease, or purchase. DMA-80 rental is available on a limited basis. Spare parts and consumable supplies can be added to the original DMA-80 order or can be received within 24 to 48 hours of order placement. Supplies and standards not provided by Milestone are readily available from laboratory supply firms. 67 ------- Chapter 9 Bibliography Anchor. Environmental. 2000. Engineering Design Report, Interim Remedial Action Log Pond Cleanup/ Habitat Restoration Whatcom Waterway Site, Bellingham, W.A. Prepared for Georgia Pacific West, Inc. by Anchor Environmental, L.L.C., Seattle, W A. July 31, 2000. Confidential Manufacturing Site. 2002. Soil Boring Data from a Remedial Investigation Conducted in 2000. Milestone, 2003.- Getting Ready for the DMA-80 Direct Mercury Analyzer from Milestone Inc., Instrument SOP. June, 2003. Rothchild.E.R., R.R.Turner,S.H. Stow,M.A. Bogle, L.K. Hyder, O.M. Sealand, H.J. Wyrick. 1984. Investigation of Subsurface Mercury at the Oak Ridge Y-12 Plant. Oak Ridge National Laboratory, ORNL/TM-9092. U.S. Environmental Protection Agency. 1994. Region 9. Human Health Risk Assessment and Remedial Investigation Report - Carson River Mercury Site (Revised Draft). December 1994. U.S. Environmental Protection Agency. 1995. Contaminants and Remedial Options at Selected Metal-Contaminated Sites. July 1995. Washington D.C., EPA/540/R-95/512. U.S. Environmental Protection Agency. 1996. Test Methods for Evaluating Solid Waste, Physical/Chemical Methods, SW-846 CD ROM, which contains updates for 1986, 1992, 1994, and 1996. Washington D.C. U.S. Environmental Protection Agency. 1998. Unpublished. Quality Assurance Project Plan Requirements for Applied Research Projects, August 1998. U.S. Department of Energy. 1998. Report on the Remedial Investigation of the Upper East Fork of Poplar Creek Characterization Area at the Oak Ridge Y-12 Plant, Oak Ridge, TN. DOE/OR/01-1641&D2. U.S. Environmental Protection Agency. 2002a. Region 9 InternetWebSite,www.epa.gov/region9/index.html. U.S. Environmental Protection Agency. 2002b. Guidance on Data Quality Indicators. EPA G-5i, Washington D.C., July 2002. U.S. Environmental Protection Agency. 2003. Field Demonstration and Quality Assurance Project Plan - Field Analysis of Mercury in Soil and .Sediment. August 2003. Washington D.C., EPA/600/R-03/053.. Welding Supply, 2003. Price List for Oxygen and Pressure Regulators, www.weldingsupply.com. July 2003. Wilcox, J.W., Chairman. 1983. Mercury at Y-12: A Summary of the 1983 UCC-ND Task Force Study. Report Y/EX-23, November 1983. www.Fishersci.com, 2003. www.milestonesci.com, 2003. 68 ------- Appendix A Milestone Comments In the present work, the method detection limit (MDL) was determined from the data collected during the Demonstration exercise. To demonstrate the versatility and to validate the performance of the instrument, samples were processed, as provided by the EPA and in the order presented by the EPA. As a result, low concentration samples were processed after high concentration samples, and vice versa. As in most analytical instruments, a memory effect can occur when low level samples are processed immediately after high level samples. This effect can be completely mitigated by operator technique. When the MDL is determined by running 7 replicates of a low level soil standard, in sequence, the MDL is found to be 8 ppb. To obtain optimum results for such low level samples in the field, good operating technique would require the technician to process several blanks prior to running the low level samples. This appendix was written solely by Milestone. The statements presented in this appendix represent the developer's point of view and summarize the claims made by the developer regarding the DMA-80. Publication of this material does not represent EPA's approval or endorsement of the statements made in this appendix; performance assessment and economic analysis results for the DMA-80 are discussed in the body of the ITVR. 69 ------- Appendix B Statistical Analysis Two separate hypothesis tests were used to compare the referee laboratory samples to the vendor tested samples. This appendix details the equations and information for both of these statistical analyses. For purposes of this appendix, we have chosen to call the test comparing sample populations using a separate calculation for each sample lot the "hypothesis test," and the statistical comparison of the entire sample set (all 28 separate sample lots) analyzed by the vendor and the laboratory the "unified hypothesis test," also known as an "aggregate analysis" for all of the sample lots. Hypothesis Test A hypothesis test is used to determine if two sample populations are significantly different. The analysis is performed based on standard statistical calculations for hypothesis testing. This incorporates, a comparison between the two sample populations assuming a specified level of significance. For establishing the hypothesis test, it was assumed that both sample sets are equal. Therefore, if the null hypothesis is rejected, then the sample sets are not considered equal. This test was performed on all sample lots analyzed by both Milestone and the referee laboratory. H0and Ha, null and alternative hypothesis respectively, were tested with a 0.01 level of significance (LOS). The concern related to this test is that, if two sample populations have highly variable data (poor precision), .then the null hypothesis, may be accepted because of the test's inability to exclude poor precision as a mitigating factor. Highly variable data results in wider acceptance windows, and therefore, allows for acceptance of the null hypothesis. Conclusions regarding this analysis are presented in the main body of the report. To determine if the .two sample sets are significantly different, the absolute value of the difference between the laboratory average XL and the vendor average. xv is compared to a calculated u. When the absolute value of the difference is greater than u, then the alternate hypothesis is accepted, and the two sets (laboratory and vendor) are concluded to be different. To calculate p, the variances for the laboratory data set and the vendor data set are calculated by dividing their standard deviations by the number of samples in their data set. The effective number of degrees of freedom is then calculated. -2 Where: f = effective number of degrees of freedom VL = variance for the laboratory results nL = number of samples for the laboratory data set • Vv = variance for the vendor results nv = number of sam pies for the vendor data set. The degrees of freedom (f) is used to determine the appropriate T value and used to calculate u at the 0.01 level of significance using the following: 70 ------- Unified Hypothesis Test For a specified vendor, let Y,, be the measured Hg concentration for the y* replicate of the /"" sample for ; =1,2 1 and; = 1,2,...,^. Let X,- = log(Yj), where log is the logarithm to the base 10. Define xaog to be the average over all log replicates for the f1 sample given by: X ilog -1 log J, s /-I Where x2M is approximately a chi-square random variable with (1-1) degrees of freedom: -' log (jr,tog - jr-;.bg ) and Denote the estimate of the variance of the log replicates for the f sample to be: Now for the reference laboratory, let Y'(Jbe the measured Hg concentration for the /" replicate of the /"" sample for / =1,2 1' and j'- 1,2 J1,. Denote the reference laboratory quantities X'f/, x/, and s'2 defined in a manner similar to the corresponding quantities for the vendor. Assumptions: Assume that the vendor measurements, Yf, are independent and identically distributed according to a lognormal distribution with parameters M/and a2. That is, X,y= log(Yj) is distributed according to a normal distribution with expected value M/and variance o2. Further, assume that the reference laboratory measurements, Y'j, are independent and identically distributed according to a lognormal distribution with parameters p',-and o'2. The null hypothesis to be tested is: H0 : & = p'i + 5, for some 5 and i = !,...,! against the alternative hypothesis that the equality does not hold for at least one value of /'. The null hypothesis H0 is rejected for large values of: !-l 2-1 Critical values for the hypothesis test are the upper percentile of the chi-square distribution with (1-1) degrees of freedom obtained from a chi-square table. Results of Unified Hypothesis Test for Milestone Milestone and ALSI both supplied multiple assays on replicates derived from a total of 28 different sample lots, either field materials or reference materials. The Milestone and ALSI data from these assays formed the basis of this assessment. The statistical analysis is based on log-transformed (logarithm base 10) data and uses a chi-square test for equality of Milestone and ALSI population means for given a sample lot. Equality of variances is assumed. A description of the statistical procedure is provided below. Initially, the hypotheses tested was that, on average. Milestone and ALSI would produce the same results within a given sample lot. This hypotheses is stated as: H10: (Milestone lot log mean) = (ALSI lot log mean) H10 was strongly rejected in that the chi-square statistic was 140.2, which exceeds the upper 99th percentile of the chi-square distribution with 28 degrees of freedom having value of 48.3. The null hypotheses was rejected in part because Milestone results tended to exceed those from ALSI for the same sample lot. To explore this effect, the null hypothesis was revised to included a bias term in the form of: 71 ------- H20: (Milestone lot log mean) = (ALSI lot log mean) +(delta). Where delta is a single value that does not change from one sample lot to another, unlike the lot log means. H20 was rejected, in that the chi-square statistic was 63.9, which exceeded the upper 99th percentile of the chi- square distribution with 27 degrees of freedom with value of 47.0. In this analysis, delta was estimated to be 0.12 in logarithmic (base 10) space, which indicates an average upward bias for Milestone of 100-12=1.318 or about 32%. Milestone analytical results for sample lot 37 were large relative to the concentration provided with the sample reference material, and the Milestone data for sample lot 37 made a substantial contribution to the chi-square Table B-1. Summary of Unified Hypothesis Test statistic. Accordingly, excluding sample lot 37 from the data set resulted in a chi-square statistic of 42.1, which does not exceed the upper 99th percentile of the chi-square distribution with 26 degrees of freedom with value 45.6. So, with excluding sample lot 37 data, one fails to reject H20 at the 99th percent level. In this analysis, delta was estimated to be 0.11 in logarithmic (base 10) space, which indicates an average upward bias of 10°'11=1.288 or about 29%. Note further, that excluding sample lots 37 and 18 resulted in accepting H20 at the 95% level, and excluding sample lots 37, 18 and 39 resulted in accepting H20 at the 90% level. Summary information on these analyses is provided in Table 6-5. Hypothesis Total Sample Lots H,0 28 Hso 28 H20 28 Hso 28 H,. 28 Excluded Lot None None 37 37,18 37. 18. 39 DF 28 27 26 25 24 s2 0 pool 0.01562 0.01562 0.01369 0.01411 0.01410 Delta 0.0000 0.1230 0.1104 0.1172 0.1228 Chi-square 140.211 63.901 42.085 36.124 31.304 P-value 0.000000 0.000079 0.024058 0.069744 0.145220 DF = Degrees of Freedom s2 = variance 72 ------- |