------- ERA'S MANAGEMENT SYSTEM FOR ENVIRONMENTAL DATA QUALITY TOTAL QUALJTY MANAGEMENT (TQM) is the process whereby an organization, led by senior management, commits to focusing on quality as a first priority In every activity. TQM implementation creates a culture in which everyone in the organization shares the responsibility for improving the quality of products and services, and for "doing the right thing, the right way, the first time." EPA's QUALITY ASSURANCE (QA) program for environmental data operations is based firmly on the principles of Total Quality Management Quality assurance is the process of management review and oversight at the planning, implementation, and completion stages of an environmental data operation to assure that the data provided by a line operation to data users are of the quality needed and claimed. The TQM concepts which the Agency's QA program has put into practice include the following: * customer-supplier relationships, especially a clear statement of the customer's (data user's) needs; * establishment of measures of performance for supplier implementation and customer evaluation; * process analysis through techniques such as process flow diagramming; and * employee development, involvement, and recognition. QA is not Identical to QUALITY CONTROL (QC), which is an aspect of the implementation phase of an environmental data operation. QC includes those activities required during data collection to produce the data quality desired and to document the quality of the collected data (e.g., sample spikes and blanks). At EPA, quality assurance Is a management system based upon the proven management philosophy of Total Quality Management The primary responsibility for implementing QA belongs to the line managers of EPA organizations which are involved in the collection or use of environmental data, whether in Headquarters, Regions, or Research and Development Laboratories. EPA managers at all levels benefit from a program which succeeds in bringing the Agency's environmental data operations Into alignment with its decision-making needs - "the right thing, the right way, the first time." ------- CONTENTS Agenda i Executive Summary iii Biographies of Speakers v PRESENTATIONS Introduction by Nancy Wentworth 1 Welcoming Remarks by Robert E. Layton 3 Keynote Address by Joe D. Winkle 5 EPA's Environmental Monitoring Management Council by Ramona Trovato 9 Harmonization of QA Requirements Across the Government Introduction by Nancy Wentworth 19 Integrity in Science by Adil E. Shamoo 23 Journey Toward Quality: The Federal Express Story by Gilbert Mook 33 PANEL DISCUSSIONS Good Automated Laboratory Practices 43 Hazardous Waste Site Remediation 63 Ecological Monitoring 89 Discussion Summaries 107 GROUP SESSIONS National Program Offices 109 Regional Offices 113 Office of Research and Development 115 QUALITY ASSURANCE MANAGER OF THE YEAR AWARD Introduction by Nancy Wentworth 119 Acceptance Speech by Marty Brossman 123 ------- AGENDA Monday, April 22 1:00 pm - Call To Order, Introduction Nancy Wentvorth, Director, Quality Assurance Management Staff 1:10 pm - Welcome Robert E. Layton, Regional Administrator, Region 6 1:30 pm - Keynote Address Joe Winkle, Deputy Regional Administrator, Region 6 2:15 pm - Good Automated Laboratory Practices (panel discussion) Kaye Mathews, National Enforcement Investigation Center (MODERATOR) Joan Fisk, Office of Emergency and Remedial Response Jeff Worthington, TechLav, Inc. Rick Johnson, Office of Information Resources Management 3:15 pm - EPA's Environmental Monitoring Management Council (EMMC) Ramona Trovato, Executive Secretary of the EMMC Policy Council, Analysis and Evaluation Division 4:00 pm - Common Interest Group Discussions, Introduction and Planning by Nancy Wentworth Tuesday, April 23 8:30 am - Harmonization of QA Requirements Across the Government Introduction by Nancy tfenttrorth 9:00 am - Hazardous Waste Site Remediation (panel discussion) Gary Johnson, Quality Assurance Management Staff (CHAIR) Marcia Davies, U.S. Army Corps of Engineers John Edkins, Naval Energy and Environmental Support Activity Duane Geuder, Office of Emergency and Remedial Response Tom Morris, Martin Marietta Energy Systems, Oak Ridge National Lab 11:00 am - Ecological Monitoring (panel discussion) Robert Graves, Environment Monitory Systems Laboratory (CHAIR) Adriana y Cantillo, National Oceanic and Atmospheric Administration James Andreasen, U.S. Fish and Wildlife Service Thomas Cuffney, U.S. Geological Survey 2:00 pm - Discussion Groups - Harmonizing QA Across Government 3:30 pm - Summary of Discussion Groups 4:00 pm - Integrity in Research Adil Shamoo, Editor in Chief, Accountability in Research ------- 5:00 pm - Adjourn 6:30 pm - Banquet - Journey Toward Quality: The Federal Express Story Gilbert Mook, Vice President for Properties and Facilities, Federal Express, Inc. 1990 Malcolm Baldrige National Quality Award Winner - Presentation of the Qualify Assurance Management of the Year Award Wednesday, April 24 Training Sessions: QUALITY AUDITS FOR IMPROVED PERFORMANCE Dennis Arter, Columbia Quality, Inc. D A one-day condensed version of the ASQC course COMMUNICATION SKILLS Dr. Linne Bourget, Positive Management Communications Systems n A half-day seminar on communication skills for positive power and influence, relationships, and organizational change TRAIN-THE-TRAINER SEMINAR Mary Ann Pierce, JWK International Corp. a A half-day seminar on how to use QA training tools effectively QUALITY CIRCLES RADM Frank Collins, Frank Collins Associates D A half-day seminar on how to organize, train, implement, and nurture Quality Circles in your organization QA CAREER MANAGEMENT SEMINAR Joanne Jorz, Conceptual Systems, Inc. D A half-day seminar designed to help QA professionals develop and manage their careers Thursday, April 25 Common Interest Group Sessions Friday, April 26 8:30 am - Common Interest Group Sessions Wrap-up 10:00 am - Summary Reports from Common Interest Groups 11:30 am - Closing 12:00 noon - Adjourn ii ------- MEETING HIGHLIGHTS Nancy Wentvorth, Director of the EPA Quality Assurance Management Staff, welcomed attendees to the meeting, touched upon the theme "Data Quality Across the Government," and stressed the need to work together with a common goal to meet environmental challenges that lay ahead. Robert Layton, Regional Administrator of Region 6, described the concept of Total Quality Management (TQM) and emphasized the importance of providing a uniform approach to assuring data quality. Joe Winkle, Deputy Regional Administrator, Region 6, discussed the benefits of standardizing quality assurance requirements in Federal, state and private sectors. Raaona Trovato, Executive Secretary for the Environmental Monitoring Management Council's policy council, described the Environmental Monitoring Management Council: how it was formed, its structure, and its functions. Adil Shanoo, Editor of Accountability in Research: Policies and Quality Assurance, discussed integrity of research: problems and their causes, the role of quality control and quality assurance, and ideas for improving data integrity. Gilbert Nook, Vice President for Properties and Facilities, Federal Express, Inc., talked about the Malcolm Baldrige National Quality Award: the goals and philosophies of Federal Express, their commitment to the customer, and their quality of service. PANEL DISCUSSION SUMMARIES: Kaye Mathevs, QA Manager, National Enforcement Investigation Center; Jeff fforthington, Quality Assurance Director for TechLaw; Joan Pisk, Deputy Chief, Analytical Operations Branch, Office of Emergency and Remedial Response; and Rick Johnson, a member of EPA's Scientific Staff in the Office of Information Resources Management, participated in a panel on "Good Automated Laboratory Practices: Recommendations for Ensuring Data Integrity in Automated Laboratory Guidance." The discussion focused on the trend toward computer automation in the laboratory. Gary Johnson, a Quality Assurance Management Staff member; Thomas Morris, QA Manager for the Hazardous Waste Remedial Actions Program; Duane Geuder, QA Manager/Program Analyst in the Office of Emergency and Remedial Response; Marcia Davies, HTW Branch Chief of the Missouri River Division of the Army Corps of Engineers; and John Edkins, QA Manager for the Navy Installation Restoration Program, participated in a panel discussion on the National iii ------- Consensus Standard that is being developed through the American Society for Quality Control (ASQC). Panelists focused on the structure of the proposed standard and ways to effectively implement the standard process. Robert Graves, Acting Coordinator for the Environmental Monitoring and Assessment Program; Janes Andreasen, from the U.S. Fish and Wildlife Service in the Division of Environmental Contaminants; Adriana y Cantillo, Manager of the National Oceanographic and Atmospheric Administration (NOAA) National Status and Trends Program/Quality Assurance Management Program/Quality Assurance Program; and Thomas Cuffnay, an Ecologist in the Hater Resources Division of the U.S. Geological Survey, participated in a panel that focused on harmonization of ecological monitoring. Panelists discussed the monitoring process in their respective agencies. iv ------- BIOGRAPHIES OF SPEAKERS Nancy Wentworth is the Director of EPA's Quality Assurance Management Staff. Prior to working for QAMS, she was a manager in the Office of Drinking Water for seven years. In 1988, she was awarded the USEPA Bronze Medal for Commendable Service; in 1989 and 1990, she was the recipient of the USEPA Special Service Award. Ms. Wentworth has a B.A. in Civil Engineering and a Master's degree in Sanitary Engineering. Robert Layton has served as the Regional Administrator of EPA's Region 6 since February 1987. He initiated a Value Engineering Program in the Hazardous Waste Management Division, and negotiated the first EPA Superfund program agreements with the Navajo Tribe. Mr. Layton is the Vice Chairman of the Policy Committee for the Galveston Bay Estuary Program and Co-chairman of the Policy Committee for the Gulf of Mexico Program. He is a member of the National Council of Engineering Examiners, and was selected Engineer of the Year for 1976-1977. Joe Winkle has been the Deputy Regional Administrator of EPA Region 6 since April 1988. From October 1982 to 1988, he served as the Director of the Disaster Assistance Programs for the Federal Emergency Management Agency (FEMA). Prior to joining the national office of FEMA, Mr. Winkle was the Acting Regional Director and Deputy Regional Director for FEMA Region 6 in Denton, Texas. He served as Regional Director of the Federal Disaster Assistance Administration (FDAA) from 1973 to 1979 and has been the Federal Coordinating Officer for more than 50 Presidentially-declared major disasters. Ramona Trovato is the Executive Secretary for the Environmental Monitoring Management Council's policy council and a division director in the Office of Water. She is a chemist who has worked for EPA for many years: in the Region 3 Central Lab in Annapolis, on the Quality Assurance Management Staff, and as a liaison between Headquarters and the regional offices. Adil Shamoo has been a professor of Biological Chemistry at the Maryland School of Medicine since 1979. He is the editor of the journal, Accountability in Research, and the textbook, Principles of Research Data Audit. In 1988, Dr. Shamoo organized the First International Conference on Scientific Data Audit, Policies and Quality Assurance. Gilbert Mook is the Vice President of Properties and Facilities for Federal Express Corporation. He joined the company in 1983 as Director of Space Operations and Advanced Programs, was named Vice President of Satellite & Video Systems in 1985, and assumed his ------- current position in 1988. Federal Express is the first winner of the Malcom Baldrige National Quality Award in the service category. Kaye Mathews serves as EPA's Quality Assurance Manager for the National Enforcement Investigation Center in Denver, Colorado, she provides evidentiary consultation to EPA's Contract Laboratory Program and oversees the contract Evidence Audit Team's involvement in the CLP. Joan Fisk, Deputy Chief, Analytical Operations Branch, Office of Emergency and Remedial Response, has been involved with the Superfund Contract Laboratory Program since July 1983. With a B.S. in Chemistry from the University of Bridgeport, Ms. Fisk has worked as an Analytical Chemist throughout her entire career. She is presently a member of EPA's Environmental Monitoring Management Council's panel on Method Standardization and chairs the "Interagency Work Group on Data Authority." Jeff worthington is the Quality Assurance Director for TechLaw, Inc. in Denver, Colorado. He also serves as the Technical Programs Coordinator for TechLaw's Contract Evidence Audit Team (CEAT) contract to EPA's National Enforcement Investigations Center. Rick Johnson is on EPA's Scientific System Staff in the Office of Information Resources Management. He is currently directing the development of EPA's Good Automated Laboratory Practices, of which he is the original author, and the integration of ORD's Data Quality Objectives with the agency's System Design and Development Guidance to establish a methodology for sharing environmental data of documented quality. Gary Johnson serves on EPA's Quality Assurance Management Staff. With a B.S. in Nuclear Engineering, he was a nuclear engineer for Duke Power Company and an environmental engineer for EPA's Office of Research & Development. In his eleven years in Quality Assurance, Mr. Johnson has received two EPA Bronze Medals for outstanding contributions. He is a member of the American Society for Quality Control (ASQC), and presently serves as an Associate Regional Councilor for the ASQC. Marc la Davies, the HTW Chemistry Branch Chief of the Missouri River Division of the Army Corps of Engineers, has a Ph.D. in Analytical/Inorganic Chemistry. Her branch of the Army Corps of Engineers is responsible for the continuing development of the ESACE Chemistry Data Quality Management Program and national oversight of its implementation. vi ------- John Edkins is currently employed by the U.S. Navy at Port Hueneme, California as a hydrologist, and is the Quality Assurance Manager for the Navy Installation Restoration Program. He has an M.A. in Geology and is registered as a professional geologist. As a Section Vice-chairman in the American Society for Testing and Materials subcommittee 0-1821 for Ground Water and Vadose Zone Investigations, Mr. Edkins is also a member of the Interagency Ad Hoc Committee for Quality Assurance in Environmental Measurements. Duane Geuder has an M.S. in Radiation Biology/Biophysics. He has worked as a chemist in the Marine Corps, as an oceanographer for the U.S. Navy Oceanographic Office, and most recently as a Quality Assurance Manager for EPA's Superfund. He has more than 25 years of experience in environmental sampling and analysis and related QC and QA activity. Mr. Geuder is a member of the Ad Hoc Panel on QA/QC services under the Environmental Monitoring Management Council, and belongs to the Association of Official Analytical Chemists. Thomas Morris is currently employed by Martin Marietta Energy Systems, Inc as the Quality Assurance Manager for the Hazardous Waste Remedial Actions Program. A member of both the Working Group and the Policy/Steering Committee for QA Harmonization, Mr. Morris has eight years of experience in Total Quality Management concepts. Robert Graves has worked for EPA's Environmental Monitoring Systems Laboratory in Cincinnati since October 1978. Prior to that, he was employed by the U.S. Treasury's Bureau of Alcohol, Tobacco, and Firearms to examine and analyze evidence from various law enforcement agencies. Mr. Graves has an M.S. in Chemistry and an M.B.A. in Finance. Adriana y Cantillo currently works for the NOAA/National Ocean Service (NOS)/Office of Oceanography and Marine Assessment/Coastal and Estuarine Assessments Branch. She is the Manager of the NOAA National Status and Trends Program/Quality Assurance Management Program/Quality Assurance Program, and the Coordinator of NOAA activities of the multiagency ocean Dumping Ban Act Research and Monitoring Program, with a Ph.D. in Chemistry, Dr. Cantillo is a member of EPA's Methods Integration Work Group, the NOS representative to the National Ocean Pollution Policy Board, and the NOS representative to the Working Group developing the Federal Plan for Ocean Pollution Research. Development and Monitoring: Fiscal Years 1991-1995. James Andreasen has 18 years of experience, in both research and operations, evaluating the effects of environmental contaminants on fish and wildlife populations and water quality. He has a Ph.D. in vii ------- Zoology/Natural History and is employed by the U.S. Fish and Wildlife Service in the Division of Environmental contaminants. As the technical advisor for the Service on the Department of Interior Irrigation Water Quality Program, Dr. Andreasen is responsible for pulling together various elements that contribute to harmonization of field methods and in formulating standard operating procedures for the operational contaminant program. Thomas Cuffney is an Ecologist in the Water Resources Division of the U.S. Geological Survey. He serves on the Technical Issues Committee of the North American Benthological Society and the 1992 Program Committee of the Society of Environmental Toxicology and Chemistry. He has a Ph.D. in Biology and seven years experience in aquatic ecology and environmental toxicology. viii ------- INTRODUCTION Nancy Wentworth Director EPA Quality Assurance Management Staff I'd like to welcome you to the llth Annual National Meeting on managing environmental data quality. The theme of this meeting is "Data Quality Across the Government." Today is Earth Day. There are probably a fair number of people here in the audience who participated in the first Earth Day 21 years ago. It was something that made a mark on my life and created a commitment to protecting and preserving the environment. I think that's what we're here to consider this week. There are people from a wide diversity of agencies and their supporting contractor communities, and we need to work together. We need to work efficiently, because we have many things that we must do and we do not have unlimited resources. So efficiency is our only hope in the long term of being able to meet the challenges that are before us in the environment. One of the things that I have most enjoyed in my six years with the Quality Assurance Management Staff (QAMS) is the opportunity to learn and to grow, and to try to build continuous improvement into EPA's Quality Assurance Program. I look at this meeting as a significant opportunity for all of us to come to a much greater understanding of our mutual concerns, our mutual problems, and our mutual goals. Your gift to QAMS this week is your knowledge. The QAMS staff who are here are going to be listening and probing and trying to gain as much as they can. Please share your experiences with us. This is what we need to hear so that we can do our jobs better. We value the diversity of the audience that's here. We have people from all backgrounds, from all geographical areas of the country, and from all of the programs that EPA is concerned with, either directly or indirectly. That is very important to us. How we are going to succeed in the long term is to bring all of these views and visions together and move forward as a group. We can't afford to go in separate directions; we must work together. I think we have a tremendous goal ahead of us this week and I look forward to achieving it. ------- WELCOME Robert E. Layton Regional Administrator EPA Region 6 Welcome to Region 6 of EPA and to Dallas, Texas. You know that Texas is the largest state in the Union. Now don't tell me Alaska is. I know it covers more area, but when all the ice melts, Texas is still the largest. Texans are proud of their quality as well as their quantity, and we always like to think of Texas as being the biggest and the best. I sound, perhaps, like a couple of fellas in this story about a Texan and a man from a Arkansas who got on a train in Texarkana to go to El Paso. The Texan began bragging about the size of the state and he said, "You know, we could ride all day on this train and still be in Texas." And the man from Arkansas quietly replied, "Well, don't feel bad sir, we've got slow trains in Arkansas, too." Of course, they weren't speaking from a uniform standard of data. The man from Arkansas wasn't looking at the train from the same standpoint as the Texan. And that is likely to occur sometimes in quality assurance (QA) management—we may not all be looking at it from the same perspective. The theme for this year's meeting is "Management of Environmental Data Quality Across the Government." All of you who are EPA QA Managers should take great pride in your contribution to providing quality data for this Agency. For all of you who are here representing other agencies, we welcome your interest and your participation, and we know you are as dedicated to helping provide quality data as we at EPA are. We look forward to continuing a dialogue that will result in strengthening all of our programs. Perhaps this week could be the beginning of a process that could change the way government does business in the field of data quality management. I challenge you to begin a communication with one another which transcends agency lines. You have a unique opportunity to lay a foundation for a uniform approach to data quality throughout the government. If you can begin to devise methods for providing a uniform approach to assuring data quality, you will provide an invaluable service to not only your particular agency, but to the entire structure of government as well—Federal, state, and local. Your ability to communicate, compromise, and come to a consensus on QA and data gathering would certainly be within the framework of what Administrator Reilly calls Total Quality Management (TQM). ------- Implementing the ideal of TQM is both realistic and beneficial, and it is mandated in EPA. And so I charge you this week to seek out each other. Discuss common values and standards as they relate to data quality. Become involved in creating new and excellent approaches for improving data quality across the lines of Federal agencies. Find similarities and build on them. Look closely at differences among existing QA programs and see if they can be reconciled. Good luck in this worthy endeavor. ------- KEYNOTE ADDRESS Joe D. Winkle Deputy Regional Administrator EPA Region 6 It's a pleasure to be here today and to address such a distinguished group of people that has such an important mission as it pertains to the job of environmental protection and science in general. F. Henry Habicht, our Deputy Administrator of EPA, in an article back in November 1989 in The Quality Manager, said that "few things are more important to our success as an Agency than credible, usable data. Indeed, virtually every regulatory decision, research activity, and budgetary action taken by this Agency is based, in significant part, on environmental data. Our Agency is in the business of making decisions—often difficult decisions. In order to achieve our shared mission, our actions must be supported by reliable environmental data." Habicht went on to say that Administrator Reilly intended to pursue top priority initiatives to ensure that the Agency had access to the best data from all significant sources. To that end, our Administrator and his Deputy have established an internal task force to explore the merits of creating a National Center for Environmental Statistics. However, the National Center will only be as good as the data it receives from the various agencies. Quality of interagency programs and projects across the government is always a critical issue, particularly in Region 6 where we work with many Federal, state, and local agencies. All are involved in environmental data operations. We have numerous facilities under our jurisdiction; sometimes we feel there are more than we can comfortably work with. Standardization of QA requirements will lessen the number of documents that must be written and reviewed, resulting in savings of time and money. Each Agency has different requirements for the generation and documentation of environmental data; therefore, it would be prudent and feasible to have a uniform approach for producing such information. Each time we standardize and streamline we save ourselves untold work. It is imperative that we address cost-saving issues in these days of budget crunch. Oversight of interagency programs and projects becomes simplified with harmonized data quality assurance standards. The framework for QA in data operations would be common across agencies, thus eliminating confusion. There could be no question about which set of standards would apply, as all would be the same. Newer, more efficient techniques would allow us to spend our time working toward constructive solutions to our problems, instead of discussing the style and scope of QA efforts which may have resulted from a variety of approaches to assuring data quality, none of which may be suitable for decision making. This ------- additional time saved in having a uniform approach to data quality would allow us to launch into new and even more challenging areas. However, providing a single set of standards applicable across the board for programs and assuring the quality of environmental data will be a formidable task. Standard definitions are a must. You must more clearly define what is being sought. At times those whom you interface with do not clearly understand what is being stated. Since environmental data is at the core of so much of what we do, decisions are only as good as the data on which they are based. Our effectiveness as regulators and enforcers depends on our credibility. Without credibility, the regulated community and the public will be reluctant to accept the decisions and actions of any governing or regulating agency. We must, therefore, ensure the quality of our data and thus the quality of the foundations on which our decisions are based. Perhaps you can build communication this week which will become a cornerstone in building the data and enable us to make sound and defensible environmental decisions. An important example of this team communication concept took place in a recent meeting in Washington between the Office of Water and the Office of Enforcement. In this meeting the participants sought to identify areas in which there were differences. The intent was to achieve a smoother operation. The participants included members from the Office of Regional Council, Office Directors from the Office of Water, and the Office of Enforcement, as well as Water Management Division Directors. The consensus was that with good communication and smoother relationships, a rapid solution of problems would result and get the job done in a timely and efficient manner. We are hoping that you QA Managers can begin to fashion a national standard for environmental data quality programs. You could format QA principles and policies which would enhance cooperation and improve the quality of environmental data in the Federal, state and private sectors. As you begin discussions this week, there could evolve a common goal—to provide a framework that would accomplish our variant but important missions in addressing the complicated task of a harmonized approach to data quality. You can make an effort to implement Mr. Reilly's mandate to become involved in TQM, for you have the unique opportunity of applying that TQM principle to a vital task. As each of you provides your own special expertise, and as you discuss and compromise for the common good, you will exercise the very principles of TQM. The key task you will face this week is to communicate on what the Federal environmental mission really is and how you, as QA Managers, can best support it. Through joint efforts you can lay the foundation for harmonization of QA requirements: planning, implementing and reviewing data collection programs. A closer look at data raises some important questions. When a manager is making a decision based on environmental data, that manager must ask: "Do ------- these data meet the needs? Does it help me to make the decision I'm facing? Can I live with this level of uncertainty? Are these data legally defensible?" Much of our data must survive the close scrutiny of court cases. These are only a few of the many questions which arise when assessing data as they relate to an environmental decision. One very important question is often raised when resampling is required: Was the resampling required as a result of too wide a variance in the data? Or, was it required as a result of unclear objectives to begin with? Exact standards, clearly defined objectives, uniform procedures and precise testing—in short, a good job of upfront planning—will eliminate a great deal of resampling. We at EPA use the term data quality objectives (DQO's). As you Know, this term refers to a planning process which determines upfront how good the data need to be to support Agency decisions. So many times, we shoot first, then aim. We have to reload, aim, and then shoot—a procedure we should have used to begin with. All data have a level of uncertainty. We need to know that the level of uncertainty is acceptable. During the planning stage we must ensure that the technical insight of the laboratory analyst who is testing the sample is compatible with the needs of the decision-makers who will use the data. This communication is essential between two cultures: management and technical. DQO's are a cross-cultural communication that provide decision-makers with the ability to manage the level of quality in data that is needed to make the important regulatory decisions they face. The astronomical costs involved in cleaning up the environment dictate that trustworthy data be assured. All these problems could be resolved with a standard, universally- accepted approach to data quality assurance. Our tools are acquiring data—and the data itself must be above reproach, for we are the regulators. Just as the police force should be uniform in the manner in which they enforce the laws, we should be uniform in the manner in which we make regulatory decisions. To gain confidence from the public and from those being regulated, the data used in making those decisions must be appropriate. One of the challenges of the DQO process is that you must strike a balance. Too much quality is as poor an idea as not enough. There should be a balance, and the emphasis must be placed upon arriving at that balance. So we set before you the task to begin communications which will ensure uniform Federal standards for environmental data quality. Provide us this week with a foundation for meeting this goal; give us the building blocks for getting and creating a standard uniform management system for environmental data quality which would begin with how to plan it; then go on to say who will do it, what methods ------- will be used, and most importantly, what standard will be applied in evaluating. This is the heart of data generation. Recent budget constraints make it mandatory for us to seek better methods for doing our job. You QA Managers are at the forefront in providing assistance in doing just that—better, more productive QA management methods, derived from universally-acceptable standards. Your challenge is to begin creating that harmony. I am confident that you will excel at this task; your challenge is great, but it can be achieved. We look forward to the results of your deliberations during this conference. ------- EPA'S ENVIRONMENTAL MONITORING MANAGEMENT COUNCIL Ramona Trovato EPA Analysis and Evaluation Division The Environmental Monitoring Management Council (EMMC) was formed by Deputy Administrator F. Henry Habicht in FY 1990. The EMMC was formed because of three events. one was that a white paper was developed on long term development of analytical methods and short term analytical method development that recommended a need for a senior management group to address monitoring methods issues. Then there was a report called the Section 518 Report that looked at the availability, adequacy and comparability of the Agency's analytical test methods under section 304-H. They were then directed to look further than just the water methods and to look at the other Agency methods and to make recommendations. One of the recommendations they made was to set up an environmental monitoring management council. That happened in 1988. Then in 1989 Region 3 got a new Regional Administrator, Ted Erickson, and he sent a letter to the Administrator that said he thought it would be a good idea if a group was set up like this; the Agency does have some issues in lacking comparability across programs and in methods and in QA and QC. As a result, the Deputy Administrator established the EMMC in FY •90. The EMMC's purpose is to recommend coordinated Agency-wide policies concerning environmental issues. Our charter covers seven areas that EMMC is supposed to look at: 1) coordinate the Agency- wide environmental methods research and development needs; 2) foster consistency and simplicity on methods across media and programs; 3) coordinate short and long term strategic planning and implementation of methods development needs and promote adoption of new technology and instrumentation; 4) coordinate development of QA/QC guidelines as they apply to specific methods; 5) evaluate the feasibility and advisability of a national environmental laboratory accreditation program; and then, other activities that influence environmental monitoring. The EMMC can focus on any issues that affect environmental monitoring which they feel are important. The bottom line is that they're supposed to recommend coordinated Agency-wide policies concerning environmental monitoring issues. All of us have recognized for a long time that we need somebody who's looking across the programs. And I think this is the first time we've had a coordinated effort where we got some senior managers involved in looking at the issues associated with environmental methods development and the QA/QC in environmental laboratory accreditation. 9 ------- I should mention that we're not just inside folks. On the ad hoc panels we have folks from states and from other Federal agencies. In one instance we're going to set up a Federal Advisory Committee so we can have folks from industries come in and talk to us and provide us with advice. Figure 1 shows how the EMMC is structured. The Deputy Administrator is a tie breaker if the Policy Council has issues they can't decide on their own. The Annual Reports also go to the Deputy Administrator. We're in the process of finalizing the first year's Annual Report which we'll be talking about at the next Policy Council Meeting later this week. Figure l Structure of Environmental Monitoring Management Council Deputy Administrator Policy Council Steering Committee QA Services Ad Hoc Panels Methods Integration Automated Methods Compendium Analytical Methods and Regulation Development National Laboratory Accreditation The Policy Council identifies the big issues of concern that need to be coordinated across programs. It meets at least twice a year; in this first year I think it met four times. The Steering Committee oversees the investigation of specific monitoring issues. It meets at least four times a year; it met more this year. The ad hoc panels are the people who really get out and do the leg work. They're the ones who identify the issues, come up with the 10 ------- options, and make recommendations to the Steering Committee which are then elevated up the chain and implemented. They meet as often as they need to meet. The folks who are on the Policy Council are: Erich Bretthauer, who is the Assistant Administrator for the Office of Research and Development, is one co-chair; Edwin (Ted) Erickson, who helped get this ball rolling, is one of the other co-chairs. I'm the executive secretary. Then we have members—one member from each office. Members include the Deputy Assistant Administrators and the Deputy Regional Administrator from the lead region for research and development. The first year it was John Wise; recently it's been Jack McGraw. This representative changes as the lead region changes—every two years. The Steering Committee reports to the Policy Council and they're the ones who make sure that things on the ad hoc panel are moving along in the direction they ought to be moving. The Director of the Office of Modeling, Monitoring Systems, and Quality Assurance is the chair of the Steering Committee. That was Rick Lindhurst until he recently moved upstairs to work directly for Erich on some other projects. Now Jack Puzack's acting in the meantime. David Friedman, whom many of you may know from his days in the Office of Solid Waste, is the executive secretary for the Steering Committee. The members are comprised of Division Directors from headquarters and the Office of Research and Development (ORD), the Environmental Systems Division (ESD) Directors from the lead region for ESD, and the ESD Director from the lead region for research and development. So we have on both of these committees good representation from ORD, the regions, and the headquarters program offices, which is what we were trying to achieve. The regional folks know where the problems are because they're trying to implement the regulations. The program offices write the regulations and need feedback from the regions. ORD folks have to go out and come up with whatever the program offices think they need. We were trying to get everybody represented on these committees so we could move ahead and no one would feel left out. There were five ad hoc panels established about a year ago which started meeting maybe nine months ago. There was one on QA Services which was supposed to look at what the Agency's needs were in terms of QA services, what it cost, and how we could fund it. There was a Methods Integration panel which was supposed to look across the Agency and see where there were opportunities for consolidating methods. I think this panel's an exciting one because, having been in a lab for 13 years where I had to do one method for drinking water and one method for the National Pollutant Discharge Elimination System (NPDES) and another method for the Resource Conservation and Recovery Act (RCRA), I was excited about the idea of doing one method for the same analyte in the same 11 ------- matrix. So I have high hopes for this group. But they also have a tough problem, because they have to make sure that they're going to get the detection limits they need, the precision and accuracy that they need, and still be able to meet the regulatory requirements of that program. The Automated Methods Compendium panel was set up so that we could have a single database that would list all of the Agency's methods. The Analytical Methods and Regulation Development panel was supposed to figure out a way to get analytical methods and QA more visible in the regulation development process. You have to have a good idea of how our regulation development process works and where you can plug things in in order to make a change. And I think this group found out how to do that. The last panel was set up to look at National Laboratory Accreditation. First, is it feasible—can we do it—and second, is it advisable—should we do it? The QA Services panel is co-chaired by Tom Hadd from the Office of Research Program Management and William Hunt from the Office of Air Quality Planning and Standards. We tried to have two folks co- chairing each ad hoc panel. The people had to come from a different organization, and there were three main organizations: ORD, the regions, and headquarter's program offices. We would pick any of the two to try to get this covered and get people who were interested in the area so that we could get good representation on the chair so things didn't go too much one way or the other. The charter of the QA Services group states that they will address the issue of sustaining adequate funding for QA services and research provided by ORD. The QA Services panel did a QA Service and Research Needs Survey to identify needs; they're in the process of trying to quantify those needs and decide how much money we're spending now so they can get a good idea of how much more we need. They're also trying to find a way, or make recommendations on a way, to fund those needs in the future. QA services includes such things as training and QA research methods. Everybody has a slightly different idea of what we need and what they're willing to pay for. So this one's turning out to be very difficult. The next step will be to develop a Final Report and Recommendations for the FY 1993 Budget Proposal. We are trying to get this budget initiative together and are working hard to figure out how we can pay for it and how we can have a stable base. I think the most important thing that's come out of this panel so far is the recognition that when it comes to funding QA services and methods development activities, they always take a back seat to everything else. It's the first time I think that has been recognized by folks up the chain as a problem. So at least we've 12 ------- made headway that far, even if we're not able to get a budget initiative going this year. We hope we'll be ready to have a budget initiative next year for that one, so that we have money actually set aside and earmarked for QA activities and QA research and development. The Methods Integration panel is chaired by Larry Reed from the Office of Emergency and Remedial Response and Ort Villa from Region 3, Environmental Services Division. This is the group that is trying to figure out how to consolidate methods. Their charter is to evaluate the feasibility of standardizing analytical methods across media and programs. They have two basic activities that they need to address. One is how to consolidate existing methods or integrate existing methods; the other is how to get a handle on all the new methods that are getting developed in such a way that you don't keep proliferating lots of methods that are slightly different but do the same thing. They've begun by addressing existing methods and they're trying to find opportunities where they can consolidate those methods with a minimum amount of difficulty. There are three method integration pilot projects underway. They began with volatile organics, metals digestions, ICAP—those are in process and they feel that they will be done by the end of this fiscal year in September. The other two that they have just decided to start looking at are semi-volatiles and microwave digestion. They feel that microwave digestion will fit in nicely with metals digestion and they're hoping to have those two done by September '91. This group is trying to figure out what the best way is for it to be organized to address these issues, because this is a big area and there's a lot of work to begin. On this panel, we have a lot of other Federal agencies represented because many other groups use EPA's methods. So we have the National Oceanic and Atmospheric Administration, the Department of Energy, the Department of Defense, and a number of others because our methods are not just for our use but for other agency's uses too. There are some other issues: How do we format the method? Should we come up with a whole new methods manual in addition to all the already existing methods manuals that we have? How do we go about making them legal, that is, including them in the Federal Register? And when we figure out how to do that, do we make them supplement or supplant existing methods? There are lots of different strongly held opinions on how we should go about doing this. So those meetings are very interesting. And if anybody has two cents they'd like to throw in, I would encourage you to talk to David Friedman, who is heading up the work group that's dealing with this issue. The Automated Methods Compendium panel is co-chaired by Fred Haeberer from the Quality Assurance Management Staff and Bill 13 ------- Telliard from the Office of Water. This is, I feel, one of the success stories for EMMC in the first year. What the EMMC decided to do was to adopt the existing List of Lists system that the Office of Water had been putting together over the years as the Agency-wide tracking system, and to include methods from Superfund and RCRA and Air and other programs that they didn't already include in their database. Each of the other programs was going to kick in some money to help fund this and keep it going. It has gone out in a read-only pilot form to Region 3 and its states. I think about 36 copies of this went out to Region 3. Another 22 got distributed to folks around the Agency that Bill and Fred knew were interested. They're expecting feedback by the end of this month on how things went and then, based on that, they'll modify the existing system and make it available for distribution. We thought it would be a good idea to have this in EPA's library in headquarters so that the reg writer folks could have access to it, and then also have it in the regional lab libraries, the regional libraries, NEIC, and the three Environmental Monitoring Systems Laboratory (EMSL) libraries. We'll look at the results from the pilot study and make any changes. So far the feedback I've heard from people who are using it is that it does address which program the methods support, what the detection limits are, and what type of method it is, and that it's been very useful to those folks so far. The Analytical Methods and Regulation Development panel started out as the Quality Assurance and Regulation Development Project at one of these national QA meetings. Anyway, the Regional QA Managers decided that they didn't feel QA was getting enough attention in the regulation development process, so they started an initiative with Carol Wood of Region 1 as their leader, figuring out how they could get QA better highlighted in the regulation development process. When EMMC got started, and they were talking about analytical methods in the regulation development process, we volunteered Carol and said, "She's already leading this effort; why don't we just let her continue on as one of the ad hoc panel co- chairs." The other co-chair is Maggie Thielen from the Office of Regional Operations and State/Local Relations. Their charter is to develop a system for.factoring methods development and validation concerns into the Agency's regulation development4process. The panel was recently successful in getting the regulation development steering committee to adopt their recommendations, which is in the Start Action Request, the document that begins all regulation writing at EPA. They would have a question that says, "Does this method require environmental measurement?" And if the answer is yes, then the work group chair must go out and find somebody who knows something about environmental measurements and include them on the work group to make sure that whatever they come up with is implementable. 14 ------- In addition, Jim Weaver, who's from the Office of Regional Operations and State/Local Relations, and is the steering committee representative from that office, will contact the programs in the region and the ESD in the lead region to make sure that we get representation from the regions on the group. Finally, once it comes to workgroup closure, the chair of the Regulatory Steering Committee, who is, I think, Tom Kelly from the Office of Policy Planning and Evaluation (OPPE), will ask the question, "Is this rule implementable from the perspective of environmental measurements?" If the answer is no, it doesn't go to red border; it stops until they can work it out. If it's yes, it can go on to red border, which is the final step within EPA in getting something ready to go into the Federal Register. This was a big step, because it's very difficult to get something included in the guidance to reg writers. So we feel that this one's quite a success story, too. They're going to try it out for a year, and when the year's over, evaluate and see how well it worked, how should they change it, and if they should leave it alone. We felt strongly that the national program office QA Officers needed to play a strong role in this, because they're the folks who are going to know what regs are coming forward and are already being worked on. We wanted them to be our first line of defense to notify others in case they feel that's not being addressed. I think one of the driving forces was not whether there were court challenges or not, but how well the regional folks felt they were getting methods in time and appropriate QA and proper quality control (QC) requirements in time to carry out the regulations as they came down the pike. In the spirit of TQM, this will be one incremental step forward in improving our whole process. The last panel is the National Environmental Laboratory Accreditation panel, which I co-chair with Jim Finger, the ESD Director in Region 4. We were charged with two things: first, to decide if a uniform National Environmental Laboratory Accreditation program for labs performing environmental testing procedures was feasible, and second, if it was advisable. After a number of meetings and a lot of disagreements, we agreed that National Environmental Laboratory Accreditation is probably a good idea and that it would benefit the programs. We thought that if we decided we wanted to do it, we could do it. What we didn't decide is should we do it. And we didn't decide that because we felt what we needed was a Federal Advisory Committee because the reason this all started was that EPA's Deputy Administrator F. Henry Habicht got a call from the private lab industry and said we really need national environmental lab accreditation. We felt very strongly that we ought to hear from 15 ------- all the users of environmental lab data and not just from EPA. So our ad hoc panel did include one state representative. Initially, we had some folks from the lab community and were told by general counsel that we were not allowed to do that. If we wanted to talk to the lab community or any of the users of laboratory data such as API, we had to set up a Federal Advisory Committee. So I'd say for about the last five or six months we've been in the process of trying to figure out how to set up a Federal Advisory Committee and get it going. That committee will consist of representatives from the states, from EPA and other Federal agencies, from the laboratory industry, laboratory associations, and trade associations to make recommendations to the Deputy Administrator on what they think would be valuable to them in terms of a National Environmental Lab Accreditation Program. We felt that one of the keys to having a National Environmental Lab Accreditation Program was getting the states on board. Without the states, it seems to me we'll just be layering another bureaucratic program on top of a lot of already existing programs. That's why we felt very strongly that we had to have a lot of state representation on this Federal Advisory Committee. If we can't get reciprocity, which seems to be one of the big problems, then I'm not sure we're ever going to be successful with National Environmental Lab Accreditation. We did invite the National Institute for Standards and Technology (NIST) to come and talk to us about their lab accreditation programs. They felt very strongly that if they were going to set up a program it would have to be centralized; they would have to have complete control of the program from NIST in Maryland, and based on the existing programs that we already have—almost every state has some kind of a lab accreditation program—we felt that wouldn't work at all. So the NIST model doesn't look good for what we want to do. We're hoping to have our first Federal Advisory Committee meeting in June. Our goals within the next year are to identify the critical elements and design of a national environmental lab accreditation program, identify the benefits to all the users of laboratory data, and provide options on how it could be funded and how it should be managed. So that's what we hope to do in the next year on National Environmental Lab Accreditation. We have talked to the Health Care Financing Administration, which has recently set up a program where they have to certify medical laboratories, and we invited them to come and tell us what they were doing and how well it was working. They declined and said it was too soon to tell and that they would have something to talk to us about later; so we're going to try them again, since it's now later. The Clinical Lab Improvement Act was passed by Congress in 1988, and they have to set up a clinical lab accreditation program. They 16 ------- proposed the requirements for that about a year ago in the Federal Register, and that's something in excess of 20,000 comments. Under court order they have, just recently I think, re-proposed because of the extent of the comments they got. But they do have some very good and interesting techniques in there for monitoring performance of the lab. Some of the states have said they want a Federal regulation that says you have to participate in this. And some of the folks we've talked to have said they want a voluntary program. So one of the things we need to recommend is whether it's going to be Federally mandated or not. What the private lab community is telling us is that if we set up a national program, they will participate because they'd rather be part of a national program than subject to each of the individual states' programs. Jeanne Hankins of the Office of Solid Waste has agreed to serve as our executive director of the Federal Advisory Committee and to handle these lab accreditation issues for us. I'm looking forward to having her join us, because this is a big issue and I haven't been able to give it as much of my time as I would have liked. We welcome her to this project. Just to sum things up: the two big achievements, I think, are the Environmental Monitoring Methods Index (EMMI), which is the automated database that Bill Telliard and Fred Haeberer were working on and the QA Services group, which is still struggling to convince program offices and others how much is really being spent, how much needs to be spent, and how we should go about funding QA needs and methods development needs in the long term. That's all I have to say about it. 17 ------- HARMONIZATION OF QA REQUIREMENTS ACROSS THE GOVERNMENT: INTRODUCTION Nancy Wentvorth EPA Quality Assurance Management Staff We're going to be talking about the differences and similarities between QA programs across the government. We have representatives from the Army, the Navy, the Department of Energy (DOE), EPA, U.S. Geological Survey (USGS), and the National Oceanic and Atmospheric Administration (NOAA)—a wide range of people dealing with many different programs and many different concerns. One of the things I've enjoyed most in QAMS is the ability to meet people from across the Agency and across the government. All of us in QAMS enjoy the opportunity we have to learn from other people. Harmonization of QA requirements is one of the areas that we have learned a tremendous amount about over the last few years. We've talked to many people across the Agency and, as part of our responsibility for oversight and review, we've looked in detail at a number of programs and found that there are many common concerns manifested across the country. We have found that the exchange of information has provided us with a tremendous resource. One of the major things we have recognized is that the guidance that we have available to the QA community within EPA is outdated. It's in need of revision to better reflect the QA program that we are now operating. Many of you are aware that the QAMS-005/80 guidance is approaching 11 years old. The QA program in the Agency when that was written was not the program that it is today. There have been many strides forward and we need to have our guidance accurately reflect the current program. One of the things that is a priority to me as the Director of QAMS is to bring our guidance up to date to the program that we're operating now, and the program that we see is needed across EPA and all of the environmental programs. We've learned a lot from you. And one of the things we've learned is that we need your help if we're going to be able to revise the guidance and improve it and make it usable across the Agency. We realize that the guidance has to be readable; it has to be easy for people to translate into their day-to-day environmental monitoring operations. It can't be written in hard tech engineering dialect if it's going to be used in laboratories where people are used to the chemistry dialect. It has to be in a common dialect that we all understand. We're trying to look at the issue from a Total Quality Management (TQM) perspective—not from a narrow EPA-only issue, but from the wide spectrum of environmental monitoring operations that go on across the government. It seems appropriate, if we are going to 19 ------- improve our guidance, that we consider all of the needs of the other monitoring programs we run that affect other agencies in the government. As long as revisions are needed, it seems appropriate to incorporate the concerns and requirements of other agencies and to harmonize the requirements. I think that's an important thing to recognize. We're trying not to have EPA stand alone, but to have EPA stand with the other government agencies—to have a logical, reasonable, usable program across the board. I think that will help significantly to eliminate a lot of the problems and concerns that you share. A few years ago when we saw TQM rising as an important way of doing business, we made a point of talking to the QA managers of other environmental programs across the government. What we found was very similar to what was in our own Agency—that the same guidance was being interpreted differently across the country and in different applications. It became clear that there needed to be an effort to take some of the differences out of the program so that we could operate more efficiently. The vehicle that presented itself to us as a way of doing this was the American Society for Quality control (ASQC) Energy Division, which has an Environmental Waste Management Committee with representatives from EPA, DOE, Department of Defense (DOD), and contractor communities. This committee provided an opportunity for a small number of people to sit down and begin to work through the issues and concerns that their agencies had. Through this vehicle we were able to begin discussions on harmonizing a standard. The EQA-1, Quality Assurance Program Requirements for Environmental Programs, is a draft that has been prepared through this ASQC committee to begin the discussion and the widespread comment within the community to try to develop this consensus standard. We have spoken about this document at a number of meetings outside of EPA— ASQC meetings and other technical meetings—and the document has been widely distributed. We are seeking your comments. For this effort to succeed we must have your input on how QA programs can be made more common across the government. We'd like to get a wide range of comments from the QA community, both from within EPA and across the DOE and DOD organizations and the contractor community. We're interested in hearing from all of you. We have assembled two panels to talk about the differences and similarities in QA programs across government agencies. I think this is a valuable opportunity for us to exchange ideas. It's important for us to talk about our differences or similarities in programs, the differences in our needs, and the differences in our management and operating circumstances. For those of you who are not EPA employees, we need to know what 20 ------- your views are on the benefits of harmonizing QA requirements, the savings that you can see in that kind of a shift, and the types of concerns you have that are not, as you read it, addressed in the EQA-1. This is an interagency activity that is unprecedented. We are committed to improving EPA's QA program, EPA's guidance across the country, and EPA's clarity in explaining what people are expected to do to meet the EPA requirements. We are hoping that this effort will respond to a lot of those points. We*re also looking at this as an important team effort within EPA. I do not want any of you to think that what's presented to you is a done deal. We are in need of supporting and improving our QA program. We have made a proposal that we think responds to the concerns that have been voiced to us over the years. We need to hear back from you whether your concerns have truly been addressed and dealt with. With that information, we can revise, modify, make more user friendly, and make clearer the document that you have so that it can work for EPA. After the panel discussions, we will be breaking into discussion groups to talk about the issue of Harmonizing QA Across the Government. The purpose of that discussion is to talk about the process of changing our approach to QA and making it more consistent across government and the processes that you, as individuals with QA responsibilities in different organizations, will have to go through to change your program. We'd like to know what support you need from us, and what support you need from within your own organization to help implement change. I know from experience that change at EPA is often difficult, because people don't want to do things a new way or a different way. What we're trying to do now is begin both the continued refinement of EQA-1 or a new QA guidance system for EPA, but also the process of people looking at it and beginning to assimilate it into their operations. So I think we'll have some interesting discussions from a process perspective, not from a technical standards perspective. I'm particularly looking forward to hearing how other agencies are handling QA issues, what they view as their major differences and major similarities with EPA's program, and the results of the discussions on what kinds of process things we need to consider as we look at implementing changes in our QA program. 21 ------- INTEGRITY IN SCIENCE Ad11 E. Shamoo Editor in Chief, Accountability in Research; Policies and Quality Assurance I want to thank the organizers for giving me this opportunity to share my views with you. Some of what I am about to say has been said, but it's going to be from different perspective. I came into this area 10 years ago, when there were a lot of people suffering on both sides as the investigator and the field in general on issues involved in integrity of research. There are two key areas that we are trying to change in order to ensure the integrity of research: policy and individual investigators. Both areas are important; however, policy changes, I believe, can have quicker and farther-reaching consequences. Therefore, we need to address policy makers and emphasize to them that these policies can foster integrity in research or breed fraud, misconduct, and sloppy work. And that's basically what we are dealing with here; that's quality assurance: to prevent and reduce sloppy work. To influence individual investigators without changes in policies is nearly impossible. However, even if policies are appropriate, changes in individual investigators will be slow. Are the current concerns regarding the integrity of research data an indication of: (a) the decline in the ethical values and conduct of research by investigators; (b) a new awareness of an old problem that is produced by a small and negligible number of sociopaths and deviants of our society; or (c) an old problem, but increasingly larger enterprise makes them appear bigger? What are we talking about? Are we talking about fraud, misconduct, careless practices or error, or all of them? In order to answer these questions, an historical perspective of the problem is needed. Researchers at the turn of the 20th century were no longer monks on top of a mountain. After the industrial revolution and specifically after the first and second World Wars, the research enterprise grew rapidly. In the United States alone there are now one million research scientists, two and a half million participants, and a budget exceeding 160 billion dollars annually. So it is a big business. Just a few examples to illustrate some of the irregularities in the research area in the 20th century; first, the famous story of the Piltdown Man. Piltdown is a city near London, where in 1908 fake skull bones were found. The bones were made to appear thousands of years old and made to look as if the skull was a mixture of a monkey and a man, thus proving that man came directly from apes. Charles Blinderman—and this is the last paragraph in his book The 23 ------- Pijtdown inquest (1986)—said something which is true at all times. He said that anyone conversant with Piltdown and history will readily, if not eagerly, agree that many of the researchers shaped reality to their heart's desire. Protecting their theories, their careers, their reputations, all of which they lugged into the ^pit with them, because that's where it was found, in the pit, in the Piltdown. This story lasted for 40 years. There were over one thousand research papers and Ph.D. theses written in support of the fake skull bones. An example now of an error that affects national policy: A 1985 report on Social Security statistics, presented by Martin Feldstein, who was a chief economic adviser for the President, said that an attempt to replicate Feldstein's construction of Social Security wealth revealed that his series was incorrect. Feldstein has acknowledged that a computer programming error was made in incorporating with those benefits provisions of the 1956 amendments to the Social Security Act. As a result of this error, his Social Security wealth series grew rapidly after 1957. By 1974 the series was 37 percent larger than the correct value. Our national policy was based on those error numbers. There are numerous other cases of alleged and proven cases of errors, misconduct, and fraud. We have been deluged recently with investigations, reports, and Congressional hearings regarding misconduct in science. What are the issues of data integrity? They are: 1) falsification of data; 2) plagiarism, which is the hardest one; 3) suppression and selection of data; 4) misuse of privileged information; and 5) what you are concerned with—poor data quality. What are the causes of these problems? In my view the causes are numerous, but they can be categorized into two broad areas: institutional in nature and generic in nature. The initial institutional response to the problems of data integrity is like a sick patient being told for the first time he has cancer—his denial, defensiveness, combativeness, then, of course, followed by reluctance. However, due to public pressures, research now has new regulations—the National Science Foundation (NSF) has, the National Institutes of Health (NIH) has, EPA has, and the Food and Drug Administration has. But none of the regulations or their basic foundation acknowledges the very generic nature of the problem. Thus, there is no development and formulation of policies and procedures to prevent, deter, deal with, and remedy the problem on a continuing basis. In his book, The Strategy of social Regulation; Decision Frameworks for Policy in 1981, Lave summarized our society's dilemma with regulations. Americans can live with social regulation despite its cost and disruption and can't live without it because of the strong public desire to curb the worst abuses of an industrial economy. That's what you're dealing with; if all 24 ------- industry were perfect, EPA would not exist. Neither of the extreme choices—putting more resources into the regulatory agencies and' increasing promulgation of rules versus eliminating the laws and agencies—is viable. Americans have no choice but to learn to accomplish these social goals with less controversy and greater efficiency. Sociologists such as Emil Durkheim, of the late 19th and 20th century, and those who applied his theories to the sociology of science in the past 50 years, such a? Merton and Zuckerman, tell us that deviant behavior is due to "breakdown in values," whether in the community at large or a sub group, such as the scientific community. Furthermore, sociologists deduced that "social control" was the guardian of norms and values in science. In 1977 Zuckerman stated: Social control and science depends partly on scientists internalizing moral and cognitive norms in the course of their professional socialization and partly on social mechanism for the detection of deviant behavior and the exercise of sanctions when it is detected....They must also provide for the detection of deviant behavior and for the exercise of sanctions when it occurs. In science the institutionized requirement, that new contributions be reproducible is the cornerstone of the system of social control. It has two functions, deterrence and detection. Zuckerman then realizes the potential in her argument when she states: Critics of the social organization of science contend that in all fields insufficient incentives are provided for regulation, and, of course, as long as reproducability of scientific results remains an ideal not often realized in practice, it cannot serve as a deterrent to the cooking of data. This is exactly one of the reasons why dependence on social control does not work. More importantly the cost of reproducing large and complex research is financially, in current day economics, prohibitive. It is also undesirable for society to wait for a very long time such as 5 to 10 years to repeat the study when the research results address an important public health or the security of our society. As a matter of fact, the majority of research, by its very nature, is not contestable, and so won't be reproduced. Not too many research findings fall in the category of cold fusion where everyone wants to repeat those experiments. Social controls also fall apart where research scientists promote a product, or a drug in which they themselves have large financial 25 ------- interests. We all are under certain kind of pressures to cut corners. Be aware of them. You in the environmental field acknowledged these problems years ago. In the past 20 years you have developed a framework for how to deal with the issue of the integrity of data. Furthermore, your policies continue to keep up with the changes in the field and the profession. We in the research field, especially in the sciences, have a long way to go since we barely admit there is a problem. This audience is easy to convince that these generic problems exist in the research field; moreover, that these type of problems exist in any complex and large enterprise. Remember, research and development is a $160 billion business per year. The generic problems are really three. one is conflict of interest. To me, that's the most important one. It exists in every aspect of what you do and what research and development does. Second is complexity and third is remoteness. Research is no longer conducted by a certain individual in a small laboratory, but rather in a large and complex group with a differing expertise. Therefore, laboratory directors rely on numerous portions of research data from different areas. Furthermore, the project leader may physically be at a remote location from the research laboratory, just like the people in Washington, as well as disconnected from the daily bench research. I want to go back to the conflict of interest issue, which I think is even more important and critical in research than in any other system. For example, in the financial world one can compartmentalize and isolate several functions that have inherent conflicts of interest. However, in research those who collect, analyze, and manipulate the data are one and the same. Furthermore, in research it's not only undesirable to compartmentalize, but rather nearly impossible. In a fertile research environment you want all those involved in research to know and have access to the data on a daily basis. Accessible research data is as crucial for each subsequent step in research. Justice Learned Hand describes conflict of interest in 1939, in the Yale Law Journals Our convictions, our outlook, the whole makeup of our thinking, which we cannot help bringing to the decision of every question, is the creature of our past, and into our past have* been woven all sorts of frustrated ambitions with their envies and hopes of preferment, with their corruptions, which long since forgotten, still determine our conclusion. A wise man is one exempt from the handicap of such a past. He is a runner stripped from the race, he can weigh the conflicting factors of his problems without always finding himself in one scale or the other. 26 ------- Let me give you a recent example of the insensitivity of the research community to the issue of conflict of interest, or the perceived conflict of interest, because it is equally important. NIH appointed a committee due to Congressional pressure to investigate allegations against Dr. David Baltimore, a Nobel Laureate. The Magazine section of the Washington Post last Sunday was about Walter Stewart and Ned readers, two of my colleagues at NIH who are called "fraud busters," who have been after it for four years. Let me know that I am not here to address the merits of these allegations. The original committee appointed by NIH consisted of three people. One member was a former post-doctoral fellow and co-authored with Dr. Baltimore 14 papers. None of these papers, of course, was the one in question. The second member co-authored a textbook with Dr. Baltimore. Because of Congressional pressures the composition was, of course, changed. As all of you know, a recent draft report by NIH has claimed and pointed a finger of misconduct by the principle author of the paper, and the paper has been withdrawn by Dr. Baltimore himself. Rough estimates of the amount of questionable research, that sloppy work, ranged from 7 to 12 percent. These estimates, as one would predict, have a tremendous immediate economic impact. In other words, we have close to $10 to $15 billion dollars a year spent on questionable research. For each expenditure of one dollar in research, we have an additional $10 to $30 dollar impact on the economy in the long range. One can see the huge annual economic impact of shoddy research on our economy. These numbers on the economic impact of continued questionable research are detached from its moral and ethical impact on our society. In this century errors, deceptions and fraud have lasted too long, as in the case of the Piltdown man—lasted 40 years or have had tragic consequences, as in the case of the Challenger disaster. It is apparent that our society cannot fully depend on the current system of extremely defused and ill-defined accountability. The primary source of all the problems with integrity of data is the conflict of interest. The other two are complexity and remoteness. There is little you could do about them, because that is the very nature of our huge enterprise. The conflict of interest you could do something about. On May 6-7, I am presenting a paper at Georgetown on Ethical Issues in Research. I have taken statistics for 10 years, showing that industry advisory councils are the top group at most Federal agencies. And I discovered that for the same merits priority score, these advisory groups have probability of getting funding twice. These are the people who write the programs, the direction of the program, and where the money ought to go. They can't be objective because the statistics for 10 years indicate they're getting twice the money for the same priority score. 27 ------- The same thing with industry. There is no way unless they are Mr Spock of Star Trek; he's all the time pre-programmed to be objective, right? People don't behave that way. I will not behave that way. No one does. We all have conflict of interests. And we've got to avoid it to have objective opinion. Try to get people with independent sources of judgment, who don't care what the outcome is because it doesn't affect them whatsoever. No one of us is devoid of all these connections. There is a strong argument toward getting the most experts to be the most objective reviewers of that project. You have two issues: one is expertise; the second is the fact that if they have a conflict of interest they will exercise censorship. A great deal of creative work might be hampered and you've got to weigh the two. I review grants for Muscular Dystrophy and the American Heart Association; I serve on the policy committee and peer review committee of several organizations; I've even been on NIH advisory for NSF. I would say that 80 to 90 percent of the proposals I have reviewed are not directly in ray area. I don't know these people. I don't eat and drink and meet them all the time in meetings and know their wives and their children. I am much more objective when I don't know the person. I believe I am a competent scientist to be able to spend an extra two hours to evaluate a project not directly in my own very, very narrow field of expertise. And I think I would be much more objective than the expert, because the joke is, in science, that you can nickel and dime any proposal to death when you are an expert. The American Heart Association uses huge councils. They review all the research grants, and 99 percent of them are not in their area. And they do a darn good job in evaluating and funding money. They've had tremendous success in funding excellent people for the past 20 years. I want now to turn my attention to the issue of QA and peer review within the environmental data quality. I have spoken at length with about 10 people from within and outside EPA. Let me first define my terms the way I have developed them for the past 10 years from their use and the literature in all available fields of finance, research, science, health care, quality assurance, and policy analysis. By the way, the field of finance is ahead of us in terms of accountability and data audit. That doesn't speak much about the S&L crisis; but the S&L crisis is basically the failure of accountability, rather than the lack of the science of accountability within the financial field. First, what do I mean by quality control? QC concerns itself with ensuring the quality of products and services during the numerous production or service steps in order to guarantee compliance with the originally agreed-upon specifications. Most of what I know from people—what they call QA is really QC. The specs are already 28 ------- predetermined and you're trying to follow them to the letter. QA concerns itself with reviewing the quality of products and services after the products or services have been used by the consumer for a certain length of time. What I perceive in the research and development of what is the value of QA: there are three and they are all related to R&D. And everything else emanates from these three things: evaluation of the current R&D activities, actual performance of R&D, and selection of future R&D projects. That's basically what you do most of the time. Selection of future R&D products helps you to evaluate where you should go next. What are the generalized methods of achieving QA? This is, again, not necessarily only for EPA. An internal QA/QC unit, in my view, should be down to a minimum number of staff. Guidelines on conflict of interest. Training and education—that's so important in QA to all of our personnel, starting from top management. Remember, all of us were trained. But we got our college degree, or master's degree, or our Ph.D. without any concern to QA. So we are almost too late. The new generation is where it's important; as for us, we have to be retrained and retrained. Retraining is harder, just like the infant when he learns a language, it's easier to teach it to him when he's less than six years old. Use of QA/QC consultants, use of outside independent data auditors, implementation of outside QA data auditors' recommendations, retraining and education of internal QA/QC staff, use of independent external QA data auditors whenever a new event occurs— that is, you have a new project, or a new direction in the company or in the agency. I want to give you my definition of what a data auditor is. I have read the entire 3,000-page document of the Food and Drug Administration on Good Laboratory Practices. They have no definition of what data audit is. None of these documents have a definition of the auditor either—they just go around it. A data audit is the systematic process upon which objective evidence is obtained and evaluated as to assertions about research data and their value to determine the degree of correspondence between those assertions and values and established or predetermined criteria which can then be communicated to interested parties. This is work that took us two years. Steven Loeb is a financial auditor, one of the top in the country, and an editor-in-chief of Accountancy in Public Policy. We searched the literature from the 1930's till now, on how the word data audit is used. John Lawrence, the Director of the Institute of Quality Assurance in Canada, said (this is going to appear in the next issue of 29 ------- Accountability in Research, which is devoted to environmental data quality assurance): "The three legs of a quality assurance program: a quality assurance management plan, quality control, and quality assessment." Here is how he defines his terms. Stringent requirements means that the entire measurement process, from initial site selection through to data interpretation and archiving must be thoroughly quality controlled. Rigorous QA procedures and protocols must be an integral part of all monitoring programs if reliable, traceable and compatible data are to be generated. The integrity of samples must be ensured throughout collection, handling, and analysis. Then he continues to define quality assessment. "Quality assessment is administered under the quality assurance management plan, preferably by a neutral third party." The whole thread of what I've been talking about is having a neutral third party—a source of an independent judgment, I call it. It has a management and a scientific component. The managerial component through interagency comparison study—and the audits—provides the necessary information and advice to managers on overall credibility of data and the suitability of field and laboratory protocols and of the effectiveness of internal QC. Based on these definitions and my rudimentary understanding of the EPA program, I would like to make a few comments, first about my understanding and appreciation of the EPA peer review program. It reviews, primarily, reports; no QA program probably is in it, and no data audit is in it. It is not a focused responsibility. I'm trying to instigate you to think and get angry and ask questions because I think you need an outsider to say these things^-conflict of interest issues are not clear. The reviewers may not be neutral on the subject or the outcome. How truly independent are they? That wasn't clear to me. In house peer review, is it truly objective? That is, once you know the staff, you know everybody in the program. Peer review may coincide with the end of the program. It could, therefore, be too late, and too wasteful. It has no value—the program is finished. I guess it will help in subsequent projects which are similar. Peer review should not be a substitute for a QA program. How binding are the peer review opinions? Otherwise, what are they there for? Is there a follow up? Is it really a single event only? How thorough could a peer review be when you only spend just a few days to examine the data? Is there truly a critical review of the data, of the raw data or original data? I would like to make a few comments on the QA program within EPA. One: Is there a philosophical commitment of QA by everyone, the entire EPA? Is there a QA agency culture that is similar to corporate cultures? Is there such a thing? Because I believe some 30 ------- pharmaceutical companies, especially the modern ones, have a true QA corporate culture within them. Two: Lack of concept of evaluation of data quality and why. Three: Does it truly audit raw data? I was not able to decipher that from all the documents I read, or from all of the 10 people I've talked to. Four: How are all QA programs integrated? Five: QA training of scientists. All these university contractors have zero QA training. They not only don't believe in it, they think it's a pile of junk. Could scientists be trained as QA officers? Six: QA, in order to be effective today, should have access to all data, including industry confidential data. The way I envision QA is after the fact—to look back and peer review and assure the data after a time has lapsed of the services or products. Again, you have to look at the literature and how it was used by all these other fields. Once you know the specs and you've put them down in writing, and you're telling all the subordinates how to do it, that is no longer a QA program; it's a QC program because you know what to expect from them. QA is you evaluate the QC program after two or three years; you change the QC program; you make suggestions. The procedures may be wrong; new methodologies may be introduced; new science may be introduced— it's much more global in nature. That's how I view QA, rather than as primarily a check list approach. This is based on the literature of the past 40 years. And part of my job in writing and urging a lot of people to write on this subject is to have a basis of sharing information. The Journal is one, the book is another, the conferences that I organized three years ago was one, the next one in Rome is another one, another one is going to be in Washington next year. This is the basis of why we share information. How many books or literature written, for example, or articles on the EPA QA program were read by everyone? Not a lot, and that's important. I'm inviting you to write your views, to disagree—that's what academic life and intellect is all about. I will end by giving some bold, overall strategic suggestions. One: Scientists and QA personnel cannot run away from the implication of the data on policy and regulation. There is no such thing called pure science without an impact on other parts of our society. I heard a lot about "we're only interested in science." They should become involved in policy and regulation—you and scientists. Two: There should be one overall QA program where peer review should be a component of QA programs. QA programs are a permanent fixture and the peer review is only a temporary event to augment an ongoing QA program. Three: QA programs should be augmented by truly independent outside reviewers who perform a truly scientific data audit. 31 ------- Four: Discovery laboratories. Exploratory laboratories that test market new ideas and are nonprogrammatic, not of the basis of any paper or a policy, but purely the private notebook of a scientist, should not be the subject of any QA or data audit. This is to protect the creative process. In the 1990's our society will demand a greater and greater accountability for the conduct of each professional group's action. Society will not give each professional group their unchecked license to do what they wish on the principle that they know best. They will not do this for the military, doctors, accountants, scientists, or even QA personnel. 32 ------- JOURNEY TOWARD QUALITY: THE FEDERAL EXPRESS STORY Gilbert Mook Vice President, Properties and Facilities Federal Express, Inc. It's my pleasure to represent Federal Express, especially to be out preaching the quality gospel. We have an opportunity to go out and talk to a number of groups about our.quality program, and I've had the opportunity to meet with a lot of different organizations, both government and non-government, and I've seen some interesting things going on—a lot of positive activity. One of the things we found is that no matter what business you're in, chances are that you have a shelf full of "how to" books. Everyone wants to know the secret of success, and we can always go to a host of self-appointed quality gurus who are willing to let you in on their secrets. Unfortunately, what works for one company or one organization doesn't necessarily work for another. And that's why many organizations, including Federal Express, employ many different techniques and methods in their ongoing efforts to define and refine their quality process. And so, when people come to us and say, "What did you guys do to win the Baldrige Quality Award and how can we incorporate the special secret into our quality process," unfortunately we don't have a one-size-fits-all answer. What we can do, perhaps, is go over three fundamentals that have guided us. First of all, customer satisfaction starts with employee satisfaction. Second, service quality has to be measured, and third, customer satisfaction is everyone's job. These points are the basis of what I'm going to talk about with you this evening. Our focus at Federal Express on quality began the day that we began. For although we were really a maverick at the time in the distribution industry, our objective then, as it is now, was to provide timed, definite delivery for high priority documents and packages. Fred Smith, our founder and CEO, was,first to apply the hub-and-spoke concept for distribution practices. At the time there were no precedents to guide us because this was a unique approach; consequently, we had to create practices and processes as needs arose. Though many of these processes have changed and our operations have expanded over the years, the focus has always been on delivering quality service. When we began operations in 1973 we shipped eight packages on our first official night of operation. From these humble beginnings, we've become the world's largest air express transportation company, delivering over 1.6 million items to 127 countries around the world each working day. Our fleet today consists of more than 33 ------- 400 aircraft, which includes Boeing 747s, DC 10s, 727s, F-27s Cessna 208s. On the ground we have more than 35,000 computer and radio-equipped vehicles. Also, we have 94,000 employees to support our worldwide operation. w*".*. To expand so far and so fast, we've had to continually reassess our policies, our procedures, our logistics, everything. In fact, our focus has always been on 100 percent customer satisfaction. From the beginning we've been guided by a simple but profound three-word corporate philosophy: People, Service, Profit. And this is as relevant today as it was nearly two decades ago. The philosophy guides the setting of our annual corporate goals, and we have one goal for each of these three elements. Our measurable people goal is the continuous improvement of our management leadership index score, which we track through our annual Survey Feedback Action program. I'm going to tell you a little more about that in a few minutes. Our service standard is 100 percent customer satisfaction. We had to guide our efforts. In order to do that, we've created a unique measurement system of service quality indicators. This index measures our weekly, monthly, and yearly progress toward achieving our 100 percent goal. Our profit goal, much like any other company's goals, is fundamental to our long-term viability. If you don't make a profit, you can't sustain growth. To sum up our People Service Profit philosophy, we believe that if you place your people first they will, in turn, deliver an impeccable level of service, which is demanded today by our customers, and that profit will be the consequence. The essence of our people-first policy is that customer satisfaction begins with employee satisfaction. Let me share with you some of the methods that we use to demonstrate our commitment to a people-first philosophy. One of our most important programs is our annual Survey Feedback Action (SFA) program. SFA has been a part of our quality process for the last 11 years. The survey gives people the chance to express their feelings about their managers, their service, and about pay and benefits. Once a year every employee within every work group anonymously fills out this surve'y. A portion of the survey includes a series of statements concerning the immediate manager's leadership abilities. One such statement may say, "My manager asks for my ideas about work." Or another: "I can tell my manager what I think." Or: "My manager tells me when I do a good job." In each case, the person filling out the form may respond either favorably or unfavorably. While the individual responses are kept 34 ------- confidential, the overall results of the survey are passed onto each manager, who then must meet with their group to develop an action plan for resolving any actions that were identified in the survey. So the survey gives useful information regarding individual managerial strengths and weaknesses. Just as importantly, all work group results are integrated into an overall corporate leadership score. These scores are then used to diagnose the corporate-wide leadership problems and in addition become part of management's overall objectives. And here's the rub: this is how they get your attention, because these scores are then tied to incentive compensation for both managers and professionals throughout the corporation. in fact, if the company's wide leadership score isn't as high as it was the year before, no one in management receives a bonus. That's incentive. So the survey feedback action encourages strong, even-handed leadership and open two-way communication. Another people program is our Guaranteed Fair Treatment (GFT) Program. The aim of the GFT process is to maintain a fair environment in which everyone who has a grievance or a concern about his or her job, or who feels that he or she may have been mistreated for whatever reason, can go through and have these concerns addressed through the management chain. A team of management weekly reviews GFT cases that have not been resolved and have progressed up through the three-step internal process to the final stage, which we call the appeals board. What we've found over the years is that this is not the fastest way to address an issue, but we think it is the fairest. Both the Survey Feedback Action program and the Guaranteed Fair Treatment program promote open communications. By creating this environment we have found that people are more apt to take part, to make suggestions for improvement, to question decisions, and to surface concerns. We work hard at keeping these lines of communication open within and between divisions, departments, management, and the front line worker. For example, we have an 8:30 a.m. operations meeting every day and this serves as an example of some effective cross-divisional communication. At these meetings divisions representatives from our various divisions around the world come together to discuss major operational problems encountered during the previous 24 hours. We run essentially an entire war plan every 24 hours. So people are gathered around the table or participating via world wide conference call—these people will determine who is going to solve each problem that's been identified, how it's going to be done, and all these action plans must be developed and implemented within 24 hours. This continuous check on quality of service enables us to find ways to reduce failures. It's also another way to communicate effectively. 35 ------- One of the things we've found over the years is that gone are the days where senior management used to be able to sit down and face the entire work force to discuss problems. Today we've gone to the high tech solution. We now have a technology to share information quickly and effectively via our satellite link television network, FX-TV. New information is quickly and effectively shared through live phone-in question-and-answer sessions between top officers of the company and employees. This front line feedback is vital to the quality process. Today all major presentations are broadcast over the FX-TV network for employee viewing. Effective communication is critical to keeping everyone aligned with our goals. Training is also fundamental to the success of our quality process. Employees must know what's expected of them and be given the proper training in order for them to be effective in their jobs. At Federal Express all customer contact people receive extensive training before they assume their jobs and deal with customers. For example, our call center agents are given six weeks of intensive training before they'll ever take the first call. Senior agents provide one-on-one coaching to help the trainees become familiar with the computer terminal and its functional screens. Every six months couriers and service agents and customer service representatives must participate in a job knowledge testing program. These tests are online and can be taken at any one of our over 25,000 terminals around the world. After the tests are completed, the computer tallies the score and stores it in the employee's training record, within 24 hours the agent or courier receives the pass/fail results of his test. Along with the test results, each person receives a personalized prescription that targets areas requiring review. It gives them a list of resources, training materials, and interactive video lessons to help them get back up to speed. We're very serious about training. If, for example, a customer service agent were not to pass the test, he or she would be relieved from duty for eight hours of remedial training. This training can be obtained through interactive video lessons. By going through this training, employees would progress at their own rate until they're ready to take the tests again. So all our training directly supports the continuous improvement in the quality of our service. We also have a variety of reward programs that encourage people to work toward providing the highest possible quality of service. Pay for performance is written into everybody's job description from myself as a senior officer, down to the courier that you may meet in your workplace. A manager can bestow what we call a Bravo Zulu Award, which is U.S. Navy jargon for "well done." This can be bestowed on the spot. This commendation is given for clearly going above and beyond one's job responsibilities, and can bring with it monetary reward in the form of a check, or a non-cash award, such 36 ------- as tickets to a dinner or theater tickets. For the non-management employee we have what we call the Golden Falcon Award, which is a recognition of service above and beyond. Teamwork is applauded through a monthly Circle of Excellence Award given to the top performing work unit or station in the field. We've found that when people have the autonomy to make decisions that affect their performance and its outcome, and they have their ideas listened to and acted upon, they have greater ownership in their job. I'll give you an example: We didn't just create a tool like the Cosmos Super Tracker and hand it out to our couriers for them to use. We asked for their input. We asked them to help us design it and work with us to make it better before we even rolled it out for the first time. By providing real time information through our Cosmos Tracking System, we are proactively trying to reduce customer dissatisfaction. One of the things we've discovered is that a customer not having information about a package is equally, if not more frustrated, than receiving a late package. According to studies that we've made, 70 percent of customers with complaints don't complain, they just go away—often for good. To avoid this consequence, we believe that you must quantitatively measure service quality. Many of us who have studied quality have read of W. Edwards Deming. He said that you cannot manage that which you can't measure. For many years we measured our service levels by measuring our success rate, the percentage of on-time deliveries. And, indeed, we found out that 99 percent looked pretty good. But in 1985 we realized that if we really wanted a true picture of our performance, we needed to begin our performance by our customer's standards and perspectives. We initiated customer satisfaction studies that afforded an in- depth look at the way our customers perceived our service. We set about interviewing, quarterly, a sample of our customers who shipped with us—some exclusively, some with our competitors. They were each asked to participate in a 20-minute interview covering about 50 areas. Two years into these studies we realized that while we were becoming more aware of customer needs, we were missing the mark because we were falling prey to the law of large numbers. One of the things we found is that even at a 99.1 percent success rate, this translates into 2.5 million failures a year. That's 2.5 million customers who may go away. By 1987 we decided we needed a more rigorous method to measure the quality of our performance. So we turned our measuring stick upside down. We stopped measuring the percentage of our success and began measuring the number of actual failures. Our tracking system provided the data base. From this we developed our Service Quality Indicators. We call it SQI. SQI is a service measurement index which tells us exactly how much 37 ------- improvement we've made in reducing our errors across 12 critical categories of service. Here's how the system works: Weights are assigned to each component, from one to ten points per failure, according to how bad these failures would frustrate a customer. For example, a late package is weighted at one point, and this, indeed, is less upsetting than a damaged package that's weighted at 10 points. The nature of the failure is just as important as the total number committed. What we've done is to go out and measure all the things that annoy our customers and then measure ourselves against that. A missed pickup is a pretty big hit. That's when someone calls and says, "Hello, I have a package." And no one shows up. That's an egregious error. We also have lost packages. What could be worse? Damaged packages. So these points are assigned to each one of these sins as we go through. One of the things we found is that often a customer may not even be aware that we've failed. For example, suppose we guaranteed a priority one service by 10:30 the following morning. If we were to deliver that package at 10:31, our system that counts the delivery knows it's a failure, even though the customer may not even notice it. Why would we go to the trouble of docking ourselves for packages delivered just one minute late? The answer's quite simple. One hundred percent customer satisfaction is our goal and nothing less. All of these efforts, which involve the setting of clear goals, the creation of a people-first environment, the cross divisional communication, the extensive use of technology, training, rewards and recognition systems, and sophisticated measurement systems, have been encouraged and developed so that people are prepared to deliver quality service. Part of improving our process has been the development of Quality Action Teams (QAT's). And through the years we've seen more and more people from all areas participating in these QAT's. The critical success factor is that no one comes closer to the expertise about a particular job as the person who is doing the job. We now have a process of measuring quality efforts, and once a quarter each division goes through an internal selection process in which they choose their best quality success stories. Teams chosen are considered to be the best of the best. We have a ceremony; team members present their quality success stories to management; and some of the results have been astounding from all levels. One of the things that I've had a difficult time preaching to people about quality is that every person in every job is an expert about his particular area, and it's those little things that add up. Those are the real expressions of quality, as opposed to coming in with some sweeping managerial changes. 38 ------- Last year, for example, we had a QAT from our Memphis Super hub. They developed a recycling plan. We are now recycling such things as steel, batteries, plastic, wood, pallets, tires, waste oil, paper, and the results in one year—just in this group—are that they saved over $200,000. Now this program is company-wide and we're recycling all over the world. We talked about our 12 service quality indicators. Well, we formed what we call 12 root cause teams to enhance the quality process which looks at each one of these service quality indicators. And although each one of these teams is led by a corporate officer, there's really no star quarterback. As with all the quality action teams, every member is equally empowered regardless of whatever his position may be in the organization. So everyone, especially our frontliners, has the option, the time, and the power, to deal with customer problems. That empowerment translates into better service and more satisfied customers. We formed quality action teams across divisions to improve responsiveness to our internal relationships. When analyzing the way people rely on one another throughout our operation, one of the things that we found is that everyone's job is to support someone else. That someone else may very well be their internal customer. So, if you can do a good job in providing service to your internal customer, when that service reaches the external customer, satisfaction is built in at every step and it then becomes a permanent fixture. What models did we use at Federal Express? We had a quality guru organization come in and give us a list of procedures. Basically it was a language. It seems to me that the most important thing is to get everybody talking the same language so that they can communicate. And that's all it was. Our corporate quality structure is an Executive Quality Board and a Quality Advisory Board. The key thing in getting the quality program going is that it has got to be bought off and demonstrated by the top management. It's not something that you can send down and say, institute this quality program. Because if the people don't see the top guys doing it, they're not going to buy it. And that, to me, is worth a thousand books on quality. You've got to preach it if you want to get other people to buy into it. Everybody wants their organization to run better. Everybody wants to have a more profitable organization. And one of the things that we found—and this is an example of why management would buy into this thing—is that we spend an inordinate amount of money ensuring 100 percent customer satisfaction. We charter extra airplanes, we run people overtime, and it costs a lot of money. The cost of quality is very expensive. What you find going through the design phase and finding something to change is that it costs maybe 10 39 ------- dollars. But if you change something once it's operational and out in the field, it costs you 100 dollars or more. The real truth is not that we were blessed with any more wisdom than anybody else, but when it gets right down to it, quality was a tool that we thought we could use to make our company more profitable. And that's what gets people's attention. Management puts the procedures in place. I am the quality disciple in my division. We do have some quality professionals but I don't have any in my division. And the reason I don't is because it's always been my theory that as soon as you have somebody else to come in and perform the task, management then assumes that they're off the hook to be accountable. So we don't have any in our organization. Everybody's got to do it. And everybody's measured by it. Not measured by the number of events, but rather by the performance of their jobs and it's got to continually improve. People talk about the quality process—that's a scary word. Quality process can be a group of people who run around with clip boards counting are you using the right language. Our quality philosophy is that everybody buys into the fact that if I communicate with somebody, and if we have a customer supplier alignment, then I know what's expected of me. We've got a form that we fill out. In one column is "How do we think we're doing in providing service to this customer?" The other side, the side the customer fills out, asks, "How is this guy really doing?" And guess what? The numbers aren't the same. So the idea of the customer supplier alignment is to get his perception and your perception the same thing. And then you make a contract and you write it up. And then you are measured by meeting the elements of that contract. And it works. It gets people's attention. Our well-communicated focus on customer satisfaction saved the day for one of our customers. A fellow by the name of Michael Davidson of Atlanta was unable to meet the cutoff date of his business service center, so he ran the Atlanta Hartsdale airport with 150 packages that needed to go out. They happened to be his company payroll and he was very interested in getting it out. He came charging in here during what was the busiest time of the shift for the young man who was accountable for getting that aircraft out and loaded. His name was Mr. Augustus. Realizing the time constraints on this one guy, he shanghaied a service agent and a couple of pilots and together they worked to code the package so they could get the correct sort designators and get the freight on its way. All this was done, by the way, without jeopardizing the on-time departure of that particular Newark to Atlanta flight that Mr. Augustus was assigned to get out. Thanks to his dedication in getting this job done, he did earn the Golden Falcon Award that we talked about earlier. For Mr. Augustus and all of us at Federal 40 ------- Express, quality is an objective that leads to 100 percent customer satisfaction. Driven by this quality-focused mind set, we then applied for the Malcom Baldrige Quality Award, even though we were struggling through a very challenging year. We had just integrated with Flying Tiger Lines and had opened up a whole new segment of our business. Going for the Award and winning it has been quite an educational experience for all of us at Federal Express. And naturally, we're all proud of being the first service company to be named as a recipient. But we're really only as good as our latest pickup. So whatever loyalty we have earned in the past it must be preserved, because loyalty in a competitive marketplace is a very fragile commodity. We found that the real value of the Malcom Baldrige Award is the opportunity to rigorously evaluate our own company. Our quality processes have improved as a result of going through this. Applying for the Malcom Baldrige Quality Award will do the same for any company that's willing to undertake the challenge. One of the other things that concerned us upon receiving this award was that we would run the risk of everybody saying that we have now solved the quality problem and let's go back to the regular way we've been screwing up the business for years. It's in many ways a mixed blessing. With all this in mind, we do keep a watchful eye on our Quality Service Indicators and an open mind as we look into the future. We in industry today are fighting for our existence in a global marketplace where our competitors are ahead in the quality arena. We can go through a long list of industries that are no longer with us or no longer viable because of this. And so, while some may see the Malcolm Baldrige Award as a culmination, it certainly is not to us—it's only just another landmark on our journey toward achieving our objective of 100 percent customer satisfaction. Fortunately, it's not a journey that we at Federal Express take alone. The example is this meeting here today. Today countless corporations have begun to align themselves with suppliers who share their penchant for quality. These organizations realize that quality performance is becoming a necessity for survival. The quality improvement process at Federal Express is aimed at 94,000 employees believing that our goal of 100 percent customer satisfaction will always be the key to our continued prosperity. Thank you very much. 41 ------- GOOD AUTOMATED LABORATORY PRACTICES PANEL DISCUSSION Moderator: Kaye Mathews EPA National Enforcement Investigation Center Panelists: Jeff Worthington TechLaw Inc. Joan Fisk EPA Office of Emergency and Remedial Response Rick Johnson EPA Office of Information Resources Management Kaye Mathews We will be focusing our attention on the trend toward computer automation in the laboratory and EPA's progress toward addressing these trends with its draft guidance "Good Automated Laboratory Practices: Recommendations for Ensuring Data Integrity in Automated Laboratory Guidance." Joan Fisk I'll be talking about Good Automated Laboratory Practices Guidance: A Perspective for Superfund Data Collection Activities. By way of background for those of you who aren't familiar with Superfund's Analytical Operations Branch, we have quite a few responsibilities related to analytical services under our cognizance. We provide environmental service assistance teams to each region. These are contractors within each region who do things such as data review, analysis of samples, and reviewing QA plans. We also are the coordinators for developing guidance for data review or data usability. We are the leads generally for Superfund methods development in coordination with the Office of Research and Development. We also maintain an extensive database called the Contract Laboratory Program (CLP) Analytical Results Database Card. We also provide large routine analytical services through the CLP and special analytical services. We are very much involved with quality assurance (QA) oversight and quality control (QC) programs for Superfund. We work with Duane Geuder, Superfund1s QA Officer, and our QA Coordinator, Jim Baron. We are involved in the QA oversight of the CLP and are also trying to get into other areas where analyses are being done outside the 43 ------- CLP. We're in the process of making every effort we can to put performance evaluation (PE) materials into the regions for all these uses. We've seen this Good Automated Laboratory Practices (GALP) as being very important in looking beyond the traditional in our QA oversight because of the changes that have happened in the industry over the past few years. in the last few years, the laboratory community has become heavily computerized. There are many reasons for this. The market pressures 'with all the environmental legislation—the Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA) or Superfund, the Resource Conservation and Recovery Act (RCRA), the Clean Water Act, Clean Air Act, and Safe Drinking Water Act—are such that there's an enormous amount of analytical work to be done out there. Fortunately the technology has marched along with the need so that the facility is there for the laboratories to be automated and therefore able to get their jobs done. However, along with all the benefits that come with automation, we have gotten a new set of problems to deal with. We have many things that affect or compromise data integrity. In the old days we used to worry about things as simple as manual transcription errors. Now we have to worry about computer errors. We're talking about things such as data entry errors, computers talking to each other, storage of data, and electronic transfer of data whether it be on magnetic media like diskettes or over the telephone lines. So we have additional sources of error such as sampling errors, analytical errors, and operator errors. In addition, we've had another problem, which is that alleged fraud is compromising the integrity of our data. It's become much easier to cheat in the laboratory community because of all these neat things you can do with computers. We find that there have been instances where laboratories have manually edited their data just to meet the contract requirements and there was no technical justification for doing this. All this has happened because it has gotten much easier. We believe that the GALP guidance will promote data integrity within the analytical laboratory community. We think it will assist us in ensuring the quality of data. It will not guarantee it, but we believe that following the guidance will give us a better chance at success. It's important that the laboratory not look at the document at face value. They can use it as a foundation for data management/ automation practices, but it's important that they look at it as a minimum set of requirements of things they must address. This does not mean that there are not other things that need to be addressed. They have to establish within their own laboratory the systems that have to be in place in order to meet the requirements they set up 44 ------- for themselves. It's important that you have a system in place to know when it's working and when it's not. You have to have checks in place. And you also have to be able to have a system where you're going to correct your problems when you find that you haven't met success. In connection with the fraud issue, we believe that if you institute some of the things in the GALP document, such as some of the ideas on security or audit trails, it's going to be more difficult to commit fraud. It may be at the point where it's easier to do it right the first time, or cheaper to do it over again than it is to go through manipulations with the data just to make it look right. While this is not going to prevent fraud, it will make it more difficult. The laboratory community is familiar with this document. Rick Johnson gave a talk at our annual data management caucus last July and the audience was very enthusiastic. The lab community has been anxiously awaiting this document ever since. We plan on having another data management caucus this upcoming December. We'd like to think there will be people using it by then, and we'll be able to have some success stories related to it by the lab community. We do plan on providing the GALP guidance to our community when it's available for release. I think it probably has to undergo a revision before that. We have included some data management requirements and tenets of the GALP in our existing CLP contracts and I believe that our special analytical contracts have also included some of these GALP tenets. We have put in hardware and software requirements; we have added personnel that were not ever listed in our contracts before; and we've clearly defined some of the security levels that are absolutely necessary. We've required something that's essential to us in looking at data, and that's the audit trail. We think that with these security things in place it may make fraud more difficult. As it makes it more difficult to access the data system, you can eliminate the number of people that have access and limit the potential for people who are going to do bad stuff to your data. Also, as far as the audit trail goes—if it's required that you identify where you've made a change, you qualify the data as having made a change and you have to signify. You may be more reluctant to go in and make that change if you do not have a technical justification. So we do think that these things will impede, though not prevent, fraud. We have added some items to our repertoire that reflect data management issues. We think we can improve on it, and we are going to be requesting assistance from Rick in helping us perfect our audit process so that we would be good at doing the QA oversight job on data management practices of our community. 45 ------- I consider the GALP guidance as a great breakthrough for EPA. It was not written specifically for Superfund, but it does provide a basis for providing data of integrity from computerized systems that are critical for Superfund decision-making. I believe it is in the mutual benefit of the laboratory community and the EPA clients to take advantage of the GALP in our mutual striving to improve data quality and integrity. Jeff Worthington I'll be providing a testimonial to some of the types of observations we've seen in laboratories over the last five or six years. First, I want to give you some information about where our experience and background came into this type of observation. TechLaw provides support to the Federal government primarily as the Contract Evidence Audit Team (CEAT) contract to EPA's National Enforcement Investigations Center. We have been doing this from 1980 to the present. We have conducted 850 evidence audits (field and lab) . We also participated in the environmental survey as well as the Love Canal habitability study. I think probably in some of the study areas is where we first saw, for instance, the use of electronic data transfer. Besides going in to audit laboratories that are providing data to EPA, we've also assisted in litigation support in over 500 cases. The type of support sometimes entails helping prepare samples and sample evidence for trials for EPA or Department of Justice attorneys. In addition, we've conducted audits of PRPS1 search reports that have been developed by other contractors. TechLaw also has some technical enforcement support, which is along the lines of litigation support as provided under the CEAT contract. With the Department of Energy we've recently spent some time designing document management systems and preparing technical and evidential audits. We've also worked as expert witnesses for the Department of Justice in assembling environmental data. One of the trends that we're seeing in laboratories is increased use of laboratory information management systems (LIMS). In 1985, when we were looking at laboratories, many of the laboratory directors would try to drag you off to the side and say, "Don't you want to see my nice new toy, my new LIMS?" And we'd say, "No, we want to see your papers." And over time we've started to see the paper disappear into the LIMS and to understand that the evidence is starting to enter into computer systems. The second trend that we've observed is the use of electronic data transfer systems by laboratories. They're doing this for several reasons. The first reason is to facilitate transfer of data. These can be used for quick turnaround, and it's a good marketing tool. Some of EPA's programs and other clients are asking for this type of service as a deliverable. 46 ------- The third trend is the replacement of handwritten records with direct computer data entry. This started first, probably, with the' analysis area. Many people who are most comfortable with computers—the analysts sitting at their GC or GCMS—began to think, "Why do I need this log book; why don't I just start putting this information directly into the computer?" You sometimes see that same activity in the preparation area and sometimes in sample receiving. But when you look at information generated by a laboratory, both on handwritten records and information inside a LIMS, sometimes the information disagrees. This is for a variety of reasons. Sometimes the LIMS is not designed as a QC or QA tool; it's designed just to track information within the laboratory. For that reason people use it to say, "Was this thing prepared or was this thing analyzed?" Often those records conflict, because maybe the laboratory manager went to the .LIMS two days after samples were prepared and then punched in all the samples that were analyzed, and maybe used different initials and names. Often when you're preparing evidence, you see that the information that's summarized in those systems doesn't always agree with the handwritten records. That's something that needs to be reconciled. It's not necessarily a problem for the evidence. Also, there's a lack of written procedures for the use of software and hardware systems, both in laboratories and in field operations. There's seldom any life cycle documentation of software development. I think that problem is universal across all businesses, not just EPA and its contractors. Often you will not find the software within a life cycle documentation. Many times there is no clear definition of the responsibilities for the functions related to software and hardware systems. There is also no check on data accuracy. That depends very much on the type of business, but often in laboratories and field operations there may not be anybody who actually checks to verify that the information is accurate. we could almost think of the evidence as actually moving into computers—the papers are disappearing in some cases and that's where the evidence is. There's more and more reliance on computers, and not just in the laboratory. And with future cleanups being the thing of the 1990's, maybe we'll see more use of electronic data transfer to transfer information into the field and people's decisions will rely on the data that is transferred. People are addressing these concerns now. In conclusion, there are several things I'd like to touch on. First of all, guidance is needed for this area. The GALP is an example of a guidance document that covers the laboratory areas. I would also suggest that other areas such as field or sampling analysis plants—anywhere there's a computer that's being used, 47 ------- either in the collection and generation of data or by any other contractor—need to address many of the same types of issues. Secondly, if you're in the process of having a computer replace your handwritten records now, or you're working with somebody who does, print out your computer record in a timely manner. For instance, if it's an analysis activity, have them print out their GC log, look through it, verify that it's accurate, and then sign and date it. Lastly, as Rick Johnson presents the GALP, I'd like to ask the audience to consider both computer software and hardware guidance for potential inclusion in any QA program or project plans in the future. It's being addressed in the EQA-1 draft document now and we strongly support that. Rick Johnson Those of you who aren't particularly computer literate hang on because the problem that I'm going to describe is not a bunch of technical issues as much as it is an overall set of procedures and understandings in management—practices in the laboratory to make things better than they currently are in a number of labs. My purpose is to describe our program for ensuring the adequacy and integrity of computer resident data. What I'll do is give you a feel for the basis of it—where we got into it, why we got into it, what it's all about, and where we think we are with it today. I'm on what's called the Scientific Systems Staff. The Scientific Systems Staff is in the Office of Information Resources Management, one level below the Administrator. There are basically five program offices that report to the Administrator. There are several others, and then all of the Regional Administrators. Our mission is to help the Agency out in the area of information technology, anything from providing the resources for hardware and software development to overseeing contractor's work in this area to developing and managing all the Agency's information assets to promoting data sharing and integration. This last one turns out to be something that's important to all of us here at hand, because this is what Good Automated Laboratory Practices are all about—the issue of data sharing and data integration. Why did we ever start on this in the first place? About two and a half years ago we started becoming clearly aware that there was a rise in automation in the laboratories. Problems were beginning to surface. The auditors that audit for the GALP program and also in the CLP programs were beginning to wonder what to do with all of the automation that was coming into the laboratory. There was no uniform set of EPA principles to guide laboratories as they automated. People were making expenses in hardware and software that clearly were not meeting perceived needs that were developing as the Agency moved forward in its information management policies. 48 ------- In addition to all this, there were a number of activities going on in the way of development of requirements for information management and information dissemination. Collectively all of these things looked as if this was something we wanted to check into and see if there was some genuine need for the assistance of our office in this area. What we did is put together a program to identify as much information as we could in various areas to be able to assess whether or not we needed to go any further with it. It was a several-pronged program. The first was to go out, and through a combination of site visits and a questionnaire survey that was sent out to about 200 labs, assess what the state of automation was in laboratories, and if there was a problem, determine if we needed to get involved with it. Secondly, we wanted to examine all the existing procedures that were around and not go about re-inventing the wheel. Automation has been in the banking arena for a number of years; perhaps there are some lessons there that we could learn. Perhaps we could also find out things from the clinical laboratory industry. Forensic toxicology and stuff like that have been around for a number of years, and they've put together a set of principles and practices in their laboratory and are also automating. Why not go to them and see what they've done, and find out if there are lessons to be learned from there? Third, we wanted to look at state-of-the-art hardware and software technology as it applied to laboratories. Perhaps there were some readily available fixes there. There are some things that can be done...capitalized on...utilized. Perhaps there's software in the labs that vendors are selling that could do a much better job, and maybe there are some hardware fixes, simple things that could make a lot of sense. Since the Good Laboratory Practices (GLP's) have been around for a number of years and the computer is already under the aegis of the GLP's, why not look into the GLP's and see, in fact, if there aren't some lessons to be learned that we could capitalize on as well. They've withstood the test of time; they've been under review by the scientific community for some 10 years and were finalized about two years ago by EPA. They'd been in place by the Food and Drug Administration for a long time. Why not look into those and see if there are things that make sense and apply to the computer when it's put into the laboratory? An important consideration was the fact that there were existing requirements already on the books that we should probably be incorporating into whatever guidance we come out with. Finally, there were a number of requirements underway that could 49 ------- also affect laboratories—the electronic data transmission standard, some of the system design and development guidance, the computer security act, and a number of other things. These were basically the areas we charted out. We had people with different capabilities looking at each one of these things over the last couple of years. And what we found out, first of all, was that the state in the laboratories, regarding their use of the computer, was not as good as we would like to see it. Physical security was typically lacking. You could easily walk into a laboratory in many cases without any ID—without any checks or balances. System access was not protected. People could get on the system, log into it, and there were no passwords in place or voice recognition. I think this was the case in about 50 percent of the labs. Probably one of the biggest areas where problems occur is in data verification. We are moving from the area of laboratory notebooks and people are beginning to key in on the computer all their information—setting aside the fact that it's also now being electronically transcribed from instrumentation. There were very few verification procedures in place in a number of the labs. There was no double check, for example, to see that data were right or blind entry. Generally, if there was a problem across laboratories that was immediate to impeding data integrity, it probably was the verification. Documentation was sketchy. A number of the labs that we went into had no idea what version of the software was used to create what data sets. In a number of cases they had none of the versions of the software available, yet they were relying on this data to make decisions and provide the Agency with environmental information. And then there were a host of other things—anything from the lack of competent staff available to the adequacy of the staff. In some cases we saw a couple of labs where people were working 70- and 80- hour work weeks. I don't consider that to be adequate resources for a lab, but some people may. When we looked at automated financial systems we learned a number of things. They've been in place for about 10 years. One of the first things they do is perform a security risk assessment. They look into the environment of their operation and determine where various breaches to the potential integrity of the data are, analyze them, and come up with an overall schematic that lays out where the problems are. Then they go through what's called a risk management program, where they effectively respond to each one of the security needs that they found out about. Some of the components of that system were access management programs: varying the password; requiring people to make periodic changes to it; when 50 ------- somebody leaves the organization, their password is automatically dropped; verification procedures; double entry; and sometimes blind keying in the second entry. Audit trails were standard, as was hard copy retention. When we looked at hardware and software technology we found that generally there is no single guarantee that ensures data integrity in either hardware or software areas. And surprisingly enough, there were no established software standards for data integrity in laboratories. There are a number of laboratory information management vendors selling a lot of different software, yet there were no standards in place. Some LIM systems, for example, had password protection, some of them didn't. Some had certain data backup and data recovery features, others didn't. They can customize software to meet whatever needs you want; the problem has been that nobody's ever had a clear understanding of what was needed or what is needed. So when people went in to determine what it was they needed they came back to the software vendors and had a prescription that in many cases was probably good for what their perceived needs were, but that lacked the requirements for an information system. I blame that in part on us as much as those folks that are dealing with us, because we never really had in one place a common compendium that lists out all of the requirements and understandings of what we feel constitutes good management practices in the laboratory to ensure data integrity. I'd say the software manufacturing business is in bad need of some overall standardization. And I think we really found it out, particularly in the laboratory environment. There've been countless amounts of investments made in off-the-shelf LIM systems that many companies have been sold a bill of goods on. Some other interesting things we found out were that some of the advancing technology that some of you bump into now in the grocery stores—optical scanners—can be adapted to automation in the laboratory and ease some of the problems in our transcription errors. Also, magnetic ink readers have become standardized across the entire banking industry, virtually worldwide. And "smart cards"—they're like the strip of information on the back of your banking card that can hold an entire portfolio of information and can be updated as you move from various points in the lab to others. That's not out of the question at all any more. We looked at our GLP regulations that are in place now and found out a couple of things. One of the GLP's that was in the toxics program and the pesticide program, which had been adopted by other EPA programs, allows for raw data to be just about anything you want it to be. "Raw data may include.. .computer printouts, magnetic media...and recorded data from automated instruments" (54 CFR 158 & 160.3 [FIFRA] and ibid 792.3 [TSCA]). 51 ------- Wherever the data are first recorded, that is considered the raw data, but there is no requirement that the raw data have to be' recorded in any one form or the other. Laboratories have the latitude to choose what media they use and how they go about it, but once they do it and decide on it, that is raw data. Then, depending on the medium that it's retained on, there are certain requirements regarding the retention of that raw data. The issue of what constitutes the raw data for the laboratory— wherever the data are first recorded and however they're first recorded are considered the raw data. That raw data then must respect the data change requirement, the data entry requirement, and the data verification requirements which are spelled out in the GALP. In the case where you have information sitting somewhere and that information generally will be considered the raw data, the computer information should not be relied on. The person who's in charge of the machine should be attached to that raw data when that raw data were entered the machine identification for that information. There are all sorts of specifications for acceptance testing of software and hardware configurations that are required. Finally, the testing of facilities encompasses those operational units that are being used or have been used to conduct studies, e.g., the computer. Therefore, if the computer is being used in the conduct of the lab, it also must come under the GLP requirements. A couple of other things—the GLP requirements apply to automated systems. There are also requirements for the documentation of personnel qualifications, oversight of QA, and standard operating procedures (SOP's). Regarding existing requirements, we had a couple of them in place. EPA's Electronic Data Transmission Standard now makes it possible for people to report data electronically to the Agency, with the standard in place to allow people to do that, and with an understanding of what needs to be in each one of the transmission records. We also have a number of things going on in the Information Resource Management Policy related to such things as system design and development to documentation. There were a number of requirements under development at the time we put this together. The Federal Electronic Reporting Standards were being pushed into all the Federal agencies and they must have a general standard for the electronic transmission of data. EPA beat Congress to this by about a year. The Federal Computer Security Act specifies various levels of security that must be in place for certain types of data as defined within the Act itself. Finally, we are continually involved in changing our System Design and Development Guidance. What all this meant collectively to us was that we probably needed 52 ------- to do something. It looked as if there was clearly a reason to do it. There were a number of things under development, and we needed to bring it all together. We decided that we needed a registry of principles—what we subsequently called the Good Automated Laboratory Practices—and then to help folks to implement the GALP, implementation guidance as well. We wanted to not just lay this out to people and tell them that they should do it, but to give them some feel for what we expect them to do, and some understanding of why it is that they should be doing it. We also wanted compliance guidance for the. auditors to enable them to determine that they are, in fact, following specifications. Recognizing all this and putting together all this information between late '89 and early %90 and collecting it all, we finally came up with the draft Good Automated Laboratory Practices. Then we went one step ahead and put into it Implementation Guidance. Figure l shows you relationships to various principles within the GALP. The bottom line here is that nothing's new. Everything that's in the GALP is already there in one place or the other. The Federal Computer Security Act is dealt with; EPA's Information Resource Management (IRM) Policy is there; the Good Laboratory Practices are incorporated in the GALP; some of the statutory requirements from several different programs are incorporated. Retention requirements that the Federal government mandates all Federal agencies to do and the types of media are all addressed there as well, and others such as the electronic data transmission standard are also included. The Quality Assurance Unit has oversight over such things as the standard operating procedures (SOP's), the operation maintenance of the computer system, and security. In addition, this Responsible Person (RP) has to report to management that everything is going according to plan. Or, management can assume the role of the RP if they want. Figure 2 shows some of the different areas of the GALP and the different citations to the different statutes so that you can cross reference back and forth. This includes areas dealing personnel and qualifications, the personnel training, such as what laboratory management should do, the role the RP should have in the laboratory, what the QA Unit should do and what operational roles it should have, the facilities and what govern them, the equipment, security specifications, standard operating procedure, software requirements, data entry, raw data definition and records and archiving. Again, these all relate back to any one of the different statutes I showed you earlier. Section 7.14 covers comprehensive testing periodically done on the system to ensure that, with any of the changes that have been done with the system over time, the system is still fully capable. I think we specify that in the GALP at least once every two years, as I recall. 53 ------- Under section 7.10, data entry, there are two areas: integrity of data and data verification. Under integrity are three separate requirements: tracking person—that's who entered the data, or who was handling the machine at the time the machine entered the data into the system; the eguipment that was being used, the identification information on the equipment, and the time and date that it was entered; and data change. This is one of the 82 principles in the GALP and it relates to changing data in the information system. There's nothing new in this requirement; it's in the GLP's. It's been in fundamental accounting principles for a number of years. What this does is let somebody know that the data have been changed, who changed it, when they changed it, and why they changed it. This is very important. It seems, on the surface, like something anybody would want to be able to do and have in place. A number of software vendors have sold software to laboratories, advertising that they have an audit trail in the system. In many cases it does not meet one or the other of these requirements. It doesn't, for example, tell the auditor or somebody going back through it that the data had been changed. It does preserve the original data; it dumps it off onto a tape somewhere and writes the new data in. But there's no indication that the data were changed. In other cases, it doesn't tell who changed it, why they changed it, or when it was changed. But it may show that it has been changed. The full compliment of requirements here are only met right now by about three (that I know of) commercial software vendors. A number of labs on their own have built systems to meet these specifications. So in an effort to help people, rather than just telling people what to do, the guidance is formatted like this: it shows the icon at the top, gives you a specific statement of the individual requirement, attempts to explain what the requirement is, why it is wanted, gives you an example of it, and then it shows a coding in here for who is suggested to take on the role of managing that particular area of the GALP. For example, with regard to the data change requirement, please look at Figure 3. Here you'll see the data change requirement restated, an explanation of what this is all about, what is meant by it, an example, and who's responsible for it, and there's an underlying principle- called audit. There are six basic principles behind the GALP and six basic operational roles that are assumed in the guidance. Finally, a special consideration—where we can we've tried to show a picture. If the picture's right, it's worth a thousand words. For example, Figure 4 shows a picture of what comprises an adequate audit trail according to the GALP principles. You'll notice that 54 ------- there are additional notes as well that can refer you back to various types of documents and various EPA requirements that are already in place. In this case you have the Toxic Substances Control Act (TSCA) and Federal Insecticide, Fungicide, and Rodenticide Act (FIFRA) regulations that are already in place and specifically require that. And you have a number of background assessment documents we've done leading up to the GALP that are also referenced. This requirement is already in place in the clinical industry; it's already in place in the financial industry; and it's a requirement in our GLP's as well. Where we are with it today: the document was sent out for review and comment in early February. We asked that the comments be back by the end of March; we're still waiting for some offices. What we intend to do over the next several months is look at the comments. I've hired a couple of folks to take a careful analysis of the comments and go out and make some laboratory visits and quality checks on them. Following the completion of that, we will come out with a paper that shows each comment, how we've addressed it, what we consider about it, what we're doing with it, and where changes to the GALP are indicated. It may be out by early December. We struggled with whether or not to require this of all the Agency's programs via Administrative Order, or to package it as recommendations to Agency programs to adopt, or call it guidance for Agency programs to adopt. We spent about a month and a half back and forth considering those two options. It was deemed at the time that the merits of the individual GALP will stand for themselves and that the EPA programs can then, by merit of the individual GALP principle, adopt those within their framework so that the integrity of the individual principle will speak for itself and the Agency programs can then pick it up. You'll see in a number of cases for a given principle that it may say a "shall" or a "will" or in some cases it speaks in very general terms. Those are specifically written that way. If somebody's going to adopt an Agency program, they're going to adopt one of the GALP principles. Then the language in there is very specific to that effect. If they're going to adopt a principleone place. Then we go on to the next site. And we can put all of the site by site planning in one place and separate the lab standard procedures and the field standard procedures and incorporate them by referencing them—get them separated from the issue of planning. I want to make a point, which is emphasizing existing terms and 55 ------- another 20 or so programs in the Agency that have already adopted many others. And there are a number of programs that have a mish- mash of them. If we went out with requirements, one of the problems we would have is that they would be viewed as a duplication of those folks who are having to establish them. They already have them on the books; why do they have to do them again? There also are individual interpretations and some of the different areas that different programs have for their statutes that we feel the integrity of which had to be maintained. So at this point we are packaging them as guidance or recommendations, providing ways in which someone can achieve compliance with them in the document and giving examples of how one can achieve compliance in a number of cases where special considerations can come into play. But again, it is going to be up at this point to the individual EPA programs to decide which of the elements they want to adopt and which ones they don't. The document, by the way, has not just been distributed to EPA. There are about 250 organizations inside and outside of EPA. I know that the U.S. Department of Agriculture, U.S. Department of Energy, the Food and Drug Administration, and the associated people in those organizations have all been given copies of it and were involved in the program from the very onset. 56 ------- Figure 1 Computer Security Act of 1987 Statutory Requirements for Environmental Programs: Superfund Resource Conservation and Recovery Act Clean Water Act Safe Drinking Water Act Others GALP EPA IBM Policy: EPA System Design and Development Guidance EPA'S Operations and Maintenance Manual EPA Information Security Manual EPA'S Data Standards (or Electronic Transmission of Laboratory Measurement Results Findings of EPA's Electronic Reporting Standards Work Group National Archives and Records Administration's Electronic Records Management Regulations EPA's Good Laboratory Practices Federal Insecticide, Fungicide, Rodentlclde Act GLP (40 CFR Pan 160, August 1989) Toxic Substances Control Act GLP (40 CFR Part 792, August 17, 1989) 57 ------- Figure 2 APPENDIX A: INVENTORY OF COMPLIANCE DOCUMENTATION "CORD PURPOSE SUBSECTION REFERENCE Organization and Personnel Personnel Record* Quality Assurance Inspection Fteporu Ensure competency of personnel Ensure QA oversqht Facility Environmental Specifications Ensure against data toss from environmental threat 7.1 7.4 7.5 FIFRAGLPs 160.29 TSCAGLPs 72929 FIFRAGLPl 160.35 TSCAQLPs 792.35 FIFRAGLPs 160.43 TSCAQLPS 792.43 Equipment Hardware Description Acceptance Testing Maintenance Records Identify hardware in use Ensure operational Integrity of hardware Insure ongoing operational integrity of hardware 7.6 7.12 7.8 7.12 7.6 7.12 FIFRAOLPs 160.61 TSCAQLPs 792.61 EPA Information Security Manual for Personal Computers System Design and Development Guidance FIFRAGLPs 160.63 TSCAGLPs 792.63 Laboratory Operations Security RM Assessment Standard Operating Procedure* • Security Procedure* • Raw Data Definition Wentfy security rtstt Ensure consistent use of system Ensure data Integrity secured Define •computer-resident* 7.7 7.S 7.8 7.8 Computer Security Act FIFRAGLPs 160.61 TSCAGLPs 792.81 Computer Security Act FIFRAGLPs 160.3 TSCAGLPs 7913 58 ------- Figure 2 cont. APPENDIX A: INVENTORY OF COMPLIANCE DOCUMENTATION RECORD • Procedures for data analyra. preceding • Procedures for date storage and removal • Procedures for backup/recovery • Procedural for main- tenance of computer system hardware Standard Operating Procedural • Procedures for Electronic Reporting • SOPs at bench/ wo natation • Historical Files PURPOSE SUBSECTION REFERENCE Ensura comment use of lyitem Ensure comment use of system Ensura consistent use of system Ensura comment use of system Ensura comment use of system Ensure comment use of system provide histoncal record of previous procedural n use 7.8 7.8 7.8 7.8 7.8 7.8 7.8 FIFRAGLPs 160.87. 160.107 TSCAQLPs 792.81. 792.107 FIFRAGLPs 160 81 TSCAGLPs 792.81 EPA Information Security Manual lor Personal Computers FIFRAGLPs 160.63 TSCAGLPs 792.63 Transmiiilom Standard! Electronic Reporting Standards Workgroup FIFRAGLPs 160.81(c) TSCAQLPS 792.81(C) FIFRAGLPs 160.81 (d) TSCAGLPs 792.81 (d) Software Documentation Description Uf« Cycle Documentation • Deign Document/ Functional Specifications Identify software In use Ensura operational Integrity of software Ensura operatnnal Integrity of software 7.9 7.9 7.9 FIFRAGLPs 160.61 TSCAGLPs 792.81 Computer Security Act System Design and Development Guidance see above 59 ------- Figure 2 cont. APPENDIX A: INVENTORY OF COMPLIANCE DOCUMENTATION RECORD PURPOSE SUBSECTION REFERENCE Ute Cycio Documentation • AcoeptanceTosting TMting • Chang* Control Preeeflurea • Procedures tor Reporting/Resolving Software Problems • Historical Fib (version numMn) Enaure operational Imtgrtty ofaoftware Ensure operational integrity ol software Enaure operational Intagrtty of software Ensure reconstrueiion of reported data 7.9 7.9 7.9 7.9 EPA Information Security Manual for Personal Computer! see above see above see above FIPRAQLPs 160.81 TSCAQLPs 792.81 Operations Records/Logs BacK-up/Recovery Logs Software Aocaptanea T«st Record Software Maintenance (Chang* Central) Records Protection from data loss Enaure operational Integrity of software Enaure ongoing integrity of software 7.12 7.12 7.12 EPA Information Security Manual tor Personal Computers System Design and Development Guidance see above 60 ------- Figure 3 7.10 Data Entry /; Integrity of Data 3) Data Change When a laboratory uses an automated data collection system in the conduct of a study, the laboratory shall ensure integrity of the computer-resident data col- lected, analyzed, processed, or maintained on the system. The laboratory shall ensure that in automated data collection systems: 3) Any change in automated data entries shall not obscure the original entry, shall indicate the reason for change, shall be dated, and shall identify the individual making the change. EXPLANATION EXAMPLE When data in the system is changed after initial entry, an audit trail must exist which indicates the new value entered, the old value, a reason for change, date of change, and person who entered the change. This normally requires storing all the values needed in the record changed or an audit trail file and keeping them permanently so that the history of any data record can always be reconstructed. Audit Trail reports may be required and, if any electronic data is purged, the reports may have to be kept permanently on microfiche or microfilm. Responsibility: Responsible Person Principle: 3. Audit SPECIAL CONSIDERATION: Laboratories may consider adopting the policy by which only one individual may be authorized to change data, rather than implement- ing a system that records the name of any and all individuals making data changes. 61 ------- Figure 4 7.10 Data Entry Integrity of Data 3) Data Change ;c* 134.7 • NAME OF PERSON ENTERING DATA •DATE OF ENTRY 1 AUDIT TRAIL • CHANGE PROCESS 144.7 " 134.7 NAME OF PERSON MAKING CHANGE DATE OF CHANGE REASON FOR CHANGE ORIGINAL DATA CHANGED DATA Notes... For additional guidance, see: FIFRA GLPs 40CFR 792J30(e); TSCA GLPs 40CFR 160J30(e); Automated Laboratory Standards: Evaluation of Good Laboratory Practices for EPA Programs, Draft (June 1990); Automated Laboratory Standards: Evaluation of the Standards and Procedures Used in Automated Clinical Laboratories, Draft (May 1990); and Automated Laboratory Standards: Evaluation of the Use of Automated Financial System Procedures (June 1990). 62 ------- HAZARDOUS WASTE SITE REMEDIATION PANEL DISCUSSION Moderator: Gary Johnson EPA Quality Assurance Management Staff Panelists: Thomas Morris Martin Marietta Energy Systems, Inc. Duane Geuder EPA Marcia Davies U.S. Army John Edkins U.S. Navy Gary Johnson I'd like to give you a little background on the proposed National Consensus Standard that is being developed through ASQC. Then we will have presentations by our panelists on their view of the issues that need to be addressed in order to make an acceptable National Consensus Standard work for their organizations. We'll tell you a little about the origin of our efforts on this. This effort occurred because of concerns expressed over QA requirements for hazardous waste management clean-up activities. We found that there were multiple sets of requirements out there. There was EPA guidance in several forms. The now infamous QAMS- 005/80 which has been out there since December 1980 has not been revised since the "interim guidelines" were published. Other organizations have other sets. Probably the most common set of requirements that has been applied to these activities has been NQA-1. This was a set of QA requirements developed primarily for nuclear facilities. In fact "N" means "nuclear" in NQA. Its application to environmental concerns has been somewhat successful in some areas, but not without some difficulty. There were not only these different requirements, but also the fact that those in the regulated community often had to respond to multiple requirements—both EPA and NQA—since they were not fully compatible. And this has led to duplication of planning documentation; it's led to rework; and both have led to added time and cost, which neither we in government nor the regulated 63 ------- community can afford to allow to continue. Harmonization began as an effort of the ASQC Energy Division. We began this effort just a little over a year ago, in the winter of 1990. The participants included EPA, DOE, DOD, Nuclear Regulatory Commission (NRC), various EPA and DOE contractors, and private consultants. This has been solely a voluntary effort by a group of your peers—fellow quality professionals who recognized that a need existed to bring more consistency and standardization into the way we go about QA in environmental work.' Our purpose as we began this effort was to harmonize the current multiple QA requirements into a single set for environmental programs. No current standard exists for environmental programs. A standard should provide a clear statement of what QA elements are needed, allow flexibility on how and by whom requirements are implemented, enable more consistency by everyone doing the same things, and enable use of the good work already done to harmonize QA requirements. When you think of a standard, probably you're thinking in terms of a performance standard. What we're talking about is a requirement standard. In the Agency we do not have a current requirement standard; we have EPA Order 5360.1. That is not a requirement standard. We have guidance, such as the now defunct 004 and 005. Those are guidelines, not requirement standards. In fact, there is no requirement in EPA right now that anyone should implement a QA project plan. All the contract regulations say is that you have to prepare it and get it approved—they do not say you have to implement it. We need a clear statement—a good solid foundation—for quality assurance in our Agency. What we've heard over the last 10 years is that you've had a lot of frustrations in trying to deal with carrying out and institutionalizing a QA program in your own organization. Our recommendation that went to the ASQC group was that a standard is a pretty good idea—that perhaps we can provide this clear statement of the things that we will do in a QA program. But at the same time, we emphasized that flexibility had to be provided to the organizations implementing this standard to determine how the requirements would be implemented and by whom. Everyone recognized from the very outset that we could not write a prescriptive requirements document that would have any prayer of working in EPA—given the diversity of our programs—nor in any other Federal agency engaged in environmental work. But we felt that if we could get everybody at least doing the same things and talking about the same things, then we could begin to bring more consistency and ultimately improve the way that we're carrying out QA activities in our respective organizations. We also recognized in this group that a lot of good work had 64 ------- already been done. We have to give credit to the NQA Committee of the American Society for Mechanical Engineers (ASME), because there are some really good things in NQA-1 that our committee felt was appropriate to include. And so our philosophy was, "steal shamelessly" because there are a lot of good ideas out there. And so we solicited ideas from whatever sources we could and we haven't stopped soliciting those ideas. Now, to briefly describe the structure of the proposed standard. It has three parts. Part A deals with management systems. The intent here is to define what you need for an effective QA program in terms of the structure and framework. In other words, what do you need to be able to carry out day to day activities on specific technical projects? Part B, characterization of environmental processes and conditions, is primarily environmental work— environmental monitoring, sampling and analysis activities. Part C deals with design construction and operation of environmental engineering systems. This part deals with the technologies that we use in pollution control, remedial design, remedial action, and Super fund. Part C addresses an area that the Agency has not completely addressed, and so we felt there was an opportunity there to address that. In particular, we found that a lot of the work that had been done by DOE and DOD provided us with a wealth of information which, through the standard, we could create good, concise, clear statements of what would be needed. Figure 1 shows the sections of Parts A, B, and C. To quickly show you how Part A is constructed, visualize this as an umbrella under which everything is done. The management commitment and organization, the first statement in that standard, says that management at all levels is responsible for quality. That is an essential statement; it's an essential part of this entire process. The QA program could document the management's systems that you are employing. So that's really nothing new there. As for personnel training and qualifications, we all recognize (and certainly from viewing the world from a TQM standpoint) that human resources are our strongest asset. We need to make sure that those concerns are addressed and that people who are doing work that affects the quality of the results, and therefore the decisions that you're making, have the necessary skills to carry out that work most effectively. Regarding management assessment, it's absolutely essential that management participate in the process and that they periodically examine the effectiveness of that process. 65 ------- FIGURE 1 Part A Management Systems 1 Management commitment and organization 2 QA program 3 Personnel training and qualification 4 Management assessment 5 Procurement of services and items 6 Documents and records 7 Use of computer hardware and software 8 Operation of analytical facilities and labs 9 Quality improvement Part A requirements apply to both parts B and C. Part B Characterization of Environmental Processes and Conditions 1 Planning and scoping 2 Design of data collection operations 3 Implementation of planned operations 4 Quality assessment and response 5 Assessment of data usability Part C Design, Construction, and Operation of Environmental Engineering Systems 1 Planning 2 Design of environmental engineering systems 3 Implementation of environmental engineering systems design 4 Inspection and acceptance testing 5 Operation of environmental engineering systems 6 Quality assessment and response 66 ------- In the procurement of services and items, we have to make sure that. the services we get from subcontractors meet our criteria, or that the equipment that we purchase to monitor polluted streams or air meets our needs. We need to do a better job in the Agency of maintaining documents and records. One of the most frequent comments that I've heard from Superfund has been the difficulty in maintaining satisfactory documentation on given sites, largely because, in many cases, we don't have a process in place to do that. So we can learn from what others have done here. We're depending more and more on computerized acquisition systems and storage systems for our data. And so it's appropriate that we address that issue in our requirements standard for QA pertaining to the use of computer hardware and software. The operation of analytical facilities in laboratories—it's important that these labs practice good automated laboratory practices, that this be an integral part of their operations. It was felt that highlighting this was very important, because so much of what we do depends on the products of those operations. Last under Part A is quality improvement. We can always learn from our experiences and next time do it a little better. In Part B we followed a very simple axiom: plan, implement, assess. The first two requirements relate to the planning that goes into designing the data collection activities. The requirements are consistent with the data quality objectives process that we have been using in EPA for several years now. Then an interesting thing: implementing what you've planned. That's perhaps an area where we haven't been as persistent as we should be. The next item is assessment. This refers not only to assessment of the work that's being done in process, in responding to it where responses are appropriate, but also the assessment of the usability of the results—recognizing that even imperfect data sets can be used for some decisions if we understand the limitations on the use of those data sets. Part C deals with engineering systems. Again, the same approach: plan, implement, assess. Plan the design of environmental engineering systems that may be required for a Superfund remedy; design those systems; implement the system's design, and make sure that the components of the technology that are being constructed pass inspection and acceptance-testing requirements. You need the assurance that this system is going to work once it's actually put out there in the field. To make sure we have thought out the operation of these systems, sufficient guidance should be provided 67 ------- in the form of operating manuals and the like for the successful operation of these systems. Last is quality assessment and response—to make sure that the systems do, in fact, perform as intended, and if they do not, that appropriate response actions to correct the situation be performed. I'm sure you are probably thinking, what's the impact on EPA? Well, for the large part, most of your programs are going to be unaffected. Because as Nancy said, over the last number of years, we've heard from a lot from you, and our responsibility has been to try to carry those concerns to the harmonization committee. I think that when you look through this document, you'll find that it really doesn't change what you've been doing very much, because the essential elements are there. Part C does add something new, because it deals with issues pertaining, particularly in Superfund, to remedial design and remedial action that we have not effectively addressed as an agency. But there's still ample time to do so. Our key interest in this is to be able to provide a foundation, a basis upon which new guidance for a variety of issues such as QA management plans, project plans, and audits can be developed and given to you and to the regulating community at large to use. The current status: we believe that ultimately we can achieve an acceptable national consensus standard. But the emphasis is on the word consensus, and consensus requires your participation. We have also presented this document to the second annual Hazardous Waste Conference in Las Vegas, which was sponsored by the ASQC Energy Division because of the particular community that would be directly impacted by this and that has a special interest in it. So the document has, at least to this date, gotten somewhat limited distribution. We have gotten a number of comments back in and I appreciate that very much. Within ASQC the standard setting process has begun. Two weeks ago in Hollywood, Florida, the Energy Division council voted to report the standard out of committee and to request that the ASQC standards committee list this standard and report it to the American National Standards Institute as a proposed national consensus standard for environmental programs. I've got to say a couple of words here about why that's important— it takes this out of the political arena so that no one organization feels that any particular organization is trying to enforce its own agenda. I can assure you that we've had some candid and lively discussions from the policy group and the work group that put together this initial draft. We've tried to accommodate everyone's concerns. 68 ------- The pleasant thing about it all is that the group coalesced very quickly. We found that the concerns were the sane. You get quality professionals together and you find out that you've pretty much got the same set of problems. This week some of our colleagues in the committee are presenting this standard to the ASME NQA committee which is meeting in Williamsburg, Virginia. We're looking for their support and comments on it and also to express our appreciation to that group for the effort that they've put into the NQA standards through ASME, which provided us with valuable information. The next step, as we said all along, is getting input from you. Please review it and let us have your comments. Sometime later this summer we hope to publish a revised draft in the Federal Register to allow for a formal public comment period. This will be used as input to the ASQC standard-setting process. And we're hoping that in the not-too-distant future ASQC will issue this standard which may be at that point known as ASQC Standard E-4. We're hoping that the standard will be accepted and endorsed by the participating agencies, by EPA, if it meets your needs. In the meantime we need your help, my fellow EPA folks, to help pull together revised guidance that more closely reflects your needs and your concerns right now for where our quality program needs to be for the remainder of this decade. For the purpose of our panel discussion we're going to assume that we can reach an acceptable national consensus standard. We're going to ask the panel to address what must be done, or what issues must be addressed to make the national consensus standard work in the organization. Through their discussion we hope to be able to bring some of these key issues to the forefront so that we'll know what areas we're going to have to identify in the months ahead as we address this. Marcia Davies I would like to talk to you about the structure of the Army Corps of Engineers and how the Corps views QA. The Corps since the 1940's has always had the same structure. We have a headquarters in Washington whose responsibility is policy, interface with Congress, and interface with the headquarters of other government agencies in Washington—trouble shooting and resolving differences that can't be resolved elsewhere. Beneath headquarters we have divisions that are geographic. There are about a dozen of these around the country. Each division has several districts and each district has area offices. As you go down that structure, the geographic responsibility gets smaller and smaller and smaller. What are the purposes of all these? The worker bees of the Corps are in the districts. The districts are the direct overseers of 69 ------- the contractors. The districts do the work in-house. The dam builders for civil works and so forth are at the district level: The area offices which are under the districts are construction representatives. And those are broken down to small enough geographic areas so everywhere that the Corps has a construction project going on, there is a construction representative on-site that reports to a nearby office. This was put together for the Corps' traditional civil works and military construction missions over the years. The sole purpose of the division is quality assurance. The division oversees the work districts—makes sure all their contracts are legal, and that they're not doing under one particular contracting mechanism what should be done under another one. There's a lot of technical support at the division level: legal support, personnel support and so forth. And each division services several districts and many area offices. I feel that the traditional structure of the Corps lends itself very well to the TQM concept. The principle of the Corps is one- step-up review. If the districts have a contractor working, the districts are supposed to have the personnel to review the work of the contractor, with general division oversight. If the districts are working in-house, than the divisions are supposed to directly oversee, technical and otherwise, what the districts are doing. Within that structure, there are special offices that are called design centers or centers of technical expertise. The division that I belong to, the Missouri River Division, is designated as a mandatory center of technical expertise for environmental works. And so we have special missions where we oversee all of the districts for some aspects of their work. To the Corps, QA is a government function. Contractors do QC and the government does QA. It is rather rare for the Corps to contract out its QA function. That's almost always done in-house by government employees. Sometimes we get into semantics differences with the folks in the regions or in the states in terms of what they think we should be doing and what we're actually doing. The Corps entered into hazardous waste remediation in 1982 to support EPA in Super fund, primarily in the area of design and construction since the Corps is an engineering outfit—but an engineering outfit which had been traditionally devoted to design and construction. Since that time, the Corp's missions and requests for assistance have expanded quite a bit, so that the Corps districts currently manage remediation problems for EPA, DOE, and the Department of the Army. The Corps manages the entire formerly-used defense sites program for DOD. We do some work for the Navy, the Air Force, the General Services Administration, the National Guard, and the Department of Transportation—whoever comes 70 ------- knocking at the door saying, "HELP" is what the Corps often gets involved in. We strongly support any harmonization effort, because working for that many different Federal agencies is very difficult under the current set of circumstances. The Army Corps of Engineers recently had an internal reorganization effort. The principles under which we're slowly reorganizing internally say that the people who are responsible for the projects and programs—the managers responsible for the schedules and for communicating with the outside agencies and so forth—are in a different reporting line from the people who are responsible for the technical adequacy of the products. And so for each project, sometime down the road in the Army Corps of Engineers—I'd say within the next six months in some districts- there will be a project manager and a technical manager. The technical manager will coordinate all of the internal QA efforts for that particular project. That will be in the environmental arena as well as in military construction and all of the other areas in which the Corps operates. The Corps QA program has all of the traditional structural elements, specific requirements for each project as well as general guidance requirements. We write QA program plans; we like to call them chemical data acquisition plans because of what our focus on QA is and we think that says better what they are. Internal review—land validation. We do land validations; we do them out of my office. We use audit samples and we generate them ourselves within the Corps. We have our own audit sample system. And in our audit sample system we have real world soil samples, and a large percentage of the CLP labs does not pass that sediment sample on the first run. We've taught them a few things about extracting. And we do not run quarterly blinds. Instead, for every project we take government splits. The Corps has nine division laboratories and those government split samples go to our own laboratories and they either analyze those in-house, or with intensively managed commercial contracts that they hold. The news is there aren't many matrix effects out there. So we've learned a lot of things and we feel that we have a good chemistry program going. Problems are essentially recognition. If it ain't the CLP, it ain't what the regulators want. Is it CLP- equivalent? Very frequently. But it's not the CLP. We like to standardize on SW 84$; we like to require all of the internal QC that has to this date been optional in SW 846. We like to see that reported. We're probably on 90 percent of our work never going to court—maybe more than 90 percent of our work never going to court. It's quite clear who did it, when they did it, and what they've got to do about it. I'm pointing up some problems in implementation that I see. I think the main problem that we see is lack of recognition that 71 ------- there are a lot of different ways to do things and they're probably mutually acceptable. And they all work well. Now speaking mostly to the investigation phase. Given all the things that we have in place, including geotechnical guidance, almost everything that we have we have ripped off from EPA, because we started in this arena in 1982 with a Memorandum of Understanding with EPA. We have very few wheels that we've invented that weren't already somewhere mentioned or suggested in existing EPA guidance, including the split sample concept,, which as far as I know, we're about the only ones who implement. Because we've been a design construction organization for a long time, we have in place a number of things with respect to specs or guide specs for building slurry walls on contaminated projects, and landfill covers that are compliant with rec river requirements, that are helpful to our field people, who are overseeing projects but don't really have the time or the technical ability sometimes to sit down and go through all of the legalese that's in the guidance in the Federal Register. So we tend to develop pieces of guidance that translate those things into field directives for our folks. And we have been doing this on the design and construction site for quite a while. I think that we may have something important to contribute in that area as an organization to that part of this new standard. A couple of the specific problems that we have are, for example, on studies: Who reviews? Who trusts whose reviews? When have we reviewed enough? How many agencies do we have to review? How long does it take, and how much money do we spend on this? I'm not downplaying review; review keeps us alive—it's a very important part of the activities of the office from which I come. But some of the projects we review, the Department of the Army reviews. We pay the states and the regions to review; the contractors review. And you've got these massive numbers of people in essence all saying the same thing, and I would bet you a dollar that if you and I sat down next to one other and you're a chemical professional in this area and we review the same document for chemistry, we wouldn't come up with the same set of comments. So I think that the review process needs a lot of work. And I think that it needs to be addressed in the standard in greater detail and I think that we need to figure out a way to get on with this process. DQO's are wonderful things; I endorse them heartily. I felt that the little two volume set on DQO's that came out of EPA really turned the lights on. That was great. But they're not implemented. You can use data quality objectives all you want, but you're going to get to a regional Federal facility project manager who's going to say, "You've got to do it over, guys. This isn't the CLP and that site's on the NPL list." Those decisions are not being made by quality professionals. If we could talk to the quality professionals I suspect we wouldn't come up with the same 72 ------- decision. There are things like that which I feel the standard needs to address—and address in greater detail in terms of what's happening and who does what. How much uncertainty are we willing to accept in data? -Well, we're not willing to accept bad data. So we want the contractor to accept the responsibility for the analytical, and if most of it is flagged, he's going to do it over. And so we don't have contracts that allow them to flag it and give it back to us as acceptable. Consequently, we've run into problems because our data's not flagged. And it's not flagged because we didn't accept the flags, whether it's SWA 46 or whatever type of analytical method that they're using. I think we need to pay a lot of attention to DQOs on all the analytical levels and that we need data validation guidance for level one, level two, level three, level four—the concept of level four needs to be expanded to include more things than the CLP. I think we need to figure out a way to do this work cheaply, smartly, on schedule, by doing a whole lot more interagency cooperating than we do now. The Corps would welcome the opportunity to do that. And I think that we need to address redundancy of effort among Federal agencies. I want to say that the brightest light in the past year, and the thing that I think will help us a great deal as it gets down the road, are the actions of the Environmental Monitoring Methods Council. I was excited when I heard about that and I think it's wonderful. And that's about all I have to say. John Edkins I'm going to agree with almost everything that Marcia said and then act as a ground wire for some rudimentary problems that we have in the Navy—perhaps a general agreement for a harmonization effort for the same reasons that Marcia described and a plea for help because of the particular resources problem that we have. I'm not a QA professional; I'm a geologist. I've been doing this for about four years, so I have had to deal with some very basic problems in starting with this, and through my friends at the Quality Assurance Management Staff, gain a basic perspective. I want to say something about our structure for environmental compliance. The Naval Facilities Engineering Command in Alexandria, Virginia, responds directly to the Chief of Naval Operations. Then that structure is broken down into eight engineering field divisions, which is the operating arm for cleanup of Navy shore facilities. Then the Naval Facilities Engineering Command has several other smaller support organizations, namely ours, where we were formulated originally as a small multidisciplinary group. As far as compliance with CERCLA is concerned, somewhere around 1980 all of the original Navy property 73 ------- assessment was done out of our office. When the engineering field divisions plunged into the problem of investigating the environment ih«?UKtltUdin°U? dri11 holes and collecting environmental data and QA became an issue, our command office then turned back to us and *aid' .a11 ri9ht» 9ive us QA, because you're the multidisciplinary group. That's how we happened to fall into this picture. We have a OA°fna?aei.f i?J?<2fCV.' ^e tend to buy everything. When I got into QA in late 1987 the first thing that I attempted to do was bring to the attention of my management the need to restructure everything that the Navy does to make it sound, look, act, and taste like EPA guidance. I realized that we were not going to get accepted on a local level unless we had those key program elements and names in our program. We now have some 2,000-plus sites. They're small sites. We identify the individual pesticide rinse aid area and a fuel spill as individual sites in our program language, although we do have the large landfill with the multi-types of waste present. A common Navy site will be a PCB waste discharge area. Typically one Navy base, a Naval air station for example, will have from 12 to 15 sites identified of that nature. To give you an idea of our resources, our Navy engineers that act as remedial project managers—the people that drive the contractors—may have as many as three Naval installations and perhaps upwards of 50 sites for which they have responsibility. These people typically will be one or two years out of graduate school and just be becoming integrated into the process by the time they are hired away by our more well- to-do industrial brethren. So my thesis today is one of a procurement-based problem and also the idea that there is indeed nothing new under the sun. I took Gary seriously about barriers to harmonization, and I could only really come up with one barrier. It was communication, without a doubt. We have a multi-disciplinary industry and terminology from many disciplines. We've got toxicologists trying to talk to engineers who are trying to talk to bio-ecologists, and now I find that we are trying to talk to procurement specialists about what all of this is that we're trying to buy in terms of environmental data and QA services for environmental data. We have a lack of general industrial standards or standard operating procedures. Many of the procedures that we use don't have very strict source referencing requirements for contractors. Many times in the Navy we have seen instances where EPA guidance documents have been used in lieu of good contract standards or good contract specifications. In other words, a statement of work will say, take these guidance documents and those regulations and comply, comply, comply. A contractor is out there performing his own interpretation of guidance, which can sometimes be ambiguous and does lack these kinds of small critical elements as the source 74 ------- referencing requirements. Guidance tends not to identify by whom individual or specific actions are to be taken. When you get in a procurement milieu you find out very quickly that that becomes important. The fact that we are trying to turn a lot of good ideas in Federal guidance into a procurement system, and then trying to adapt to sometimes changing a revolving guidance and revising our procurement system to address that, becomes a problem for us and creates a lot of redundancy and waste. Certainly Federal guidance has a lot to say for it. It does tend to standardize and provide a lot of good elements for us. The ideas of redundancy and streamlining have been recognized and these guidances are continuously improved. The very concept of DQO's I heartily applaud—the think before you leap approach. That was a badly needed concept early in the industry when "characterize the site" meant go out and get data on just about everything and then figure out later what you were going to do with it. Also, on the very idea of risk-based decision making I heartily approve. The concept of harmonization itself—the standardization of national programs for environmental clean-up and compliance—is something that's going to benefit everyone. We have the same structural problem that Marcia alluded to in that our Navy field divisions are roughly the same in size and number of geographic acreage as EPA regions, but they don't coincide. So we'll have one engineering field division with possibly as many as three different EPA regions that they are responding to, not to mention state and local governments. The work that's been done by QAMS and DOD-sponsored annual QA meetings and the ad hoc committee on quality and environmental measurements, ASQC Energy Division work, has resulted in the EQA-1 effort. All of these things are definite progress, and I want to not focus too much on those but go on to something that I'm very interested in, which is the individual method standards, because I'm a firm believer in building blocks. Standards for data collection and interpretation include EPA guidance and standard operating procedures (SOP's), state standards and SOP's, and contractor and other agency in-house SOP's. Some other people that I have talked to in these meetings have indicated a joy in collecting SOP's—we call this piracy—but actually many of these procedures the government's paid for probably 16 times, so we ought to be able to pirate them, especially in the contractor area. Looking for all sources of standards, I sifted through EPA and came up with a batch several years ago and then turned around and plugged those into ASTM Committee D18 Soil and Rock, for which I'm an active member. I got sucked into that process with a great deal of joy, mixing myself with industrial interests as well as other government agencies and local government agencies. 75 ------- I'd like to focus just a little on D18.21 standards. These are consensus standards and my initial charge was to develop field QA *? correspond with lab QA, and it turned out that we got into a lot of different things. But just to mention a few: the surface and borehole geophysics and vadose zone monitoring, everybody's favorite .."how to put in a monitoring well," soil sampling practices, pump testing for hydraulic purpose and ground monitoring well design and construction, monitoring well maintenance, and rehabilitation and decommissioning. We typically decommissioned monitoring wells by cutting them off at ground level with a backhoe or bulldozer, becoming lost in the records and also conduits into subsurface lower layers, a real problem. The American Society for Testing and Methods (ASTM) is looking at this. This one, which I'm particularly involved with, I think is very important because this gets into some of the aspects of DQO's: how to design an experiment for a site and how to interpret the data. These are very important issues and I wanted to raise our involvement in the building blocks aspect of It • I'm going to talk about what more we can do. We need to recognize and focus on QA as a procurement-based problem. We also need an increased recognition by people of data as the primary product. We need more model contracts requirements documents and standards for contractor action, right down to the building block level. We also need to look at some long-standing technical terms and determine how we might clarify existing program language. QA/QC seemed a very simple concept, at least originally. As I said, I'm a geologist, not a QA expert. It took me a while to figure it out, because when you ask people what QA/QC is you get as many different answers as you ask people. QA in terms of procurement is inspection and acceptance. As for the in-house problem, it seems that a motor company can have a staff that decides they want to sell this to our perceived customers and set up a series of specifications, and then another engineering staff will design those into tools and set up a manufacturing line. QC has a responsibility for making sure that the product that comes off the line conforms somehow to the specifications that were set ' up by QA. As soon as the QA expert leaves the motor company and goes to work for the Federal government, there gets to be a very sharp line between the two. It's a procurement line; it's a contractual line; it's a no conflict line—buying products on behalf of the American people. I did some checking around on definitions and there are a couple of things that were interesting. QA implies inspectioning acceptance on behalf of the government, confirming that a product conforms to contract specifications, whereas QC is ensuring that it conforms to specifications. But in both cases we have the implicit specification already there. We presume that we have 76 ------- specifications in both QA and QC. Also, the interesting thing here too is that it's very clear that QA is a government responsibility and QC is a contractor resnonsibilit-v. u«w j.a unau At -» very ciear tnat QA is and QC is a contractor responsibility I did some reading on inspection acceptance in the Federal Acquisition regulation. They basically defined higher level contract quality requirements as those which apply to critical and complex procurement, and the definitions of complex and critical were given. Complex means that the product has quality characteristics not wholly visible in the end item. Critical means that the failure of the item could injure personnel or jeopardize a vital agency mission. I looked at those and said, That looks like data to me. it's very hard to say that I've got a high quality piece of data when I look at it. And it is critical, too, because we have human lives that are at risk in this environmental cleanup business. The thing about a higher level contract requirement is that it puts an increased onus on the government to not use contractor inspection of product. It puts an increased onus on the government to have a non-conflict inspection except as a process. In wrestling with the problems of QA, I ran into the fact that we didn't have any standards. In talking to contracts people, they presumed that we had standards and specifications for what we were trying to buy and I said, no, we don't have any. And they said, that's not QA. And I said, well, what is it? Well, that's acquisition support—to develop those specifications and standards. One of my co-workers, Barbara Johnson, helped me with this one. I think that this appears in EQA 1 under the heading of Quality Improvement. It says that you look at what's wrong with what's coming off the assembly line and decide how to re-specify it so that you continuously improve. You've got to have all of these things happen in order to get quality data in a new industry. Yet we've got a procurement system that's already written up in a Federal Acquisition regulation that we have to communicate with whoever thinks QA is just inspection/acceptance. So we've got to look at these issues and determine how to communicate with these people and get them on board to our issues and on our side. I want to emphasize one last thing: the acquisition process has nothing to do with the people who are generating the data. The Navy buys everything—we buy QA program plans; we buy QA project plans; we buy the whole business. The contractors are over here; they're producing. We're over here figuring out, well, we'll get their product in here and we'll inspect it. I wanted to touch on the recognition of data as a primary product, and I argued this in our Naval Facilities Engineering Command. I have people that want to refer to this as construction, and I said, 77 ------- That's fine, but you don't know where to clean up; you don't know to what level to clean up; and you can't verify that you indeed cleaned the area up without data. How do you reassure the neighbor across the border line that you got rid of a part per billion of benzene out of the ground when it's an invisible situation? Data supports your decision. You stand or die by data. We all know this, but it's something that I have to keep hammering away at in my own shop. * We need an increased focus on these contracts requirements documents, individual methods standards and SOP's, and model work statements, which could be almost in any form for every step of the process, for planning documents, for reporting, for execution of work for feasibility studies, for remedial designs, you name it. We need standardized planning and reporting formats. In a way, a lot of our problem is on format. We can provide the same information with the same substance, but if we call it something slightly different, or label it in a slightly different order because we're buying these things so quickly that one engineer out there procuring a million dollars worth of contracts services in a very short time provides that to a region, and the region looks at it and says no, that's not acceptable, that's not what we're asking for, you don't have a QA program plan here or whatever it's called. And so we're required to redo the effort or revise it. That's very expensive. Referencing is something that I think we need to have in guidance documents, in our model work statements, and in our contracts requirements. We all track information. We buy a lot of redundant information in the Navy. I'm sure you've all seen that problem before. It was very complex for me to try to convert a very good piece of EPA guidance into a streamlined procured work plan for the Navy. One of the things that I noticed was that the QA project plan is very similar to something that we designed also. The project plan guidance 005/80 consists largely of standard operating procedures. When I looked at some of the things that had to do with sites and the objectives for measurement, I had to sweep all of these things that were site-specific over into one place, because, as I said, often we have 15 sites on one Navy base, but that usually will be one contractor as well. The same procedures for drilling wells will be the other contractor on all sites. What we have here is site one, with all of its conditions, its data quality objectives and scoping, and its risk-associated elements put in one place. Then we go on to the next site. And we can put all of the site by site planning in one place and separate the lab standard procedures and the field standard procedures and incorporate them by referencing them—get them separated from the issue of planning. I want to make a point, which is emphasizing existing terms and 78 ------- trying to pull in as many existing terms to define program language as we can—in terms that other people in the industry might recognize. There's nothing inherently wrong with acronyms in themselves, but they tend to get a life of their own if we don't pull them back to long-term technical terms. DQO is one of my favorite acronyms because I really like the concept. However, when we get on a small site scale, I think a lot of people would recognize these sorts of things: hypothesis development, hypothesis testing, experimental design, and errors and controls. We need to go for all of these and we need to use these words to tie together these concepts. I've now defined QA as government responsibility in a procurement- related system, with planning being something different from standard procedures. We had a contract provide a QA plan which named these program plans and project plans, but since it planned to have QA, this very much looked like a plan to plan the planning plans and we had no actual planning there at all. We can spend a lot of money on these kinds of documents. So I'm advocating a terminology problem where we need to tie these back together somehow against the Federal Acquisition Regulation and the problems of procurement, and that will help. We can keep these terms if we give them a little more clarification and definition. From the Navy perspective we are looking at standard procedures; we're looking at building blocks; and we're looking at procurement tools. We have a need to provide engineering field divisions with things that buy good services just by the plug-in sort of factor, so we have a very baseline sort of attitude on this. Any guidance that comes out that will be useful to us will help name those relationships, especially for contracts and government responsibilities. QA is procurement to us. Duane Geuder My major concern in being almost last on a panel like this is that almost everything has generally been said. Fortunately I'm not last, so I still have a chance. For any number of issues that I address, approximately half of them have already been addressed either by Nancy or Gary or John or Marcia. What I want to start out with is to give credit to Gary and anyone else involved in the generation of this document because I think it's a quantum leap forward. Obviously a lot of time has gone into it. I hope we can, in fact, get it to a point of implementation. Two things jumped out at me from the document: first, the advocacy of plan, implement, and assess; and second, to foster a no-fault attitude. Many times the QA community is looked upon as the policeman and we have difficulty doing our job because people think we're looking for fault as opposed to trying to improve the 79 ------- process. The plan, implement, and assess was at least in Superfund. We're pretty good at implementing but we really haven't done a good job in the planning and assessment. With the fraud issue in the CLP, or the fraud issue in the lab community-,- what I'm hoping is that the diplomacy among the federal agencies will continue and that that diplomacy will pervade not just the Federal agencies but all of the organizations within the Federal agencies, because we in Superfund and elsewhere have to deal with numerous organizations within Superfund that all support the project. And likewise, with implementation of this guidance that it would minimize and reduce all opportunity and inclination for fraud in the lab community or elsewhere. Anyway, I understood my mission in a slightly different way from Gary. I think it's not terribly off track, but I was attempting to answer two questions: What could a single QA standard mean? and what are the issues for implementation? I think we mostly dwelled on the second, but the first, which was addressed by Robert Layton, was the question of terms, definitions and acronyms which the Navy is so fond of—EPA and Superfund are enamored with acronyms. When I first started with EPA, it took me three weeks just to learn the acronyms and that was in water enforcement. Superfund has another order of magnitude of acronyms. At any rate, this standard guidance would certainly help to standardize our language, our terms, and our definitions. The guidance will provide a standard process that will ultimately facilitate review and audit by EPA or by Superfund or Federal facilities or contractors and most importantly, Potentially Responsible Parties (PRP's). Again, more acronyms. By definition it will eliminate barriers, both in its development and in its application. The barriers among the agencies are already falling; I hope that the barriers within the agencies will tumble in a similar manner. The buy-in by the agencies and the sub units will, in fact, eliminate the barriers. If each entity has a buy-in, some ownership of the product, which is what Gary has emphasized, then certainly it can't fail. It'll provide a yardstick, especially in Superfund, for cost recovery. We need some kind of yardstick, some measure of what the PRP's or any other entities should be doing at a remedial site—what is an appropriate level and how many samples, because we go after cost recovery. If EPA says you should do more than is necessary, then we're not going to get all our costs. If EPA says less, then the project won't be done properly. So, definitively a yardstick for cost recovery and in some instances we can go for triple damages. In that instance we definitely need a very viable, documentable yardstick to measure the activities. You can't get away from it. Clearly, resource implications come out from this. Increased resources are unlikely; therefore we're going to have to do things smarter with the resources we have and reallocate those resources. 80 ------- It's been mentioned by Nancy, Gary, and others that we're going to need tools and guidances. But they're going to have to be" customized; they're going to have to be user-friendly. Hopefully they'll be automated. We at headquarters have heard from the regions numerable times that there's too much guidance—we've heard this in some of our audits and reviews. Therefore, whatever guidance is forthcoming, we'd better be careful to keep it terse, to the point, and as user-friendly as possible. Last, but not least, this will provide a significant improvement to EPA's, and I hope to the other, QA programs—the other Federal agencies. Significant improvement, because it emphasizes a cradle- to-grave approach where you do QA through the upfront planning, you do through implementation, and then at the tail end you do the proper assessment to make sure everything has worked and what you can do to improve it. These are the issues for implementation. Resources—we can't get away from it. I mentioned it already. I can see that if we have a full fledged audit or review process within Superfund, the resource implications are dramatic. We currently don't do a lot of field and on-site auditing. If we are to do so, significant resources are going to have to be shifted from some other activity if we're not going to get additional resources. The other thing: we plan two or three years in advance, so they are now providing input to the '93 budget process. I've done so already. If we're going to change, or get additional resources, we won't see anything until '94 at best, based on this particular process. Nancy mentioned that there's always resistance to change. This will require dramatic change, especially in where we have our QA resources applied presently. We've emphasized the analytical component; we've minimized the upfront planning and scoping; and we've minimized the review and audit process. Marcia mentioned that data quality objectives (DQO's) are great, and they are. But within Superfund and within EPA there's a DQO syndrome. It carries a bad taste, unfortunately, with too many of the engineers because of some of the problems in its infancy with implementation. We also have a QA syndrome that we have to deal with, where the engineers and a lot of the data users are not QA advocates; they want to get out with their shovel and re-mediate the site. So we have those two syndromes to deal with. We also have the issue of true implementation. We can write program plans and project plans until they come out our ears; if they are not fully and accurately implemented, they're of minimal value, except as an exercise and profit for the contractors. We need more than lip service and we need true upper level management commitment to this guidance. Once it's in place, management has to understand they're going to be wed to this and it's going to cost. Others have mentioned education and training. I put it as 81 ------- enlightenment education and training, especially for the engineers and the data users that we have to deal with in the QA community, so that they understand that QA is not a necessary evil, a hurdle, a barrier, but a useful management tool and a mandatory management tool. Last, but not least, are analytical facilities and labs. I think we need to develop among the Federal agencies a minimum standard that the labs will ascribe to all data. If the user needs more, the DQOs will get it. If you can accept the bare minimum then it's taken care of within all of the laboratory community. Tied in with that, we need to do a lot more in addressing performance-based results as opposed to method-based results. We often encumber ourselves by being wedded to particular methods that don't always work. The CLP is a good example. We have very rigid requirements that can't work in every instance. There has to be some flexibility. There are two related projects going on in Super fund presently that will dovetail very well with this standard. One is called DAS for those who love the acronyms; it's tied in with the long- term contracting in Superfund and it is the Delivery of Analytical Services. That component was left out of the strategy; it's currently being addressed. Clearly the chemistry in the analytical services required by Superfund will be a major component of this standard. The other is that we presently have an ongoing review of the Superfund QA\QC to look at the program in its entirety, to find out where there are shortcomings and where there can be improvements made. This is being done in a TQM approach, so we have regional as well as other input. Thomas Morris In the Hazardous Waste Remedial Actions Program (HAZRAP) we have lots of customers. It's difficult because we've got the DOD people, and even within DOD, we have the Army, the Navy, the National Guard, and then we've got EPA and DOE. Everybody has a different agenda so we have a pretty difficult time of it. We also have all the subcontractors who do work for us that we oversee and do project management for. I've worked in about four different areas—all within Martin Marietta—and I fight the same battles, time and time again, so maybe I can shed some light on areas that will help in.implementation of quality standards. There are five keys to effective implementation of any quality standard: understand cultural resistance to change; demonstrate top management commitment; provide adequate education and training; be willing to define the overall objectives and process requirements at a global level; and ensure availability of 82 ------- electronic user-friendly implementation techniques, tools, and processes to help project team participants. Respect for QA I'll talk about first. I expect that you all have about the same level of respect by your project managers and your upper level managers that I have, and I don't know how we get through that. I've struggled with it for several years. I think it's changing slowly. One of the problems was that they put QA people into QA positions when they had nowhere else to put them. There are some exceptions to that—we"re probably those exceptions. We have different ways we put things, and I think managers have trouble understanding it sometimes. To understand cultural resistance to change, I think you have to analyze and pay attention to the basis for resistance. You've got to actually look back in time and say, The managers that are in power, what's their real heritage? What year did they grow up in? I think you have to get to the point of understanding what their real motivational drive is all about. Our personnel people have done a lot of activities over the past four years or so, such as using Myers-Briggs and those kinds of tools. I had the benefit of knowing the guy in training who does those and he's plotted all of our senior managers and they're all power hungry. In seriousness, I point that out. It's real and you have to learn to understand that it's there. I think part of the reason why managers have difficulty with QA people is that we're threatening to those managers. In many cases, of course, we really do understand what we're talking about and it's not easy for them to accept that. Break down barriers that were established as a result of ego and insecurity. That's tied to the real bottom-line needs. I think we need to start using some psychology-oriented people to better understand what drives people. You've got to deal with it. Empower the people doing the work—that's embodied in TQM—and listen to the people who actually do the work. Involve mid-level managers. Even when people get on board with TQM and even when the top managers speak the right words, whether they practice it or not, we seem to somehow miss the mid-level managers. They finally find out about it somewhere down the road, by hearsay, and that's threatening for them because they haven't had a part in it. The real commitment is when you say and do. And that's a pet peeve I have about presentations. People get up and say what needs to be done and never get down to actually doing it. Encourage change for the better; emphasize continuous improvement. Demonstrate top management commitment. Practice values implementation. There's a lot of politics involved in decisions, 83 ------- and sooner or later politics is going to drive you down. I don't think you're going to get anywhere until you treat people like people and listen to what they say and let them listen to what you have to say. Reward those willing to take risks and accept accountability. Nobody wants to accept accountability. I understand why, too; there's risks associated with it. There's legal liabilities associated with it. We have major problems getting people—whether it's EPA or whether it's the states—to accept accountability. Say, okay, I'll put my signature down and I'll accept the accountability for whatever method you're proposing, or accept a plan that's been submitted. And you've got to reward the people who take the risks and if they fail, they have to suffer some of the consequences for that, but on the other hand, you've got to take into account that they're willing to try and make some decisions to move forward. Make the QA function a key step in the upward mobility chain. I think one of the things that will really demonstrate top management commitment is when they start having the QA function be a place where "fast trackers have to move through." Actually, once people get into it, they do develop an understanding of it. But when they're out on the side and you're bugging them about management systems that need to be in place, it's just a pain for them. You put them in that position and let them go through that and then let them take it out and implement it. If they actually learn it and you put them in a project position, then all of a sudden you have it infused within where the work actually takes place. Integrate methodologies for assuring quality into a defined line organization procedural system. Within HAZRAP for our QA program we're getting all of our procedures in place for doing business. With our procedures manual, we've got a section called "Administration" and of course it's got all of the "how do you hire and interview people and how do you orient them," and that sort of thing. But in the project management system section we've got not only the things about "how do you plan a project and how do you use a team concept and how do you estimate your costs and schedule it and go through what procurement acquisition strategy you're going to do," but also things such as quality management and DQO's, "how do you do readiness reviews, how do you document non-conformances, how do you resolve problems and get to root cause, how do you apply lessons learned." Those kinds of things have always been pushed off on the QA people, but they're really a part of doing the project. And so, we're trying to move as much of QA into the project as we can, because document control, and all of the things that go along with document control, is not a QA function but a records management function. That's probably not even a project function. We don't have document control in our project management section; 84 ------- we've got it over in a section called documentation; and we let people who understand all the myriad of regulations associated with keeping documents for 75 years and microfiching take care of it. That's not a QA responsibility. QA is an -integral part of a structured and disciplined project management system, not something that some QA people do off to the side. I think you have to adequately staff and fund a QA function and train the project participants, depending on the level that you involve QA into the project. You need to say, we're going to have a QA function, we're going to recognize them as a viable entity, and we're going to fund that operation because we believe those are the kinds of management practices that need to be taken if we're going to do work, regardless of what kind of work it is. Or, if you're going to let project participants be the people who document your non-confonnances and do your problem reporting and problem resolution, then you've got to fund training the project managers to do those things. QA people have a better understanding of that, but nothing says you can't train project managers to get to what caused the problem and document it. Respect the QA professionals' credibility and input equally. I daresay that if it's like it is in some of the organizations and programs that I've worked in, the QA person gets called in when there's a problem, or when there's something that has to be resolved or when there's been a non-conformance, and you've got to fill out a form and the project person doesn't know how to fill it out because he's never been trained to fill it out and never respected the fact that you needed the form in the first place. Provide responsibility, accountability, and authority for function- specific decisions, our documentation doesn't yet give us specific enough responsibility or the accountability and authority for the things that QA people are responsible for, or for the things that the program people are responsible for. Provide adequate education and training. I think we have to teach management and project personnel what QA is, not from the standpoint of the QA requirements, but from the basic practical elements of QA. This is what document control is all about; it's about making sure that the right work plan's out in the field. I think you ought to teach the QA professionals a little more about basic project management principles so we'll have a better understanding of what the project people are facing. They're facing a lot of problems, too, with controlling costs and schedule, and overhead rates. Teach everybody the basics of regulations, terminology, and statistics. We've got instances where states require absolutely knowing that there's not one part per billion anywhere on a site, and we'll have other states who actually will say, you can do a 85 ------- statistically-designed random number-generated statistical sampling for anything. You can give me a 95 percent competence level, or 99' percent competence level or contaminant is below a certain limit then we'll accept it. I don't think we use nearly enough statistics. Teach DQO's. I'm a strong supporter of DQO's. We've got several hundred projects but we probably had one or two DQO's that were actually used. Yet it's a wonderful process. We're having difficulty in teaching our contractors about what DQO's are. We have to sit with them in the hotel room—they've got a contract to do the DQO's for us—and we have to say, here's some examples of how DQO's need to be done. Then they feed it back to us. It's reality. I think we need to teach assessment and oversight to everybody, because one of the things we're moving toward is not only independent assessment such as the QA people typically do, and audits and field surveillance, but we're also trying to teach the project people to do self assessments all the way along where they're looking at themselves, even if it's just for a small piece of the activity. I think the training courses need to be structured 25 percent lecture, 75 percent practical. I think we need a lot of DQO training and that training needs to be set so that even if it's a simple example, you can work through the forms and fill in the blanks and learn by doing. And last, staff the training function with personnel qualified to teach the subjects. We typically have training departments and they're staffed with trainers and you can give them the slides and they can go through the slides, but if they don't understand the material that they're teaching, then they have no credibility with you. The audience probably has more knowledge than the trainer before they even begin to take the course. Be willing to define the overall objective and process requirements at a global level. Set the program structure up in the top, but formulate the procedures from within the organization. We formulated our list of procedures for doing business in HAZRAP at the staff and department level, but now that we're assigning the preparation of these procedures for doing business, we've got secretaries doing the ones on secretarial work; we've got project managers doing the one on DQO's and quality planning—not the QA person, but the project people. We've got the people who actually do the work doing it. Now we have that separate bit of oversight in assuring that once they get a procedure together the implementation of it does fulfill the requirements that have to be fulfilled. But we try to get it written at the project level. Drive for consistency, at least in approach. Maybe there are 86 ------- slightly different requirements in CERCLA versus RCRA for the kinds of information that need to be in a QA program plan, but the general approach that you go about in doing a program plan should be standardized. Trying to strive for consistency is a difficult thing to do. But the only way that you can eat an elephant is one bite at a time. It would be nice, though, if we had a consistent approach for eating the elephant we started maybe up here instead of back here all the time. But on the other hand, I think the harmonization effort is starting to try and set up at a global level, and from then begin to build all the parts. Build any graded application into the procedures themselves. If there's a need for a graded approach for a small site versus a large site, or a site that may be either contaminous or not as critical, you can try to build that into the procedure and have a procedure that has a graded approach in it, rather than saying that we have this QA program for this kind of site, and we have this kind of QA program for this kind of site, because controlling a document is controlling a document. If you're really talking about QA principles, at a minimum coordinate the building of the pieces of the program to maximize the strengths of each part. There's great work that we face in all of these areas. CERCLA has a work plan, a sampling analysis plan, a field sampling plan, QA program plans, and health and safety plans. The Underground Storage Tanks (UST) program has things such as initial response measures and site characterizations. RCRA, CERCLA, UST, and NPDES all overlap, and many of the elements are similar enough in nature that I think if we could get people together and use the best parts of them, we could come up with generic statements or generic work plans. As a matter of fact, we have a generic statement of work. We may use it internally, but we have a generic statement of work that you can pull the pieces from. It's on a computer and you can go in and fill in the blanks. As you go through it you pick the things that are applicable and pull them out, and the computer goes in and through some WordPerfect manipulation puts together a statement of work in a couple of hours instead of a couple of weeks. We're also working on a generic work plan that leads to ensuring availability of electronic user-friendly implementation techniques, tools, and processes to help project participants. Our next step in this generic work plan is to build an expert system where the screen asks a series of project related questions about a specific project in the language that a project manager can understand, and then obtains his or her answers to those questions. For example, what kind of a site is this? Is there water contamination? Is there soil contamination? Is there both? As the manager answers, the expert system automatically says okay, because there's soil contamination, these are the potential applicable standards that go with it. The project manager gets the particular words that need to be in the statement of work from a database that already exists. 87 ------- Structure software to respond transparently to project manager language. We're working toward generic QA program plans, generic statements of work, generic work plans, and health and safety plans. Share lessons learned with everyone. I don't think we share lessons learned enough and we ought to share then in a user- friendly lessons-learned system of some type. One of our agenda items is to work with our contractors. We have 11 major general order contractors who work with us. We'd like to get a lessons- learned system that they're willing to share among each another. They all face the same things; they-just face them in a different area of the country. We have one region that tells us we've got to use a certain kind of decon procedure that calls for specific decontamination fluids to be used, and we have other regions that say you can use whatever is applicable including steam. So we run into the same kinds of problems that have been mentioned in some of the other presentations. Those are going to happen unless we work together and we stand to benefit, because you are our customers too. We have to deal with the regions and the states and get decisions made, so I know it would benefit us. We want to work with you, because we want to be in this business for a long time too. Appreciate it. 88 ------- ECOLOGICAL MONITORING PANEL DISCUSSION Chair: Robert Graves EPA Panelists: Janes K. Andreasen Fish and Wildlife Service Adriana y Cantillo National Oceanic and Atmospheric Administration Thomas F. Cuffney U.S. Geological Survey, Water Resources Division Robert Graves I am the Acting QA Coordinator for EMAP, which stands for Environmental Monitoring and Assessment Program. This talk is not going to be about EMAP per se, but rather about ecological monitoring. Nancy had asked for my talk to cover a few different objectives. One is to define what ecological monitoring is; two is to describe why harmonization standardization of QA is important in ecological monitoring; three is to give a brief overview of EPA's QA program; four is to link EPA's QA program to ecological monitoring; and five is to link EMAP's QA structure to EPA's QA requirements. To give you a little background: from my perspective, EPA has two broad objectives: one is to protect human health and the other is to protect the environment. To achieve these objectives, EPA works within the area of risk assessment and risk management. The Agency has in the past focused mainly on risk assessment as it applies to human health. That is, we regulate contaminants for the most part, to protect the health of human beings. However, we are now, I believe, re-emphasizing that other objective we have within the Agency, and that is to protect the health of the environment. I think we're beginning to realize that our biological resources sustain our existence, and that to protect our own health, we have to protect the biosphere in which we live. Accordingly, ecological monitoring research within the Agency is starting to get more emphasis than it had in the past, and that is one of the reasons I believe that EMAP has developed. EMAP is probably the Agency's largest ecological monitoring program. 89 ------- Let me begin by giving you a brief definition of what ecological monitoring is. It's the measurement of abiotic and biotic factors in the environment to assess current conditions, that is, status, and to identify and warn of changes in biological resources, which is trends. What are the QA objectives of any ecological monitoring program? One is to ensure that the data generated are of sufficient quality to meet program needs. Two is to ensure that the procedures and processes used are such that they will produce the desired results. Three is to ensure that all procedures and processes and data are sufficiently documented. Fourth, which I think is the crux of harmonization, is to ensure that data generated in one program, albeit EPA, are defined well enough so they can be validly compared to those data generated in other programs, that is, programs such as the National Oceanic and Atmospheric Administration's (NOAA) Status and Trends, U.S. Fish and Wildlife Service, U.S. Geological Survey (USGS), and a lot of other Federal agencies and states and other research efforts. The purpose behind harmonization, I think, is to have data comparability. If we can have ecological monitoring programs that give us data comparability, we can then put all the data together and make some real integrated assessments. Harmonization/standardization, the way I look at it, touches on various methods. One is sampling methods. The one thing I think is true with ecological monitoring, and it's probably true of other programs as well, is that when I speak of methods, I don't necessarily think of the laboratory methods. I think the biggest crux of ecological monitoring is to harmonize sampling methods. For example, when you go out in the field, how do you collect your samples so that the way EPA collects its samples is compatible with the way the Geological Survey collects its samples? If we're looking at fish tissue, for example, are we looking at the same tissue? Are we looking at the edible portion of the fish, are we looking at the whole fish, are we looking at livers? How do we take the fillets if we're looking at the edible portion; is it skin on or skin off? All these factors will have a great impact on whether we can pull that data in the end. That, I think, is more important than the analytical techniques used. The analytical techniques are comparable—and you'll listen to Adriana talk about NOAA's program where they don't really specify standardized methods--in other words, what they do is specify performance criteria of the laboratory so that laboratories can use any laboratory methods they' want as long as they generate data of a certain precision and accuracy. I think that's legitimate. But if we don't have sampling protocols that are standardized, we'll still never be able to pull that data in the end, and I think that's where we need to focus a lot of our attention—not only sampling methods in that sense, but also things like statistical 90 ------- methods. How do we design our protocols? Are we using randomized samples or non-randomized samples? Again, I think we need to get our statisticians involved upfront to see if we take our samples in different fashions and whether or not we'll be able to integrate that data together in the end. QA/QC techniques is another big issue. There are certain QC techniques that are mandatory, such as blanks or calibration standards. I think that from a laboratory perspective, if all the Federal agencies could agree on a certain QC program, it would make it a lot more cost beneficial for us to run samples rather than what happens right now. EPA is probably one of the worst offenders in this realm. No matter what program you're working for, it has completely different QC requirements. And most of these requirements are not requirements because they're really necessary for that particular program, but they are requirements that some manager puts in because that's their particular bias. Therefore, what happens is in the laboratory they've got to shift, as they do samples for the drinking water program and they've got to shift again, as they do samples for the NPDES programs, and then they've got to shift to QC techniques a third time when they do for Superfund, or for RCRA or for any of the others. Method Detection Limits (MDL) determinations—I think that's another important thing. For a lot of our methods, we do list MDL's, but the way that we calculate these MDL's varies from method to method in some cases; therefore you can't get a good feel whether method A is really comparable to method B, or at least go down to the same levels. Definitions—I think that was mentioned the other day; laboratory certification on a national scale was another one. Indicator selection with respect to ecological monitoring and computer hardware/software standardization, which I think was addressed when they were talking about GALP's. If we achieve this, what will this buy us? What it will do is buy us integrated programs; it'll give us an integrated database. We can then take inputs from either the Federal agencies that are doing ecological monitoring, whether it be EPA, NOAA, the Bureau of Land Management (BLM), U.S. Department of Agriculture (USDA), or any other agency, or any of the state programs that are doing ecological monitoring, and other international programs, and feed it into a common database that will allow us to do integrated assessment reports. ' Let's take a quick look at the QA components within EPA. The way I look at what QAMS has dictated to the rest of us within the Agency is that we have an umbrella document called a quality assurance program plan (QAPP). A QAPP more or less sets the general guidance of how we're going to operate a program. Under that we have various other documents that we're mandated to put 91 ------- together, such as data quality objectives (DQO's), quality assurance project plans (QAPjP's), audits, and audit reports. Why do we require these particular programs? The rationale is that the QA program establishes the principles and guidance. The DQO's establish what the customers needs are, and I think it's important that we go back and determine who our clients are and exactly what it is that they expect out of these programs. The QAPjP then establishes the process for satisfying customer needs. The audits, in turn, evaluate that process to* improve it and to judge its applicability to satisfy customer needs. Finally, the reports are the product to meet customer needs. How does this all tie in with respect to ecological monitoring programs? What steps does one go through when one does the planning for an ecological monitoring program? This is the scenario that I came up with, and I tried to tie that into the various elements of what's required under EPA. First, you have the initial preparation where you define the objectives of the program. Basically, that is when you're asking questions such as what data are required, what is the end use of data, and what is the total allowable error? That is the vernacular of EPA's DQO's. Then you always have that thing called resource allocation, which is the budget—that seems to be the roadblock in most of our programs. We never have enough resources to really do what it is that we plan to do. That's a reality check and we have to pare back so that we can do our program within our prescribed budget. The system design is where you do your research plans, your QA project plans, your design plans, data processing (you have your computer hardware/software needs), and field operations, which describes your QA project plans and your logistics plans. Analytical laboratory operations from the QA perspective is described in the QA project plans. You have data reduction and analysis and validation, and again, the procedure for that should be described in your QA project plan. You have your audit reports and your corrective action memos. System certification—you go out and do pilot studies and demonstration projects. Finally, you conduct your full scale study which is implementation. Then you have your data reporting, which is your final report. How does this process fit into ENAP, which is the Agency's ecological monitoring assessment program? And when I say the Agency's ecological monitoring assessment program, I think I should step back, because that is not quite true. EPA is making a concerted effort to make EMAP not an EPA program but a Federal program. We're looking for partners out there in the Federal sector to work with us on EMAP. And I think we're doing a fairly good job at working with the other Federal agencies. 92 ------- We did a pilot study this spring in Virginia Province which is an estuary system, and we worked very closely with NOAA's Status and Trends program. We're working closely with the BLM when it comes to arid lands, and we work with the USDA Forest Service with respect to the forest monitoring effort that was done in the East Coast this last year. We are also forming partnerships with other Federal agencies. The EMAP process of going through this is that first we do a pilot study. The pilot study entails designing your study, conducting your study, and of course, interpreting your results. If it was successful, then you move on to a demonstration. If it's not, you go back and do a pilot study again. A pilot study is a very small study where you test all your processes and all your hypotheses to determine how you should design your study and choose the indicators to tell you what it is that you want. A demonstration project, on the other hand, is a large-scale pilot study done on an eco-regional basis. It allows you to test things such as spatial variability to make sure that it isn't so great that it swamps out the variability of your measurements. And again you go through the same scenario, and if it's successful you move on to implementation, which is where you take it on a national scale. One of the things, though, with implementation is that we're never quite satisfied. We don't implement it and just stick with it. We have this thing called continuous improvement, and we keep recycling back. So even though you're in the implementation phase, you've got to continually reassess what it is you're doing, relook at your indicators, see if it's telling you what you want it to tell you, and redesign on a year-to-year basis. How does that fit in with the QA techniques that are mandated by EPA? The QA tools—DQO's, QAPjP's, and audits—are required in each stage of the process, at the pilot stage, the demonstration stage, and the implementation stage. The difference is that the content of these various documents differs depending on which stage of the project you're in. For example, let's take the QA project plan. You may at the pilot stage make it more extensive than it is at the implementation stage, because at the pilot stage what you also want to do is to test all those QC techniques out there to determine what QC techniques are needed to control those parameters that you're measuring. That way, when you go to the implementation state you can do the most cost effective QC on a national basis. When you're doing basically thousands of samples, you don't have to be doing all that same QA/QC that you did at the pilot stage when you were trying to learn and assess. If we achieve our goals, we'll end up with a national integrated ecological monitoring program where many different Federal agencies will be contributing to the EMAP database. You'll have EPA 93 ------- ?,™ri »UrJ-ngJ, but you U also have the Fish and Wildlife Service, USGS, BLM, NOAA, the states, the local government, and other research efforts that can all be feeding into that. If we're really successful—maybe another decade out—we'll end up with an international integrated ecological monitoring program? EMAP has already made some strides in that direction. We've talked to Brazil and they do have the rain forests on one of the EMAP grids; we've also been over to Australia with the arid-lands people over there and Australia is very interested in the EMAP concept and with becoming part of the same grid system. So we are moving into the international ecological monitoring program. In closing, I'd like to reiterate what Robert Layton, the Regional Administrator for Region 6, said yesterday. He charged this group to develop data collection processing and reporting techniques that transcend agency boundaries. I think that's really what harmonization is all about, at least on the ecological monitoring scale. We need to develop these types of techniques that transcend across the agencies so that data from NOAA and the Fish and Wildlife Service can all be pooled and put into a common database, and we can really get around to doing an integrated data assessment. Thank you. Adriana y Cantillo I'm here to describe briefly the National Oceanographic and Atmospheric Administration (NOAA) National Status and Trends Program, which has just started its sixth year of sampling. The QA function of status and trends is an integral part of the program and was included in the program plan from its inception. I must add that this was done to the chagrin of the potential contractors because we're dealing with a community in the marine sciences field that is used to QA, and initially the comments were: we don't need to do QA, we know what we're doing. But now they have decided that this is an excellent idea and in presentations that I have seen, they make a point of saying that they participate in the status and Trends QA Program. So they're actually publicizing the fact that they take part in it. The Status and Trends Program measures the current status of and any changes over time in the environmental health of Eastern and Coastal waters of the United States. It has six major pieces— benthic surveillance project, mussel watch project, biological surveys, the QA program, the specimen bank, and historical trends assessment. The major pieces that I'm going to discuss are benthic surveillance, mussel watch, and the QA program. If anybody's interested in further information on the program, feel free to drop me a line. We have almost 100 publications on results and interpretations from the program—we'll be happy to get a list of publications to you. 94 ------- Benthic surveillance collects sediments bienially in bottom- dwelling fish from around 75 sites in the United States. All analysis is done by NCAA's National Marine Fisheries Service. The sampling sites are determined by where the fish are. So over time we may have to go to different places in an estuary. The fish are collected -by trawling, so we are dependent upon where the fish happen to be on that particular day. We are also looking at the incidence of fish disease and tumors. Mussel watch is a continuation of the 1970's Mussel Watch Study that was conducted by EPA and other groups. In this case sediments and bivalves (mussels/oysters) are sampled around the country yearly at about 220 sites. Texas A & M University and Battelle are the contractors for this study. The sampling sites are governed simply where the mussel or the oyster beds are. We collect natural specimens so where they live is where we go pick them up. We don't have, again, any choice in where the sampling sediment sites are. The QA program will document all the sampling protocols and analytical procedures used in the Status and Trends Program. This becomes very important from the fact that we have to pool samples together to get enough material to analyze. And so it's very important that the various contractors know that they have to prepare composite samples of X animals and what part of the animal they're supposed to collect and be consistent about it over time. We also want to reduce variations inside a particular laboratory and between laboratories. We want to use the QA program to eventually compare our data to the data generated by other programs. In methodology, we're quite different from EPA. We do not specify any analytical methodology. The laboratories are free to use whatever method works. This frees the laboratory—especially the contractor laboratories—to use analytical instrumentation that they may have on hand and have the expertise on knowing how to use expertise that perhaps is not common to everybody. This is quite all right, as long as they can produce results in the intercomparison exercises that are similar to everybody else's. This is very much appreciated by the contractors and it eliminates a lot of headaches for us. We require the use of standard reference materials in control samples. We contract with the National Research Council of Canada, and with the National Institute of Standards and Technology (NIST) in the preparation of control materials and standard reference materials (SRM's). Some of the materials that have been prepared for Status and Trends have become SRM's. All the methodology and sampling protocols are documented. We're in the process of doing that now. It's difficult because the contractors don't seem to want to provide the detail of information that we want for the documentation documents we're preparing. We 95 ------- do want to put on paper the level of details down to the part number of the materials used in the plastic ware, because we do not know if 10 years in the future that may be something that will be critical. So we're doing a very detailed documentation of methodology. And we're looking at the addition of all the results of intercomparison exercises into the national Status and Trends database, so that anybody looking at our Status and Trends data will have available the results of intercomparison exercises, things such as detection limits, what the precision was, how the laboratories compare with each other, and how their accuracy was compared with SRM's. This will all be part of the electronic database. The most important part of the QA program is that each year there is an intercomparison exercise that takes place and the NOAA contractors, including the National Marine Fisheries Service, are require to participate. The organic intercomparison exercise is prepared and run by the National Institute of Standards and Technology (NIST), and the inorganic by the National Research Council of Canada (NRC). NOAA does not do anything in the laboratory; we do not have a lab in the office; we contract this to NIST and NRC. So they act as an independent sort of judge or evaluator of how our contractors are doing. Sample types in the past have included free strike sediments and extracted sediments, so that you take out that variable of the extraction, homogenized frozen tissues. They now want to explore the possibility of tissue samples from matrices, such as mussels other bivalves, fish, and also sediments from clean areas and contaminated areas. The way this works is that every year at the beginning of the year, the labs get the samples from NRC and NIST; they analyze the results; and then they send back the results to the originating organizations. They, in turn, get all the results together, and every year they have a meeting with all the laboratories and the results are presented. Our core laboratories have been participating in the program since it began five years ago. They are doing very well in the analysis of the intercomparison exercise samples. We just opened up the QA program to other laboratories; this is a way of showing that participating in these exercises has a beneficial side effect. Our core labs do very well compared to NRC, and the new labs still have some work to do. Every year there is a meeting in the late fall or winter once the results of these intercomparison exercises are sent back to NRC and NIST, where the laboratories that have participated and would like to go to the workshop can attend. The purpose of the QA workshop is to show everybody's results and discuss what problems there may be in common and what the probable solutions are. It's almost like a teaching situation. During that workshop the laboratories and 96 ------- NIST or NRC decide what kind of samples they want to do next. So the choice of sample for the next intercomparison exercise is, most of the time, the choice of the laboratories. In other words, they may say, we're having problems with analyzing a real clean sediment, or we have analyzed clean sediments before but we're not sure how we would do with a real contaminated sediment, so we'd like to do that. Or, we would really like to test and see what our extraction is . like. Can you send us something that is not extracted so we can add that into the degree of difficulty? It changes from year to year and that's why it's not possible to show you an improvement of one lab over time. The samples change and the degree of difficulty changes as the labs get better and better. In future developments, some of the EMAP labs are now going to take part in our QA program, especially the ones that will be working in the coastal area. We also are going to open up the intercomparison exercises to non-NOAA contractor labs, or any Federal or state government lab that would like to participate. The response to this has been overwhelming. We have about 30 labs already that are going to participate in the inorganic intercomparison exercise. The organic exercise is completely full and we have a waiting list of about 25 labs. A lot of the comments from the field are that we're doing such and such analysis but we're not sure whether we're doing it right. So the response has been very good. We haven't been able to open it up to more labs simply because of funds. It's very expensive to do this, but probably slowly we'll be opening this up. We have found that for marine environmental samples there isn't that much QA available. The marine chemists are an independent breed and a lot of them have become chemists because they need to determine something in the marine environment. So they don't see the need for QA and don't like to do it. But very slowly the point is getting across. We're also beginning to work with international groups. Right now we have the International Mussel Watch that is scheduled to begin in the Caribbean, probably this year or next, and they will be part of an intercomparison exercise since we also do Mussel Watch. Slowly we're spreading and trying to get the word out. If anyone would like to participate in the trace metal intercomparison exercise, please give me a call. James K. Andreasen The mission of the Fish and wildlife Service is to conserve, protect, and enhance the fish and wildlife resources of our country 97 ------- and their habitats. I want to briefly describe an ongoing marketing program that we've had, tell you a little about some of the results of that program, and then discuss a new effort that we have started. During all this I'd like you to consider how this all interacts with the QA requirements. Consider the complexity of the environment, the number of different species that are out there that should be monitored in order to say something about the health of the environment. Think about how diverse those habitats are and the difference between the life history of each of those critters we're trying to collect, different collecting methods that are needed, storage techniques, and everything else that goes along with that. Maybe you'll get some idea of why we think there is a tremendous need for having harmonization within environmental sampling. Like EPA, we're a strongly decentralized organization. We have seven geographical regions in the country plus a separate research division. It makes another level of complexity on top of the whole thing. The Service has responsibility under several Federal laws and international treaties to manage migratory birds, threatened and endangered species, anadromous fish, and certain marine mammals and lands that are under the control of the Service, such as our National Wildlife Refuges and National Fish Hatcheries. Currently there are about 475 wildlife refuges around the country and this consists of some 89 million acres of land. We are in the process of developing a monitoring program for each one of those sites. These are some of the resources that we're charged with protecting: migratory birds, endangered species, and brown pelicans. We have our various wildlife refuges and they're all being assaulted by various kinds of environmental impacts. All you have to do is open up any newspaper and you'll see that there are daily new problems coming along that are going to affect one of those species. A lot of things we found in our previous monitoring efforts that showed the impact of environmental contaminants on natural resources were from our irrigation drainwater program which we've had functioning throughout the west. We saw impacts from the Alaskan oil spill, not only the critters being killed by the oil directly, but the secondary effects on bald eagles from eating those oiled birds. We're looking at a lot of these things, not strictly for the direct impact on the critters, but rather the fact that a lot of these birds and other fish eat smaller things that are in the environment, so we're concerned about the impact on the birds consuming contaminated prey in some circumstances. I want to describe briefly a program that was originally called the National Pesticide Monitoring Program (NPMP). It was set up by Congress in 1967, right after Rachel Carson's book hit the stands. 98 ------- Originally there were many agencies involved in the project. Gradually, over the years, most of them dropped out. The Fish and Wildlife Service kept the thing going and we're still collecting fish—not primarily pesticides any more—but we're looking at the whole suite of environmental contaminants similar to the list that the National Marine Fisheries Service is using for the Mussel Watch Program. We're looking at that same suite of contaminants still. We're going for pesticides, PCB's, and a suite of metals. When we originally set up the statijons for this program, we tried to look at selecting areas that would be fixed stations, there over time, and that would integrate a large watershed area in the fish part of the program. We originally also looked at whole body starlings, and we looked at duckwings. Those programs, because of the extreme migratory nature of birds, were not as successful as the fish program so we've dropped those in the last few years; but we are going to reinstitute another sampling in that work for birds. We've tried to pick sites so we could integrate a lot of watershed areas but yet not be impacted by a point source. This is what the network ended up looking like. There are presently 112 stations in the program. Samples are collected on a two-year cycle, and we analyze these samples in-house or at the Columbia National Fisheries Contaminant Research Lab in Columbia, Missouri. It takes about two years to get this many samples analyzed. Here again we can only sample the species that are there, but we tried to establish criteria for each one of the different kinds of sites. It was cold water, warm water; we tried to look at a predator species and another species that was more tied to the bottom. Over the years we've decided that we would only sample fish for whole body residues. We don't do fillets for this program because we're more interested in what the consumer organisms are eating and the impact on their health, not so much on human health, although this data set has been used in several publications to try to get some idea of what the impact might be on human health. We look at composite samples. Our directions specify that at each station we would collect two samples of a bottom dwelling fish and one sample of a predatory species. There will be five fish in each composite of the sample. They were to be selected for uniform size. We established the criteria; the fish weren't supposed to be flopped around on the bottom of the boat where they got a lot of oil and grease on them. We took them wrapped, froze them immediately and then shipped them off for analysis. All the samples in the field are collected by our own biologists. We have about 60 field stations around the country and 110 or so biologists that work in the program. They do things other than just collect these samples. So we are collecting in-house; we do the analysis in-house also. Common carp is one of the species that 99 ------- we've collected at almost all of the sites in the country. They're an introduced species, but they're now ubiquitous in the country; Also large mouth bass. Over the years, looking at the number of stations we've collected from and the number of samples, we have well over 3,500 individual composite samples of fish in this program, and cumulatively we've collected at over 1,000 locations. One of the problems with QA in the program has been that there's been a difference. Analytical techniques have improved over the years. When we first started, people didn't know there were PCB's in the environment; therefore our early organic chlorine residue data also included PCB's. We were later able to separate that out. Same thing with toxaphene—we didn't know about toxaphene originally. Toxaphene has been detected at stations in watersheds where it was never used, indicating aerial transport of the chemical. Generally, over the program, there has been a decline in the concentration of some of these things in the environment. Isomers of DDT and dieldren show a good decline over the years. The same thing with PCB's. We've seen a pretty good decline in PCB's. They're not quite as dramatic; but as the technology has improved, we're more sensitive to what's out there. Some of the elements, though, don't show this same kind of a decline. We haven't seen a decrease in selenium concentrations across the country; its residue seems to be staying about the same. The program has been very successful in showing declines of some elements and of some compounds. There's been a statistically significant decline in DDT, PCB's, and dieldrum between 1976 and 1986; this indicates that the ban on these chemicals has been effective in removing them from the environment. The data also indicate that there's a geographic spread of PCB's and toxaphene in the environment. Toxaphene was never applied around the Great Lakes, yet we're finding significant residues of toxaphene in Great Lake fishes. I think this is primarily due to aerial transport, something that we hadn't considered before. PCB residues continue to be a problem in parts of the country. In Hudson River fish, concentrations are greater than 10 parts per million, while PCB residues in other parts of the heavily industrialized northeast are somewhere between 2 and 5 parts per million. The data set has pointed out the need to do more intensive studies in some areas where residues have been high over the years. We've gone back in and done more intensive sampling. In order for a monitoring program to remain effective, it has to be able to adapt to new and previously undetected compounds. It has to be able to respond to new locations that are needed, and we need to be able to adapt to new management perspectives for what kind of 100 ------- data are needed, with that in mind, the Fish and Wildlife Service has designed a new program that we're calling Biomonitoring Environmental Stats and Trends, or the BEST program. We have completed a design document. This has gone out for review across the country. BEST is going to be the overall umbrella for our program which will be site specific sampling, it'll be probability based, but within fixed stations. We're hoping to be able to integrate all these different things into one program that will answer the questions that are being asked of us by Congress. The Fish and Wildlife Service has also signed Memoranda of Understanding (MOU) with EPA for the EMAP program. We're involved with EPA on several other programs that involve monitoring and doing assessments for nonpoint source pollution through our drain- water irrigation and quality program. We're partners with USGS in their National Water Quality Assessment (NAWQA) program, which you'll hear about next. And we see a need for having harmonization of these collection programs across all the different programs in the government. We're hoping that there's a new day dawning for the environment. Through all these various programs we will have a greater concern for our resources. Thomas F. Cuffney It's a nice feeling to be able to come and visit with EPA and give a presentation. But I must confess that I was a little uneasy to come here and address this group. For one thing, I am not a Quality Assurance Officer, though I am and have been involved with our branch of QA, particularly Dave Erdman, Bill Shampine, and Tom Maloney. In addition, I'm not a chemist. And beyond that, I'm not an engineer. So the question comes up, what am I? Well, I'm an ecologist. I am one, as I understand it, of only two people with that job description currently in the U.S. Geological Survey (USGS). I'm also going to be talking about something I think is very different from what has been talked about in terms of QA here— QA applied to biological programs, specifically programs that are not involving chemical measurements. In addition, I come from an agency which has no regulatory responsibility and which neither owns nor manages any lands. This actually has a lot of advantages associated with it. I do come from an agency that has a long history of providing high quality and long term data and interpretations for its customers. My function in the National Water Quality Assessment (NAWQA) program is the development of protocols for the ecological surveys. These QA programs that I'm going to talk about are interesting in the fact that they're being driven by the science and not by any type of regulatory requirements. I put this talk together before 101 ------- I had an opportunity to read the harmonization document. I ask that you focus on the intent and content of this talk relative to the harmonization document and not on the vocabulary. The objectives of NAWQA are to provide nationally consistent descriptions of current water quality conditions, to define long- term trends in water quality and finally, to identify, describe, and explain, to the extent possible, the major natural and human factors affecting water quality conditions and trends. So three themes are running through this: description, trends, and cause and effect studies. There are three large scale components of NAWQA: surface waters, ground waters, and biology. The surface water and ground water components each have chemistry and hydrology involved in them. The biology is basically tissue surveys, which most people here are probably comfortable with, and also the ecological surveys. NAWQA is a large scale, long term program. We have 60 basins across the United States, including Alaska and Hawaii, which are being studied. The focus of the study will take at any one time 20 basins under intensive studies. It will take 10 years to complete an entire cycle involving all 60 study units. And each basin in there is a study unit. We're dealing with a hierarchy of water quality issues within the NAWQA program. At the study unit, or the basin level, it's the level at which the data collection occurs. Data collected at that level will be used to look at issues at a variety of levels increasing in spatial content: local water quality issues, regional water quality issues, and national water quality issues. So in terms of QA, we need to be producing data which can be applied across those different levels. In addition, looking at harmonization across Federal agencies, it will be nice that we can provide information compatible with other programs that want to use this information at different levels. The goal of the ecological surveys is to characterize the distribution and relative abundance of biological communities. We're looking at three community types. One is the benthic invertebrates, which I am most involved with; the second one is the algae; and the third is the fish communities. In these communities, looking at the distribution and relative abundance of these communities in .terms of water quality parameters, there are three elements involved—distribution, which implies spatial characterization of communities; relative abundance, which implies enumeration; and the biological communities, which imply taxonomy. The QA/QC needs for benthic invertebrates are that we need a consistency in field collections. If we're going to compile data across regions, across basins, we need to be dealing with consistent field collection methods. We need consistent lab 102 ------- processing—in other words, sorting. A lot of what goes into processing these samples is physically removing organisms from a matrix of organic material and sediment. We need accuracy and consistency in identifications, not only at the time that the samples are identified, but also because the science of taxonomy is constantly, changing. What is listed as species A this year may be species B next year. You have to keep on top of those changes if you're going to have data which are truly timeless. We also need accuracy and consistency in the enumeration, in the actual processing or counting of these invertebrates as they are collected. The way that we are trying to get to these objectives within the NAWQA program is to develop consistent approaches for field and laboratory processing. There are at least five parameters listed for QA for our field collections. The first item is standard protocols. That's what I'm involved with at this point, with developing sampling protocols which can be used across the nation in teams of different sizes. We typically use the term protocol and don't use SOP's, though they pretty much serve the same function. I want to make a point about harmonization across Federal agencies and that is that we should be more concerned with the intent and content and less with the vocabulary of what we're using. It's probably very difficult to get another agency to change what they're calling an item. It's probably far easier to get them to include extra material in a protocol that would satisfy your requirements for an SOP. We're also involved with formal training through our Denver Training Center, where we will train the project personnel who will be collecting the samples. Training involves more than just teaching these individuals how to go out and physically collect the sample. We need to train these individuals in the philosophy behind why the sample is being collected, and also in some of the theories about the operation of streams so that we can obtain from them feedback in terms of what's going on in the field and better design the program as we go along. Documentation, of course, is a key issue. We take copious field notes in the survey. The problem with field notes is they're seldom seen again. We have standard data forms which we are using, which also tend to get away. This lack of accessibility for field notes is a problem. Under our new computerization scheme, which is our National Water Information System II, we are working toward incorporating documentation into the database itself, even to the point where we're discussing the possibility of scanning field notes into the database so that if you were working with this data and you needed to call up that information, you could access it. In this current day and age, if it's not in the computer, it doesn't get used. And we need to work toward better computerization of our support 103 ------- information for the data that actually goes in the database. Another important QA issue here is our interstudy unit communication. We have regional biology teams which serve as advisory boards for each of the study units; we have a Fish and Wildlife person who will be part of that biology team. In addition to the biology teams advising the study units, there will be study unit liaison committees, which will be made up of people from Federal, state, and local governments, as well as university people who will help the projects develop their work plan. Sample processing QA/QC for these benthic invertebrate samples becomes somewhat tied up with the whole issue of contracting. What we are looking for is to achieve some standardized contracts which involve putting into those contracts QA/QC checks that are provided by the contractor and can be used by all the study units. We also are working toward implementing a contractor evaluation or certification process so that we can identify laboratories which can provide us the services that we need and also develop a formalized process by which we can weed out those organizations which bid on a project, but do not have the capability to provide us with the services we need. We're also looking at regionalizing contract laboratories so that contract laboratories in a specific region with a specific expertise in a taxonomy will be receiving the samples. This will help us, I think, both in terms of logistically controlling the contracting process and oversight for the laboratory, but also ensure that these people are familiar with the fauna across the country, which varies considerably. We're instituting a USGS QA/QC laboratory oversight, which will involve actually putting together a laboratory with people to monitor the contractor compliance of the samples so that we can monitor how well these contractors are doing in processing the samples and ensure that we're getting accurate and consistent information. We're working with the computerization of taxonomy, also QA/QC checks and sample tracking, so that as information is entered into the computer, it will tell us that that's a valid name, or flag the information to tell us that that name is not valid, the authority with that name is not valid, or that that species may not even occur in that region of the country. This is an important check for us. This, along with the QA/QC checks and sample tracking, will be in our National Water Information System too. We're also trying to develop a standardized taxonomy database employing the NOAA National Oceanic Data Center Species Codes. There's an interesting story about harmonization here. This body of taxonomic information is maintained by NOAA. It contains information on everything from antelope down to protozoa. The 104 ------- information related to marine taxonomy is very good and very well maintained. The information related to fresh water biology is not very well maintained. As you can imagine, NOAA does not have as much interest in the fresh side as the marine. But the USGS and EPA are both interested in using this database and incorporating it into their databases to keep track of the taxonomy. So the USGS is funding a NOAA person half-time to update this list. EPA is also cooperating with that. Another way to ensure quality is to deposit voucher specimens in outside collections. The voucher specimens are physical organisms which represent the name that you're attaching to them. And you put the speciman out there in an outside laboratory, a museum, or university somewhere, where other people who are specialists in this field can look at that specimen and decide whether you've named it right and inform you of any current changes that have occurred in the nomenclature. I think that this is a very important component in getting credibility for your identifications. In addition, we will be curating study collections within the USGS, which will be available both to the survey personnel and to people outside of the survey who wish to look at these. When you put this all together, the striking thing is the Biological QA Group of NAWQA. NAWQA is making a big effort to ensure the quality and timelessness of this biological information so that when we meet again, 20 or 30 years from now, that information will be just as good then as it is now. I think there's room for a lot of harmonization and a lot of interagency cooperation in developing these types of protocols. 105 ------- DISCUSSION SUMMARIES After hearing the preceding panel discussions, participants at the meeting met in small groups to discuss how they felt about implementing a national consensus standard for EPA and other Federal agencies with environmental programs. Each group then chose a spokesperson to report on the nature of the discussion. Quality Assurance Management Staff member Gary Johnson summarized the break-out discussion sessions. Question 1: What are the indicators for success in this effort? What benefits could vour organization anticipate bv implementing a new quality framework/standard? Group 1 determined the Indicators of Success to be: 1. General acceptance (stated in writing, use of other agency's data, outside group accepts and uses data) and 2. Reduced litigation. The benefits of harmonization were catagorized in four sections: 1. Cost/Schedule/Performance (improved allocation of resources, increased efficiency, reduced frustration level of multi-agency requirements) 2. Cost/Schedule (less repetition of work, less time and money spent, reduced time for inter-agency negotiations) 3. Cost/Performance (large data base) 4. Performance (common language, better quality products, consistent products, defined set of compliance requirements, better staff buy-in, clear definition of objectives across agencies, enhanced assessments). Group 2 concluded that Indicators of Success were: 1. Interagency Acceptance/Recognition (at all levels, regulator accepts product, regulated community accepts guidelines and standards) and 2. Elimination of redo's (realization that Accrual of Benefits = Indication of Success). Benefits of harmonization were seen as: 1. Data Comparability through application of standardized criteria (review, scoring, index values); 2. Reduced Redundancy/Paperwork (project and program); and 3. Translation to Management/Management Buy-in (data quality = data utility). A data quality "hierarchy" was discussed in the following order: Data (measurements and numbers); Information; Interpretation; Knowledge; and Decision-making. Panelists also determined that NIST has a role to play in harmonization. Question 2: What are the roadblocks to implementation of a new standard in your organization? What actions/changes will need to 107 ------- take place within your organization in order to implement a new standard? Group 3 regarded the roadblocks as the absence of a common language, displacement of old guidance, resistance to change, and "not invented here." Suggested actions included educating affected users, keeping open communications, and tailoring marketing strategies. Group 4 discussed six roadblocks to implementing new standards: i. Changing how people think—priority changes; 2. Ineffective communications; 3. Lack of Standard Data Comparability; 4. Fixed budget needs a good DQO, but lacks funds; 5. Management takes holistic stance; and 6. Resistance to change. Panelists agreed that several act ions/changes needed to take place: the addition of training programs, "selling" the idea (P.R. "benefits" and payoffs, marketing, cost savings), more effective use of communications, general acceptance of change, and having states "buy in" on the idea. Question 3: What factors within vour organization will facilitate acceptance of a new standard? What information/interaction from other groups or agencies would help to pave the way? Group 5 agreed that the degree of acceptance is proportional to the perceived benefits, that user implementation must be brought into the decision-making process, that the standard must be a practical tool for the user/implementer, and that senior management must be involved in a substantive manner. Panelists talked about the need for "consensus" from agencies other than EPA, the need for a mechanism to establish true interagency harmonization, and the requisite for a consensus to set performance criteria for environmental monitoring. Other observations included: 1. the document needs to be written in plain English, not technical jargon; 2. there must be sufficient specificity to minimize the range of interpretation; and 3. enforcement and litigation efforts must not be restricted. 108 ------- NATIONAL PROGRAM OFFICES From a summary by Marty Brossman Quality Assurance Officer, Office of Water The National Program Office Sessions included: the joint session with Regional Offices on the EMMC, a session on Quality Assurance Program Plans—guidance and implementation, and a session on Qualifying Data for Specific Uses. A tentative panel was proposed on TQM—concept and implementation, but time constraints precluded this session. The issues to be discussed in this session are of importance to the continuing success of TQM and QA in the Agency; accordingly, they are briefly summarized in discussion of the sessions. Quality Assurance Program Plans fOAPPl The QAPP is recognized as an important management level document at each Office, Regional and Laboratory level of EPA. It describes the QA policy, roles, responsibilities, and plans for implementation of the Agency and specific Office, Regional, and Laboratory programs. The QAPP, as signed off by a top management official, the QA manager, and with concurrence by the Quality Assurance Management Staff (QAMS) of the Agency, is equivalent to a EPA contract. Meeting that contract effectively is important to the success of a QA program. This session was designed to address issues of QAPP development and implementation to improve effectiveness. The difficulties in developing an effective QA Program Plan were addressed in detail. They included: the need to gain command of widely diverse studies and programs; management's lack of acceptance of its responsibility for the quality and defensibility of its data; the isolation of some program offices from the data acquisition tasks imposed on Regions and States; and a lack of understanding of the need to define the customer and data use. Two effective tools to reinforce the QA program were discussed. The utilization of Management System Reviews (MSRs) has proven a useful tool in evaluating a QAPP in place. Short falls become apparent when performance is evaluated against specific commitments and responsibilities. In addition, the QA Annual Report is also a useful tool. This report, prepared by the QA Officer as a part of his QA Program Plan, provides a vehicle to discuss accomplishments and short falls with management, and gain agreement for the next year's plan. While action items were not developed as a result of this session, opportunities were described to provide recommendations to improve the QA Plan development and implementation process in the new guidance being developed by QAMS. 109 ------- Qualifying Data for Specific Uses This session was designated as an "umbrella" session to cover a wide range of issues related to data qualification and use. Issues raised by Larry Keith of Radian Corporation and Wendy Blake- Co leman. Immediate Staff, Office of Water, brought forth extensive discussion. Larry pointed out that with the commonly used Limit of Detection (LOD) or Method Detection Limit (MDL) set at Standard Deviation (SD), there is a 50% probability of a. false negative detection, but less than 1% probability of a false positive detection. He suggested that whenever false negatives were important to the Agency, the LOD or MDL should be set at 6 SD, where the probability of false negative and false positive detection is equally low (less than 1%). Because of the high reliability of detection assignments at 6 SD, it was suggested that this be called the Reliable Detection Level. Mr. Keith also suggested that a distinct difference be made between laboratories reporting data and users/requesters taking laboratory data and presenting it in final form. He recommended that laboratories report all data unless requested not to do so by the user/requester of the data. He further pointed out the potential liability of the sometimes current practice of reporting "Not Detected" (ND) rather than "less than MDL or LOD" when an analyte is detected, but less than the MDL or LOD. Mr. Keith therefore suggested' that "ND" should only be used when there is no measurable value and not when the value is between 0 and 3 SD. Technical issues of this type are extremely important to the Agency since many toxic materials are harmful and below our ability to detect them. Accordingly, agreement on detectability definitions and related issues directly impacts the Agency's ability to regulate and defend its controls. Wendy Blake-Coleman addressed a range of data issues-some impacted by the Office of Water's recent reorganization. As a preface, she described the reorganization and the new locations of the Office's major environmental data bases. She also described the new role of the Policy and Resources Staff in data base and information resources management. Issues being addressed include the STORET Modernization program, integration of CIS into the data systems, and development of QA data and usability guidance. The concept of minimum data set was then discussed with particular reference to the ground water data base. The ensuing discussion centered on whether the data elements were primarily descriptive and not quantitative. The range of issues addressed were so broad that individuals decided to follow up specific areas of their interest with other discussions. 110 ------- Total Quality Management Time demands precluded follow-through on this session. Some of the issues planned for discussion in follow-up sessions, however, are summarized here. The current Agency and National focus on Total Quality Management provides great potential for progress in the QA and TQM programs— if managers understand the direct relationship. TQM has developed from within the QA/QC field represented by such professional organizations as The American Society for Quality Control (ASQC). Principles formerly applied to customer requirements in hardware and data have now been applied to other goods and services. Thus, it is apparent to many that managers cannot "buy-in" to a TQM program without having automatically endorsed the QA program. We should explore opportunities to market the QA program to Agency managers as a good example of TQM in practice. Ill ------- REGIONAL OFFICES From a summary by Dale Bates Chief, Environmental Services Support Branch This afternoon session focused on drinking water regulations, field quality assurance and quality control, the Environmental Monitoring & Assessment Program, statistical support needs, and Superfund's Enforcement Survey. Highlights of. the discussions are presented below: Drinking Water Regulations Al Havinga and Herb Brass briefed the group on the following topics: - Draft corrections to the January 1, 1991 notice in the Federal Register that will be published in July. - Regional concerns on regulatory issues. - Participants in the workgroups on lab issues (with representation from Regional Offices, Environmental Monitoring & Support Laboratory - Cincinnati, and Office of Ground Water & Drinking Water (OGWDW). - OGWDW will provide the Office of Regional Operations (ORO) with a listing of current workgroups. - Comments on the new Chapter 5, Laboratory Certification Manual should be forwarded to Nancy Wentworth by May 1. Nate Malof explained the Privatization & Cooperative Research & Development Agreements (CRADAs). Field OA/OC Llev Williams identified and discussed issues of concern and training on Field QA/QC. Environmental Monitoring & Assessment Program Marcus Kantz identified the need for coordination of activities with Regions. Bob Graves volunteered to encourage EMAP personnel to enhance communications. Statistical Support Needs Kent Kitebingham discussed the need for expertise and training in Regions, and the need to develop strategies with positive impacts. Superfund Enforcement Survey Region 2 revealed significant resources utilization in support of the Superfund Enforcement Program. 113 ------- OFFICE OF RESEARCH AND DEVELOPMENT From a summary by Allan Batterman Quality Assurance Manager Environmental Research Laboratory Quality Assurance in Modeling Projects This session focused on methods used by QAMS and modelers to address QA in modeling projects. A draft document was presented for review, and panelists edited the document for general ORD input and approval as a guidance document for modeling QA/QC. The key question for review was: How is modeling QA/QC addressed in Agency model development and use? Five presentations were given, with discussions of each following. Definitions of Modeling (overhead transparency) provided two acceptable definitions of modeling: l. Experimental data gathering that provides information to test hypotheses—computer programs implement the models which are environmental compartments. Processes and water relationships; 2. A mathematical equation or set of equations that describe relationships between variables of interest. OA of Models (overhead transparency) suggested that modelers consider up front and in writing, points that are necessary to assure validity of the model. Panelists determined that the validation of a model is easier when a written plan with guidelines and expectations is developed beforehand. Purpose of Modeling Document (overhead transparency) addressed topics such as providing documentation of procedures, applicability, and reliability. Points of discussion were: making sure creativity is not stifled when implementing QA into model development; changing the QA approach to appropriately reflect a changed model; asking the questions, Is this to satisfy the QA officer? Is QA suitable for the scope of the project?; the importance of stating the intended purpose/objective of the study up front; realizing that endpoints are descriptive of a process and interpolation may be necessary; understanding that suitable modeling for risk assessment may require specific QA; and the importance of research planning. Modeling Project (overhead transparency) discussed the process of developing a concept, validating and establishing an existing model, applying the model and verifying it, and using the model as part of data analysis or interpretation. Panelists raised several issues: documenting the accountability—determining if the modeling project should proceed; the necessity of avoiding bad models by deciding the purpose of the modeling study and evaluating and examining all possible variables of the model; and that QA should apply to the computer issues of the model, including the computer codes. The U.S. Geological Survey determined that 115 ------- modeling should be different than other investigative processes in the QA arena because creativity is required and examples of good case histories are needed. The last presentation, Computer Code. (overhead transparency) focused on evaluating models, suggesting corrective action if necessary, and testing computer models. Panelists determined that the computer code, how the computer manipulates the data into subsets, is subject to human error and must not be assumed to function properly at all times. Error checks must be made, as well as a range of acceptance set up in which samples should fall within. In addition, all errors must be documented through the use of "acceptance reports." Acceptance reports determine what data fall out of range, and adjust calibration if values are higher or lower. Members of the panel also agreed that models taken from scientific documents and publications may not have been through QA; therefore, it was important to rely on the professional judgment of project personnel to determine that procedures were addressed properly. If the model did not work, new variables would have to be studied, reassessments figured, and corrective actions taken. QA PLAN (Documentation) Types of Information Considered A. Project Description/Resources - scope and goals - project personnel B. Model Description - model equations - purpose/relevance - limits/parameters C. Data Quality - calibration/parameterization - types/quality data needed - evaluation of data - source/acceptance/format/data uncertainty D. Model Uncertainty - sources of uncertainty - assumptions - sensitivity analysis - comparisons E. Computer Program - code verification 116 ------- - records of code versions - programming documentation - user documentation Privatization of Quality Assurance Reference Materials Using the Federal Technology Transfer Act (FTTA) Cooperative Research and Development Agreements (CRADAs) The focus of this session was on EPA-certified QA reference materials: what is available, how to make them accessible on a continual basis, and where to obtain them. The discussion also included EPA's oversight role, if materials would be EPA-traceable, and questions about accreditation. The goals of the session were to: l. Minimize disruptions in existing operations; 2. Reduce ORD expenditures but assure dependable study, timely development of new products, high quality, and a reliable distribution system; 3. Encourage competition through accreditation; and 4. Maintain a long-term program to meet Agency needs. EPA's oversight roles were determined to be: 1. If wholesaler does not produce compound, EPA will furnish to the wholesaler, who will distribute; 2. Two analyses in-house; 3. Analyzing and verifying the lot before a sale; 4. All "EPA certified samples" will be stability tested. $12 million, 7 Years - What Have We Learned? (Applying DQOs to a National Survey) The purpose of the session was to examine a case study for potential areas of improvement in applying the DQO process; the outcome was to get recommendations for successfully applying the DQO process to large surveys. Lora Johnson, QA Manager, National Pesticides Survey, discussed the survey results and two approaches for evaluating the success of the survey in achieving DQOs. Dean Neptune then led a discussion among participants on the difficulties of implementing the DQO process. for large surveys. 117 ------- QUALITY ASSURANCE MANAGER OF THE YEAR AWARD Introduction by Nancy Wentworth We have a significant honor to bestow, the Quality Assurance Manager of the Year Award. For those of you who are new to the EPA community, we have, for the last three years, presented an award to an individual from EPA's QA community who has been nominated by one of his or her peers, supervisors, friends, or someone who has seen the value of the work. This year we received nine nominations. We had individual nominations and group nominations, or a number of people nominated as a group, I should say. We invited three senior managers from the Agency to serve as the award review panel. This is what we consider to be an Agency award and therefore we sought advice on the selection from senior managers. The reviewers were Richard Guimond, the Office Director for the Office of Radiation Programs and in transition to the Deputy Assistant Administrator in the Office of Solid Waste and Emergency Response; Tim Oppelt, the laboratory director at the Risk Reduction Engineering Lab in Cincinnati; and Bill Rice, the Deputy Regional Administrator in Region 7 in Kansas City. They were, as a group, very impressed with the nominations that we received from across the agency. I'd like to take a moment and credit the people who were nominated and give you a little background on who nominated them and what their accomplishments were that warranted the nomination. Then I will make the presentation of the award. First, Jeanne Hankins, who was the quality assurance manager in the Office of Solid Waste and is now on a one-year assignment on the EMMC Laboratory Accreditation Panel, was nominated by the Office Director of the Office of Solid Waste, Sylvia Lowrance, for her work in Chapter One of SW 846, which is the methods and field manual for the Solid Waste Program. She was also credited with her work in the formation of the RCRA Advisory Committee for Environmental Data, the RACED as it's known, which has been a real affirmative step toward improving the relationships with the regions and improving communication within regions on issues relating to the implementation of the solid waste law. Second, Rick Johnson from the Office of Information Resource Management, was nominated by Jeff Worthington from TechLaw, for his work on the Good Automated Lab Practices program. Third, Dr. Henry Kahn was nominated by Ramona Trovato, Office of Water Regulations and Standards, for his work in Fl Guidelines in the application of statistical methods to the development of industrial water pollution control regulations and his work in the statistical design of the national sewage sludge survey and the 119 ------- dioxin and pulp and paper mill effluence surveys. Fourth, Brenda Grizinski and Don Sandifor were nominated by Bill Fair less, the Environmental Services Division Director in Region 7, for their combined effort in the development and implementation of a statistically-based sampling procedure for use in dioxin cleanup sites. This is a project which many of you have heard about, which was an application of the DQO process in Region 7 at the request of Region 7 and which has resulted in a very significant savings of money in a waste cleanups on a dioxin contaminated site in the region. Fifth, a group of individuals from the Office of Underground Storage Tanks in Region 6 were nominated by Bob Layton, the Regional Administrator. These people were William Ray Lindale, Mike Scoggins, John Sernaro, Herb Sharo, Jim Duck, and Audrey Lincoln. They were nominated for their use of TQM principles in the development of the regional program and of the state programs for underground storage tanks. They set up quality action teams; they worked with each of the states to develop an underground storage tank program that was best suited for their legislation, their politics, and their environmental circumstances. Also from Region 6, Mary Ann LeBarr and Judith Black were nominated by Bob Layton, the Regional Administrator, for their work in implementation of field and lab quality auditing programs in the Superfund program in the region. Seventh, Bob Graves, who was nominated by John Winter, the quality assurance division director at the Environmental Monitoring Systems Lab in Cincinnati, who nominated Bob for his work in the development of the QA program for the Ecological Monitoring and Assessment Program. This is the largest ecological monitoring program that the Agency has ever undertaken and is an extensive program requiring coordination of quality programming among and between about five to seven of EPA's ORD laboratories and a number of other agencies including NOAA, Fish and Wildlife, and USGS, so it's been a significant effort on his part to bring all of the quality community together for the common goal of the EMAP program. Next, Marty Brossman of the Office of Water Regulations and Standards, who was nominated by two of his QA compatriots, Barry Towns, the QA manager in Region 10 and Jerry McKenna, who is now the lab chief in Region 2, both of whom are previous winners of this award. Marty was nominated for his work on customizing QA documentation for Office of Water Regulations and Standards projects, for his work as a regional QA liaison, taking special time to work with the regions on QA issues, for his work in the development of data quality objectives in their regions in the water programs, for looking into the issue of quality in computerized data bases and his work in the design and implementation of the National Dioxin Survey and National 120 ------- Bioaccumulation Survey. And the last nominee was Elizabeth Leovey who is in the Office of Pesticide Programs. Elizabeth was nominated by Susan Wayland, the Deputy Office Director in the Office of Pesticide Programs; Michael Cook, the Office Director in the Office of Drinking Water; and Gene Briskin, the director of the National Pesticide Survey. She was nominated for her work in the National Pesticide Survey, including encouraging and setting up a pilot survey to see if the planned activities would work, the invitation for management system review to see if the processes they had in place for the survey would yield the data they needed for their decisions, and for working to assure that the follow-up field and analytic QA programs would, indeed, meet the needs that they had defined at the beginning of the survey. I think we owe a round of applause to all of these people. The winner of this award receives a check and a plague. And there is a historical plaque with all of the award winners and the individual that wins it will also get his name engraved on the plague outside of my office in Washington. The winner: 1990 Quality Assurance Manager of the Year, Martin W. Brossman, in recognition of outstanding accomplishment in the field of guality assurance planning and management. 121 ------- ACCEPTANCE SPEECH BY MARTY BROSSMAN QA Manager of the Year I'm just flabbergasted and thrilled. And I think you know this is the real highlight of my work for EPA. So I might as well stop now. But I do feel very strongly, and I know all of you I've worked with know that I mean this, that this award is an award I share with QAMS, who've been my guiding light through all these difficult years and with, primarily, the regional QA Officers who've made it all possible for me to see things happen. That combination is really the reason I was able to do anything. And I'm inspired by the type of people that we've been able to work with over the years: this has been a tough game, to convince management that this was worthwhile and necessary. The spirit and the competence of the people that I've been able to work with, QAMS and particularly all my regional pals who have helped me carry out the kind of things that we wanted to see happen in the real world, are really the recipients of this and I thank you all. Thank you. 123 ------- |