&EPA *' '('*'" , EnvironmenfaJ'RrotecUo^ Office of Afr and Radiation (ANR-443P; ' OOOR91100 *- 20460 Air Network "Type: Effects On Emission Reductions, Cost, and Convenience Ť. - ;,oO".*>*'* Technical Information Document -I '.'<--- By Eugene J. Tierney A'-' ------- < U.S. Environmental Protection Agency Region 5, Library (PL-12J) 77 West Jackson Boulevard, 12th Floor Chicago, It 60604-3590 ------- I/M Network Type: Effects On Emission Reductions, Cost, and Convenience l^v. rVi r~, EXECUTIVE SUMMARY TX" VS- The Clean Air Act Amendments (CAAA) of 1990 require iŁt that EPA revise and republish Inspection and Maintenance * (I/M) guidance in the Federal Register addressing a variety 7- of issues including network type, i.e., whether the program ^' is centralized or decentralized. ^ ""^ Nearly every State that must operate an I/M program will need to obtain legislative authority to meet revised guidelines. This need provides an opportunity to reassess the effectiveness of current I/M program designs and make changes that will lead to improved air quality over the next decade. Network type is the most obvious and influential factor at work in an I/M program. This report attempts to analyze how emission reduction effectiveness, cost, and convenience vary by network type. Major factors that influence the emission reductions in an I/M program are test procedures, analyzer accuracy, quality control, inspector competence and honesty, and quality assurance. One of the fundamental features that distinguishes centralized programs from decentralized is the number of analyzers, stations, and inspectors. Decentralized programs range from hundreds to thousands of stations, analyzers and inspectors. By contrast, large centralized programs have 10-20 stations in a given urban area with 3-5 testing lanes each. Consequently, maintaining a high level of quality assurance and quality control over analyzers and inspectors is very difficult and costly in decentralized programs while readily available and easily implemented in a centralized system. The results of a wide array of program audits, studies, and other analyses by EPA and local governments show that decentralized programs suffer from improper testing by inspectors, less accurate analyzers, and from more variation in test procedure performance. The magnitude of the differences is difficult to accurately quantify, but evidence indicates that decentralized programs may be 20-40% less effective than centralized. Obtaining additional emission reductions from in-use motor vehicles is essential to meeting the goals for enhanced I/M programs in the CAAA of 1990. To achieve this, I/M programs need to adopt more sophisticated procedures and equipment to better identify high emitting vehicles. This would result in higher costs, which may make it difficult to justify low volume decentralized stations, and longer test times, implying the need for more lanes and stations in a centralized system. Further, alternatives such as loaded transient testing may not be practical in decentralized ------- programs. Thus, centralized programs, which are already more effective, may also have greater potential for improving effectiveness. The cost of I/M programs is also important Two components of cost are addressed in this report: the cost to the motorist and the cost of program oversight. There is a distinct difference in cost between centralized and decentralized programs; centralized programs are substantially cheaper, and the gap is widening. In most cases centralized programs have had decreasing or stable costs while decentralized programs have gotten more expensive due to improved equipment and the demands by inspectors for higher compensation. This trend is likely to continue given the additional inspection requirements imposed by the CAAA of 1990. Convenience is named as the major advantage of decentralized programs. There are many more inspection locations, and waiting lines are generally non-existent. Vehicles which are identified as high emitters can be repaired by the testing facility, which is also authorized to issue a compliance or waiver certificate after retest . Centralized testing facilities, on the other hand, can be visited without appointment, but the motorist may encounter waiting lines during peak usage periods. Vehicles which fail the inspection must be taken elsewhere for repair and returned to a testing facility afterwards. This report discusses the results of a consumer survey on decentralized program experiences and also discusses inspection facility siting and reported waiting time in centralized programs. Individual's experiences in decentralized programs vary, depending on whether they seek an inspection without an appointment, and whether the nearest licensed facility is also capable of completing complex engine repairs. So do the experiences in centralized networks; most vehicles will not need repair and one well- timed inspection visit will suffice. On the other hand, inadequately sized networks can produce serious problems for consumers. In summary, the majority of centralized programs have been found to be more effective at reducing emissions at a lower cost than decentralized programs. Improvements to decentralized programs are needed to assure that objective, accurate inspections are performed. These improvements are likely to be costly, and desired levels of sophistication may be beyond the capabilities of most private garages. Consumer convenience continues to be an important factor in network design and public acceptance. -11- ------- Table of Contents Executive Summary i Table of Contents iii List of Figures iv List of Tables v 1.0 Introduction 1 2 . 0 Background 2 3.0 Emission Reductions from I/M Programs 5 3.1 Formal Test Procedures 5 3.2 Emission Analyzer Accuracy and Quality Control 12 3.3 Emission Testing Objectivity and Accuracy 15 3.3.1 Background 15 3.3.2 Supporting Evidence 21 3.3.2.1 Failure Rate Data 22 3.3.2.2 Overt Audit Data 24 3.3.2.3 Covert Audit Data 26 3.3.2.4 Anecdotal Evidence 27 3.3.2.5 Waiver Rate Data 2 S 3.3.2.6 Test Record Data 3; 3.3.3 Conclusions 32 3.4 Visual and Functional Inspections of Emission Controls 2~ 3.4.1 Background 25 3.4.2 Supporting Evidence 2-1 3.4.3 Conclusions 4 j 4 . 0 Program Costs -44 4.1 Inspection Costs 44 4.2 Repair Costs 4 - 4 . 2 Conclusions 4 .- 5 . 0 Convenience i . 6.0 Future Considerations Affecting the Comparison of Network Types 7.0 Predicted Emission Reductions From MOBILE4 and an Additicr.a. Assumption Regarding Waivers 8.0 Conclusions -111- ------- List of Figures 3-1 Vehicles Passing California I/M After Extended Preconditioning 9 3-2 Old and New Emission Test Failure Rates in I/M Programs ....23 3-3 Failure Rates in Manual and Computerized Stations in New Hampshire 24 3-4 Fraction of Covert Audits Finding Improper Tests 27 3-5 Overall Tampering Rates in Select I/M Programs 37 3-6 Catalyst and Inlet Tampering Rates in Select I/M Programs . . 38 3-7 Tampering Rates in Decentralized and Centralized Programs . . 39 3-8 Aftermarket Catalyst Usage in Anti-Tampering Programs 40 3-9 Tampering Rates in I/M and Non-I/M Areas in California 41 3-10 Frequency of Proper Tampering Tests 42 4-1 Cost Per Vehicle of I/M Programs 45 4-2 Cost Per Vehicle By Network Type 46 4-3 Cost By Program Age 47 4-4 Nationwide Inspection.and Oversight Cost of I/M 48 5-1 Average Daily Waiting Times in Illinois' I/M Program 51 7-1 Benefits From Various Potential Changes to I/M Programs .... 57 -iv- ------- List of Tables PaC 2-1 Currently Operating or Scheduled I/M Programs 2 3-1 Test Procedures Currently Used in I/M Programs 6 3-2 Error of Commission Rates Using High Speed Preconditioning . . 7 3-3 Change in Failure Rates From First Idle to Second Idle 8 3-4 Model Year Coverage of Anti-Tampering Inspections 11 3-5 Gas Audit Failure of Emission Analyzers By Program 13 3-6 Early Emission Test Failure Rates in I/M Programs 17 3-7 Waiver Rates in I/M Programs in 1989 29 3-8 Model Year Switching Between Initial and Retests 31 3-9 Vehicles With Dilution Levels Below 8% 32 -v- ------- 1.0 INTRODUCTION At any point*in time, a certain percentage of vehicles on zhe road are emitting pollutants in excess of their design standards due to repairable causes. Motor vehicle emissions inspection and maintenance (I/M) programs employ a short screening test to identify high emitters and a retest after repairs to confirm their effectiveness in reducing emissions. It is the mandatory initial screening test and the retest after repairs that differentiate I/M from a public information campaign about motor vehicle maintenance or a program to train automotive mechanics. Where and how those tests are conducted has been one of the fundamental choices in designing an I/M program. This choice is generally made by the elected officials who establish the necessary authorizing legislation. Two basic types of I/M systems exist: 1) inspection and retest at high volume, test-only lanes (a centralized network), and 2) inspection and retest at privately- owned, licensed facilities (a decentralized network) . A combination of centralized and decentralized inspections is also found, the latter usually being for retests only. This report discusses how the choice of network design can affect the quality, cost, and convenience of I/M inspections. This report examines operating results from inspection programs acrcss the United States to compare the effects of network design. Eacr. section will present available information and attempt to come to a conclusion regarding the relative merits of each network option. The discussion of network choice is particularly important ir. light of the Clean Air Act (CAAA) Amendments of 1990, which induce significant changes for I/M programs. The CAAA amendments require centralized testing in enhanced areas, unless the State car. demonstrate that decentralized testing is equally effective. The phrasing of this requirement implies a desire on the part of Congress for getting the most out of I/M, while not wishing -_ _ close out the decentralized option altogether. This report la_. _ the groundwork for discussing what may be necessary to .~a--~ decentralized as effective as centralized testing. ------- 2.0 BACKGROUND The first emission I/M program was established in the State of New Jersey with mandatory inspection and voluntary repair in 1972. In 1974, repairs became mandatory for vehicles which failed the inspection test. New Jersey added the emission inspection to its existing, centralized safety inspection network. The States of Oregon and Arizona followed suit by establishing centralized inspection programs in 1975 and 1976. Oregon, like New Jersey, established a State-operated network, while Arizona was the first State to implement a contractor-operated network. Following passage of the Clean Air Act Amendments of 1977, which mandated I/M for areas with long term air quality problems, other State and local governments established inspection programs. Table 2-1 illustrates the choices which have been made regarding network design in the United States as of January, 1991. Table 2-1 Currentlv Ooeratina or Scheduled I/M Procrrams Centralized Contractor Operated Cleveland 1 Connecticut Arizona Illinois Florida 3 Louisville Maryland Minnesota 3 Nashville Washington Wisconsin (11) Centralized State/Local Operated D.C. Delaware Indiana Memphis New Jersey Oregon (6) 1 Texas and Ohio are counted 2 Committed to 3 Scheduled to switching to begin in 1991 Decentralized Computerized Analyzers Anchorage California Colorado Dallas/El Paso Fairbanks Georgia Massachusetts Michigan Missouri Nevada New Hampshire New Mexico New York Pennsylvania Virginia (15) Decentralized Decentralized Manual Analyzers Davis Co,UT 2 Idaho N.Carolina 2 3 Rhode Island Provo, UT Salt Lake City (6) as one program each but are listed Parameter Inspection Houston - Louisiana N . Kentucky Ohio Oklahoma (3) in 2 columns . decentralized computerized analyzers. -2- ------- Decentralized networks are more abundant due to a variety zi factors. In seventeen areas, the emission inspection requirement was simply added to a pre-existing decentralized safety inspection network. Another ten areas implemented decentralized systems because they were perceived to be less costly and more convenient to the vehicle owner. Those areas choosing to implement centralized systems cited improved quality control and avoidance of conflict of interest as the main reasons. Eleven areas selected to have private contractors own and operate the inspection network so as to avoid the initial capital investment as well as the ongoing responsibility for staffing and maintenance. Following New Jersey and Oregon's lead, four States chose to use State-owned centralized facilities and government employees to operate the test system. The U.S. Environmental Protection Agency (EPA) did not attempt to specify network choice when it first established I/M policy in 1978. EPA assumed that decentralized programs could be as effective as centralized programs, if well designed and operated. At the time that the policy was developed, there were no decentralized I/M programs to observe. EPA did, however, anticipate the increased need for oversight in a decentralized network, and the 1978 policy required additional quality assurance activities for such systems. In 1984, EPA initiated I/M program audits as a part of the National Air Audit System. The audit procedures were developed jointly by EPA, the State and Territorial Air Pollution Program Administrators (STAPPA) , and the Association of Local Air Pollution Control Officials (ALAPCO). Since the audit program's inception, EPA has fielded audit teams on 96 different occasions totaling 323 days of on-site visits to assess program adequacy. Audit teams are composed of trained EPA staff specializing in I/M program evaluation. Every currently operating I/M program has been audited at least once and many have been audited several times. The major elements of I/M audits include: 1) overt visits to test stations to check for measurement instrument accuracy, to observe testing, to assess quality control and quality assurance procedures instituted by the program, and to review records kept in the station; 2) covert visits to test stations to obtain objective data on inspector performance; 3) a review of records kept by the program, including the history of station and inspector performance (e.g., failure rates and waiver rates), enforcement actions taken against stations and inspectors found to c-r violating regulations, and similar documents; -3- ------- 4) analysis of program operating statistics, incl-air.r enforcement rates, failure rates, waiver ra~es, and similar information; and, 5) entrance and exit interviews with I/M program officials and a written report describing the audit findings and EPA recommendations on correcting any problems found. Further detail on audit procedures can be found in the National Air Audit System Guidance.1 EPA has pursued several other means of monitoring I/M program performance. Roadside emission and tampering surveys2 are conducted in I/M areas throughout the country each year. These surveys provide information on how well programs are correcting tampering and bringing about tailpipe emission reductions. EPA requests certain I/M programs to submit raw test data for in-depth analysis. These data are analyzed in a variety of ways to determine whether programs are operating adequately.3 Studies conducted by individual States and other organizations also provide information on program effectiveness.4'5'6 In preparing this report, all of these sources were considered in the discussion of network design impacts. -4- ------- 3.0 EMISSION REDUCTIONS FROM I/M PROGRAMS From an environmental perspective, the most critical aspect in evaluating an I/M program is the emission redaction benefit it achieves. There are three major factors related to network type that must be considered in making such an assessment: the prescribed test procedures, instrumentation and quality control, and actual test performance (including the issuance of waivers). 3.I Formal Test Procedures In order to assure that air quality benefits are achieved, it is necessary to assure that high emitters are identified by the emission test and are properly repaired. The Federal Test Procedure (FTP), which is used to certify that new vehicle designs meet emission standards, is not practical for use in the field, because it is very expensive and time consuming. Short emission tests were developed for I/M programs to allow rapid screening of in-use vehicles for emission performance. It is desirable, however, that an I/M test be able to predict about the same pass/fail outcome that the FTP would, and especially that it not fail a vehicle which could not benefit from repair. Generically, there are three short tests currently in use in I/M programs: idle, two-speed and loaded/idle. Which programs use which test is shown in Table 3-1. There are some variations between how programs define and carry out these tests, but the basic approach is the same for each test . There is no direct relationship between network type and test type, although loaded testing is currently only done in centralized programs. The use of preconditioning, or preparation of the vehicle for testing, is another test-related variable illustrated in Table 3-1. Preconditioning is performed to assure that the vehicle is at normal operating temperature, and that any adverse effect of extended idling is eliminated. Some I/M programs utilize a period of high speed operation (2500 rpm for up to 30 seconds) to precondition vehicles; some operate the vehicle on a chassis dynamometer prior to conducting the idle test; others do no preconditioning. Most decentralized programs, especially those employing computerized analyzers, conduct high speed preconditioning. To reduce costs and test time, many centralized programs did not adopt preconditioning. As with loaded testing, only centralized programs are doing loaded preconditioning. The importance of properly preconditioning a vehicle for the short test has become increasingly apparent to EPA during the last few years due to concerns about false failures of newer model vehicles. As a result, the merits of high speed preconditioning and loaded preconditioning have been studied to assess their effectiveness. Two recent EPA studies7 gathered data related to this issue, one which recruited vehicles which failed the short test in Maryland's centralized I/M program, and one which recruited vehicles from Michigan's decentralized I/M program. In both -5- ------- Table 3-1 Test Procedures Currently Used in I/M Programs Program D.C. Delaware Indiana Memphis New Jersey Oregon Arizona Connecticut Florida Illinois Louisville Maryland Minnesota Nashville Washington Wisconsin Alaska California Colorado Dallas El Paso Georgia Idaho Massachusetts Michigan Missouri Nevada New Hampshire North Carolina New Jersey New Mexico New York Pennsylvania Rhode Island Utah Virginia Network Type1 CG/D CG CG CG CG CG CC CC CC CC CC CC CC CC CC CC D D D D D D D D D D D D D D D D D D D D Test Type2 Idle Idle Two Speed Idle Idle Two Speed Loaded/Idle Idle Idle Two Speed Idle Idle Idle Idle Two Speed Idle Two Speed Two Speed Two Speed Idle Idle Idle Idle Idle Idle Idle Two Speed Idle Idle Idle Idle Idle Idle Idle Two Speed Idle Preconditioning None None High Speed None None High Speed Loaded Loaded Loaded High Speed High Speed High Speed Loaded None High Speed Loaded High High High High High High None High High High High High High High High High High None High High Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed Speed 1 CG = Centralized government, CC = Centralized contractor, D= Decentralized 2 Idle = pass/fail determined only from idle emission readings Two Speed = pass/fail determined from both idle and 2500 rpm readings Loaded/Idle = pass/fail determined from loaded and idle readings. -6- ------- studies, vehicles that had failed their regularly schea-le^ test were tested on the full FTP and carefully examines anything that needed repair. Both studies revealed the exists of errors of commission (i.e., failed vehicles which had low emissions and no apparent defects but nevertheless failed the I/M short test) in these programs which utilize 15-30 seconds of high speed preconditioning. The results are shown in Table 3-2. Since owners of incorrectly failed vehicles are subjected to unnecessary inconvenience and repair expense, the elimination of incorrect failures is a priority. To address the need for improved test procedures, EPA issued a technical report detailing recommended alternative short test procedures in December, 1990.8 The procedures include three variations of the idle test, two variations of the two speed idle test, and a loaded test procedure. In all but the loaded test procedure and the idle test procedure with loaded preconditioning, a second-chance test which includes three minutes of high speed preconditioning is recommended upon failure of the initial test. Table 3-2 Error of Commission Rates Using Hiah Speed Total Number Vehicles Incorrect Program Recruited Failures Michigan 237 70 Maryland 178 55 * The Michigan study used FTPs to verify whether each correct. The Maryland study gave vehicles a second as incorrect failures the vehicles that passed. No these vehicles, but they were assumed to be passes. Preconditioning* Incorrect Failure Rate 30% 33% I/M failure was short test and counted FTPs were conducted on Twenty-four of the thirty-five programs currently operating use high speed preconditioning. Seven programs do no formal preconditioning. Three centralized programs utilize loaded preconditioning and Florida and Minnesota will soon begin such testing. EPA has not done a definitive study of loaded preconditioning or no preconditioning programs but there is some revealing evidence in the data from the programs that use loaded preconditioning. Ir. these programs, the idle test is repeated on initially failed vehicles after 30 or more seconds of loaded operation. The change in pass/fail status is shown in the Table 3-3. Although these data are not accompanied by FTP results, the in-depth studies ir. Maryland and Michigan indicate that vehicles which pass I/M cutpoints after loaded operation are more likely to be -7- ------- emitters. The seven programs that use no tormal precc may be experiencing failure patterns similar to tnose c: idle of the loaded preconditioning programs, i.e., they large number of incorrect failures. Chancre in Model Years Pre-1981 Post 1980 Table 3-3 Failure Rates From First Idle to Failed First Failed Both Idle Test Tests 46% 21% 52% 39% Second Idle Delta* -54% -25% * Delta is the percent of unpreconditioned failures eliminated by loaded preconditioning. Fifteen to thirty seconds of 2500 rpm operation does not seem adequate to precondition all vehicles. No preconditioning at all is probably worse. Thirty to ninety seconds of loaded operation seems to work well but may only be feasible in a centralized network because the purchase and installation of a chassis dynamometer is considered to be beyond the financial capability of most private garages. The Motor Vehicle Manufacturers Associaticr. has endorsed three minutes of 2500 rpm operation as adequate to precondition recent model year vehicles, and EPA's analysis shows that extended unloaded preconditioning reduces incorrect failures. This type of preconditioning is feasible in a decentralizes network, and several programs that have recently switched to BAR 9C type analyzers are currently pursuing this approach. Figure 3-1 shows the results of second chance testing on 42,555 vehicles covering 1968 and newer model years in California using 3 minutes of 2500 rpm preconditioning on vehicles that fail the initial test. The data show that 37% of the vehicles that fail the initial test pass after extended preconditioning. Based on this evidence, improving preconditioning is a .-.. ; priority. The relative cost of loaded preconditioning and exter. >r: high speed preconditioning is an important question-. Lea:-: preconditioning requires the installation of dynamometers. : centralized programs, that cost can be spread over tens : thousands tests. The typical decentralized station inspects a:, . 1000 vehicles per year. Loaded preconditioning seems to accorr.c. _ the task of readying the vehicle for testing in much less time -.:.-.: high speed preconditioning, i.e., 30 seconds vs. 180 seconds. 1: : decentralized station, the time factor is important in terr.s wage labor. In centralized programs, minimizing test tire essential to keeping throughput high and test costs low. ------- Figure 3-1 Vehicles Passing California I/M After Extended Precondit-i Vehicles Tested 24% of Initial Failures SBBB^BBl Vehicles Tested Pass After Preconditioning § ^7,"^ Loaded mode operation is definitely necessary if an I/M jurisdiction wishes to test for emissions of oxides of nitrogen (NOx) . There has been little attention focused on NOx in recent years. Only Los Angeles, California experiences violations of the ambient air quality standard for nitrogen dioxide, and emission control planners elsewhere have historically concentrated on hydrocarbon-only control strategies for reducing ozone levels. There has been growing interest lately, however, in a strategy which combines HC and NOx controls to reduce the potential for ozone formation. As discussed above, loaded mode testing may be practical from a cost standpoint only in centralized networks. Tampering checks may also provide NOx emission reductions. Section 3.4 addresses such tampering checks. Transient loaded mode operation may also be essential for implementing tests that better identify vehicles that should get repairs. The steady-state idle, two speed and loaded tests currently used in I/M programs identify only 30-50% of the vehicles that need repair. EPA has developed a prototype short transient: test that identifies essentially all vehicles that need repair. This procedure is being evaluated in the centralized I/M program in Indiana. Results thus far show that the test accurately identifies vehicles that need repair, while minimizing the number of falsely failed vehicles. A transient loaded test is a more complicated, expensive and time consuming test. It involves the use of a more sophisticated dynamometer than those in use in most loaded testing programs. -9- ------- More sophisticated sampling and measurement equipment is d_o_ involved. The test has many advantages, however, over the currer.- tests. First, by accurately predicting FTP results it identifies more vehicles for repair, which should lead to greater emission reductions from the program and better cost effectiveness. It also allows for NOx testing and a functional test of the evaporative control system. Evaporative emissions represent a very large portion of the total hydrocarbon emissions from motor vehicles. Having an effective functional test to identify vehicles that need evaporative system repair is essential to reducing in-use emissions. Transient testing may also allow the elimination of other tampering checks frequently performed in I/M programs. It is accurate enough to not require this "double check." Anti-tampering inspections are another part of the formal test procedure in I/M programs. As Table 2-1 shows, some programs only inspect for the presence and proper connection of emission control devices (parameter inspection), while others do both emission tests and tampering checks. Table 3-4 shows the current tampering test requirements in I/M programs. Centralized programs typically do no anti-tampering checks or fewer than decentralized programs. Again, the concern for rapid throughput in centralized lanes has been an important factor in this choice. Recently, centralized programs have been adding anti-tampering checks to their emission only programs. Maryland, Arizona, and New Jersey are examples of this. Their experiences show that anti-tampering inspections can be done quickly if limited to the items that are not in the engine compartment: the presence.of a catalyst, the integrity of the fuel inlet restrictor, and a test for leaded fuel deposits in the tailpipe. Inspection of underhood components is possible in both centralized and decentralized networks, but only two centralized programs have so far adopted such inspections. Decentralized programs, in which throughput is less of a consideration, have more readily included underhood checks in their regulations. To conclude, the choice of network type affects the relative costs of improvements in preconditioning and test procedures. Decentralized programs spread the equipment cost across as few as 3 or 4 tests per day. Centralized programs have shied away from preconditioning and anti-tampering checks in the past because of lack of clear need (at the time of original program adoption) and the impact on throughput, not because of the hardware cost. Anti- tampering checks are currently more prevalent in decentralizea programs (the effectiveness of decentralized anti-tampering checks is an important factor which will be discussed in a later section) Centralized networks are probably the best choice for loaded rcae emission testing for NOx control and for improved identification of vehicles needing repair. An alternative that has not yet been tried but might be practical from a cost perspective would be a limited-participation decentralized network, in which a small number of stations would be licensed to insure a high volume o: business at each. -10- ------- Table 3-4 Model Year Coveraae of Anti-Tamoerina (oldest Network Program Arizona Connecticut D.C. Delaware Florida2 Illinois Indiana Louisville Maryland Memphis Minnesota2 Nashville New Jersey . Oregon Washington Wisconsin Anchorage Fairbanks California Colorado Dallas/El Paso Georgia Idaho Massachusetts Michigan Missouri Nevada New Hampshire New Mexico New York North Carolina Pennsylvania Rhode Island Davis Co., UT Provo Salt Lake City Virginia Louisiana N. Kentucky Ohio Oklahoma Houston 1 C = Centralized Type1 C C C C C C C C C C C C C C C C D D D D D D D D D D D D D D D D D D D D D A A A A A model year Catalytic Converter 75 75 77 76 75 75 75 75 75 75 75 80 78 84 81 81 85 75 84 77 84 77 84 75 80 80 80 79 80 , D= Decentralized, checked Fuel Inlet 75 75 77 76 75 75 75 75 75 75 80 78 84 81 81 85 75 84 77 84 77 84 75 80 80 80 79 80 Insoect ions is listed) Lead Test 75 75 75 80 75 77 80 80 80 79 80 Air Pump 67 80 75 75 67 75 68 78 84 71 81 75 84 77 84 77 84 73 80 80 80 79 68 PCV 80 75 75 67 68 71 64 77 84 73 80 80 80 79 68 Evap Canister 80 70 68 71 84 84 73 80 80 80 79 68 A = Ant i -tampering only 2 Not currently operating -11- ------- 3.2 Emission Analyzer Accuracy and Quality Cor.rrol I/M tests are conducted with non-dispersive infrared analyzers which can measure hydrocarbons (HC) (as hexane), carbon monoxide (CO) , and carbon dioxide (C02) , although not all equipment includes the latter capability. The concentration of pollutants in vehicle exhaust is measured by placing a steel probe, attached to the analyzer by a sample hose, in the exhaust pipe and drawing a sample of exhaust into the instrument. The concentration in the exhaust stream is compared to the I/M program's pass/fail cutpoints (1.2% CO and 220 ppm HC for the majority of the fleet) . Accurate measurement capability is essential for several reasons. First, although cutpoints are high enough that small errors do not critically affect emission reductions, any significant error affects equity and public confidence in the testing process. Naturally, public support of the program is critical to the ultimate success of the effort. Second, leaks in the analyzer dilute the emission sample and could lead to false passes of high emitting vehicles. Leaks can quickly and easily become large enough to cause large amounts of dilution. This results in a loss in emission reduction benefits to the program. Finally, bad data obscure what is happening in the program. Tracking the performance of vehicles in the fleet over time is one indicator of the success of an I/M program. Inaccurate readings could mislead planners and decision makers. The accuracy of the reading depends on periodic calibration checks and adjustments, regular maintenance of filters and water traps, and detection and repair of leaks in the sampling system. Each I/M program has requirements for maintaining the analyzers and also has an independent audit program in which analyzers are checked on a periodic basis. EPA recommends that program administrators use a +5%/-7% tolerance when auditing inspection equipment. That is, if an analyzer reads a known concentration of gas more than 5 percent too high or 7 percent too low during an audit, then it should be taken out of service until it can be adjusted or repaired. EPA audits of operating I/M programs also include independent instrument checks with a gas bottle that EPA names and ships to the site from its Motor Vehicle Emission Laboratory in Ann Arbor, Michigan. Table 3-5 presents the findings from EPA audits regarding the accuracy of different programs' instrumentation. The findings show mixed results depending upon program type. Centralized contractor-run programs show very strong quality control. Centralized government-run programs and decentralized programs are comparatively weak. -12- ------- Gas Audit Program D.C. Delaware Indiana Medf ord Memphis Subtotal Connecticut Illinois Louisville Maryland Washington Wisconsin Subtotal California Colorado Georgia New Hampshire North Carolina Pennsylvania Subtotal 1 CG = Centralized Failure of Network Type1 CG CG CG CG CG CG CC CC CC CC CC CC CC D D D D D D D government, Table 3-5 Emission Anal Number of Analyzers Tested 13 9 24 4 9 59 32 47 6 11 12 11 119 22 21 21 14 17 11 106 CC = Centralized yzers By P Rate of HC Failures 23 % 67 % 29 % 50 % 44 % 37 % 0 % 0 % 0 % 0 % 25 % 18 % 4 % 9 % 29 % 14 % 29 % 35 % 27 % 23 % contractor, roaram, Rate of CO Failures 15 % 22 % 8 % 0 % 33 % 15 % 0 % 0 % 0 % 0 % 8 % 9 % 2 % 14 % 14 % 10 % 21 % 65 % 18 % 23 % D= Decentralized Minimum quality control requirements are specified in t.v Emission Performance Warranty Regulations promulgated pursuant : Section 207(b) of the Clean Air Act Amendments of 1977. Amor other things, these regulations require weekly gas span and le; checks. Centralized programs go far beyond these minimi requirements, typically conducting these checks two or more time per day. The best centralized programs, in terms of quail: control results, conduct leak checks and recalibrate equipment : an hourly basis. Frequent multi-point calibrations, dai: preventative maintenance, and careful monitoring of equipmer performance also characterize these quality control program: While such activities are possible in a centralized program with limited number of analyzers and the economies of scale purchasing large quantities of high pressure calibration ga: -13- ------- decentralized stations simply could not afford c control. Also, the low average test volume in stations limits the number of vehicles affected by an equicrer.*: error. As a result, decentralized programs all require stations t :> conduct these checks only weekly. Some quality control practices, such as monthly multi-point calibrations, are never done in decentralized programs. Centralized government-run programs tend to have problems with calibration gas quality. In fact, most of the quality control failures in these programs can be traced to gas problems. This is also true of the two centralized contractor-run programs that had quality control problems. Finally, limited operating budgets in centralized government-run programs usually result in less frequent quality control checks and less technical expertise available to maintain test equipment, when compared with centralized contractor- run programs. Decentralized inspection facilities suffer from both problems to one degree or another. Each individual station usually is allowed to purchase gas on the open market and has no way of knowing if the gas is accurately labeled unless the State operates a gas naming program. The program manager also purchases gas for auditing station analyzers. These audits serve as the only tool for .ensuring gas accuracy in decentralized programs. However, with a few exceptions, quality assurance audits occur only two to four times per year, which limits their effectiveness. Computerized analyzers used in decentralized programs are programmed to lock out from testing if the weekly calibration check and leak check by the station owner are not conducted and passed. In theory, this ensures a nominal level of instrument accuracy. Before computerization, these quality control checks depended upon the operator to remember to do them and do them correctly. However, EPA auditors frequently report finding computerized analyzers with leaking sample systems, even though the analyzer has recently passed a leak check conducted by the owner. The leak check mechanism is easily defeated (thus saving repair costs and analyzer down time) by temporarily removing the probe or the entire sample line and tightly sealing the system to pass the check. A simpler approach is to kink the sample line to accomplish the same objective: excluding the most susceptible portions of the sample system from the leak check. EPA auditors have observed this happening during overt station audits. At this point, no one has devised a solution to this problem that could not be defeated. To conclude, quality control in centralized programs can more easily meet the highest standards achievable given current technology since economies of scale allow the use of. better practices. Decentralized programs may always have to settle for lower quality control and the associated burden of extensive quality assurance, and less emission reduction benefit due to the number of stations and inspectors involved. -14- ------- 3.3 Emission Testing Objectivity and Accuracy 3.3.1 Background Apart from the design considerations related to test procedure choice and instrument quality control, two crucial factors in the effectiveness of the inspection process are whether the test is actually conducted and whether it is done correctly. Even the best design is defeated by failure to address these factors. Obviously, simple error or inspector incompetence is a problem, but more significantly, malfeasance is a problem, as well. Thus, the term "improper testing" will be used in this report to refer to a deliberate failure by inspectors to "fail" vehicles that exceed idle emission cutpoints or which violate anti-tampering rules, either at initial test or at retest. The degree of machine control of the testing process and the level of human supervision of the inspector influence the degree to which improper testing is possible. Obviously, decentralized programs do not allow for the same level of human supervision of the testing process as centralized programs. Overt quality assurance visits occur, at most, once per month and rarely involve the observation of the testing process. Even if testing were observed during these visits, only the most inept inspectors would deliberately perform the test incorrectly while being observed by the State auditor. In the early 1980's, almost all decentralized I/M programs employed manual emission analyzers, and the inspector/mechanic was responsible for every step of the process: selecting the proper cutpoints based on a vehicle's model year and type, recording the results from an analog (needle on dial) display onto a test form, and making the pass/fail decision. In short there was no machine control of the inspection process, so the inspector's proficiency and honesty were key to objective and accurate tests. Across the board, decentralized manual analyzer programs were reporting operating statistics which did not agree with those reported by either centralized programs or decentralized programs using computerized analyzers, or with the available information about in-use emission performance. In most cases, the reported failure rate was only 20 to 40 percent of what would be expected. Table 3-6 shows failure rates from all programs reporting data from the early 1980's. Table 3-6 might be hurriedly interpreted to mean that the manual programs were bad and the computerized and centralized programs were good. The reported failure rate, however, is only one indicator of I/M program performance. It would be rash to conclude that a program is succeeding or failing based only on that statistic. There are several reasons why this statistic is not a completely reliable indicator. The failure rate in a program is defined as the number of vehicles that fail an initial test. In order to accurately calculate this statistic, it must be possible to distinguish initial tests from retests. The computerized analyzers used in decentralized programs include a prompt that asks -15- ------- whether the test is initial or a retest . In some cases, ir.sc^czjrs are not aware that a vehicle has been inspected before at another station and code a retest as an initial test. Another more common problem is that the inspector simply chooses initial test (the default prompt) for any test. This confuses actual failure rates. Some centralized programs have the same kind of prompt but require the owner to present a document (e.g., vehicle registration or test form) which is marked when a test is performed. This reduces the problem but EPA audits have found instances where coding is not accurate. Many centralized systems use a computerized record call up system, based on the license plate or VIN, that automatically accesses the pre-existing vehicle record in the database. Thus, it is nearly always clear which is the initial test and which is the retest. It is possible to sort out the miscoding problem in any database that includes reliable vehicle identification information but this dramatically complicates data analysis, and I/M programs have not systematically pursued this. Another reason why reported initial tests failure rates are not reliable indicators is the problem of false failures discussed previously. The quality of preconditioning and the sophistication of the test procedures vary from program to program. Inadequate preconditioning or less sophisticated test procedures may result in exaggerated failure rates. Most importantly, even if the reported failure rate on the initial inspections were entirely accurate, there could still be problems with retests. In a manual program, the quickest approach to completing any test is to make up a passing reading, so it is not surprising that few failures were ever reported. In a computerized program the analyzer must at least at some point be turned on, vehicle identification data entered, and the test cycle allowed to run its course or no compliance document will be printed. Since most cars will pass their first test with no special tricks, the least-time approach for a dishonest inspector may be to do a reasonably normal first inspection on all vehicles, producing a respectably high failure rate report, and devote special attention only to the retest of the failed vehicles. So, failure rates were useful in identifying the severe problems that existed in manual I/M programs, but they are not useful in drawing finer distinctions between better run programs. EPA examined in more depth six possible causes of the low failure rates in the manual programs [see EPA-AA-TSS-IM-87-1]. This study analyzed and discussed roadside idle survey data, reported I/M program data, and data collected during I/M program audits. EPA concluded that five of the explanations - quality control practices, fleet maintenance, fleet mix, differing emission standards, anticipatory maintenance, and pre-inspection repair - did not sufficiently explain low reported failure rates. By process of elimination, EPA concluded that the major problem contributing to low reported failure rates in decentralized, manual I/M programs was improper inspections by test station personnel. Anecdotes and observations reinforced this explanation. -16- ------- Table 3-6 Early Emission Test Failure Rates in I/M Programs1 Centralized Arizona Connecticut Delaware Louisville Maryland Memphis, TN Nashville, TN New Jersey Oregon Washington, D.C. Washington Wisconsin Reported % 20.2 17.2 13.7 15.7 14, 8, 24, 26, 24.0 18.4 19.0 15.3 6 1 5 1 Expected2 % 36.8 33.0 7 .7 16.2 14 .0 3.7 25.4 27.8 38.3 13.4 28.1 19.3 .55 .52 .00 . 97 .00 .00 .97 . 94 . 63 .00 .68 .79 Average Decentralized Computerized Analyzers .85 Fairbanks, Alaska Anchorage, Alaska California Michigan New York4 Pennsylvania 19.4 15.7 27.7 15.8 5.1 17.6 Average Decentralized Manual Analyzers 22.7 24 .7 28.7 12.9 33.4 19.5 .85 .63 .96 .00 .15 .90 .75 Georgia 6.6 Idaho 9.8 Missouri 6.7 North Carolina 5.6 Clark Co., Nevada 9.5 Washoe Co., Nevada 11.0 Davis Co., Utah 8.7 Salt Lake Co., Utah 10.0 Virginia 2.3 Average 25.0 16.9 20.5 21.1 29.4 29.4 21.3 21.3 15.6 .26 .58 .33 .27 .32 .37 .41 .47 . 15 .35 1 1983-1985 data, for all model years, including light-duty trucks. 2 Expected rates are based on data from the Louisville I/M program for 1988. They vary by area due mainly to differences in cutpoints. 3 Values greater than 1.00 are reported as 1.00. 4 New York's analyzers are only partially computerized. -17- ------- -J L In part because of the poor experience of the manual cr;gr; that had started earlier and because of encouragement frcrr. EPA, : next round of six decentralized programs that started in the rr.: 1980s chose to require computerized analyzers. By the begir.nir.g 1989, four of the previously manual programs had required inspection stations to purchase and use new computerized analyzers to reduce improper testing. Two more programs are scheduled to switch in the near future. Thus, most decentralized programs now utilize computerized analyzers to control the testing process. The timely question is whether decentralized programs using computerized analyzers in fact get as much emission reduction as a comparable centralized program, as Table 3-6 might suggest could be the case. The next section examines the evidence that is available, including that from the newer computerized centralized programs. The remainder of this section discusses hypothetical problems as a useful preliminary to an examination of the evidence with some comparison to centralized programs. The new computerized analyzers reduce many of the errors likely with the use of manual equipment by reducing the number and difficulty of the decisions an inspector has to make. The inspector enters vehicle identification information (make, model year, vehicle type, etc.), and based on this, the computer automatically selects appropriate emission standards, performs the test, makes the pass/fail decision, prints the compliance document, and stores the results on magnetic recording media. However, much of the process still relies on the inspector to correctly conduct the manual steps of the process: determine vehicle information, key in information correctly, properly precondition the vehicle, properly insert the probe in the tailpipe for the duration of the test, maintain idle speed at normal levels, etc. Thus, improper testing is still possible, and is not limited to cases of simple human error. There are a variety of motivations that can lead an inspector to intentionally perform an improper test. First and foremost, some inspectors may not be interested in deriving revenue from the repair of vehicles but are primarily in the business to profit from testing. In many cases, I/M tests are performed by service stations or other outlets that have test analyzers but do not stcc'< a full range of emission parts and do not employ fully qualifier engine mechanics. The service such an outlet offers may be t, provide as many customers with a certificate of compliance with ;:; little hassle as possible. In addition to improper testing on ~- initial test, improper testing on the retest of a failed ver.:._.- also occurs. One example of the motivation for this is when . mechanic attempts a repair but fails to resolve the excess emissi-- problem. It puts the mechanic in a difficult position to te-1 : customer that the vehicle still failed after the agreed _L- repairs were performed. To save face, the mechanic falsifies results and the vehicle owner believes the problem was correcte-: Finally, some inspectors may not hold the I/M program in hi--.-. regard or may doubt its technical effectiveness, and may want - -18- ------- help out customers by passing them since chey perceive no r.arr. _r. doing so. The mid-1980s experience in manual decentralized programs amply demonstrated that many licensed inspectors were willing to report test results they had not actually obtained, for the reasons given in the previous paragraph or others. To determine whether this willingness still exists and the degree to which it is affecting computerized programs, it is useful to enumerate the ways in which it is possible to conduct a computerized inspection improperly. This will suggest the type of evidence one might expect to be able to see. The most basic form of deliberate improper testing is no test at all, i.e. the inspector provides the customer with a compliance document without inspecting the vehicle. The accounting systems in most programs require that the inspector somehow produce an electronic record of an inspection for each certificate or sticker issued. This electronic record can be produced by testing a "clean" vehicle to achieve a passing score but entering the required identification information as if it were the subject vehicle. Another approach is to intentionally enter erroneous vehicle identification information to make it easier for a vehicle to pass. All I/M programs have looser emission standards for older technology vehicles and most have looser standards for trucks compared with cars. Model year input to the computerized analyzer governs automatic selection of emission standards used by the system to decide pass/fail, outcome. Thus, by entering an older model year or truck designation into the computer, the system automatically selects looser standards. The compliance document will then show the improper model year, but may never be examined closely enough to be questioned. The decentralized computerized analyzer requires a minimum amount of CC>2 in the sample stream in order to consider a test valid. Most programs use cutpoints of 4-6%, well below the 10-14% COz found in most vehicles' exhaust. One way to improperly test is to partially remove the probe from the tailpipe such that the sample is diluted enough to pass the HC and CO standards but not enough to fail the COj check. Similarly, computerized analyzers allow engine speed during the idle tes-t to range between 300-1600 RPM. Improper testing may be accomplished by raising the engine speed above normal during the idle test. This usually lowers apparent emission levels leading to a passing result. To EPA's knowledge, no I/M agency tries to detect such actions on a routine basis. Another technique is to make temporary vehicle alterations, e.g., introduce vacuum leaks or adjust idle, to get the vehicle co pass and then readjust the vehicle after the test. This type of improper testing is nearly impossible to detect. -19- ------- Finally, vehicle owners can obtain inspection documents cr. tr.e black market in some programs. A major element of quality assurance in a decentralized program is to determine whether all stickers or compliance certificates can be accounted for during the periodic audits of stations. A routine problem is the loss or theft of certificates or stickers, despite regulatory requirements that such documents be kept secure. Recently, Texas, New York, and other I/M programs have experienced problems with counterfeit I/M stickers. As with the comparison of decentralized and centralized programs, the failure rates in Table 3-6 should not be taken to be reliable indications of which centralized programs were working the best. The low failure rate programs may have been doing a better job at getting cars fixed one year so they do not appear as failures the next, and in recording only one test on each vehicle as the initial test. A closer look at hypothetical possibilities and at the supporting evidence is necessary. Improper testing is also possible in centralized networks, with some reason to suspect that such problems could be more common in government-run systems than in contractor-run systems. The lane capacity in some government-run programs was not designed or upgraded over time to handle the growing volume of vehicles demanding inspection. Also, the test equipment used is essentially identical to the equipment used in decentralized computerized programs. (This has been a cost-based decision, not one inherent to government-run stations.) Thus, all of the pitfalls associated with the use of those analyzers are possible. Finally, because enough aspects of government-run programs continue to be manual operations, they are subject to both error and malfeasance, making close personal supervision essential. As a result of these problems, vehicles have, at times, been waved through without emission testing in government-run lanes as a result of routine problems like equipment breakdown, and sloppy testing practices have been observed. Through the audit process, EPA and the State and local jurisdictions have made some progress in resolving these problems. It should be noted here that intentional violation of program policy or regulations by individual inspectors (i.e., deliberate improper testing) is not evident in these programs. Shortcuts have been observed on visual inspections, cut infrequently. On occasion inspectors have been caught stealir.g cr selling certificates or accepting payments but the supervision typical of centralized programs generally prevents this. Centralized contractor-run programs can be expected to suffer few if any problems with improper testing for several reasons. The level of machine control in these programs is such that the inspector has almost no influence over the test outcome. In fully automated systems, the inspector only enters a license plate number to call up a pre-existing record that contains the other vehicle identification information used to make the pass/fail decision. Individual inspectors are held accountable for certificates c: compliance just as a cashier is held responsible for a bala.-.c-: -20- ------- cash drawer. The actual test process is completely computerizea. In addition to machine control of the inspection process, the level of supervision in centralized programs is very high. Tr.e contractor is under pressure and scrutiny from both the public ar.a the government agencies responsible for oversight. This pressure has led to high levels of quality control and quality assurance in every aspect of the inspection system. Two other problems that may affect emission benefits of a program regardless of network type are readjustment after repair and test shopping. Readjustment after getting a "repair" and passing the test probably happens to some extent in all programs. A survey9 of mechanics in Arizona showed a significant percent admitting to have made such readjustments. This problem stems from the fact that poor quality repairs may sacrifice driveability in order to reduce emissions to pass the test. Readjustment occurs after obtaining a passing reading to improve driveability. The improvement may be real or imagined. Some owners still have a 1970s mind set when it comes to emission controls and will not believe that a readjustment is not necessary. This problem is somewhat limited to older technology vehicles since computerized vehicles are not adjustable, per se. Test shopping occurs when an individual fails an emission test at one I/M test station (either centralized or decentralized) and goes to another station to get another initial test. EPA auditors hear complaints from inspectors in decentralized programs that sometimes they fail a vehicle but the owner refuses repair and finds another station to get a passing result. In some cases these complaints have been verified by follow-up investigations conducted by program officials. Test shopping can result in a small number of dishonest garages "inspecting" a disproportionately large number of cars that should be failed and repaired. However, detecting or preventing this type of problem has only been systematically pursued in one decentralized program. In a centralized program, the main effect of test shopping is that cars with variable emission levels sometimes pass on a second or third test. 3.3.2 Supporting Evidence The foregoing background discussion described what can ac wrong with testing in each program type, and why it is reasona'c 1- to suppose a special problem may exist in decentralized progrars. However, the closest possible scrutiny is appropriate given -:. stakes involved: air quality benefits and basic network choice. Ideally, an in-depth field study of the issue would be useful -_ . quantify the extent of each of these problems and measure tr.e-: impact on the emission reduction benefits of the program. However, such a study would require years of investigation and cost milli;- of dollars. Further, the results may not be clear cut because .: the difficulty of observing I/M program operations withe-t affecting the normal behavior of those observed and the difficult.. obtaining a vehicle or station sample that is not biased by -..-- -21- ------- non-participation of owners and mechanics who know they have evaaea program requirements. Short of doing such a study, there are a variety of sources of information that shed light on the extent of improper testing in I/M programs. One major source is the audit program EPA initiated in 1984. Another source is the auditing conducted by programs themselves. Data analysis by both programs and EPA have also provided information. Complaints and anecdotal information from consumers and others involved in test programs are also useful. Finally, EPA has conducted some in-depth studies of testing programs that contribute additional data on improper testing. 3.3.2.1 Failure Rate Data It was expected that switching from manual analyzers to computerized analyzers would solve the problem of low reported failure rates on the initial test, and that appears to have been the case. Figure 3-2 shows failure rate data from I/M programs in the 1987-1988 time frame. Figure 3-2 also compares failures rates from the 1986 time frame and failure rates from the 1988 time frame. The data indicate that centralized program failure rates have decreased in most cases. This is expected as more new technology vehicles, which fail less often, enter the fleet and as the program effectively repairs existing vehicles in the fleet. Some centralized programs, Wisconsin and Louisville, for example, do not show this trend because they regularly increase the stringency of their cutpoints to maintain high levels of emission reduction benefits. The decentralized manual analyzer programs show little change or small increases in failure rates in this time period. The increases may result from increased pressure on these programs to perform. It is clear that the failure rates in these programs remain lower than expected. A good example of this is the New Hampshire program which, at the time, used manual analyzers in about 20% of the stations and computerized analyzers in the rest. Figure 3-3 shows the failure rate data from each segment of the program, with the manual analyzer stations reporting a failure rate approximately 25% that of the computerized stations. -22- ------- Figure 3-2 Old and New Emission Test Failure Rates in : '"M Prngrar Centralized Programs AZ CT DC MD NJ WA WI NC V A Decentralized Manual Programs Decentralized Programs Recently Switching to Computerized NV GA m Decentralized Programs: Computerized Analyzers From Inception 10 Old Data 15 20 Recent Data 25 30 * The recent data is from 1987 - 1988 and the old data is from 1985 - 1986. -23- ------- Figure 3-3 Failure Rates in Manual and Computerized Stations in New Hampshire' Computerized ||||||^^ 19% Manual Analyzers * About 20% of the stations in the New Hampshire program used manual analyzers while the rest used computerized equipment. The decentralized computerized programs now report high failure rates, in the range of what would be expected based on the emission standards and vehicle coverage of these programs. Thus, the operating data from decentralized computerized programs would suggest that more initial inspections are being performed properly than was the case using manual analyzers. As discussed previously, there are problems with relying solely on failure rate as an indicator. Increased failure rates are certainly a precondition to successful emission reductions, but not sufficient. The central factor is whether the final test on each vehicle is performed objectively. 3.3.2.2 Overt Audit Data Quality assurance efforts in decentralized I/M programs always include overt audits of licensed inspection stations, typically on a quarterly basis or more often in systems using manual analyzers. Overt audits generally consist of collecting data from recent tests (either magnetic media or manually completed test forms), checking analyzer accuracy, and observing inspections in progress. These audits serve to maintain a presence of program management and control in the station. With computerized analyzers, software prompts inspectors through each step of the procedure, so overt audits of the test stations rarely find problems with inspectors not knowing how to correctly perform an emission test. Of course, overt audits still find problems unrelated to the emission analyzer: missing stickers, certificate security problems, disarray in record-keeping, and similar administrative details that coula lead to a finding of improper testing. Overt audits c: decentralized I/M programs by EPA generally involve observations at only a small fraction of the licensed test stations and no attempt is made to obtain a statistically accurate sample. As a result, the following discussion will speak in general terms about overt audit findings without citing specific rates. Nevertheless, E?A -24- ------- believes that the findings from these station visits are fairlv representative. Manual operations observed during overt audits of decentralized programs are done incorrectly so often that this is almost always identified as a major problem. This is true of both anti-tampering checks, discussed in more detail in Section 3.4, and emission tests. Inspectors fail to conduct the emission test properly by eliminating important steps, such as preconditioning or engine restart, or by not strictly following criteria, such as when to take the sample and at what speed to precondition. The properly conducted test is the exception rather than the rule when it comes to manual operations. For example, the most recent audit of a manual program by EPA included overtly observing 10 different inspectors test vehicles. Only two inspectors followed procedures completely. Computerized analyzers prevent some but not all of these problems. Sloppiness is also observed in steps such as properly warming up the vehicle, avoiding excessive idle, and properly inserting the probe in the tailpipe. Overt audits find evidence of improper testing based on review of paperwork and test records. Auditors find that stickers or certificates have been issued or are missing but no test record is available to document proper issuance. Inspectors doing improper inspections issue the certificate or sticker and then at a convenient time enter the data into the computer and probe a passing vehicle to create both the magnetic and paper test record. In programs in which stations are required to keep detailed repair records, suspicious information has been found. For example, a station might charge the same exact amount of money for repairing and retesting most vehicles that come to it. This is an extremely unlikely occurrence. Another example is where the station supposedly documents the same repair on different vehicles again and again. The chances that any one repair is the one actually needed in so many cases makes the records dubious. Centralized programs tend to vary considerably with regard to manual operations. In some cases, the only manual operations are verification of vehicle identification, probe insertion, and pressing a button. In other cases, especially where anti-tampering checks are conducted, manual operations can be significant. In these cases, more variation has been observed in the consistency of testing. This is especially a problem in programs that do uncontrolled high speed preconditioning. Subjective judgment must be used by the inspector in instructing the motorist to raise engine speed without the use of a tachometer. Centralized inspectors can also get sloppy with probe insertion, although the carbon dioxide check prevents excessive dilution of the sample. The evidence from EPA and State overt audits indicates that some degree of improper testing occurs in every decentralized program, whether manual or computerized. The findings in centralized programs indicate that high levels of quality control are usually achieved, especially in the contractor-run systems. -25- ------- 3.3.2.3 Covert Audit Data In addition to overt audits, many programs and EPA's audit program have initiated covert audits. These audits involve sending an auditor to a station with an unmarked vehicle, usually set to fail the emission test or the anti-tampering check. The auditor requests an inspection of the vehicle without revealing his or her true identity. Similarly, some programs have initiated remote observation of suspect inspection stations. Auditors use binoculars to watch inspectors working and look for improper practices. If problems are found in either of these approaches, enforcement actions are taken against stations. These audits serve to objectively assess the honesty and competence of licensed inspectors. The findings from covert audits in 19 decentralized programs are shown in Figure 3-4 . Note that the sample sizes vary considerably. The small samples represent covert audits conducted during an EPA audit. I/M program quality assurance staff and EPA auditors formed teams to conduct the audits. The large samples represent data on covert audits conducted by local I/M program staff. The stations visited are randomly selected, but the small samples may not be statistically significant (no attempt was made to obtain a statistically valid sample). Nevertheless, the samples reported by California, North Carolina, New York, and Utah represent large fractions of the station population and are likely to be representative. In any case, the results indicate a consistent pattern: that improper testing is found during covert audits in which the inspector does not know that his or her performance is being evaluated by an auditor. Improper testing has always been found to one degree or another whenever EPA audits have included covert efforts. The problems discussed in Section 3.3.1 are those encountered during these audits. As a result of the ability of covert audits to identify problems in I/M programs, covert auditing is being given a higher priority during EPA audits and EPA is encouraging programs to conduct such audits on a routine basis. It should be clear, however, that these are covert audits on the initial test. EPA is concerned that there does not appear to be a practical means by which covert audits can be done on retests. Covert audits have other limitations, as well. Covert vehicles and auditors must be changed frequently to avoid recognition. Experience indicates that inspectors are getting wise to covert audits and perform perfect inspections for strangers. Thus, the effectiveness of covert audits as a long-term quality assurance tool is questionable. Covert audits have not typically been done in centralized programs. EPA included covert audits of the centralized and decentralized stations in the hybrid New Jersey program because overt audits showed problems with quality control. In 2 out of 8 State-run stations covertly audited in 1989, the vehicle was not failed for not having a catalyst, as it should have been. The same vehicle was used for covert audits of the licensed decentralize?. -26- ------- stations in New Jersey and 6 out of vehicle for a missing catalyst. Figure 3-4 Fraction of Covert Audits Finding Improper Tests* in Decentralized Programs Virginia Michigan New Hampshire Northern Kentucky New York New Jersey Houston North Carolina El Paso Nevada Salt Lake City Colorado Provo, UT Georgia Dallas California Davis County, UT Tulsa New Mexico Overall 36 Sample Size 11746 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% * Improper tests includes the full range of problems, including failure to conduct the test and use of improper procedures 3.3.2.4 Anecdotal Evidence Anecdotal evidence and complaints derived from decentralize program managers and auditors, consumers, inspectors, and other: are also primary sources. Reports of improper testing cover th- full range of possibilities. Common anecdotes include failure t test the vehicle, probing another vehicle, raising the engine spee and not fully inserting probes into tailpipes. EPA and I/M progri -27- ------- management are often alerted to improper testing at a particular I/M station by a concerned citizen. Another source of ccrrplair.ts and anecdotes is inspectors themselves. Inspectors complain accut people who are test shopping, a practice which hurts the hor.est inspector. Inspectors have explained to covert auditors how they can get around the test program requirements. Forthcoming inspectors also tell of customers who try to bribe them or convince them to falsify results. When the inspector refuses, the customer simply announces that another station will surely comply with sucn a request. Sometimes, sharp inspectors have noticed vehicles they recently failed showing a current sticker but which are still in violation. EPA has not been privy to similar anecdotal evidence in centralized programs. It is not clear whether that is because it does not exist or that because of the tighter control, EPA auditors are simply not given that information. 3.3.2.5 Waiver Rate Data Most I/M programs issue waivers to vehicles that fail a retest but have met certain minimum requirements. Programs vary widely in what they require motorists to do before qualifying for a waiver, but the most common requirement is to spend a minimum amount of money on emission-related repairs. Table 3-7 shows the percentage of failed vehicles that get waived in programs that allow waivers and the minimum cost expenditure required. A waiver represents lost emission reductions to the I/M program, so high waiver rates mean substantial numbers of vehicles that are high emitters are not getting adequate repair. Conversely, a truly low waiver rate indicates that maximum emission reduction benefits are being obtained. However, improper testing is, in a sense, a form of unofficial waiver that may be an alternative to the legitimate waiver system employed by programs. A low reported waiver rate by itself is therefore ambiguous v^tr. respect to program success. Waiver rates tend to vary by program type to some degree . Centralized programs typically have higher rates and manual analyzer decentralized programs typically have lower rates. The centralized programs in Arizona, Delaware, Illinois, Indiana, an^ Wisconsin probably represent reasonable waiver rates. This is r.c: to say that these limits are acceptable but all of these progra-.:. have good to excellent enforcement. Enforcement is importer: because there is no need to apply for a waiver if there is : threat from driving without an I/M sticker. These programs na-. - established substantial procedural requirements for receivirg : waiver,and try hard to limit waivers to only those vehicles tn have met the requirements, in most cases. Given this, the wai--.- rates seen in these programs might be what one should expect fr. other programs. -28- ------- Table 3-7 Waiver Rates in I/M Programs in 1989 (percent of initially failed vehicles) Program Pre-1981 Vehicles Waiver Cost Rate1 Limit2'3 Decentralized Manual Davis Co., UT Idaho North Carolina Provo, UT Salt Lake City, UT 13% 7% 0% 3% 4% Decentralized Computerized Alaska 1% California 29% Colorado 2% Dallas/El Paso, TX na Georgia 14% Massachusetts na Michigan 10% Missouri 11% Nevada 4% New Hampshire na ' New York na Pennsylvania 2% Virginia 8% Centralized Government Delaware 3% Indiana 10% Memphis 1% $60 $15 $50 $15 $15 $150 $50 $50 $250 $50 $100 $74 L $200 $50 L $50 $60-$175 $75 $100 $50 Post-1980 Waiver 7% 26% 0% 4% 2% $150 $30 $50 $100 $100 1% 9% 1% na 12% na 9% 14% 2% 1% na 1% 3% 1% 13% 2% $150 $175-$300 $200 $250 $50 $100 $74 L $200 $50 L $50 $200 $200 $100 $50 Centralized Contractor Arizona Connecticut Illinois Louisville Maryland Seattle, WA Spokane, WA Wisconsin 12% 5% 11% 17% 20% 21% 9% 12% $200 $40 L $35 $75 $50 $50 $55 12% 4% 11% 12% 19% 22% 9% 9% $300 $40 L $30-$200 $75 $150 $150 $55 1 na = data not available 2 Some programs vary cost limits by model years and by pollutant failure; thus, some of the limits listed here are only typical of the model year group. Some programs do not have a set cost limit but required specific repairs, indicated by the letter L. 3 Except in Alaska and Utah, cost waivers are not given for tampering. -29- ------- In decentralized programs, low waiver rates occur when it is inconvenient for either the motorist to obtain or the inspector to issue a waiver. In Colorado, an appointment must be made witn a State representative to visit the station when the vehicle is present to issue the waiver. In North Carolina, the motorist must go to a State office to get the waiver. Alternative ways of avoiding an objective test become very attractive when this level of inconvenience is involved. Similarly, if the mechanic has to do extra paperwork, e.g., in Pennsylvania, the improper testing alternative is again more attractive. Effective January 1, 1990, California switched to a centralized waiver processing system. The data shown in Table 3-7 represent the combination of easy to get waivers and the aggressive enforcement against stations in California lead to waiver rates that would be considered typical given the cost limits. Centralized programs are able to control the issuance of waivers to a high degree, but end up with higher waiver rates than are found in decentralized programs. This paradox results from the fact that it is usually very difficult to escape the test requirement in centralized programs. Lower rates are sometimes found when the enforcement system is not adequate and motorists faced with tough requirements simply choose to operate in a non- complying mode, e.g., Connecticut. Another factor that may be operative in some centralized programs (e.g., Seattle and Spokane, Washington) is that the program's geographic coverage is so constrained that it is relatively easy for an otherwise subject motorist to provide an .address, for the purpose of vehicle registration, that places it outside the I/M area. Thus, the potential for waiver control is very high in centralized programs that have high quality enforcement programs and adequate geographic coverage. High costs limits and other rigorously enforced qualification requirements are needed to keep waiver rates below 10% of failed vehicles. Waiver control in decentralized programs may not accomplish much since the alternative of avoiding an objective test is readily available. 3.3.2.6 Test Record Data In an ongoing effort to identify techniques for finding and correcting operating problems, EPA has begun analyzing computerizes test records from a variety of decentralized programs. The analyses involve looking at test histories of vehicles, especially comparing initial tests and retests. At this point, the analysis is still underway and the results discussed in this section represent a small part of the complete project. It is often difficult to calculate similar statistics because programs do net all record and store the same data. Thus, some of the analyses that follow were done on some programs but could not be done en others. Three programs were analyzed to assess the rate of switches model year between the initial test and a retest . An incorre;- -30- ------- model year entry could mean that the wrong outpoints are select;;: for a vehicle, since the computer automatically selects cutccin~3 based on model year. I.i decentralized programs, audits have four.a cases where the model year entry on an initial test was changed to an older year, yielding a less stringent cutpoint, and resulting in a test pass decision. Table 3-8 shows that a small amount of model year switching between initial and retest occurred in all three programs. Naturally, some level of error is to be expected in any program. However, higher rates of switching were found in the two decentralized programs as compared with the centralized program. There is no reason an inspector must wait for a retest to input an incorrect model year, so Table 3-8 may not reflect the true problem. The incidence of incorrect model year entries on first tests was therefore also analyzed for Pennsylvania by comparing the vehicle identification number with the model year entry. Only 1981 and later vehicles were analyzed because these are easily distinguished by the standardized 17 character VIN that went into effect for that model year. The data revealed that 6.4% of the 1981 and later vehicle records had model year entries that were pre-1981. This means that looser emission standards were being applied for these tests. Table 3-8 Model Year Program Connecticut New Hampshire Pennsylvania Switchincr Network Type C D D Between Sample Size 1949 3632 4958 Initial and Number Switched 20 67 156 Revests Percent Switched 1.0 1.8 3.1 Another data element analyzed was the CO + CC>2 scores f r - ' initial test failures and retest passes on the same vehicles. Test histories of vehicles failing the initial test were constructed ar. .1 the dilution levels were compared between initial and retest in New Hampshire and Wisconsin. (Pennsylvania was not included beca-s- te'st time is not recorded, making it difficult to determine tr.-.- actual final test). These I/M programs accept as valid any CO-^:: measurement above 6%. A vehicle with no exhaust leaks and no a.r injection system has a CO+COa level of about 14%. Table 3-9 sh.. .- the fraction of vehicles in the sample that were below 8% CO+CC; . the initial test and the retest. On the initial test about 12* the vehicles in Wisconsin and about 10% of the vehicles in :.- - Hampshire scored below 8%. On the retest, the number of vehic.- scoring below 8% nearly doubled in New Hampshire to 19% of -31- ------- vehicles, while the number was essentially unchanged in Wiscc This analysis, while in no way conclusive, is consistent wiin audits and anecdotes are telling us aboun intentional 3: dilution in decentralized programs. The time that elapses between an initial test on a vehicl a retest is another aspect that EPA is investigating. An analysis of New Hampshire data indicates that about half of all failed vehicles get retests within ten minutes of the initial test. The overall average time between initial and retests is only 11 minutes. Presumably, if an inspector fails a vehicle, and then does diagnosis and repair on that vehicle, one might expect a longer amount of time to have elapsed. Covert audit experience indicates that inspectors in decentralized stations have performed multiple tests on a vehicle that initially fails in an attempt to get a passing score without any diagnosis or repair. Vehi Program New Hampshire Wisconsin cles With Sample Size 390 503 Table 3-9 Dilution Levels Below Initial Test Failures Below 8% 10% 12% 8% Retest Passes Below 8% 19% 11% 3.3.3 Conclusions The available evidence shows that objectivity and quality of testing and the accuracy of instrumentation differ by program type. It was previously found that decentralized programs using manual analyzers suffered from severe quality control problems both in testing and instrumentation. At this point, only one manual program has not committed to switching to computerized analyzers. Decentralized programs using computerized analyzers represent a substantial improvement over manual systems. Analyzer quality control is better, but EPA audits still fail about 20-25% of the analyzers checked. The gross level of errors made by inspectors and the influence of incompetent inspectors are far less because software controls some aspects of the test and determines some of the major decisions about the outcome. Computerized decentralized programs seem to have substantial failure rates, much closer than before to what we would expect based on the emission standards being used and the vehicle mix being tested. Nevertheless, we observe improper testing during audits, and program records describe in detail cases discovered by -32- ------- program auditors. Improper retests are certainly problematic sir.c-i these are vehicles that have already been found to need reoair. Thus, improper testing of these vehicles directly impacts tr.e emission benefit of the program. At this point, the information available on improper testing in decentralized programs is sufficient to conclude that a problem definitely exists. Where waivers are cheap and convenient, the waiver rate is typically about 10-20% in both centralized and decentralized programs. Improper testing is a cheap and convenient alternative in the decentralized programs where waivers are not readily available, and for some vehicles easier than a true repair. It may be optimistic to think that more than 60-80% of high emitting vehicles are actually repaired in any decentralized computerized analyzer program. The actual percentage may be more or less, but it is difficult or maybe impossible to accurately determine. Detecting improper testing is extremely difficult because of the ways in which it occurs. It is relatively easy to catch the gross violators using covert vehicles set to fail the test, as EPA and State experience shows. But, some stations simply will not test vehicles for other than regular customers. Cautious inspectors may not do an improper initial test for strangers. Doing covert audits at new car dealerships presents formidable problems; usually, only regular customers who purchased a vehicle from the dealer would get tested at the dealership. Conducting an evaluation of improper retesting would require purchase of repairs and subsequent retesting, driving the cost of quality assurance higher. Given these problems, it may be too difficult for most State programs to adequately quantify the incidence of improper testing. The question arises about what can be done to deal with the vulnerability of decentralized computerized programs to improper testing. It is unlikely that test equipment can be developed that will completely prevent improper testing, at least not at a reasonable cost . Improvements are being made in the form of the "BAR-90" equipment and software, but programs that have adopted the updated equipment have experienced costs about 2-3 times the price of "BAR-84" equipment of about $6000 - $7000. Nevertheless, even these improvements will not prevent improper testing in decentralized computerized programs. Another way to address the problem is through more aggressive quality assurance and enforcement. More intensive covert and overt auditing and more sophisticated data analysis will enhance identification of problem stations and inspectors. Obtaining funding for additional auditors, covert vehicles, computers, programmers, analysts, judges, etc. will have to be given 3 priority even in the face of tight government budgets. Streamlined administrative procedures and broader legal authority for suspending or revoking station licenses and imposing fines will also help rid programs of problem stations ar.a inspectors. Most decentralized programs face major obstacles .r. -33- ------- trying to get rid of problem stations once they nave iae them. Administrative procedures requirements saddle managers with difficult and expensive barriers. Convincing judges to impose significant penalties or to suspend stations for a substantial amount of time are considerable problems. Furthermore, permanently barring specific individuals from a program is more difficult. Decentralized programs will have to put more financial and good will resources into this aspect of quality assurance. Some, but not all, centralized government-run programs have suffered from improper testing although for different reasons than decentralized programs. Unlike the situation in decentralized programs, it has been easier to gather sufficient evidence to quantify the emission reduction loss from poorly run programs. Ultimately, the problems found in these programs can be resolved and high quality results can be obtained. The solution is better management and equipment, more computer controls and expanded test capacity. EPA is working with the remaining problem government-run programs to upgrade their systems to achieve better results. Contractor-run centralized I/M programs do not seem to suffer from serious improper testing problems as far as we can tell. The efficient management and thorough computer control of these programs eliminate nearly all opportunities for improper testing. Even with all the possible improvements, decentralized programs will have a more uncertain measure of their own effectiveness than centralized programs, due to the greater possibility of continuing but invisible test irregularities. -34- ------- 3 .4 Visual and Functional. Inspections of Emission "or,-.:--'.-. 3.4.1 Background EPA provides State Implementation Plan emission reduction credits for the inspection of certain emission control components. The inspection involves determining whether the emission control device is present and properly connected. The components for which credit is available are the catalytic converter, air injection system, PCV valve, evaporative canister, EGR valve and gas cap. Additional credit is given for misfueling checks: requiring catalyst replacement if the fuel inlet restrictor is tampered or if there is evidence of lead deposits in the tailpipe. Implementation of anti-tampering inspections has varied widely among I/M programs, as can be seen in Table 3-4. Of course, anti- tampering inspections are a completely manual operation. They depend upon the skill and the persistence of the inspector to find components and make the determination that they are properly connected. While this is fairly easy for the catalyst and misfueling checks, finding underhood components can be difficult. The configuration of components varies from make to make and even among models from the same manufacturer. Thus, inspector training is an essential component of an anti-tampering program. With the cooperation of State and local governments, EPA has conducted roadside tampering surveys of motor vehicles in cities throughout the country each year for the past 10 years. These surveys provide information about national tampering rates and trends in tampering behavior. The surveys also provide EPA with the basic information needed to estimate the emission reduction potential of a program. In addition, they provide information on local tampering rates. There are two components to the emission benefit attributed to I/M programs that relate to tampering. Anti-tampering inspections are supposed to identify and correct existing tampering when the program starts. Also, tampering programs are supposed to deter new intentional tampering both on existing vehicles as well as new vehicles entering the fleet. As it turns out, emission testing programs have the latter effect as well. Because of the difficulty in determining whether a vehicle actually exhibits tampering, the emission reduction benefit includes an assumption that not all tampering will be identified and corrected. In the past, this concept applied equally to centralized and decentralized programs. Until 1984, when California and Houston, Texas began comprehensive anti-tampering programs, Oregon was the only program that inspected the complete range of emission control components (see Table 3-4 for details). A few other decentralized programs were doing comprehensive emission control checks, but only as pare of pre-existing safety inspection program and no emission reduction credit was being claimed. Texas and California were formal -35- ------- systems was based on the experience in Oregon. Improper anti-tampering testing is as much if not more of a problem in I/M programs than improper emission testing, since the process is completely manual. The same opportunity exists for irregularities as with tailpipe testing, with no greater ability by the program to detect them. Decentralized inspectors have many of the same motives for improperly inspecting a vehicle for tampering as they do for omitting or improperly conducting an emission test. Centralized programs are also subject to improper anti- tampering checks. Unlike emission tests, the tampering check is completely manual and relies on the honesty, attention to detail, and competency of the inspector to be performed correctly. Centralized programs may benefit from the presence of on-site supervision, the importance to the contractor or agency of maintaining a public image of being accurate and impartial, and by the opportunity for inspectors to become more familiar with underhood geometries due to their constant exposure. At this point, Oregon is the only centralized program that has been conducting comprehensive tampering checks long enough to be fairly evaluated, in terms of fleet-wide tampering rates. Nevertheless, observations of tampering checks in the centralized lanes in Arizona and New Jersey provide additional information about potential effectiveness and will be discussed where appropriate. 3.4.2 Supporting Evidence Tampering surveys conducted by EPA are the main source of tampering rate information. Comparing programs is difficult since the model year and emission control component coverage varies widely among programs. Even a perfect program would not eliminate all tampering since inspections are spread over one or two years (in biennial programs) and the fleet is constantly changing. Immigration of vehicles from outside the program area is one source of "new" tampering. Also, as vehicles age the likelihood of either passive or intentional tampering increases. Thus, an ongoing program is needed to control these problems, and we expect to see at least low levels of tampering whenever a survey is conducted. -36- ------- Figure 3-5 Overall Tampering Rates in Select I/M Programs* 24 . 3* D Centralized I/M + ATP Decentralized I/M + ATP Decentralized ATP Only No Program * The rates shown here are for catalyst, inlet, air, PCV and evaporative system on 1980 - 1984 vehicles. Programs listed inspect and have been doing so for at least two years at the time of the survey. Figure 3-5 lists all programs that were operating for at lease two years at the time of the survey and inspecting at least rr.e 1980 - 1984 model years for catalyst, inlet, air injection system, PCV and evaporative canister. The analysis was limited to these model years and components to establish a fair basis for comparing results. Two no-program areas are included for reference and tir.e rates are all from 1987 and 1988 surveys. Decentralized anti- tampering-only programs still show high overall tampering rates ar2 appear to be at most about 65% effective (Dallas) compared to tne extremes of San Antonio and the Oregon sites. Decentralized I/M - anti-tampering programs have lower tampering rates (except for El Paso) but not as low as in the Oregon cities. -37- ------- Given the model years and survey years involved, i- ra ce that the I/M tailpipe test requirement has played a ro.e ir. deterring tampering in the combined programs, rather than nr.e tampering check being successful in getting tampering fixed or.ce i- has occurred. At this point, the Oregon program is the only one that serves to represent centralized I/M anti-tampering effectiveness. The low rate in Portland may also reflect deterrence more than detection but that cannot be said of Medford since the program there started much later, in 1986. When the analysis is limited to catalyst and inlet tampering, decentralized programs appear to be more effective at finding and fixing this tampering, as shown in Figure 3-6. This may reflect the lower skill and effort required to detect these types of tampering, or the higher cost and therefore profit associated with repair. Figure 3-6 Catalyst and Inlet Tampering Rates in Select I/M Programs* 14.0% Centralized I/M + ATP Decentralized I/M + ATP Decentralized ATP Only No Program * The rates shown here are for catalyst and inlet tampering 1980 - 1984 model year vehicles only. -38- ------- Figure 3-7 Tampering Rates in Decentralized and Centralized Program: Decentralized S3 Centralized 15% 13% 12% PCV Catalyst Inlet Air System Fuel Evap Switching Canister * The rates shown here are for 1975 - 1983 model year vehicles only in the 1987, 1988 and 1989 tampering surveys.. Tampering rates may be an indicator of the effectiveness of emission testing, with or without tampering checks. Figure 3-" shows the combined tampering rates by component from the surveys in 1987-1989 for model years 1975-1983. The tampering rates are lower in the centralized I/M programs for catalyst, inlet and fuel switching. Underhood tampering shows a less dramatic difference but is still lower in centralized programs, despite the fact that only two of the seven centralized programs represented in the survey do underhood tampering checks. Only one decentralized program represented here does no tampering checks at all. -39- ------- Figure 3-8 Aftermarket Catalyst Usage in Ant, 1-Tampering Phoenix Medford Tucson 408 5.6% 259 5.4% 378 4.5% Tulsa |474 Oklahoma City |48l 1.5% 1.5% El Paso 5442; 1.4% San Antonio Covington |433l0.5% Dallas |482i0.4% New Orleans 0.2% 493 mmm LJ Centralized I/M + ATP ss Decentralized I/M + ATP Decentralized H No Program ATP Only * All of the operating programs listed here started in January 1986 except Phoenix, Tucson, and Oklahoma City which started in January 1987. The sample size is the number listed inside the bar. Another indicator of the effectiveness of anti-tamper ir.g programs is the frequency with which aftermarket catalysts are found during tampering surveys, i.e., evidence of actual replacement of catalysts. Since aftermarket catalysts are .TIUCT. cheaper than original equipment manufacturer parts, one wou.u expect them to be the replacement of choice in all programs . Figure 3-8 shows the findings for aftermarket catalysts from ar~- with catalyst inspections that started operation after 1934. Programs that started earlier mostly have been causing owners r.ct to remove their original catalysts, not making them replace or.e; already removed. Also, prior to 1985, anyone wanting to replace a catalyst would not have been able to buy an aftermarket ' catal> s: The three centralized programs show relatively high rates . : -40- ------- af terrnarket catalyst installation. On the otner hana, ir.~ decentralized programs show relatively low rates of aftermarKe- catalyst usage, in some cases no different than non-I/M areas. Figure 3-9 Tampering Rates in I/M and Non-I/M Areas in California 60 50 40 30 20 10 Pre 1975 1975-1979 1980-1983 Model Years Overall I/M NON I/M A variety of sources of data on decentralized anti-tamperi: programs are available. California has done extensive review a: evaluation10 of its decentralized biennial program, which start- in the major urban areas around the State in 1984. The study uŁ- a variety of techniques including roadside tampering surveys. .: of the many important findings of this study was that roads: tampering rates for the items checked in the I/M test did .-. differ substantially between the vehicles that had already ce- subject to I/M and those that had not. It should be noted - California uses a broader definition of the term "tampering" : both its survey and I/M checklist than that used by EPA; thus, i overall rates are not comparable to EPA's national survey rat- These results are illustrated in Figure 3-9. -41- ------- In addition to the survey data, audits of decentralized ar.-_- tampering programs find improper inspections. Covert investigations continually find that inspectors fail to checK fcr components, fail to fail tampered vehicles, and sometimes fail tc do the inspection at all. California's covert auditing work indicates that licensed inspectors neglect to fail tampered vehicles in the majority of cases. Figure 3-10 shows the results by component. During EPA overt audits in decentralized programs, inspectors have been asked by auditors to demonstrate an inspection, and are frequently unable to do the check correctly, either neglecting to check for one or more components or improperly identifying components. 100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0% Figure 3-10 Frequency of Proper Tampering Tests in California's Covert Audit Program Full Effectiveness Level !Observed Effectiveness Level ť f f-^ : : A . -.wb Inlet Evap Canister PCV Air System Catalyst In contrast to the apparent ineffectiveness of decentralized tampering programs, the success of the centralized Oregon program can be seen during audits of the program. The inspection system is well run and the inspectors, employees of the Department of Environmental Quality, are highly trained and perform the inspections diligently. By contrast, EPA's observations of the -42- ------- centralized lanes in New Jersey have found that inscectors sometimes neglect to check for the presence of catalysts, the or.lv component they are supposed to check. Management oversight ar. ^ motivation of inspectors is not adequate in that program. Xew Jersey has historically low tampering rates, so roadside surveys are of only limited use in evaluating program effectiveness tnere. Arizona, Wisconsin, and Maryland also recently started anti- tamper ing inspections, the first contractor-run systems to do so. 3.4.3 Conclusions Based on the evidence that decentralized anti-tampering inspections were not resulting in tampered emission controls getting repaired to the degree expected/ EPA reduced the emission reduction benefit (in MOBILE4, for purposes of approving post-1987 SIPs) associated with them by 50%, unless a specific program demonstrates better performance. Centralized tampering programs, except in New Jersey, seem to be working about as well as expected, but additional survey work needs to be done in centralized programs that have recently started anti-tampering inspections. The potential for effective anti-tampering checks in centralized and decentralized programs is influenced by basic constraints. In the decentralized systems, there is less concern about throughput, but inspection quality seems inherently limited by the difficulties of imposing high levels of quality control. In centralized systems the time it takes to conduct the inspection is a major constraint leading most programs that have started anti- tampering checks to do only the catalyst, inlet and sometimes tr.e air system. It should be noted that these three checks obtain 5a- of the potential HC and 85% of the potential CO benefit of visual anti-tampering checks. Furthermore, the remaining available checks are all for underhood components which require considerable expertise and care in inspection. Given this, limited but effective centralized checks may result in greater emission reduction benefits than the comprehensive but largely ineffective decentralized checks. It is not known if decentralized program performance could be significantly improved by limiting the tampering check to the easily found items and focusing enforcement and educational resources on those items. The magnitude of training differs sharply between centralized and decentralized programs. Depending on the size of the system, the number of licensed inspectors in a decentralized program car. range from a few hundred to over 10,000. Centralized programs range from a few dozen to a couple of hundred inspectors. So, just the physical magnitude of the training requirements for anti- tampering inspections can be daunting in a decentralized systerr, which partly explains why EPA audit findings show that inspector proficiency is an ongoing problem in decentralized programs. -43- ------- 4.0 PROGRAM COSTS 4.1 Inspection Costs Just as there is a wide range of I/M program designs, there is also a wide range of program costs. In every case, there are expenditures related to establishing inspection sites, purchasing equipment, labor associated with conducting the inspection, and program oversight. But the actual level of expenditure seems to be most related to the age of the inspection network, and to the centralized/decentralized choice. To compare the full cost of different I/M designs, EPA collected and analyzed data from as many I/M programs as could provide four basic pieces of information for calendar year 1989: I/M program agency budget, number of initial tests, the fee for each test, and the portion of the test fee returned to the State or local government. Using these parameters, EPA calculated an estimated cost per vehicle in 33 I/M programs currently operating around the country. The results are displayed in Figure 4-1. Decentralized computerized programs have the highest costs, averaging about $17.70 per vehicle, as shown in Figure 4-2. Removing the two highest cost areas (Alaska and California) reduces the average to $13.41. These two programs are much more expensive due to larger fees retained by the two States for aggressive program enforcement, higher labor and material costs, in Alaska especially, and a more involved and complicated test in California. Decentralized programs with manual inspections incur lower costs at an average of $11.60. Most decentralized programs cap the test fee, which may not represent the full cost to the station or, eventually, the public. Centralized contractor-run programs average $8.42 per vehicle, and centralized government-run programs claim the lowest costs at $7.46 per vehicle (the latter figure is somewhat uncertain because detailed budget information is not often available). -44- ------- Figure 4-1 Cost Per Vehicle of I/M Programs I Centralized ..'. Decentralized Anchorage Fairbanks California Nevada New York El Paso New Mexico Massachusetts New Hampshire Virginia North Carolina Indiana Idaho Michigan Wisconsin Georgia Connecticut Colorado Salt Lake City Provo Washington Missouri Davis Pennsylvania D.C. Illinois Maryland Arizona Oregon Nashville Louisville Memphis New Jersey $32 -45- ------- Figure 4-2 Cost Per Vehicle Bv Network Tvoe Computer ^ Manual Contractor Government $11.60 $7.46 There is also a relationship between the age of a program and its inspection fees. Older programs that added I/M to existing safety systems have amortized the infrastructure costs and tend to utilize old technology in the inspection process. These programs come out cheapest, but at a price. They have neither the capacity nor equipment capability that matches the demand created by a growing population of vehicles. The middle aged programs generally came on line after computerization was in full swing and tend to be more sophisticated. In this category, centralized and decentralizea systems tend to experience similar costs - around ten dollars. In the newest programs, and in those where a transition is occurring from manual to computerized analyzers, a divergence appears. The increasing sophistication and growing expertise in operating centralized testing networks, along with growing competition among centralized contractors, has tended to keep costs stable, or even decreasing in some cases. At the same time, the increase in mechanic labor costs and the requirement that garages purchase new, more sophisticated testing equipment has caused the fees in upgraded decentralized programs to rise significantly. A comparison of the average fees for programs of different ages is shown in Figure 4-3. -46- ------- Figure 4-3 Cost Bv Program Acre Upgraded Decentral Middle Aged New Central Contractor Old $15 $13 The ability of "new" centralized programs to provide inspections at a lower cost is well illustrated by the following example. The cost of inspections in Arizona's centralized, contractor-run program recently dropped from $7.50 to $5.40. The decrease resulted from the competitive bidding process that ensued when Arizona issued a new RFP for the program. The decrease occurred despite substantial improvements in the quality, quantity, and range of equipment and services called for in the new contract . The changes include all new test stations and test equipment . Expansion of the network to include an additional station and 14 additional lanes. Test station hours are expanded on Saturdays. There are three lanes for testing heavy duty vehicles rather than two. Finally, the contractor built and staffed a new referee station, and all of the other State referee stations were upgraded with new equipment. The open market process associated with contractor operated systems has forced suppliers to innovate technically, allowing these reductions in cost. If we assume that all I/M programs in the country were centralized and that the inspection and oversight costs would be the same as the average of current centralized programs, che national cost of I/M would be about $230 million less than the cost in 1989. Figure 4-4 shows the total current cost of I/M to ce about $570 million. If all programs were central, at today's average cost per vehicle in centralized programs, the national cos_ of I/M would be $340 million. It may be the case that the per vehicle cost in some decentralized programs, such as the one ir. California, would be higher than the current national average switched to centralized testing. Thus, the potential savings not be as high as $230 but likely would be substantial. i- -47- ------- Figure 4-4 Nationwide Inspection and Oversight Cost of I/M Currently and if All Programs Were Centralized $600,000,000 $500,000,000 $400,000,000 $300,000,000 $200,000,000 $100,000,000 $0 1 Centralized 3 Decentralized Current Cost All Central 4.2 Repair Costs Repair expenditures are also a legitimate cost of I/M. But, regardless of the inspection network, the repairs will be performed in the same way - either by the vehicle owner or by a professional mechanic. Any difference in the cost of similar repairs should be attributable to the relative efficiency of different mechanics cr to differences in shop labor rates, rather than where the initial test was conducted. Repair cost information is collected only sporadically IT. decentralized I/M programs, and is unreliable. Generally, program do not require cost data to be entered into the record unless tr.- vehicle is to get a waiver. Only a few centralized prcgra- collect cost data. These programs generally require motor in- coming in for a retest to provide cost information. Thus, wr. i.-/ some reliable repair cost data exists for centralized programs, .: analysis of the difference between centralized and decentral i .:- : repair costs is not possible. It may be that total repair costs are higher in central! ;:: programs, since improper testing in decentralized programs all;*- some vehicles owners to avoid needed repairs. One should bear . mind, however, that decentralized programs put more vehicle owr-r -48- ------- in a situation in which they may be persuaded to obtain repairs ar.i maintenance services they do not need or would otherwise nave purchased elsewhere. In a 1987 public opinion survey1"- , sixteen percent of motorists living in four decentralized program areas reported that while their vehicle was being tested, service technicians recommended or suggested other services such as tune- ups, brakes, or tires. Forty-three percent of the motorists who had services recommended believed that they were not really needed. 4.2 Conclusions Inspection and oversight costs of I/M programs differ widely among programs. Decentralized programs are more expensive than centralized programs, in nearly all cases. The average cost of decentralized computerized analyzer programs is about double the cost of centralized contractor systems. The national cost of I/M could be substantially lower, on the order of $200 million less, if all programs were centralized. -49- ------- 5.0 CONVENIENCE The relative convenience of decentralized vs. centralized I-M is an issue that should concern policy makers because inconvenience and lost time can be the most negative impact of an I/M program on the public, and it is important that the program be accepted by the public. The factors influencing the convenience or inconvenience of an I/M program are station location, hours of operation, required number of visits, waiting time, and certainty of service. Decentralized programs typically offer numerous test stations scattered throughout the program area and are open to the public during the week and on weekends. These features form the basis for providing a convenient inspection network. Theoretically, I/M tests can be combined with other planned vehicle services, or a vehicle owner can simply drop into the corner garage for a quick test. In practice, the situation is not as simple. According to survey work done in decentralized programs11, the majority of the public experience one or more "inconveniences" in getting their vehicles tested. About 60 percent of the vehicle owners surveyed reported waiting anywhere from 25 to 50 minutes to be tested. Thirty four percent had left their vehicles at a station as long as a day for testing. Some of these may have chosen to do so in order to have other services performed, however. About 20 percent of vehicle owners surveyed were turned away by the first decentralized station they visited for a test. Also 20 percent reported being told to make an appointment and return at a later date. (The responses total more tha-n 100 percent, because some motorists surveyed had had multiple experiences with I/M inspections.) In a decentralized program, all vehicles must get service at a licensed station, not just those which will need repair. Decentralized I/M does not guarantee that a vehicle owner will be able to get a quick emission test on demand at the first station visited. Centralized test networks appear less convenient because there are a limited number of test stations operating during fewer hours than at a typical service station. Further, centralized test stations are not as conveniently located as service stations. Nevertheless, centralized testing has been shown to be reasonably convenient when the network is well designed. A good example is the Milwaukee, Wisconsin system which imposes an average travel distance of 5-6 miles on vehicle owners and a maximum of 10-15 miles. The Wisconsin program is also a good example of providing rapid service. About 98 percent of all vehicle owners wait less than 15 minutes for a test. In the busiest month in Wisconsin, only 4 percent of the vehicle owners have to wait more than 15 minutes and maximum waiting time is 30 minutes. In other centralized programs, average waiting times are generally comparable to the Wisconsin experience. Maximum waiting ti~es vary, however, due to the rush at the end of the month for those who wait till the last minute to get tested. Figure 5-1 shows -r.- daily and overall average waiting times in Illinois. Towards - - - -50- ------- end of the month, the deadline for compliance, waitir.cr iL: increase, but overall Illinois achieves an average waiting -irre under ten minutes. So, good network design means short v;ait times and short driving distances for vehicle owners. Figure 5-1 Average Daily Waiting Times in Illinois' I/M Program 30 25 WRFSTWRFSTWRFSTWRFST Operating Days in the Month of August Some centralized I/M programs have experienced problems with waiting times. In one instance, the problem related to enforcement of the program which was very uneven. On occasion the police would suddenly start to enforce the testing requirement and long queues would develop of people who neglected to get tested when they were scheduled. This has happened in a decentralized program, too. Ir. another case, population growth was not matched by expansion ;: testing facilities. As the area grew, the program administrati;r failed to install new lanes and test facilities and eventually ~ crisis developed. It is obvious that good management and adeq-at-j funding of the program are needed to avoid these kinds of problems For the majority of centralized I/M areas where that is the case, however, waiting times are not a problem. For some motorists whose vehicles fail the initial inspect i ~r, a centralized program may, in fact, be considerably less conver.i-.-r~ than a decentralized program, because they will need to return - an inspection facility for a retest following repairs. But - . inconvenience will be limited to the portion of the population :: fails the initial inspection. At present, this portion is abc-~ -51- ------- to 20 percent. In the survey mentioned previously, a much greater percentage of the respondents in decentralized programs needed to make two trips just to accomplish an initial inspection. This is in part because the type of automotive service facility that can provide on-demand inspections most readily (retail gasoline stations) in a decentralized program is also most likely to lack the repair expertise and parts inventory to repair many vehicles. This will become more likely in the 1990s and as a result, the "extra" trip for a repair in a centralized program would often occur in a decentralized program anyway. The biggest potential convenience problem in a centralized program is where owners are "ping-ponged" between failing tests and ineffective repairs. On-site State advisors and steps to improve repair industry performance can help. Well run centralized programs do these types of things. For example, Wisconsin monitors repair facility performance and visits shops that are having trouble repairing vehicles to pass. The visits include providing assistance to the mechanic in proper retest procedures, providing calibration of the test equipment the shop might own, and referrals on where additional training can be obtained. -52- ------- 6.0 FUTURE CONSIDERATIONS AFFECTING THE COMPAR A number of future developments may affect the relative performance and cost of different network types. For the most part: these developments cannot, at this time, be subject to the kinds of evidential examination produced above. What follows is a discussion of the potential outcomes based on known constraints. Biennial inspections are becoming an increasingly attractive alternative to annual inspections. The largest portion of the cost of I/M is in the inspection process. Thus, reducing inspection frequency will cut the overall cost of I/M, although it may increase the per test cost. EPA has found that the loss in emission reduction benefits is less than the savings. Thus, in a typical switch, about 10% of the benefits will be lost but a minimum 20% of the dollar cost will be averted. Owner inconvenience will be reduced by essentially 50%, since a test is required only half as often. As a result of switching to biennial inspections, some existing centralized networks will have extra capacity that can be used to absorb growth, provide shorter lines in peak periods, or to allow a longer inspection process. However, in the short run, there could be an increase in the per test cost in some networks unless excess existing inspection facilities or lanes are closed. Biennial inspections in decentralized programs mean fewer tests per year per station and analyzer. This means that overhead costs (training, staff, equipment, etc.) must be spread over fewer tests, unless sufficient numbers of stations drop out . In order to maintain profitability, an increase in the test fee will likely be required. The CAAA of 1990 require EPA to establish a performance standard for enhanced I/M programs using an annual program as a model. This means that testing on a biennial basis will require programs to make up for the associated emission reduction loss. Given the reduced effectiveness of decentralized programs, achieving enough reductions will be more difficult, if not impossible. Another option being considered by some and pursued by at least two programs is exempting new vehicles until they reach three or four years of age. The most recent three model years represent about 22% of the vehicle fleet. However, because they have haa little time to develop problems, inspecting them produces very little emission reduction. Exempting these vehicles would have effects similar to that of biennial testing, reducing test voiu.-r.e and revenue. The same kinds of impacts in terms of each networ< type would follow as well. However, it should be noted that exempting new vehicles can cause owners to miss the opportunity to exercise the 2 year/24,000 mile emission performance warranty provided under Section 207 of the Clean Air Act. -53- ------- Another major change needed in all I/M programs in the r.e:-;- few years will be adoption of advanced test procedures, improved preconditioning methods, better emission sampling algorithms, r.cre sophisticated computer systems, and more extensive data recording. All of these improvements are needed as a result of refinements in our understanding of existing vehicle technology and what we expect in terms of future technology. The need for preconditioning was discussed in detail in section 3.1. The adaptability of I/M programs to changing conditions becomes a major issue in light of these needs. Most centralized programs use mainframe or mini computers to operate the analyzers and record data. These systems can easily be reprogrammed whenever a change in the inspection process is desirable. Such a change can be debugged and implemented quickly with a minimum of expense and difficulty. The existing computerized emission analyzers in decentralized I/M programs are for the most part hardware programmed or closed-architecture systems, with computer chips rather than software governing operation of the analyzer. Thus, making changes to these analyzers requires installation of new computer chips, a more costly proposition than a simple software update. The latest developments in decentralized analyzers call for the use of an open-architecture system that will allow reprogramming; however, these analyzers cost $12,000-$15,000 for the basic machine as compared to $6,000-$8,000 for the existing technology. This additional expense will further the demand for increased test fees in decentralized programs. The existing analyzers in most programs are aging, although still good for repair work if properly maintained and calibrated regularly. Many decentralized programs will be faced with the choice of replacing computerized analyzers to meet new testing requirements, or making a switch to centralized testing. On the newest fuel injected vehicles, current short tests leave room for improvement. As time passes, the fuel injected portion of the fleet will grow more and more important, making it more important to improve testing of these vehicles. However, improved test procedures may require the use of steady-state or transient dynamometers. At this point, the use of either type of dynamometer is most feasible in centralized programs where the cost can be spread over many tests. Use in a decentralized program would likely result in fewer test locations. Another testing frontier relates to the use of "on-board diagnostics" or OBD. Starting in 1981, some motor vehicles were manufactured with computers that monitor engine performance during vehicle operation, detect any malfunctions in the system, and store diagnostic information in the computer memory. Usually, the motorist is alerted through the use of a malfunction indicator light on the dashboard of the vehicle. OBD is currently require:: for all new vehicles in California, and EPA is developing regulations to standardize the systems at the federal level, a; -54- ------- required by the CAAA of 1990. I/M programs will be perform OBD checks once these vehicles are in use . 03D has grean potential to enhance repair effectiveness and provide an alternative or an add-on to emission testing. In the near term, decentralized I/M programs may suffer from improper testing of OBD since checking for the malfunction light, while simple, is essentially a manual process. Even in the long term when the vehicle computer will "talk" directly with the analyzer computer, there will be ways to defeat a decentralized test, for example, by connecting the analyzer to a known "clean" vehicle instead of the subject vehicle, as is done currently with the emission test. -55- ------- 7.0 PREDICTED EMISSION REDUCTIONS FROM MOBILE4 AND AM ADI ASSUMPTION REGARDING WAIVERS The emission reduction impacts of various potential I/M programs were analyzed for their relative effects and to assess the difference between centralized and decentralized networks. The most recent version of EPA's mobile source emission model, MOBILE4, was used for the mechanics of the calculation. MOBILE4 assumes that anti-tampering inspections in decentralized are only half as effective as centralized programs; it assumes emission testing is equally effective in either program type. For the purpose of this analysis, an additional "what if" assumption regarding the outcomes of a higher waiver limit was made, as explained below. Figure 7-1 serves mainly to put the known and possible differences between centralized and decentralized programs into perspective with the emission reduction effects that are possible via other program changes. The first bar in each of the charts in Figure 7-1 shows the VOC emission reduction benefit from a typical emission test-only I/M program. While no two programs are exactly alike, a typical program was taken to be one which covered all model years, all light duty vehicles and light duty trucks, failed 20% of the pre-1981 vehicles, used an idle test, and annual inspections. The waiver rate is assumed to be 15%. The next bar shows the level of benefit this program design would achieve if it switched to biennial inspections. The third bar shows the benefit from a typical program with biennial inspections that exempts vehicles up to four years old. The next bar shows the impact of reducing waivers to 5% of failed vehicles by increasing the cost limits and tightening procedures. For the purposes of this bar, decentralized programs are assumed to achieve no additional benefit on the theory that improper testing would be substituted when waivers are constrained. Other judgmental estimates can be interpolated visually. The next two bars show the benefit from adding catalyst and misfueling checks. The differential impact is based on the assumptions built into MOBILE4. The assumptions for decentralized programs are that detection is 50% of centralized, but deterrence is assumed to be equal that of a centralized program. The addition of an improved emission test (i.e., a transient test) would not be feasible in a decentralized program, so no additional benefit is credited there. Finally, underhood checks are added and the differential impact is once again based on the assumptions built into MOBILE4. The figure shows the benefit in 1990, 1995 and 2000. These scenarios do not take into consideration any loss in benefit associated with improper emission testing in decentralized programs (except as noted in the reduced waiver scenario). MOBILE4.1, which will be used to construct base inventories required by the CAAA of 199C, will do so. Benefits for carbon monoxide reductions are not shown but are similar except that tampering checks are not as important. -56- ------- Figure 7-1 Benefits From Various Potential Changes to I/M Pr Centralized s-.-.j Decentralized Typical Program Switch to Biennial Exempt New Vehicles Reduce Waivers Add Catalyst & Inlet Test Add Lead Test Add Better Emission Test Add Under hood Checks .000 .100 .200 .300 .400 .500 Grams Per Mile VOC Reductions in 1990 .600 Typical Program Switch to Biennial Exempt New Vehicles Reduce Waivers Add Catalyst & Inlet Test Add Lead Test Add Better Emission Test Add Underhood Checks .000 .100 .200 .300 .400 .500 Grams Per Mile VOC Reductions in 1995 .600 Typical Program Switch to Biennial Exempt New Vehicles Reduce Waivers Add Catalyst & Inlet Test Add Lead Test Add Better Emission Test Add Underhood Checks .000 .100 .200 .300 .400 .500 Grama Per Mile VOC Reductions in 2000 -57- ------- 8.0 CONCLUSIONS The first I/M programs, established voluntarily by tr.ree States in the 1970s, were set up as centralized networks. Decentralized networks appeared in response to the requirements of the 1977 Clean Air Act Amendments and currently comprise the majority of I/M programs in the United States. The early thinking was that decentralized inspection would be less costly and more convenient to the vehicle owner. It was acknowledged that more quality assurance would be required to ensure equivalent emission reductions, but it was also assumed, at least officially, that success could be achieved with a reasonable level of resources. This report discusses the relative effects of centralized and decentralized I/M programs. It focuses on three key issues: emission reduction effectiveness, cost, and convenience. It presents information derived from EPA testing programs, EPA and State audits of I/M programs, and analyses of I/M operating data. Recent studies have found that vehicles require adequate preconditioning before a test to assure that they are at normal operating temperature, and that any adverse effect of extended idling is eliminated. A period of loaded operation on a chassis dynamometer has been found to be most effective. Most decentralized programs, especially those requiring the use of computerized analyzers do unloaded, high-speed preconditioning. This can work if I/M programs extend the length of the preconditioning, as EPA has recommended. A chassis dynamometer would allow a shorter period but the purchase and installation of a chassis dynamometer is considered beyond the financial capability of most private repair facilities. Some centralized programs have avoided pre-conditioning because it was not thought essential and to keep costs and test time as low as possible. The trend now, however, is to provide loaded preconditioning and a second chance test to vehicles which fail an initial idle test. Centralized programs typically do few or no anti-tampering inspections. Decentralized programs typically require comprehensive checks on at least some portion of the vehicle population. The effectiveness of decentralized tampering inspections is highly suspect, however. As with preconditioning, centralized programs are starting to add anti-tampering checks t . the normal test routine. EPA audit findings show that centralized contractor-r ^r. programs have very high levels of instrument quality control. Centralized government-run systems and computerized decentralizei programs are comparatively weak. The quality of calibration gaj that is purchased, the frequency with which checks are perforre:, the easy opportunity to defeat the checks, and the le- sophist icated instrument technology are to blame. Man-.:. decentralized programs have, with a few exceptions, h : unacceptable levels of quality control. This has led to rr.o^- manual programs changing over to computerized analyzers. -58- ------- The available evidence shows that objectivity and quality of testing - the keys to emission reduction effectiveness - differ greatly by program type. It was previously found thai: decentralized programs using manual analyzers had a very high rate of inspectors conducting tests improperly, either intentionally or inadvertently. For the most part, inspectors were passing vehicles which should have failed and been repaired. The use of computerized analyzers in decentralized programs has reduced the level of inadvertent improper testing. Correspondingly, the initial test failure rates have risen substantially in programs that have computerized. Audits have found, however, that improper testing, both on the initial test and the retest, sometimes occurs despite the use of computers. Current inspection costs per vehicle were determined for a number of operating programs. The earliest decentralized programs did, in fact, charge lower inspection fees, because State legislatures imposed fee caps to protect vehicle owners. The trend has reversed, however, since requirements for computerized analyzers have been imposed. Sophisticated, high through-put centralized systems are now "outbidding" the local garage, much as franchise muffler and tune-up shops have overtaken their respective aspects of the repair business. The trend occurring in programs which are new, revised, or reauthorizing is for increasing fees in decentralized programs and decreasing fees among competing contractors. Decentralized computerized programs now have the highest costs, averaging $17.70 per vehicle. Centralized contractor-run programs average $8.42 per vehicle. Centralized government-run systems claim the lowest cost at $7.46 per vehicle, on average. The factors influencing the convenience or inconvenience of an I/M program include station location, hours of operation, waiting time, required number of visits, and certainty of service. Decentralized programs offer stations located conveniently to most motorists. However, a significant number of respondents to a recent survey reported being turned away without an inspection, or having to wait 25-50 minutes for an inspection. Because there is a growing scarcity of qualified mechanics in the automotive aftermarket, and since many gas stations have converted repair bays into convenience stores, it has become increasingly difficult to obtain "on demand" automotive services. Centralized programs have a limited number of facilities, and may experience long lines if the system is not well designed. Most new programs, however, have taken steps to minimize the travel distance and the waiting time, such that centralized programs are just as convenient or more convenient than decentralized systems. The overall conclusion is that centralized I/M will usually offer greater emission reduction benefits than decentralized I/M, unless the decentralized program makes special efforts that may border on the unreasonable. It has been shown that this greater benefit can be achieved at a lower cost and with limited -59- ------- inconvenience to the motorist. These advantages also dovetail vitr. trends in I/M technology, which all point in the direction cf increased sophistication, leading to higher cost unless economies of scale can be achieved. There is a growing need to assure that all of the emission reduction potential of I/M programs is achieved in actual operation. Quality is a must, if I/M is to play its part in achieving the ambient air quality standards. Reglr -60- ------- References 1 National Air Audit System Guidance Manual for FY 1983 and FY 1989, USEPA, Office of Air Quality Planning and Standards, EPA-450/2-88-002, February 1988. 2 Motor Vehicle Tampering Surveys - 1984 -1988, USEPA, Office of Mobile Sources. 3 A Discussion of Possible Causes of Low Failure Rates in Decentralized I/M Programs, USEPA, Office of Mobile Sources, EPA-AA-TSS-I/M-87-l, January 1987. 4 Evaluation of the California Smog Check Program, Executive Summary and Technical Appendixes, California I/M Review Committee, April 1987. 5 Report of the Motor Vehicle Emissions Study Commission, State of Florida, March 1988. 6 Study of the Effectiveness of State Motor Vehicle Inspection Programs, Final Report, USDOT, National Highway Traffic Safety Administration, August 1989. 7 I/M Test Variability, EPA-AA-TSS-I/M-87-2, Larry C. Landman, April 1987. 8 Recommended I/M Short Test Procedures For the 1990s: Six Alternatives, USEPA, Office of Air and Radiation, EPA-AA-TSS-I/M-90-3, March 1990. 9 Vehicle Emission Inspection Program Study, Behavior Research Center, Inc., Phoenix, AZ, 1989. 10 Evaluation of the California Smog Check Program, April 1987 11 Attitudes Toward and Experience with Decentralized Emissicr. Testing Programs, Riter Research, Inc., September 1987. U.S. Environmental Protection Agency Region 5, Library (PL-12J) 77 West Jackson Boulevard, 12th Floor -61- Chicago, IL 60604-3590 ------- |