UNITED STATES ENVIRONMENTAL PROTECTION AGENCY
                                   WASHINGTON D.C. 20460
                                                             OFFICE OF THE ADMINISTRATOR
                                                               SCIENCE ADVISORY BOARD
                                  September 28, 2006
EPA-SAB-06-011
The Honorable Stephen L. Johnson
Administrator
U.S. Environmental Protection Agency
1200 Pennsylvania Avenue, N.W.
Washington, D.C. 20460

       Subject:  SAB Review of the EPA Region 6 Geographic Information System Screening
               Tool

Dear Administrator Johnson:

   The EPA Region 6 Compliance Assurance and Enforcement Division requested Science
Advisory Board (SAB) review of the Region's Geographic Information System Screening Tool
(GISST). The GISST is a geographic information system-based tool used for rapid preliminary
environmental assessment and  reconnaissance. EPA Region 6 has applied the GISST to develop
information for evaluating environmental impact statements required under the National
Environmental Policy Act (NEPA).  An SAB panel reviewed the strengths and limitations of the
GISST. The enclosed SAB report addresses EPA's charge questions to the Panel.

   The SAB commends Region 6 for developing the GISST.  Geographic Information System
(GIS) capabilities and data layers provide essential support for efficient, timely, and proactive
NEPA evaluations and other Regional responsibilities. The SAB notes that several elements
make GISST evaluations different from other GIS evaluations.  Unique GISST elements include:
1) the criteria scoring process, 2) criteria subset selection, and 3) the process that highlights
important drivers of concern for further analysis. These elements make the GISST an objective,
spatially explicit tool for conducting, broad-stroke preliminary evaluations in a timely fashion.
The SAB finds that it is reasonable and appropriate to evaluate individual criteria or suites of
criteria in the GISST to detect potential  environmental impacts of certain types of projects.

   The SAB also has identified limitations in the methodological approach used in the current
version of the GISST. Because of these limitations, the aggregate GISST vulnerability or impact
score  should not be used at this time to conduct assessments for decision-making, such as setting
priorities or considering the impacts of project alternatives. The SAB has recommended
improvements to make the GISST suitable for future use.  In this regard the SAB finds that:

-------
   •   The current version of the GISST provides only a single vulnerability or impact score that
       can mask important differences in individual data layers used for an assessment.  Such
       differences must be considered when evaluating the potential environmental impacts of
       project alternatives as part of an overall strategy to achieve assessment objectives.

   •   The current version of the GISST does not adequately utilize modern decision analytic
       and statistical science in its development of numerical scoring. These issues must be
       considered when combining and  evaluating data layers in the GISST.

   •   EPA should draw upon the available expertise and large literature on multi-attribute
       decision-making to further develop the GISST.

   GIS-based assessment tools are needed to provide essential support for many EPA activities.
Various Agency program offices and regions have developed key components of these tools. In
addition to the Region 6 GISST, other examples include the Region 4 Southeastern Ecological
Framework, the Region 5 Critical Ecosystem Assessment Model, the Office of Water's Index of
Watershed Indicators, and the Office of Research and Development's Regional Vulnerability
Assessment methods. In reviewing all of these assessment tools, the SAB has noted common
problems of inadequate involvement of statistics and decision analytic expertise, weighting of
metrics, and procedures used to combine data.  Furthermore, despite these initiatives, the Agency
still does not have a unified single accepted approach for using spatially explicit information for
environmental decision-making. The compartmentalized development of GIS-based tools and
data by EPA program offices and regions is inefficient, given budgetary constraints and the high
value of these tools for environmental decision-making. Using spatially explicit information for
decision-making is a critical need that EPA must address.  The SAB therefore strongly urges
EPA to undertake an initiative to define a unified framework for the development of these types
of tools across the Agency. A national effort to develop such a unified framework will require
decision analytic expertise, and the resulting tool should be easy to modify to meet local and
regional needs.
                                         Sincerely,
       /signed/

  Dr. M. Granger Morgan, Chair
  EPA Science Advisory Board
         /signed/

Dr. Virginia Dale, Chair
Geographic Information System
    Screening Tool Review Panel
EPA Science Advisory Board

-------
                                     NOTICE

   This report has been written as part of the activities of the EPA Science Advisory Board,
a public advisory group providing extramural scientific information and advice to the
Administrator and other officials of the Environmental Protection Agency.  The Board is
structured to provide balanced, expert assessment of scientific matters related to the problems
facing the Agency. This report has not been reviewed for approval by the Agency and,
hence, the contents of this report do not necessarily represent the views and policies of the
Environmental Protection Agency, nor of other agencies in the Executive Branch of the
Federal government, nor does mention of trade names or  commercial products constitute a
recommendation for use. Reports of the EPA Science Advisory Board are posted on the EPA
website at http://www.epa.gov/sab.

-------
                     U.S. Environmental Protection Agency
                             Science Advisory Board
         Geographic Information System Screening Tool Review Panel
CHAIR
Dr. Virginia Dale, Corporate Fellow, Environmental Sciences Division, Oak Ridge National
Laboratory, Oak Ridge, TN
MEMBERS
Mr. DeWitt Braud, Director, Academic Area, Coastal Studies Institute, Louisiana State
University, Baton Rouge, LA

Dr. Ivan J. Fernandez, Professor, Department of Plant, Soil and Environmental Sciences,
University of Maine, Orono, ME

Dr. Carol Johnston, Professor, Center for Biocomplexity Studies, South Dakota State
University, Brookings, SD

Dr. William Mitsch, Professor, Olentangy River Wetland Research Park, The Ohio State
University, Columbus, OH

Dr. Thomas C. Mueller, Professor, Department of Plant Sciences, University of Tennessee,
Knoxville, TN

Dr. Michael C. Newman, Professor of Marine Science, School of Marine Sciences, Virginia
Institute of Marine Science, College of William & Mary, Gloucester Point, VA

Dr. James Oris, Professor, Department of Zoology, Miami University, Oxford, OH

Dr. Charles Rabeni, Leader, Missouri Cooperative Fish and Wildlife Research Unit, U.S.
Geological Survey, Columbia, MO

Dr. Mark Ridgley, Professor and Chair, Department of Geography, University of Hawaii at
Manoa, Honolulu, HI

Dr. Amanda Rodewald, Assistant Professor, School of Natural Resources, The Ohio State
University, Columbus, OH

Dr. James Sanders, Director, Skidaway Institute of Oceanography, Savannah, GA

Dr. David Stoms, Associate Research Scientist, Institute for Computational Earth Systems
Science, University of California at Santa Barbara, Santa Barbara, CA

-------
Mr. Timothy Thompson, Senior Environmental Scientist, Science, Engineering, and the
Environment, LLC, Seattle, WA

Dr. Robert Twiss, Professor, University of California-Berkeley, Ross, CA
SCIENCE ADVISORY BOARD STAFF
Dr. Thomas Armitage, Designated Federal Officer, U.S. Environmental Protection Agency,
Washington, DC
                                         in

-------
                     U.S. Environmental Protection Agency
                             Science Advisory Board
CHAIR
Dr. M. Granger Morgan, Carnegie Mellon University, Pittsburgh, PA
MEMBERS
Dr. Gregory Biddinger, ExxonMobil Biomedical Sciences, Inc, Houston, TX

Dr. James Bus, The Dow Chemical Company, Midland, MI

Dr. Trudy Ann Cameron, University of Oregon, Eugene, OR

Dr. Deborah Cory-Slechta, University of Medicine and Dentistry of New Jersey and Rutgers
State University, Piscataway, NJ

Dr. Maureen L. Cropper, University of Maryland, College Park, MD

Dr. Virginia Dale, Oak Ridge National Laboratory, Oak Ridge, TN

Dr. Kenneth Dickson, University of North Texas, Denton, TX

Dr. Baruch Fischhoff, Carnegie Mellon University, Pittsburgh, PA

Dr. A. Myrick Freeman, Bowdoin College, Brunswick, ME

Dr. James Galloway, University of Virginia, Charlottesville, VA

Dr. Lawrence Goulder, Stanford University, Stanford, CA

Dr. Rogene Henderson, Lovelace Respiratory Research Institute, Albuquerque, NM

Dr. Philip Hopke, Clarkson University, Potsdam, NY

Dr. James H. Johnson, Howard University, Washington, DC

Dr. Meryl Karol, University of Pittsburgh, Pittsburgh, PA

Dr. Catherine Kling, Iowa State University, Ames, IA

Dr. George Lambert, UMDNJ-Robert Wood Johnson Medical School/ University of Medicine
and Dentistry of New Jersey, New Brunswick, NJ
                                         IV

-------
Dr. Jill Lipoti, New Jersey Department of Environmental Protection, Trenton, NJ

Dr. Genevieve Matanoski, Johns Hopkins University, Baltimore, MD

Dr. Michael J. McFarland, Utah State University, Logan, UT

Dr. Jana Milford, University of Colorado, Boulder, CO

Dr. Rebecca Parkin, The George Washington University, Washington, DC

Mr. David Rejeski, Woodrow Wilson International Center for Scholars, Washington, DC

Dr. Joan B. Rose, Michigan State University, E. Lansing, MI

Dr. Kathleen Segerson, University of Connecticut, Storrs, CT

Dr. Kristin Shrader-Frechette, University of Notre Dame, Notre Dame, IN

Dr. Robert Stavins, Harvard University, Cambridge, MA

Dr. Deborah Swackhamer, University of Minnesota, Minneapolis, MN

Dr. Thomas L. Theis, University of Illinois at Chicago, Chicago, IL

Dr. Valerie Thomas, Georgia Institute of Technology, Atlanta, GA

Dr. Barton H. (Buzz) Thompson, Jr., Stanford University, Stanford, CA

Dr. Robert Twiss, University of California-Berkeley, Ross, CA

Dr. Terry F. Young, Environmental Defense, Oakland,  CA

Dr. Lauren Zeise, California Environmental Protection Agency, Oakland, CA
SCIENCE ADVISORY BOARD STAFF
Mr. Thomas Miller, Designated Federal Officer, U.S. Environmental Protection Agency,
Washington, DC

-------
                           TABLE OF CONTENTS







      EXECUTIVE SUMMARY	vii





1.     INTRODUCTION	1





2.     CHARGE TO THE REVIEW PANEL	2





3.     RESPONSE TO THE CHARGE QUESTIONS	2




      3.1   Question 1.1	2




      3.2   Question 1.2	7




      3.3   Question 1.3	11




      3.4   Question 2.1	16




      3.5   Question 2.2	18




      3.6   Question 3.1	20





4.     CONCLUSION	22





5.     REFERENCES	24





APPENDIX A. SPECIFIC COMMENTS ON INDIVIDUAL GISST CRITERIA	A-l





APPENDIX B. SPECIFIC COMMENTS ON THE GISST USER'S MANUAL	B-l
                                    VI

-------
                               EXECUTIVE SUMMARY

    The EPA Region 6 Compliance Assurance and Enforcement Division requested that the
Science Advisory Board (SAB) review the Region's Geographic Information System Screening
Tool (GISST). The GISST is a geographic information system-based tool for evaluating the
potential environmental impacts of large projects such as the construction of roads and the
issuance of permits for water treatment plants. A mathematical algorithm is used in the  GISST
to evaluate various impact or vulnerability criteria data layers and derive an aggregate score that
has been called the "potential for significant environmental risk" of a project. The GISST has
been used to provide information for preparing and reviewing environmental assessments and
impact statements required under the National Environmental Policy Act (NEPA).  The  SAB
GISST Review Panel met in December 2005 and deliberated on six charge questions.  These
questions focused on: 1) whether the GISST methodology is reasonable and appropriate for use
in conducting initial assessments of potential environmental vulnerability and impacts; 2) the
strengths and limitations of the GISST as a tool for use in prioritizing and comparing
environmental vulnerabilities and impacts for decision-making; and 3) recommendations to
improve the GISST User's Manual and documentation.

   The Panel commends Region 6 for developing the GISST and providing strong GIS support
to its environmental managers.  Geographic Information System (GIS) capabilities and data
layers provide essential support for efficient, timely, and proactive NEPA evaluations and other
Regional responsibilities. GISST is an exploratory tool that can improve understanding  of how
potential decisions can affect the systems being considered.  Three types of potential NEPA
issues can be explored using the GISST:  1) consideration of whether impacts are likely to occur
(the scoping process); 2) consideration of whether impacts have been adequately addressed in an
environmental document; and 3) consideration of alternatives that appear to have the least
impact. The benefits of compiling data layers for the GISST have undoubtedly extended to other
applications within Region 6.  The Panel finds that it is reasonable and appropriate to use the
scores of individual GISST criteria, or suites of criteria corresponding to different types  of
vulnerability, for conducting broad-stroke preliminary evaluations to detect potential
environmental impacts of certain types of projects. However, the Panel has identified limitations
in the methodological approach used in the current version of the GISST. Because of these
limitations, the aggregate GISST score should not be used in assessments for decision-making.

   Concerted EPA effort is needed to develop assessment tools like the GISST.  Various EPA
program offices and regions have  created screening tools similar to the  GISST since GIS
technology became widely available in the 1990s.  Examples include the EPA Region 4
Southeastern Ecological Framework, the EPA Region 5 Critical Ecosystem Assessment  Model,
the EPA Office of Water's Index of Watershed Indicators, the EPA Office of Research and
Development's Regional Vulnerability Assessment methods, and the NEPAssist web-based
mapping tool developed by EPA's Office of Federal Activities.  Leibowitz et al. (1992) utilized
similar methods in the synoptic approach developed for cumulative impact assessment of
wetlands.  The principal goal in all of these developments was to harness the power of geospatial
data as an aid to environmental decision-making. However, the utility of these GIS-based
decision assistance tools is limited by the amount and quality of the underlying geospatial
                                          Vll

-------
environmental data sets and the lack of suitable indicators of biological and ecological effects.
   Many EPA program offices and regions have common needs for data sets such as those
developed for the GISST.  Particular examples include the geographically aggregated summaries
of point data from EPA's Storage and Retrieval System (STORET), the Toxics Release
Inventory (TRI), and National Pollution Discharge Elimination System (NPDES) databases. The
compartmentalized development of GIS tools and data by EPA program offices and regions is
suboptimal, given the universal need for such tools. It does not help the development of a
national perspective, nor is it an efficient use of scarce resources.  The Panel urges EPA to make
a concerted effort to develop a unifying framework for the creation of these tools and suggests
that they could be provided by EPA's Office of Environmental Information.  The SAB
Ecological  Processes and Effects Committee made a similar recommendation in its review of
EPA's Index of Watershed Indicators (U.S. EPA Science Advisory Board, 1999). In that review
the SAB recommended that, "the Agency should add more indicators of biological and
ecosystem effects to the Index of Watershed Indicators."  The panel also urges that future
development of the GISST be consistent with the principles embodied in EPA's  Guidance on the
Development, Evaluation, and Application of Regulatory Environmental Models (U.S. EPA
Office of Science Policy,  2003).

 GISST Mathematical Algorithm

   The Panel was asked to comment on the reasonableness and appropriateness of using the
GISST algorithm for conducting screening-level1 evaluations. The algorithm is simple and
straightforward, providing a reasonable way to conduct initial evaluations. However, the
algorithm generates an aggregate vulnerability or impact  score that may mask important
differences in individual criteria scores. Such differences must be considered when evaluating
project alternatives. The panel suggests that the algorithm score itself be de-emphasized and the
GISST be used as part of a screening process that clearly defines the limits of integrative indices
and promotes the subjective scientific evaluation of numerical output and supporting
information.  In order to further explore uses of the GISST for rapid environmental assessment
and preliminary reconnaissance, EPA should develop  optional models for aggregating the criteria
used in the algorithm and weighting2 them according to their relative importance to decision-
makers.  The Panel emphasizes, however, that in summing criterion scores, high scores can
offset or compensate for low ones and hence mask potential problem areas. Such compensation
reflects implicit tradeoffs, functions of the coefficients and units of the scores being summed,
regardless of the intent and preferences of the users and clients of the analysis. Since such intent
and preferences are unlikely to be known, or known fully, the summation of criterion scores is
unlikely to express participants' acceptable tradeoffs.  In  addition, to be mathematically
legitimate,  one must assess or rate an alternative's (e.g., site's) performance on the relevant
criteria on measurement scales appropriate for the arithmetic operations - addition, subtraction,
multiplication, and division - subsequently used in the algorithm.  Thus, summation of impact
and vulnerability scores should only be used in  an exploratory vein and as a preliminary
reconnaissance.  The potential  utility of summing scores lies with uncovering potential problems
1 In this context a screening-level evaluation is a rapid preliminary assessment of potential environmental and
societal impacts of certain types of projects.
2 Weighting is the process of assigning multipliers to a set of quantities to indicate the relative importance of each
quantity within the set.
                                           Vlll

-------
derived from the interaction of multiple impacts.  Summation should not be used to verify the
absence of adverse consequences. It is recommended that the GISST algorithm be re-evaluated,
taking advantage of additional expertise in decision science and statistics to address these and
other concerns expressed in this report.

   The Panel was also asked to consider the reasonableness and appropriateness of the method
used by EPA Region 6 to evaluate environmental vulnerability in the Interstate Highway 69 case.
The published GISST algorithm was not used in this case. Instead, the vulnerability within the
highway  corridor was evaluated by summing the vulnerability scores within 1 km2 areas. The
Panel finds that the approach described in this case can be used to explore potential
vulnerabilities or impacts if the criteria scores are averaged and as long as they represent only
impacts or only vulnerabilities.  Other concerns regarding mathematical operations for
aggregating scores are described below.

GISST Criteria

   The Panel was asked to comment on whether the GISST criteria (the kinds of data used to
score vulnerability and impact) were reasonable and appropriate for use in evaluations of the
potential impacts of projects and vulnerability of project areas. The Panel finds that the
individual GISST criteria are intuitive, but there is a need to better describe how to use groups of
criteria. The criteria have competing purposes with varied thresholds that can result in criteria
being combined in illogical ways (e.g., combining noise and odor criteria scores with scores for
use of energy efficient appliances). Criteria categories need to be defined consistently so that
combinations or groupings make sense in an overall strategy. Instead of focusing on the
appropriateness of a particular group of criteria, it would be better to define a general process for
selecting groups of criteria for use in various kinds of evaluations. The GISST could be
improved by developing impact templates. These templates  could identify the kinds of impacts
that might be associated with particular project types and the key criteria relevant to evaluation
of those impacts.

   The Panel is also concerned that EPA has not clearly described the differences between the
criteria used to determine vulnerability and those used to determine impact. Some criteria reflect
both vulnerability and impact but others reflect only one of these indices. More detailed
descriptions and explanations of the criteria and supporting databases are needed in the GISST
documentation. A statistical examination of the supporting data is needed to:  1) determine
relationships between the criteria data layers, 2) identify potentially redundant measures, and 3)
increase the soundness of the rating scales.

GISST Scoring System

   The Panel was asked whether the GISST 1-5 scoring scale was reasonable for use with
different data sets and data coverages to develop an initial assessment of the potential cumulative
impacts of proposed projects. The GISST scoring system is reasonable for use in detecting
potential individual concerns where no mathematical operations are required for aggregating
scores. The advantage of the GISST scoring system is that it allows evaluators to simplify a
combination of diverse criteria and show contrasts in the assessment of potential impacts among
                                            IX

-------
project alternatives. A disadvantage of the scoring system is that assessors lose the ability to see
specific information that may be provided by the underlying data for a particular assessment.

   A number of issues must be considered if mathematical operations are required to aggregate
criteria scores.  If the GISST criteria values are ordinal, it is not appropriate to average or
multiply them.  The GISST criteria values seem to have been binned or scaled using different
techniques or functions so it may be  inappropriate to equate their scales during summing. A
number of statistical issues must also be considered. Many  of the criteria seem to overlap and
possibly interact.  The potential consequences of interactions are unclear and must be evaluated.
The Panel suggests that the GISST criteria scoring system could be substantially improved by
involving more decision analytic and statistical expertise in  the development of an expert
opinion-based ranking system.  Explicit decision rules and criteria weightings should also be
defined.

Use of the GISST in the NEPA Process to Prioritize Project Impacts for More Detailed Analysis

   The Panel commented on the strengths and limitations of the GISST as a screening tool to
prioritize project impacts in the NEPA process. If sets of core criteria were identified for
evaluating certain kinds of projects, the GISST could be used in the NEPA process to assist EPA
in reviewing and scoping environmental assessments and environmental impact statements.
However, the GISST is inadequate for prioritizing impacts unless the following limitations are
addressed: 1) the GISST criteria must be weighted so that scores represent the relative
importance of various impacts to decision-makers; 2) the GISST criteria must reflect the
concerns of stakeholders or decision-makers for specific problems being addressed; 3) a scoring
system must be developed to reflect impacts and vulnerabilities in specific ecoregions and
physiographic regions where scores will be applied; 4) inaccuracies resulting from imprecise
data scales must be addressed;  5) the map classes selected and used in the GISST should be
transparent and well documented.

Use of the GISST in the NEPA Process to Evaluate Environmental Impacts of Project
Alternatives

   The Panel was asked to comment on the usefulness of the GISST for evaluating the
environmental impacts of project alternatives to help inform NEPA-related decision-making.
The Panel finds that the GISST could be used to detect potential environmental  impacts of
project alternatives. However,  to make the GISST useful for evaluating impacts of project
alternatives and for making decisions, EPA must address the following limitations of the tool: 1)
the GISST algorithm cannot identify specific impacts; 2) the GISST scoring system assigns
equal weights to the criteria and may mask environmental impacts; 3) the GISST does not
include  a process for identifying relevant and meaningful criteria, thresholds, and impact levels;
4) spatial dependence of the cells in the GISST should be explicitly considered; 5) vulnerability
and impact criteria can be confused;  6) the GISST cannot map projects simultaneously to
illustrate advantages and disadvantages of alternatives;  7) the sum of the GISST average does not
provide information about the pros and cons of alternatives;  8) the GISST will not be helpful in
designing new alternatives unless criteria and objectives are defined a priori.

-------
GISST User's Manual Enhancement

   The Panel recommends enhancement of the GISST User's Manual.  The GISST User's
Manual provides a useful introduction to the tool, but it does not contain adequate instructions on
how to operate the tool or interpret outputs.  The User's Manual could be enhanced by including
the following additional material:

    •  Background information about how the GISST supports NEPA assessments;
    •  Information explicitly describing the types of decisions that the GISST supports;
    •  Information describing the conceptual model underlying the GISST;
    •  The basis and process for selecting GISST criteria to use in an evaluation;
    •  Suggested approaches for integrating spatial data;
    •  The mathematical boundaries of output parameters and guidance on interpretation of
       results;
    •  Representative applications of the GISST;
    •  Definitions of key terms.

   The Panel also recommends reorganization of the User's Manual and provides editorial
comments.
                                           XI

-------
1.     INTRODUCTION

   This report was prepared by the Science Advisory Board (SAB) Geographic Information
System Screening Tool Review Panel (the "Panel") in response to a request by the EPA Region 6
Compliance Assurance and Enforcement Division to review the Region's Geographic
Information System Screening Tool (GISST).  The GISST is a descriptive geographic
information system tool that has been used to evaluate the potential environmental impacts of
large projects such as the construction of roads, the permitting of water treatment plants, and
confined animal feeding operations.  The GISST has also been used to evaluate the potential
environmental vulnerability of the proposed sites of such projects. Such evaluations have been
used by EPA Region 6 to facilitate decision-making and to prepare environmental impact
statements required under the National Environmental Policy Act (NEPA).

   Users of GISST assess the potential environmental impacts of proposed projects and the
potential environmental vulnerabilities of project sites by visualizing various sets of
geographically referenced data.  These data sets underlie selected vulnerability and impact
"criteria." For example, the GISST developers state that rainfall at a project location is evaluated
as a vulnerability criterion because more rainfall can be associated with more infiltration to
groundwater and runoff to surface water. The density of federally managed lands at a project
location is evaluated as an impact criterion because federally managed lands tend to support a
variety of ecological services and greater project impacts are anticipated in areas with higher
densities of these lands. To facilitate decision-making, a scoring system (with a scale of one to
five) is used in the GISST to evaluate data sets associated with each criterion. A lower score
equals a lower level of potential impact or vulnerability.  The GISST scoring system was
developed by using arithmetic groupings to evaluate natural breaks in the data and applying the
best professional judgment of EPA Region 6 staff.

   In the GISST the degree of potential vulnerability of a watershed subunit, project area, or
other appropriate geographical unit is defined as the average of the vulnerability criteria scores
within the geographic  unit. The degree of potential impact produced by the project is defined as
the average of the impact criteria scores.  A mathematical algorithm is used in the GISST to
derive the "potential for significant environmental risk" associated with a project.  This
algorithm considers the ratio of the cumulative area affected by a project to the total area
evaluated, the degree of potential vulnerability of the area evaluated, and the degree of potential
impact produced by the project.  The results of GISST analyses can be displayed on maps that
can include data overlays generated using different criteria. The Panel sees GISST as an
exploratory tool that can improve understanding of how potential decisions can affect the
systems being considered.  It is the use of GISST and not particular numbers calculated by it that
most contribute to the  decision-making process.  However, the understanding provided by
GISST is only as good and as complete as the underlying data layers.  In the world of
environmental decision-making, incomplete information is common, and decisions are often
made using only available information. In cases where the scientific evidence is incomplete or
contradictory, decisions are often made without scientific input.  Wiens (1996) found such a case
in the aftermath of the Valdez, Alaska oil spill. GISST provides a means to assimilate available
information for rapid assessment. Because this tool is used for preliminary environmental

-------
reconnaissance, it is appropriate to consider alternative weightings of data layers and
interpretations of results. Such analysis of alternatives with GISST should be quite insightful.

   The Panel reviewed the Region 6 GIS Screening Tool (GISST) User's Manual provided by
EPA Region 6. The document contained: 1) an introduction in question and answer format that
described the uses of the GISST; 2) background information on concepts underlying the GISST;
3) information on the development of the GISST algorithm and criteria; 4) case studies
illustrating how the GISST has been applied; 5) the finalized GISST criteria; 6) additional
GISST criteria that  are under development; 7) the geographic information system program used
in the GISST; and 8) a peer review history of the GISST.

   The Panel notes  that various EPA Regions and program offices have developed GIS-based
assessment tools similar to the GISST.  Examples include the EPA Region 4 Southeastern
Ecological Framework, the EPA Region 5 Critical Ecosystem Assessment Model, the EPA
Office of Water's Index of Watershed Indicators, the EPA Office of Research and
Development's Regional Vulnerability Assessment methods, and the NEPAssist web-based
mapping tool developed by EPA's Office of Federal Activities. Despite these initiatives, the
Agency still does not have a unified single accepted approach for using spatially explicit
information for environmental decision-making. The compartmentalized development of GIS-
based tools and data by EPA program offices and regions is suboptimal, given budgetary
constraints and the high value of these tools for environmental decision-making.  The SAB
therefore strongly urges EPA to  undertake an initiative to define a unified framework for the
creation of these types of tools.  The panel also urges that future development of the GISST be
consistent with the principles embodied in EPA's Guidance on the Development, Evaluation,
and Application of Regulatory Environmental Models (U.S. EPA Office of Science Policy,
2003).

2.     CHARGE TO THE REVIEW PANEL

   EPA Region 6 sought comment from the Science Advisory Board on the following issues:  1)
whether the GISST  methodology is reasonable and appropriate for use in conducting initial level
assessments of potential environmental impacts and vulnerability, 2) the strengths and limitations
of the GISST as a tool for use in prioritizing and comparing environmental vulnerabilities and
impacts for decision-making, and 3) steps that can be taken to further develop the GISST User's
manual and documentation. Six detailed charge questions were provided to the Panel.

3.     RESPONSE TO THE CHARGE QUESTIONS

3.1     Question 1.1.  The GISST mathematical algorithm (presented in Chapter 3 of the
       GISST User's Manual) for determining the "potential for significant environmental
       risk" of projects is a multiplicative formula using the watershed as the base unit.
       Please comment on the  reasonableness and appropriateness of using this algorithm
       for conducting screening-level evaluations as described in the GISST User's
       Manual.

       In the Interstate Highway  69 case study, the GISST algorithm was not used because

-------
       it was not beneficial to obtain one cumulative vulnerability score for the entire
       highway corridor.  Instead, vulnerability within the corridor was evaluated by
       summing the scores of vulnerability criteria within 1 km2 areas in a grid system.
       Please comment on the reasonableness and appropriateness of this method for
       conducting an initial screening-level evaluation.

Reasonableness and Appropriateness ofGISSTfor Conducting Screening Level Evaluations

   The GISST algorithm is straightforward, consisting of three multiplicative parts. The first
deals with the overall area affected, the second with potential for vulnerability, and the third with
potential for impact.  The GISST algorithm has the advantage of being a simple tool that allows
the aggregation of layers of information for environmental assessment screening.  In this regard,
geographic information system (GIS) based approaches to screening such as the GISST are very
useful. They should be further developed, and their use should be promoted.  While the GISST
algorithm provides a single score, the score itself should be de-emphasized.  Much of the value
of the approach is in the process itself and the real strength of the tool is in the individual data
layers. Focusing on the algorithm alone can mask some of these details and their value. The
GISST tool should be part of a screening process that clearly defines limits of integrative indices
and promotes the subjective scientific evaluation of the numerical output and information that
supports it. This screening process can include other visual representations of GIS-based data
layers that allow for visual identification of underlying layer information (e.g., "radar plots" that
enable visualization of the criteria values).

   The GISST algorithm provides a reasonable way to conduct an initial evaluation of potential
environmental impacts, but the limitations and caveats associated with the use of this simple
algorithm and discussed below must be considered and addressed. The multiplicative nature of
the algorithm is appropriate since it accords with the Risk = Probability x Consequences
structure of the most commonly used definition of risk. It also makes sense that if either
vulnerability or impacts are nil (zero), then their product should also be nil and no further
attention should be given to assessing the risks of the project. The scaling of the potential  project
impact by project size relative to watershed size is simplistic but reasonable. However, the
specific algorithm used in the GISST needs to be re-evaluated taking advantage of the available
literature and expertise in decision science and statistics, particularly in the area of multi-attribute
decision-making, to address the following concerns raised during this review as well  as other
concerns that may be identified.

   •  Compensatory effect. The algorithm is an example of the "compensatory" class of multi-
      attribute evaluation methods, in that high scores on some dimensions (criteria) can
      compensate for low scores on others.  This relationship arises from the fact that the  overall
      vulnerability of a project alternative is measured as an average of the vulnerability scores
      of individual vulnerability criteria, and likewise for overall impact. One caveat thus
      pertains to this compensatory structure: it could be undesirable that very low scores on
      some dimensions can be countered by high performances on others. Such tradeoffs are
      commonly not acceptable in many biological, ecological, and ethical situations: an
      organism cannot substitute more energy intake (calories) for less water; a greater
      abundance or biomass of a common species may not offset losses of a rare or endangered
      species; more very rich people may not compensate for more dire poverty (cf.,  the

-------
computation of environmental justice scores as the average of three different individual
criterion scores).  This compensatory scheme masks potential differences between project
alternatives, since the same or similar averages can arise from very different constituent
scores.  Unless one scrutinizes all individual scores, one will not necessarily notice such
differences. Such compensatory structures are common for utilitarian welfare functions,
where one attempts to maximize total or average welfare, or utilitarian damage or impact
functions, where the total or average cost (pollution, risk) is minimized.  Compensatory
methods do not address distributional aspects like equity. Users need to be aware of these
characteristics of compensatory, utilitarian structures.

Weighting the criteria.  In the GISST algorithm, the criteria scores are averaged to derive
either the degree of vulnerability (Dv) or the degree of impact (D;), although the authors
point out that sometimes it might be best to sum the criteria in each group.  Three
problems argue against such averaging and/or summing.  First, to be mathematically
legitimate, criterion scores must be measured on an interval scale  (i.e., on the 1-5 scale,
unit differences between scores are all of the same magnitude) or ratio scale (i.e., on a 1-5
scale the scores are stated with respect to a rational zero) if they are to be averaged; if they
are only summed, and not averaged afterward, then scores must be on a ratio scale. It is
not clear that all of the criteria scores meet these requirements. Many of the criteria scores
appear to have ordinal properties (i.e., on the 1-5 scale the scores are  only indicative of
more or less of a characteristic), and it is not mathematically legitimate to sum or average
ordinal  scores. Second, the average of a number of different scores fails to reveal the
range of those scores. Extreme values (e.g., those that fall above or below some important
threshold) will thus be concealed, and the power of the tool to illuminate ("flag")
potentially problematic effects will be diminished.  Third, summing scores on different
dimensions, even if they have the same units (e.g., kg of phosphorus and kg of nitrogen),
implies a value tradeoff between any two scores and can mask critical impacts. Since this
tradeoff is equal to the negative  of the ratio between their coefficients, even "intentionally
unweighted" scores (i.e. with  coefficients  equal to 1.0) are in fact implicitly weighted
equally (changing units, for example from kilograms to tons, reveals this immediately).
Hence, weighting and tradeoffs are unavoidable in the additive value (impact) function.
The tradeoffs can mask critical impacts because, in the summation, desirable effects
(whether many small ones, or a few very large ones) can compensate undesirable ones
(whether small or large), since it is only the sum that matters.

The implication of this unavoidable weighting and potential masking of impacts is that
summed criterion scores should  not be used as the only signal of potential adverse impacts
unless the weights actually represent clients' (decision-makers', stakeholders' and other
relevant actors') desirable tradeoffs between different criterion scores. Given the
difficulty of determining such tradeoffs (weights) for public problems, and the likelihood
that a linear (weighted-sum) function would represent them, summed scores should only
be used in an exploratory vein: extreme sums can indicate that there might be a problem,
but only examination of the individual scores will confirm this and reveal where and on
what dimensions. Nevertheless, any analysis that moves from separate measures to
summary measures must contain normative judgments, and if the analysis is used in an
exploratory way, it must be easy for users to employ different judgments. It will therefore

-------
   be useful if the GISST can be modified to allow assignments of different weights to
   criteria scores since a differently weighted function might reveal potential problem areas
   otherwise hidden.  Nonetheless, it is crucial that the user be aware of "false negatives":
   because of the compensatory nature of the summation, absence of extreme sums does not
   imply absence of significant adverse impacts.

•  Measurement of risk. The conceptual nature of GISST algorithm core, vulnerability x
   impacts, needs refinement if it is to measure something akin to risk (or as stated in the
   GISST User's Manual, "the potential for significant risk").  The uncertainty reflected in
   the standard definition of risk is represented by a probability term.  In the GISST,
   uncertainty is apparently represented by vulnerability, and impacts are synonymous with
   consequences. However vulnerability is not precisely defined in the GISST User's
   Manual.  A number of concepts and operational definitions of vulnerability are found in
   the literature, but few would equate vulnerability with probability, likelihood, or some
   other notion of uncertainty, and unless vulnerability is measured on a scale of 0-1 it would
   not reflect probability. Thus,  some measure of uncertainty of consequences would have to
   be added to the GISST algorithm if it were to be used to determine the potential for risk.
   Such a modification would not be necessary if EPA continues to use the GISST as a tool
   to identify the potential for impacts.

•  Choice and number of criteria. Another problem associated with the GISST algorithm lies
   in the choice and the number of criteria to be used in the analysis. Users can select and
   apply criteria deemed appropriate for their particular project.  This adds specificity and
   provides flexibility.  However, it also provides a mechanism by which the impact of
   important or "problem criteria" can be masked by relatively benign ones.  For example, if
   5 criteria are chosen, and their scores are 1, 1,  1, 1, and 5, the averaged score is 1.8. If
   only two of these criteria are chosen, their scores could be 1 and 5 resulting in an average
   score of 3. The GISST authors are aware of this important issue, and point out in the
   GISST User's Manual that it is also important to examine the scores of each criterion to
   look for potential hot spots. This in essence is the weighting procedure discussed above.
   It is also important to always keep in mind the number of criteria used to determine Dv
   relative to D;, so that each of those groupings have equal weight, if that is desired.

•  Alternate  algorithms. In order to conduct sensitivity analyses and further explore uses of
   the GISST for rapid environmental assessment and reconnaissance, EPA should develop
   some optional models of aggregating and weighting the criteria, not just the one shown in
   the GISST User's Manual.  EPA Region 6 has not used the algorithm in the User's
   Manual for the major test case of the Interstate Highway 69 Corridor.  It is therefore not
   clear why this algorithm is published as the only available choice.  Other possible
   integrative techniques should  be provided in the GISST User's Manual.

•  Modifiable Area Effect. The GISST criteria can be applied in a number of different ways
   depending upon what "appropriate geographic unit" (AGU) is selected for an evaluation.
   The selected unit can be a physiographic unit watershed, a U.S. Geological Service
   hydrologic cataloging unit (HUC), or a square kilometer, and the results will vary
   depending on that  selection. If the AGU is very large relative to the affected area, then the

-------
      GISST score will be small regardless of what the Dv and D; scores are because the
      percentage of area affected relative to AGU will be very small.  While the GISST is a
      screening tool and there is no "safe" score, small scores can still leave the impression that
      there are few problems associated with a project, when in fact there might be significant
      ones.

   •   Landscape dynamics.  System or landscape-level dynamics are not addressed in the
      GISST algorithm and should be considered.  For example, the beginning of a highway
      construction project may cause a big disruption of urban lives and affect wildlife but
      things may improve after a period of time (e.g., in the Florida Everglades, black panthers
      now use "underpass ramps" which connect the landscape made  disjoint by the highway).
      Conversely, nitrate pollution from confined animal  feeding operation (CAFO) systems
      may not show up for decades in groundwater. EPA should be cognizant that the problem
      should define the tool, not vice versa.

   •   Quality of data. The presentation of GISST analyses as numerical output should include
      the underlying "metadata" as part of the overall report in order to provide the user or
      reviewer information on: data source, dates, extent, quality, completeness, processing
      steps, responsibility, and contact information. In addition, sensitivity analyses and
      "goodness-of-fit" measures could be provided to describe how well the index and its
      components can address the questions asked. It is apparent that some GISST analyses
      clearly provide answers to questions, but others could result in quite different answers
      depending on the judgment of the user. The Panel emphasizes that numeric indices can be
      very useful but they should not be used in a vacuum to make  decisions.

   •   Math errors in the algorithm should be corrected. For example, the expressions describing
      D; on page 21 and in Appendices A and B of the GISST User's Manual mean different
      things. It is also not clear in the  algorithm what is being added with the summation sign
      (the Panel questions, for example, whether Aw is  fixed or a variable).

Interstate Highway 69 Case

   The Panel considered the reasonableness and appropriateness of the method used by EPA
Region 6 to evaluate vulnerability in the Interstate Highway 69 case. The published GISST
algorithm was not used in this case because a cumulative vulnerability score for the entire
highway corridor was not useful. Instead, vulnerability within the corridor was evaluated by
summing vulnerability scores within 1 km2 areas.  The Panel notes that application of the GISST
to evaluate the proposed highway system significantly limited the size of the appropriate
geographic unit in the analysis.  The comparison of possible highway routes, rather than
determining a cumulative score for the  highway corridor, was most valuable to planners, and
hence the approach used by Region 6 is reasonable. Adding rather than multiplying indices (as
EPA has done using the vulnerability criteria scores in the Interstate Highway 69 case) is
acceptable if the indices are averaged, and as long as they represent  only impacts or only
vulnerabilities.  The Panel recommends that the approach  used in the Interstate Highway 69 case
be refined and provided to GISST users as an optional algorithm. However, as discussed in the
response to charge question 2.3 below,  the GISST should  only be used to conduct initial

-------
evaluations of potential environmental impacts.  Summary criteria in the GISST should not be
used to make project decisions.

3.2    Question 1.2. Appendix A of the GISST User's Manual identifies the impact and
       vulnerability criteria that are used in the GISST to evaluate environmental impact
       and vulnerability. A subset of these criteria3 is frequently used by EPA Region 6 to
       conduct GISST evaluations. Are the criteria in this subset reasonable and
       appropriate for use in GISST evaluations of the potential degree of vulnerability of
       a project area and the potential degree of impact produced by a proposed project?
       Please provide similar comments for the other criteria in Appendix A. Are there
       additional categories of criteria that should be developed for use in GISST
       evaluations?

   The Panel finds that, given limited alternatives and available resources, EPA Region 6 has
developed the GISST thoughtfully and implemented it with good balance. Comments
concerning the use of the criteria and recommendations to improve the criteria are provided
below. The Panel emphasizes that these comments do not imply that the GISST is inadequate
for use in initial evaluations to detect potential environmental impacts.  However, with the
information currently available in the GISST User's Manual, it is difficult to answer the question
of whether the criteria are "reasonable and appropriate."  The Panel notes that several elements
make GISST evaluations different from other GIS evaluations. Unique GISST elements include:
1) the criteria scoring process, 2) criteria subset selection, and 3) the process that highlights
important drivers of concern for further analysis. These elements make the GISST an objective,
spatially  explicit tool for conducting initial, broad-stroke evaluations in a timely fashion.  The
Panel notes, however, that better definition and description of these processes is needed in the
GISST documentation.

   Many criteria are described in the long appendices of the GISST User's Manual.  Each
criterion  is described in a one-page summary. The Panel understands that including detailed
descriptions of the criteria in the User's Manual would make the document very long. However,
the User's Manual currently contains insufficient information to provide a complete
understanding of the basis for screening-level actions or decisions. The Panel suggests that more
detailed explanations of the criteria and examples illustrating their use could be placed on a
GISST website. Interested persons could obtain information from the website to gain a better
understanding of a decision or discussion that was based on a GISST evaluation. As additional
information for such a website is developed it will be important to keep the perspective of vested
or interested parties in mind.

   As the Panel deliberated on the charge questions it became apparent that decision analytic and
statistical support for development of the GISST criteria and scoring system had been
3 The subset of criteria most frequently used in GISST evaluations includes: Stream Density (surface water
quantity), Population Density, Minority (environmental justice), Economic (environmental justice), Agricultural
Lands, Density of Managed Lands, Hazardous Waste (Other Industries or Pollution Sources), Impaired Stream
Segments (Clean Water Act 303(d) Segments), Wetlands, Floodplain, Ozone Nonattainment, Texas Ecological
Assessment Protocol (TEAP) Diversity, TEAP Rarity, TEAP Sustainability, TEAP Composite, Wildlife Habitat,
Federally-listed Species, and State-listed Species, and Ecologically Significant Stream Segments.  The TEAP
criteria were derived using a tool developed by EPA Region 5, the Critical Ecosystem Assessment Model
(CREAM). The SAB has reviewed the CREAM.  The SAB report on the CREAM is available at:
http://www.epa.gov/sab/pdf/cream_sab-05-011 .pdf
                                            7

-------
unavailable to EPA staff. The Panel strongly recommends that decision analytic and statistical
support be provided to the GISST developers to revise and improve the criteria. Many questions
and issues raised by the Panel could be remedied through discussions with statisticians. For
example, a statistical examination of the data layers underlying the criteria could identify
redundant measures and increase the soundness of rating scales.

Appropriateness of "most frequently used" GISST Criteria

   The  19 "most frequently used" GISST criteria listed in the footnote to charge question 1.2
were used in the Interstate Highway 69 corridor evaluation summarized in the GISST User's
Manual. These criteria appear to have been selected on the basis of some requirements for the
federal highway impact assessment process and seem to be appropriate for use in this case, but
the GISST documentation does not describe how or why these criteria were chosen for the case.
There is an extremely heavy emphasis on vulnerability scores within this group of criteria and it
is not clear how criteria that appear to cross the vulnerability and impact categories are used. Of
the 19 frequently used criteria, 12 are indices of vulnerability, 3 are indices of impact, and 4 are
listed as indices of both vulnerability and impact. Sorting these criteria into GISST groups that
define the type of impact or vulnerability shows the following distribution among groups: 11 are
ecology, 3 are water quality, 3 are socioeconomic, 1  is toxicity, and 1 is air quality.

   The Panel notes that different subsets of criteria will probably be selected for use based on the
context of the screening assessment to be conducted.  The Panel recommends that, instead of
focusing on the appropriateness of a particular group of criteria, it would be better to define a
general process whereby groups of criteria could be selected for use in various kinds of
evaluations. Different impact templates  could be created and used as guides for selecting criteria
to conduct different screening activities.  Templates might be developed to describe the impacts
from energy facilities, transportation projects and CAFOs.  Each of these general types of
projects would share a set of predictable impacts and the key criteria relevant to evaluation of
those impacts could be identified.

   Overall, the Panel finds that the criteria need to be better described in the GISST User's
Manual. Information describing the criteria databases, references, definitions, limitations, and
uncertainties is not clearly presented in a form that can be easily understood by general users.
An exception is the Texas Ecological Assessment Protocol (TEAP) criteria group. Within the
TEAP group, justification and definition of how the criteria are used is provided.  Criteria
definitions are especially important for GISST users who may be asked to define new specific
groupings of criteria for future analyses.  In the appendices of the User's Manual, the criteria are
broken into groups that are defined according to types of impacts (e.g., water quality, toxicity).
Criteria in these groups are then broken into vulnerability and impact categories.  The criteria are
presented in a somewhat arbitrary order within the impact groups.

Appropriateness of other GISST Criteria

   In addition to the 19 frequently used GISST criteria, 50 other criteria are described in the
GISST User's Manual criteria appendices. The Panel finds that these other criteria appear to be
reasonable for use in some cases.  However, the context of an assessment will drive a
determination of whether the selected criteria are reasonable and appropriate. Without a better
rationale for use of the criteria, and a more detailed description of how they may be used, it is

-------
possible to provide only an intuitive response concerning the appropriateness of the criteria.
Specific comments on some of the individual criteria are provided in Appendix A of this report.

   The EPA Region 6 presentations to the Panel have illustrated an important feature of the
GISST that highlights the usefulness of the criteria. This feature is the power of the GISST for
use in conducting "scoping" studies to identify potential issues of concern.  Such large scoping
studies requiring use of all the criteria are likely to be very informative to applicants seeking
approval of projects.

 Comments on the  Vulnerability and Impact Criteria Categories in the GISST

   The Panel finds that the term "criterion" used in the GISST documentation might not be ideal
because it often implies a threshold. The Panel suggests that "attribute" or "factor" would be
better terms to use in this context.  Also, the uses of the terms "vulnerability" and "impact" in the
GISST documentation are questionable.  As discussed below, different kinds of vulnerability and
impacts are considered in the GISST (e.g., environmental justice and ecology).  The Panel
suggests that it would make sense to separately track broad categories of criteria (e.g., water
quality, socioeconomic) with clearly stated objectives for conditions that are desirable. The
Panel recognizes that it is important to balance competing thresholds in a NEPA assessment, but
those thresholds need to be set out separately in order to make the best judgments.

   A key issue to be addressed is the way groups of criteria are utilized in the GISST. The basic
GISST algorithm is heavily dependent on the potential area affected by a project.  As noted
above, the number of criteria lumped into vulnerability or impact areas can affect the ability  of
important criteria to drive the assessment outcome.  For example, the CAFO case study in the
GISST User's Manual is driven by the areal extent of potential impact and not by vulnerability
scores or impact scores.  In addition, the way composite GISST scores are broken into categories
seems to be arbitrary or not well described (e.g., the CAFO case study presents  information on
the composite score in a non-linear fashion with no justification or rationale). Many of the same
comments can be made relative to the Interstate 69 Highway study scaling issues, composite
scores, and the use of composite or group data.

   Evaluating the degree of vulnerability is compromised to some extent in the GISST because
the criteria aim to identify fundamentally different types of vulnerability in a single measure. In
a very general sense, there seem to be three kinds of vulnerable systems with which EPA is
concerned:  1) stressed, heavily disturbed systems that have already sustained relatively high
cumulative impacts and have limited (or no) capacity to absorb additional stressors ("type 1
vulnerability")4, 2) relatively unaltered or "pristine" areas that may be regionally important
because they represent the few remaining systems or places with high ecological value ("type 2
4 The following GISSTcriteria are examples of measures that primarily reflect "type 1" vulnerability because higher
scores are associated with more contaminated, altered, fragmented, and disturbed environments (note that for these
criteria, "type 2" vulnerable areas would receive low Dv scores): water releases, ground water quality,
channelization, agricultural lands, habitat fragmentation, road density, population density, total population, other
industries, pollution sources, or protected lands, hazardous waste, impaired waters, and ozone nonattainment areas.

-------
vulnerability")5, and 3) areas that have unique characteristics (hydrologic, human, etc.) that
would make these areas or population in them particularly vulnerable to contamination ("type 3
vulnerability")6.  The first two types of vulnerability are especially problematic because they are
on the opposite ends of the spectrum in the GISST scoring process.  For example, the
vulnerability measures for an agricultural landscape with fragmented habitat, high population
density, and many roads would be higher for many of the criteria than for an ecologically
important forested landscape with large blocks of forest-interior habitat, little anthropogenic
disturbance, and high water quality.  Several alternative approaches might be considered to
address this problem.  One alternative is to calculate three different vulnerability indices.
Another alternative is to separately consider the ecological value of an area (i.e., what is present
that we want to protect?), vulnerability to negative impacts from project (i.e., what is the
likelihood that the system will be impacted by the project?), and magnitude of change in system
(i.e., how much and in what way is the system likely to be altered?).

General Comments on the GISST Criteria

  The Panel provides the following general comments on the criteria.  Additional comments on
specific GISST criteria are included in Appendix A of this report.

  There is a need to better delineate how to use the groups of criteria.  It is presently unclear
how one decides to include or exclude particular criteria in the  analysis. The GISST
documentation should include more quantitative analyses to support the basis for choosing
criteria and their weights.  Substantial information concerning selection of indicators is available
in the literature (e.g., Cairns  et al, 1993; U.S. EPA Science Advisory Board, 2002; National
Research Council, 2000). Cairns et al. (1993) have reviewed how different types of indicators
are used in decision-making.  This paper may be particularly relevant when considering how
socioeconomic criteria should be used in the GISST.  As discussed below, the  criteria weights
and scoring system could be further developed through statistical analysis of information
provided through an expert opinion process.

    •  The individual GISST  criteria each make intuitive sense but, as noted above, they can
       have competing purposes and counteract one another. They can also have different
       thresholds (i.e., different values) and may be combined  in illogical ways (e.g., combining
       noise and odor criteria scores with scores for use of energy efficient appliances).  There is
       a need to define criteria categories in a consistent manner whereby combinations or
       groupings make sense in an overall strategy to achieve assessment objectives (e.g.,
       assessing current status, different kinds of vulnerabilities, future impacts).
5 The following GISST criteria are examples of measures that reflect "type 2" vulnerability because higher scores
are assigned to environments with greater ecological "value" based on measures of diversity, rarity, and
sustainability (note that for these criteria, type 1 vulnerable areas would receive low Dv scores): wildlife habitat,
TEAP diversity, TEAP rarity, TEAP sustainability, and TEAP composite.

6 The following GISST criteria reflect type 3 vulnerability because higher scores are assigned to areas with unique
or defining features likely to make the area or human population residing in the area particularly vulnerable to
negative impacts from the project: wetlands, floodplains, stream density, surface water, surface water quantity,
distance to surface water, ground water probability, average stream flow, sole source aquifer, aquifer/geology rating
individual well water, septic tank and cesspool use, soil permeability, high school education, colonias, educational
achievement ranking, economic, minority, age, children, older population, pregnancy, houses lacking plumbing,
ability to speak English, foreign born, telephone communication, and linguistic isolation.
                                              10

-------
   •   As EPA continues to develop the GISST criteria, the Agency should consider
       incorporating measures of runoff potential and nonpoint source pollution.

   •   A fundamental concern of the Panel is the relevance of the GISST criteria to specific
       projects. EPA has stated that some of the criteria were incorporated into the GISST
       because data were readily available, not necessarily because the criteria were relevant to a
       problem or management objective. As suggested above, it might be possible to improve
       the GISST by developing alternative "impact templates" that identify the key criteria
       relevant to particular types of projects.  Impact templates may reflect different models
       that are applicable to specific types of projects. Such models might be developed by
       conducting expert assessments of impacts and vulnerabilities for various types of projects
       and by using statistical methods to evaluate the explanatory variables (i.e., criteria).

   •   Although a large number of spatial overlays representing different criteria can be viewed
       in the GISST, it is important that analysts use only the most  significant layers in an
       assessment. The significance of truly critical factors can be  diminished if so many are
       equally considered, and it may become difficult to determine how a particular result was
       obtained. In its application of the GISST, EPA should keep  in mind the Principle of
       Parsimony and not increase, beyond what is necessary, the number of entities required to
       explain anything.  EPA should select and use criteria that are significant and critical to
       decisions and processes.  The selection of criteria that are "significant" can be validated
       by conducting a statistical analysis of data on the outcome variable of interest (i.e.,
       vulnerability or impact) for a random sample of plots for which the explanatory GIS data
       and values of the outcome variable are available.

   •   As discussed above, the Panel is also concerned that EPA has not clearly described
       differences between criteria specified to determine the Dv and/or the D;.  Some GISST
       criteria reflect both Dv and D; and other criteria reflect only  one of these indices. There
       are several cases where it is unclear whether impact or vulnerability is the real focus of a
       particular criterion.  For example, the toxicity-weighted water releases criterion on page
       A-62 of the GISST User's Manual is identified as an impact criterion, but it is not clear
       why it is not also a vulnerability criterion.

   •   As stated above, the Panel finds that the GISST User's Manual, particularly Appendix A,
       would greatly benefit from additional detail regarding the rationale for, definition of, and
       assumptions underlying specific criteria.  Definitions are needed in some of the
       descriptions of individual criteria in the User's Manual. For example, the criteria
       descriptions of ecologically-significant stream segments and density of managed land
       make references to conditions within an "area" but fail to specify the area.  In some cases,
       the rationale for scoring is not obvious to the reader/user.

3.3    Question 1.3. The GISST uses data sets (in Appendix A)  with different coverages
       generated for different purposes (e.g., point sampling of water quality, census data,
       and land cover data gathered by satellite). Is the GISST  1-5 scoring scale on these
       coverages and datasets reasonable for developing an initial assessment of the
       potential cumulative impacts of proposed projects?

                                           11

-------
   The Panel discussed the use of different criteria and coverages in the GISST 1-5 scoring
system. An advantage of the GISST 1-5 scoring system is that it allows (or forces) the
evaluators to simplify the combination of diverse criteria and show sharp contrasts in the
assessment of potential impacts or comparisons among project alternatives. A limitation of the
scoring system is that assessors lose the ability to see what specific information the underlying
data provide for a particular assessment. In EPA's description of the scoring system, it is not
clearly stated that some criteria scores are based on graded responses (on a true 1-5 scale), some
criteria scores are based on intermediate scales (1,3,5), and some are based on binary scales (1,
5).  The type of scale used to score each of the criteria is not clearly defined so it is not possible
to evaluate the impact of the scoring system on the final outcome of an assessment.

   The Panel questions whether the GISST criteria can be generically scored without considering
the type of project being evaluated. For instance, the influence of rainfall or stream density on
vulnerability may be quite different in a coal-fired electrical generation plant evaluation and a
swine CAFO project evaluation.  As discussed above, EPA may need to develop a set of scoring
templates that incorporate criteria associated with the general types of projects that are evaluated.
Each template would better incorporate scientific understanding of the processes that the criteria
represent for a specific type of project.  The Panel has identified the following specific
limitations and concerns regarding the GISST 1-5 scoring system.

Scoring system limitations and issues of concern

    •   As stated above, the Panel is concerned that some of the criteria in the GISST User's
       Manual may not be relevant to specific projects. Similarly, it is not clear that the 1-5
       scores assigned to criteria are always applicable for any project evaluated.  The
       appropriateness of some of these criteria scores may be "project-dependent."

    •   The Panel questions whether all of the criteria should always be scored with a maximum
       value of 5.  It seems that a score of 5 assigned to any criterion should "raise a red flag"
       and be indicative of a problem.  A score of 5 may not be appropriate for criteria that are
       less important in an evaluation.  This is not to suggest that the potential range change,
       rather it means that the value of '5' may not always be observed.

    •   As stated above, the 1-5 GISST  criteria scores are summed and averaged.  Averaging
       helps bring the scores into a common range when different numbers of criteria are used,
       but it also ultimately reduces the dynamic range of the results.  In some sense this reduces
       the level of information, but the values could be rescaled to accommodate this concern. It
       should be noted, however, that changing scales may add uncertainty to the analysis. It is
       helpful to preserve information about the distribution of component values that make up
       the averages. The  GISST  1-5 scores appear to serve the purpose of numerical bins rather
       than simple rank-order ordinal measurements.  The only permissible arithmetic operation
       using ordinal values is transitive relationships.  Addition,  subtraction, multiplication, and
       division using ordinal numbers do not produce valid or reliable results.  Consequently,
       use of ordinal values in the scoring process is inherently invalid. The interval scale
       scoring method is reasonable for use in scoping level assessments to identify
       environmental concerns that are likely to require detailed analyses in subsequent
                                            12

-------
   environmental impact assessments.  For this use, the scores are reasonable for detecting
   potential individual concerns where no mathematical operations are required for
   aggregating scores. Scoring methods similar to the one used in the GISST have been in
   widespread use in suitability analyses since the  1960's (McHarg, 1969) for quantifying
   the relative desirability (or vulnerability) of sites for project alternatives.  However, such
   an application of the GISST scoring system would require the use of a ratio measurement
   scale rather than an interval measurement scale.

•  Most of the scaling of GISST criteria is based on physical values with true ratio
   quantities. Therefore, binning these values to the GISST  1-5 scale may be acceptable.
   However, the values seem to have been binned or scaled using different techniques or
   functions: linear, interval, stepped, logged, and natural breaks.  An assumption in the
   GISST is that the scores of different criteria are equivalent, but this may not be  the case.
   If the values  are scaled using different techniques or functions, it is invalid to equate the
   scales of different summed values. There is no easy way to address this problem unless it
   can be demonstrated that the values of scaled variables are equivalent (e.g., a value of 2
   for one scaled variable is equivalent to  a value of 2 for another).

•  There is probably a large difference in the scales of spatial data layers used in the GISST.
   The combination of spatial data that are highly variable in scale may create a mixture of
   incompatible spatial frequencies since phenomena are scale-dependent.

•  A scoring system with a scale of 0-5 might be more appropriate than the 1-5 scoring
   system because some features represented by the scores may be associated with no
   vulnerability or impact. These features should receive a score of zero.

•  Scoring the 19 criteria listed in charge question 1.2 on the GISST 1-5 scale provides the
   potential for  a cumulative score of 95.  Any one criterion is therefore insignificant
   because it can only influence the total score by 5 points.  As noted above, it is important
   to identify the criteria that are most critical for an evaluation and make sure they do not
   "get lost" in the scoring process.  Various methods for doing this, such as weighting, have
   been discussed.

•  The weighting of GISST criteria has been discussed above.  It is the use of GISST, and
   not particular numbers calculated, that most contributes to the decision-making process.
   Therefore, the weighting of individual criteria must be performed correctly.  Typically
   this is accomplished by developing specific weighting factors (e.g., multiplying by 1, 2,
   3, etc.).  There are inherent problems associated with this  approach, and they must  be
   considered. For instance, in the GISST vulnerability associated with a particular criterion
   is scored on a scale of 1-5 where 1 represents the lowest vulnerability and 5 the highest.
   If scores on this  scale are multiplied by a weight of 2 the bad score of 5 becomes two
   times worse,  but the better score of 1 also becomes worse by a factor of two (since higher
   values represent more vulnerability). This is problematic  because assigning a weight of 2
   to the GISST score of 1 should make it two times better. Normalizing the weights  on a
   scale of 0-1 will alleviate this problem.
                                        13

-------
   •   The GISST does not account for the possible interaction of data layers used in the tool.
       Some layers may be correlated (e.g., soils and geology) or have compound or inverse
       relationships (e.g., wetlands are typically not found at high elevations).  The decision of
       whether to include a data layer in an evaluation process should be based on the
       environmental context of the data layer and its statistical relationship to other data layers.
       Furthermore, without differential weighting the GISST approach runs the risk of
       improper double counting of information. Methods for constructing relational trees
       (diagrams) may be helpful in evaluating the interaction of data layers.

   •   When landscape features are rescaled or aggregated, the Modifiable Areal Unit Problem
       (Openshaw,  1984) is introduced.  EPA should be aware that values representing variation
       in an attribute can be significantly modified by cell size, resolution, and aggregation
       based on boundaries, categorization, or grouping. The following potential  effects should
       be considered and discussed when the GISST data are grouped or categorized in areal
       units to assign scores to the criteria:

              Scale effect (raster resolution). This effect is a variation in numerical results
              caused by grouping data in different zone sizes.
              Aggregation or zonation (raster resampling).  The effect is a variation in
              numerical results caused by grouping zones, methods used, or differences in zonal
              boundaries.

Suggestions to improve the scoring system

   The Panel provides the following suggestions to improve the GISST scoring system.

   •   As suggested above, rather than combining the scores of the all GISST criteria, it may be
       more informative to group the scores  on the basis of similar characteristics such as air
       quality or water quality.  The six broad categories of ecological indicators identified in
       the SAB Framework for Assessing and Reporting on Ecological Condition (U.S. EPA
       Science Advisory Board, 2002) could be separately and independently scored (i.e.,
       landscape, biotic, ecological processes, chemical & physical, and hydrology &
       geomorphology). Perhaps the vulnerability (Dv) and impact (D;)  criteria groups could be
       scored independently and separately,  not summed or averaged with each other. This
       approach would give the decision-maker more information and multiple scores instead of
       a single score based on different criteria.  It may also be  useful to highlight the scores of
       "critical" criteria.

   •   The process  of summing the scores of the GISST criteria is similar to counting the
       number of criteria that have been assigned various scores.  Rather than using summed
       scores, it may be more informative to combine the criteria using such score counts and
       generate maps showing the number of features that have been assigned various scores.
       This approach preserves information about the distributions of scores across the different
       criteria that make up a group.

   •   Critical features could be mapped using criteria functions with Boolean operators.  In this
                                            14

-------
       sense, the user defines values that identify significant vulnerability or impact for each of
       the criteria and they are combined on this basis as a cumulative map.

    •   Where no data are available, GISST assigns criteria values differently among the criteria.
       The "no-data" areas should be treated consistently, perhaps flagging such sites as
       "potentially vulnerable".

    •   With respect to the three vulnerability types mentioned previously, alternate scoring
       systems could be applied to the same criterion. For example, the scoring system for the
       criterion "habitat fragmentation" currently assigns higher scores to more fragmented
       landscapes. This is appropriate for identifying systems that have already sustained high
       levels of impact, but the criterion could also identify areas with high ecological value
       (type 2 vulnerability) if the scoring system were reversed (i.e., areas that are not
       fragmented receive high scores).

Statistical issues concerning the GISST scoring system

   Several statistical issues emerge relative to the use of the GISST criteria and scoring  system.

    •   Many of the criteria seem to overlap and the potentially important consequences of this
       are unclear.  A formal analysis of these consequences is needed. Dv and D; are derived
       by multiplying or summing the scores of "mixtures" of criteria that are often selected
       based on the availability of underlying data. The criteria are quite variable in the content
       of underlying data, validity, and utility. Some of the criteria are quite specific, and some
       are quite broad.

   •  The combining of criteria scores may violate assumptions of independence of the data or
      be influenced by data correlations that should be investigated.

   •  The criteria scoring system could be significantly improved by involving more statistical,
      expert elicitation, and experimental design expertise.  Natural breaks in the information
      were used to develop the GISST scoring system based on a software decision algorithm.
      That approach might not be as effective or desirable as using an expert opinion-based
      ranking system.  This is because an expert opinion process could potentially provide more
      insight and meaningful measures of uncertainty. The Panel  notes that it is important to
      carefully consider how such expert opinion would be gathered. EPA may want to
      consider asking each expert to independently develop judgments  and scores.  A central
      tendency could then be used instead of gathering experts together to obtain a collective
      assessment.  If independent expert opinions were collected,  some measure of the
      uncertainty in the scores could be generated and potentially used  to discard or retain
      various criteria, or to stimulate further attention to areas with inadequate information.
      Experts can develop judgments about the impact or vulnerability  associated with actual
      plots of land in a random sample of plots for which the GIS  variables are also available.
      These independent ratings could be used as dependent variables to calibrate statistically
      derived criteria weights. The Panel recommends that EPA Region 6 staff explore the use
      of expert opinion elicitation methods, pattern recognition techniques, and visualization
                                            15

-------
      tools to revise the criteria scoring process.

   •   It became apparent during EPA's presentations to the Panel that the Agency had used
      implicit rules and informal criteria weightings in their GISST analyses. However, the
      rules for decisions involving appropriate "show stoppers" or moving the decision process
      out of a straightforward GISST scoring system were unclear. The Panel notes that the
      lack of clear, formal rules or criterion weightings could produce stakeholder or EPA
      confusion in the decision process. The Panel therefore recommends that explicit decision
      rules and criteria weightings be clearly defined.

3.4    Question 2.1. EPA intends to use the GISST in the NEPA process as an initial
       screening tool to prioritize potential single, direct, and cumulative environmental
       impacts of projects for more detailed analyses. Please comment on the strengths
       and limitations of the GISST as it applies to this purpose.

   The Panel finds that, with the following caveats and recommendations, the EPA Region 6
Geographic Information System and some of the GISST criteria can be used in the National
Environmental Protection Act (NEPA) process to assist the Agency in reviewing and scoping
environmental assessments (EAs) and environmental impact statements (EISs).

   •   The GISST should be used solely as an initial environmental impact identification tool.
       In this regard, the GISST should only be used as a tool to detect potential impacts
       commonly associated with certain types of projects.

   •   Sets of core criteria should be identified (from the list of frequently used criteria provided
       in charge question 1.2) for use in evaluating certain industrial activities or projects. For
       example, it is likely that sets of core criteria could be identified to evaluate projects such
       as CAFOs,  highway construction, coal-fired power plants, and other kinds of projects.
       These industry-specific lists should be made publicly available.

   •   Decisions based upon single, combined criteria scores should be avoided for all GISST
       applications. Summing the scores of specific groups or categories of criteria may be
       appropriate (e.g., toxicity, water, environmental justice/socioeconomics), but scores
       should not be summed across these categories.  For example, it would be inappropriate to
       combine "Noise and Odor Thresholds" with "Energy Efficient Appliances."  Depending
       on the project, water quality impacts are likely to be on a different scale of importance
       from odor/noise or energy efficient appliances. These groups of criteria scores should not
       be combined, or at least not without establishing some weighting system (other than the
       system of equal weights that is currently applied). The GISST  summary algorithm
       should never be used in and of itself to make a screening decision.

   •   Cumulative scores,  if used, should be very clearly defined and communicated to NEPA
       applicants.  The GISST should be made available to the public  so that applicants  can
       evaluate their data in advance of permitting activities. EPA's NEP Assist program may be
       the place to develop this option.
                                           16

-------
•  While the GISST may be useful for a preliminary evaluation of environmental impacts, it
   is not adequate for prioritizing impacts unless the following elements can be changed or
   addressed:

   1.  Weighting of the criteria.  The GISST must in some way reflect how the scores
       represent the relative importance of each impact (i.e., the values of the decision-
       maker).  Since the criteria, and hence the criterion scores, are equally weighted (each
       is multiplied by 1) it is all but assured that such equal weights do not reflect the
       relative importance that decision-makers or stakeholders would accord them, and they
       do not reflect society's values.  Since prioritization reflects the values of an individual
       or group of individuals, without deliberate weighting of the criteria the GISST should
       not be used for prioritization. It is impossible to make decisions about "impacts"
       without linking them to goals and objectives

   2.  Stakeholder concerns.  GISST criteria must reflect the concerns of stakeholders or
       decision-makers for the specific problem being addressed. The 19 frequently used
       criteria in the GISST do not necessarily reflect the problems or projects to which they
       will be applied and the concerns (values) that people may have about these projects.
       Some of the 19 criteria may not be relevant at all.  Since data sets are selected for an
       evaluation before the problem (objectives and concerns) is defined, priorities
       identified would be highly suspect.

   3.  Eco/physiographic regions. A scoring system must be developed to reflect the
       impacts and vulnerability  in specific homogeneous ecoregions and physiographic
       regions where the scores will be applied. The scoring system  cannot be applied
       universally across EPA Region 6.  For example, the scale developed for ranking
       stream density in the arid  Texas panhandle or Oklahoma would not be appropriate for
       use in humid southeastern Texas or Louisiana (Land Evaluation, 1968).  Prioritizing
       impacts on the basis of those  value scores would likely be inaccurate and misleading.
       In this regard a classification system different from the hydrologic units (HUCs)
       could be developed or adapted. An approach that might be considered for some uses
       is the ecoregion system (Omernik, 1995).

   4.  Inaccurate representation  of cumulative impacts.  Cumulative  impacts are represented
       in the GISST by averaged criterion scores (for vulnerability and impacts separately),
       summed criterion scores (for  the Interstate Highway 69 case),  or the scaled product of
       average criterion scores (i.e.,  the total "CRIA" score yielded by the GISST formula).
       The compensatory trade-off nature of these computations, the  lack of weighting, and
       the use of inadmissible operations on ordinal or interval level  scores (discussed
       above) will mean that cumulative impacts will be inaccurately measured and
       priorities may be misguided.

   5.  Inaccuracies of modeled impacts arising from the use of imprecise data scales. EPA
       should be aware of inaccuracies that can result from using imprecise scales of data.
       The scale of resolution of the information and analysis should be appropriate to the
       scale at which the decision is being considered.
                                        17

-------
       6.  Selection and documentation of map classes. Because different map classes will
          show different patterns and hence indicate different impact priority areas, map classes
          selected should be transparent and well documented. Through graphical pattern
          display, geographic information systems have the potential to effectively and
          efficiently divulge differential spatial impacts that can highlight high priority areas of
          environmental impact.  In the GISST such patterns depend on the class intervals used
          in choropleth maps. The GISST produces and generates those intervals automatically
          by a data-determined algorithm (Jenks & Coulson, 1963) (rather than by using
          people's values and concerns) that users are not likely to understand. While this
          provides a good tool for exploratory impact identification, users may not employ it in
          this way and no guidance is provided to interpret the GISST outputs.

       7.  Consideration of spatial interdependence of processes.  The manner in which GIS is
          used in the GISST fails to exploit its capabilities as  a spatial analysis tool, limiting its
          use mainly to spatial display or mapping.  Spatial interdependence and interaction of
          environmental and socioeconomic processes and consequent effects appear to be
          ignored in the GISST. Computations in the GISST  are done on a cell-by-cell (or
          polygon-by-polygon) basis without considering the  effects of the same or similar
          processes operating in nearby cells.  Thus the determination of cumulative impacts
          due to interaction effects is likely to be missed. The Panel notes that truly accounting
          for spatial interaction of environmental processes is equivalent to modeling and it
          would require involving additional experts.  However, this may be beyond the scope
          the GISST, which is not intended to be a predictive  modeling tool.

3.5    Question 2.2. EPA also intends to use the GISST in the NEPA process to evaluate
       environmental impacts of project alternatives to help inform decision-making.
       Please comment on the usefulness of the GISST as a tool for this use.

   As stated above, the Panel finds that it is reasonable and appropriate to use the GISST as a
tool for broad-stroke preliminary assessments of large complex projects.  In these cases,
individual criteria, or suites of criteria, could be evaluated to detect potential environmental
impacts that might be associated with project alternatives.  Such preliminary evaluations can help
inform decision-making. However, proper evaluation of project alternatives requires comparison
of the projects with respect to different dimensions of importance (the criteria) and determination
of relative advantages and  disadvantages of each project. Limitations of the GISST discussed
above impede or prohibit such a comparison. The GISST is further limited for use in evaluating
the environmental impacts of project alternatives because the criteria selected for use, and
scaling of the underlying data, must be specific to the ecoregion in which they are applied. For
example, in the Interstate Highway 69 case the stream criteria scaling that was applied to the
corridor in Brownsville, Texas, is not an appropriate scaling for application in the Tyler, Texas,
corridor.  These scaling issues limit the usefulness of the GISST in large scale  assessments for
decision-making such as the Interstate Highway 69 corridor study.

   To make the GISST useful for evaluating the environmental impacts of project alternatives to
inform decision-making the tool should be tailored to address the following questions and issues.
                                           18

-------
   Question 1: Does a project have potential environmental impacts?

   Issues of potential  concern to decision-makers:
•  Summary scores that do not include weighting would potentially mask environmental
   impacts.

•  All criteria that are relevant to the stakeholders and decisions makers must be included in
   the analysis.

•  Irrelevant criteria must be excluded.

•  Relevant and meaningful thresholds or impact levels should be considered (e.g., stream
   criteria applied to Brownsville vs. Tyler, Texas).

•  Spatial dependence of the cells in the GISST must be explicitly considered.

   Question 2: Is it easy to see why a project has an impact?

   Issues of potential concern to decision-makers:
•  The current GISST summing algorithm cannot identify the impacts. Decision-makers
   need to examine individual criteria scores and their associated maps to identify impacts.

•  The vulnerability and impact  criteria can be confused.

   Question 3: Is it easy to see the advantages/disadvantages of alternative projects?

   Issues of potential  concern to decision-makers:
•  It might be important to map two project alternatives simultaneously. Although GIS
   clearly has such capability, it is not evident that the GISST allows one to create
   "difference maps"  of the criterion scores pertaining to two different alternatives.
   Likewise, with three or more  alternatives, it is not clear that the GISST could make a
   synthesis map showing which project had extreme criterion scores for any particular
   geographic region  (e.g., show which project would have the worst water quality in each
   grid cell). There is a need for improved graphical presentation.

•  The sum of GISST averages does not provide complete information about the pros and
   cons of alternative projects. Two projects may have equal cumulative scores but it is
   important to understand how low scores on individual criteria may affect the cumulative
   score. The GISST does not enable one to conduct sensitivity analyses easily. Since
   differential weightings are not allowed, one cannot investigate how different value
   tradeoffs would lead to preferences for different projects. Likewise, the 1-5 value scores
   assigned to the raw score intervals cannot be changed easily, so differences in expert
   judgment regarding vulnerability or impact cannot be investigated.

   Question 4: Can the method be helpful in designing a new alternative?
                                        19

-------
       Issues of potential concern to decision-makers:
   •   The method will not be helpful in designing a new alternative unless the criteria and
       objectives are defined a priori.

3.6    Question 3.1. Please provide recommendations on steps that can be taken to
       enhance the usability of the GISST User's manual and documentation.

   The Panel finds that the GISST User's Manual contains a good introduction to the GISST
and the case example in the document helps the reader understand how the tool works.
However, the GISST User's Manual does not contain step-by-step instructions on how to use the
tool. The User's Manual provides documentation of the GISST, but it is not a user's manual. It
appears to vacillate between providing specific details about how the GISST has been deployed
and information about the general usefulness and applicability of the tool.  The Panel finds that it
would be very useful to develop a true user's manual for the GISST and offers the following
suggestions and recommendations to improve the existing document. Additional specific
comments on the User's Manual are provided in Appendix B of this report.

   •   Identify GISST users. It is not clear who a GISST user might be.  The User's Manual
       does not provide adequate instructions on how to operate the tool nor interpret the
       outputs.  The User's Manual provides a moderate amount of background description
       about the general approach, what it does, and the specific criteria that may be used. If the
       EPA team in Region 6 is the only user of the GISST, the existing User's Manual may
       provide them with enough information.  However, if others will be using the tool,
       additional information must be provided. It is important to clearly identify GISST users
       and  organize the document appropriately.

   •   Describe the need for the GISST, explicitly identifying the types of decisions EPA
       managers make in the NEPA process and how the GISST assists them  in making those
       decisions.  The User's Manual should provide background information describing the
       pressing need for this kind of tool to support NEPA assessments. The  Manual should
       state that development of the GISST is a first step in meeting this need. Chapter 2 of the
       existing User's Manual contains criticism of approaches like GISST. For example page
       13 provides criticism of the simple type of data integration applied in the GISST (Suter,
       1993). Chapter 2 also lists some properties of GIS assessment tools (Leibowitz et al.,
       2000). However, the Manual does not respond to the criticisms nor identify how GISST
       was designed to achieve the desirable properties. In other words, it cites literature
       without interpreting it in the GISST context.  The document needs to respond to these
       issues. The User's Manual does contain a very helpful comparison of the GISST criteria
       to indicators in the SAB framework for reporting ecosystem condition  (U.S. EPA Science
       Advisory Board, 2002).

   •   Provide the underlying conceptual model.  The User's Manual would be vastly improved
       if the authors began with a diagram/description of the conceptual model underlying the
       GISST, such as dose-response or EPA's risk assessment model. The criteria could then
       be related to the conceptual model.  It is unclear why some GISST criteria represent
                                          20

-------
   vulnerability and some represent impacts.

•  Provide the basis and process for considering and selecting criteria.  The basis and
   process for considering and selecting criteria for evaluation of different kinds of projects
   should be provided.  EPA should consider organizing the criteria hierarchically into
   themes rather than a large number of unrelated criteria that make interpretation and
   synthesis difficult. The rationale for criteria scaling and binning should also be provided.
   For example, the criteria in Appendix A of the User's Manual could be organized
   according to vulnerability and impact, and then by criteria groups (e.g., water quality,
   ecological, air quality, socioeconomic, toxicity,  etc.)

•  Provide definitions.  As noted above, vulnerability and impact are never defined in the
   User's Manual.  For many of the criteria it is not clear why they are classified as impact
   or vulnerability measures.  In fact, some criteria are listed as both types. Definitions of
   acronyms should also be provided in a glossary.

•  Provide an index to the User's Manual.

•  Clearly present the mathematical expressions. As noted above there are errors and
   ambiguities in the mathematical expressions describing the GISST algorithm.

•  Define the mathematical  boundaries of output parameters and provide guidance on
   interpretation of the range of expressed parameters.

•  Provide suggested approaches for integration of spatial data.  As noted above, the User's
   Manual provides criticism of existing approaches to integration  of spatial data.  It would
   be helpful to suggest useful approaches that might be applied to integrate the data.

•  Name the criteria in terms of impact or vulnerability factors.  The criteria are named in
   terms of spatial data rather than as impacts or vulnerability factors.  As noted previously,
   the spatial  data themselves are not impacts or vulnerability factors.  This makes it
   difficult to understand the purpose of each criterion  and whether a larger value in the raw
   data is good or bad.  The description of each criterion should start with the relationship of
   the criterion to impact or vulnerability and not just the measurement units of the data
   variable. For instance, a criterion might be named "level of potential infiltration from
   rainfall" rather than simply "rainfall." Sometimes such a description is provided in the
   Definitions/Assumptions, but not always. As an example, the sole source aquifer
   criterion description on page A-13 does not explain  the basic issue, which is that "a sole
   source aquifer makes the communities it serves vulnerable to project impacts."  This
   explanation should be prominent at  the beginning of the section. The section could then
   describe how spatial data are interpreted or analyzed to measure the criterion and provide
   the assumptions and limitations of the data or the mismatch between the data and the true
   meaning of the criterion (i.e., if the  data are an indirect estimate of the criterion value).

•   Provide examples of different applications of the GISST and relate them to the types of
   decisions mentioned above. It would be useful to provide at least one example for each
                                        21

-------
       type of decision.  Such examples could be organized around the uses of the GISST (e.g.,
       scoping, designing alternatives, reviewing environmental impact statements, etc.) This
       would inform users of the appropriate use of the tool for each level of review.  The
       GISST methods, criteria, and algorithm are intermixed in the User's Manual without
       identification of when and where their use is appropriate. The CAFO swine case study
       presented in the User's Manual is not very informative. The conclusions about specific
       impacts, monitoring, and denial of the application are not clearly supported by the brief
       description and maps, nor does the case study describe how to use GISST to make these
       determinations. The Interstate Highway 69 case study may not be the best one to use as
       an exemplar since it is stated on page 48 of the User's Manual that the methodology was
       changed for the analysis (i.e., the area and impact portions of the algorithm were not
       included, only vulnerability was considered, and scores were calculated for individual
       cells).

   •   Clarify the context of the NEPA Document Preparation and Review Section of the User's
       Manual. This section of the document was included after the conclusions.  Although the
       section is listed separately in the table of contents, it is not clear what the section is
       supposed to illustrate. It does not appear to be a new case study nor does it appear to be
       part of the Interstate Highway 69 case study. The section needs an introduction with
       background material referring to a particular case study and showing how the information
       provided is different from material in previous sections. Table 3 in this section should be
       organized by vulnerability and impact criteria and criteria groups (see recommendation
       above concerning  criteria  names and descriptions) rather than providing a long list of
       criteria. The items included in the table are described as spatial data, not as criteria, and
       it is not clear how the data are interpreted (e.g., Is percent of population under 7 years of
       age a good or bad  thing for a wildlife refuge?  What does rainfall total mean?).

   •   Edit the final  document. The final  GISST User's Manual should be reviewed by a
       technical editor to correct  such problems as inconsistent use of plurals (e.g., criteria vs.
       criterion), incomplete sentences, and typographical  errors.  The document should also be
       reviewed to ensure that technical words such as normalize,  standardize, significant, and
       rank have been used correctly.

   •   Reorganize the GISST CD.  The GISST  CD should be reorganized to clearly indicate
       where the reader could locate the files containing the entire document and files
       containing various figures in the document.  It is recommended that the entire document,
       including figures, be provided in one file for users who wish to print it. The document
       should also be placed on the GISST Internet website.

4.      CONCLUSION

   The Panel commends Region 6 for developing the GISST. Geographic Information System
capabilities and data layers provide essential support for efficient, timely, and proactive NEPA
evaluations and other Regional responsibilities.  The benefits of compiling data layers for the
GISST have undoubtedly  extended to other applications within Region 6.
                                           22

-------
   The Panel has identified a number of limitations of the methodological approach used in the
GISST. These limitations restrict the usefulness of the GISST and must be considered in any
application of the tool. It is reasonable and appropriate to use the scores of individual GISST
criteria, or suites of criteria representing types of vulnerability, in preliminary evaluations to
detect potential environmental impacts associated with certain types of projects. However, the
aggregate GISST score should not be used in assessments for decision-making.  The Panel has
recommended improvements to make the GISST suitable for this use.

   GIS-based assessment tools are needed to provide essential support for many EPA activities.
Various EPA program offices and regions have developed screening tools similar to the GISST
since GIS technology became widely available in the 1990s.  Examples include the EPA Region
4 Southeastern Ecological Framework, the EPA Region 5 Critical Ecosystem Assessment Model,
the EPA Office of Water's Index of Watershed Indicators, the EPA Office of Research and
Development's Regional Vulnerability Assessment methods, and the NEPAssist web-based
mapping tool developed by EPA's Office of Federal Activities. The compartmentalized
development of GIS tools and  data by  EPA program offices and regions is inefficient, given the
universal value of these tools.  The Panel  strongly urges EPA to undertake a concerted effort to
develop a unifying framework for the development of assessment tools based on spatial
information technologies.
                                          23

-------
5.     REFERENCES

Cairns, J., P.V. McCormick, and B.R. Niederlehner.  1993.  A proposed framework for
developing indicators of ecosystem health. Hydrobiologia 236: 1-44.

Jenks, G.F., and M.R. Coulson.  1963. Class intervals for statistical maps. International
Yearbook of Cartography 3:119-134.

Land Evaluation. 1968. Edited by G. A. Steward. MacMillan. Melbourne.

Liebowitz, S.G., B. Abbruzzese, P.R.  Adamus, L.E. Hughes, and J.T. Irish.  1992.  A Synoptic
Approach to Cumulative Impact Assessment: A Proposed Methodology.  EPA/600/R-92/167.
Office of Research and Development, U.S. Environmental Protection Agency, Washington, D.C.

Leibowitz, S.G., C. Loehle, B-L. Li, and E.M. Preston. 2000. Modeling landscape functions and
effects: a network approach. Ecological Modelling 132:77-94.

McHarg, I.L.  1969. Design with Nature. Natural History Press. Garden City, NY.

National Research Council. 2000. Ecological Indicators for the Nation.  Committee to Evaluate
Indicators for Monitoring Aquatic and Terrestrial Environments, Board on Environmental
Studies and Toxicology [and] Water Science and Technology Board. National Academy Press,
Washington, D.C.

Omernik, J.M.  1995. Ecoregions: A spatial framework for environmental management. Pages
49-69 in W.S. Davis and T.P. Simon,  eds.  Biological Assessment and Criteria: Tools for Water
Resource Planning and Decision Making.  Lewis Publishers, Boca Raton.

Openshaw, S.  1984.  The modifiable  areal unit problem.  Concepts and Techniques in Modern
Geography 38:41.

Suter, G.W.  1993. A critique of ecosystem health concepts and indices. Environmental
Toxicology and Chemistry 12:1533-1539.

U.S. EPA Office of Science Policy. 2003. Draft Guidance on the Development, Evaluation, and
Application of Regulatory Environmental Models. U.S. EPA Office of Science Policy, Office of
Research and Development, Washington, D.C.
(http://www.epa.gov/ord/crem/library/CREM%20Guidance%20Draft%2012_03.pdf)

U.S. EPA Science Advisory Board. 1999. An SAB Report: Review of the Office of Water's
Index of Watershed Indicators.  EPA-SAB-EPEC-99-014.
(http://www.epa.gov.sab/pdf/epecl4.pdf)
                                          24

-------
U.S. EPA Science Advisory Board. 2002. A Framework for Assessing and Reporting on
Ecological Condition: An SAB Report. Edited by T. F. Young, and S. Sanzone, EPA-SAB-
EPEC-02-009. U.S. Environmental Protection Agency, Washington, D.C.
(http://www.epa.gov/sab/pdf/epec02009.pdf)

 U.S. EPA Science Advisory Board.  2005.  Review of the EPA Region 5 Critical Ecosystem
Assessment Model (CrEAM).  EPA-SAB-05-011. U.S. Environmental Protection Agency,
Washington, D.C. (http://www.epa.gov/sab/pdf/cream sab-05-011 .pdf)

Wiens, J. 1996. Oil, seabirds, and science: The effects of the Exxon Valdez oil spill. Bioscience
46:587-597.
                                          25

-------
Appendix A.  Specific Comments on individual GISST Criteria

    •   Rainfall (page A-4).  The rationale for increasing vulnerability score with increasing
       rainfall is not clear. Might some xeric areas actually be more vulnerable to projects and
       consequent changes in hydrology?

    •   Water releases (page A-5).  TRI releases only consider releases onsite, not flows from
       upstream units.

    •   Surface water quantity (page A-6). The water quantity criterion is only based on length
       of streams, not on volume. This should be validated where data are known for actual
       volume.  Otherwise this criterion may be a poor proxy.

    •   Distance to surface water (page A-7). EPA could use GIS flowpath algorithms to get
       distance instead of using straight-line distance.  Again this criterion is a poor proxy.

    •   Ground water probability (page A-8).  It is not clear what probability this criterion
       measures. Is this the proportion of cells where the water table is less than 6 feet from
       surface?  It would be helpful to include an explanation of the rationale for the 10-acre
       scale and the 6-8 foot threshold in the definition of the criterion.

    •   Unified watershed assessment (page A-10).  This criterion is defined exactly like the
       surface water criterion described on page A-2 but with different scoring. Are they
       correlated? Differences should be explained.

    •   Floodplain, and others where a zero value indicates no data available (page A-14). The
       use of zero values for missing data is acceptable if the only use  is  to flag the lack of data.
       It is inappropriate to actually use the zero value in the analysis since it unbalances the
       composite scores whether they are summed or averaged.

    •   TRI reported water releases (page A-19).  TRI releases are listed as both an impact and
       vulnerability criterion.  This appears to be a repeat of the water  release criterion provided
       on page A-5.

    •   Soil permeability (page A-20). The definitions and assumptions are copied from the
       water table depth criterion and are not relevant for this criterion.

    •   Agricultural lands (page A-21). This measure is very coarse because it combines all
       agriculture types that may differ greatly in associated vulnerabilities.

    •   Agricultural land (page A-21).  It is not clear how farmland can be both a vulnerability
       and impact criterion. Is it even useful to include this criterion without mapping prime
       farmlands? The U.S. Department  of Agriculture and/or states have mapped prime
       farmlands and other classes for the U.S. It is not necessary to rely on National Land
       Cover Data (NLCD).
                                           A-l

-------
•  Wildlife habitat (page A-23).  This criterion also applies an extremely broad definition of
   wildlife habitat. Is open water really appropriately considered to be "wildlife habitat"
   given that it likely is an artificial reservoir in this region? The wildlife matrix measure
   also ignores effects from the landscape matrix.  Adjacent land uses can reduce the quality
   and use of habitat by wildlife.

•  Wildlife habitat (page A23) and wildlife habitat quality (page A-24). The differences
   between these criteria are not clear.

•  Habitat fragmentation (page A-25). In extremely fragmented small-patch landscapes,
   average patch area and patch-area-ratio (PAR) can actually increase due to attrition of
   small patches.  Additional description would be helpful. For example, is PAR calculated
   for all patches of any land use, or only for wildlife  habitat? What will the scoring be for
   landscapes where wildlife habitat is the matrix and there are patches/perforations of
   disturbance? Is this measure appropriate for landscapes with long/linear riparian
   ecosystems that one would expect to see in xeric areas?  Many of these important and
   critical ecosystems have high PAR. Also, given the pixel size used, are patches less than
   10 ha in size omitted given the scale of the land cover data and analysis? This is
   important because the index might under represent the small patches and edge.

•  Habitat fragmentation (page A-25). This criterion  assigns a high value of vulnerability
   only to unfragmented lands. Habitat is simply lumped together as all undeveloped land
   types. Highly fragmented lands are also highly vulnerable too.

•  Ecologically-significant stream segments (page A-29). It is not clear why stream
   segments defined in this criterion are "ecologically significant."  The criterion is loosely
   and subjectively defined. A better description is needed.

•  TEAP criteria (pages A-30 - A-34).  Because these criteria are derived from EPA's
   Critical Ecosystem Assessment Model (CrEAM), the SAB assessment of CrEAM should
   be carefully reviewed (EPA Science Advisory Board,  2005) and appropriate
   recommendations should be applied to TEAP criteria.  For example,  the CrEAM review
   panel strongly cautioned against heavy reliance on "diversity." In GISST, diversity  does
   not always indicate ecological vulnerability. In fact, some heterogeneous landscapes that
   are naturally fragmented may be more resistant/resilient to disturbance.

•  TEAP criteria (pages A-30 - A-34). Using individual TEAP scores and the TEAP
   composite (which is the sum of the individual TEAP scores) is redundant.

•  TEAP sustainability (page A-32).  Sustainability is defined here as resistant to
   disturbance. If that is so, then why should the most sustainable sites be also considered
   the most vulnerable? Also, the value is based on the average of 30 m cells within 1km2
   polygons. This will create artifacts from the resampling problem.  This may be seen in
   other criteria as well but averaging rather than summing reduces the problem.

•  Density of managed lands (page A-3 7). Density of managed lands includes
                                       A-2

-------
   presence/absence only but does not reflect density or amount as suggested by the text. It
   is not clear how vulnerability would be measured.

•  Colonias (page A-41).  Colonias are included by counting numbers, but size is
   disregarded.  In terms of understanding potential human health vulnerability, which is
   more meaningful - size or number?  Is a single large colonia less vulnerable than several
   small ones? It seems that the actual population number is more appropriate.

•  High school education and Educational achievement ranking (pages A-42 and A-43).
   These criteria are probably highly correlated, along with some other socioeconomic
   criteria.

•  Environmental justice (pages A-44 - A-45). Why is a four square mile radius from a
   boundary of a facility applied for economic (environmental justice) measures? A
   statement of the rationale would be helpful.

•  Minority, and other criteria where the state average is used, (page A-45). The use of state
   average data for the low end of the ranking score and two times the average as the high
   end does not allow a full examination of the factor.  This approach truncates the
   distribution of data and in some cases cannot be used (e.g., if the  state average is >50%).
   It would be more appropriate to use  quantile scores.

•  Socioeconomic measures that rely on percentages of populations  showing a particular
   trait (pages A-42, A-44, A-45, A-46, A-48, A-50, A-51, A-53, A-56, A-57, A-58, A-59,
   A-60). In large metropolitan areas or heavily populated counties, there could be large
   numbers of affected people but this may represent a low percentage of the total
   population. Would absolute number be a better indicator in these cases? Examples of
   these criteria include:  well water, telephone communication, linguistic isolation, and
   houses lacking  plumbing.

•   Age (page A-46). The criterion name does not reflect the purpose of the criterion.

•  Age, Children,  and Older population (pages A-46, A-48, and A-50).  These criteria are
   closely related and probably highly correlated so why include all  three?

•  Phosphorus budget (page A-71).  Phosphorus budget seems to be missing the first
   definition of how it is calculated (as included in nitrogen budget).
                                       A-3

-------
Appendix B. Specific Comments on the GISST User's Manual

Page 1, sentence 1: Improve the definition of the tool and be consistent in the use of this
definition throughout the document. Be sure to make it clear how the GISST differs from
traditional GIS use.

Page 2, line 4: Improve the description of the tool. For example EPA might want to say
something like, "GISST is a prioritization tool.  That is, given several options for a decision that
affects the way land is used, the tool can be used to display a map with scores indicating
potential ramifications of alternative decisions on the environment so the user can assess
potential impacts of each decision and where vulnerabilities occur."

Page 2:  Specify who can use GISST.  Identify the kind of data, computer system, and computer
expertise needed to use GISST. Specify the geographic region(s) where the tool can be used.
Consider defining who the "users" are (e.g., EPA staff or stakeholders who examine the GISST
output.)

Page 3, Line 5:  Allow the user to establish the rules for combining the scores.

Page 4:  While GISST may be transparent to the EPA staff that developed the tool, it is not clear
whether the GISST is transparent to others. This point is made on page 70 where it is said, "for
people not familiar with GISST or the output, this [initial spreadsheet] was not very user
friendly."

Page 4:  Make it clear if the user must specify the rules by which the criteria are combined.

Page 5:  Each of the drawbacks should be discussed in more detail. The emphasis on the use of
GISST as a screening tool is important and should be mentioned earlier in the document as well
as in this section.  Is the scale of resolution of the data a constraint?

Page 7, second sentence: Be consistent throughout the document in the way the tool is defined
and described. The document might say, "The tool is an environmental assessment tool
developed to provide  more systematic approach to using spatial data to consider cumulative
impacts in making environmentally sound decisions."

Page 13, second paragraph:  The choice to use equal weighting of the criteria affects the outcome
as well.  A discussion of the implications of weighting is  an important part  of the document.

Page 14: The first paragraph needs expansion.  Why are these properties key for assessment? It
is not clear why the ability to manipulate spatial data is not mentioned as a  key property.

Page 16: While the Panel was glad to see Table 1 included in the document, it was disappointing
because the Table makes it clear that the GISST includes no information on ecological processes
and disturbance regimes, and only limited  information on biotic condition and chemical and
physical features.  Perhaps this long table should be an appendix and the key results should be
reported in the text.
                                          B-l

-------
Page 20: Chapter 3 is much more than "Criteria Development."

Page 21: It is not clear how At is defined. Is it the actual area of the project (i.e. the road) or is
there some buffer around the actual project?

Page 21: De-emphasize the primary algorithm as being a general case.

Page 21: It is confusing to use Dvto represent both individual criteria and the average of those
criteria (and same holds  for DI).

Page 21, line 16:  It is not apparent why "five options" are included in the comparative risk.
What are these options?  This paragraph is not clear.

Page 22: The report correctly points out the bias caused by using more than one criterion for a
resource but does not suggest the basis for selecting the number of criteria to use, and how they
are distributed among the resources. Perhaps the SAB Framework (EPA Science Advisory
Board, 2002) can be useful in this decision by providing an approach for grouping sets of
criteria.

Page 24: The concept of "flipping" criteria is important, but it is also important that criteria have
a direction of impact with a lower number always implying less impact.

Page 25: It is surprising that the air quality section has few criteria. One would think that this
section could be better populated with criteria.

Page 27: The application examples are important to include, but the peer review appendix is
inadequate.  It is not clear how the peer review was conducted. What was the basis for selecting
reviewers? How were they engaged in the process?

 Page 29 Last paragraph: It would be useful to know how the stakeholders were involved, not
just who was involved.

Page 40, Table 2: Identify those data that were used in GISST in this application.

Page 48: The following statement on this page is unclear "The method described in the GISST
User's Manual needed to be modified." The GISST User's Manual is describing a modification
of the method in the User's Manual. Clarification is needed.

Page 48: The least cost path analysis sounds interesting, but the quantitative procedure is not
fully described. How does this approach relate to the Transportation Problem of Mathematical
Programming?

Page 76: The Help Sheet begins to provide information that is needed for a User's Manual, but it
does not go far enough.  More detailed information is needed.
                                          B-2

-------
The following changes are suggested for Appendices A and B of the GISST User's Manual:
   •    Clearer organization of the criteria and a preamble to explain this
       organization is needed.
   •    Hierarchical organization of the criteria is recommended.
   •    The peer review process is poorly described.  The basis for selecting criteria is not
       provided.  The quality assurance and control process is not documented. Additional
       information is needed in these areas.
   •    It would be better not to group "definitions, assumptions, limitation and uncertainties"
       but instead to provide each separately.  It would be useful to provide a single list of the
       criteria in Appendices A and B.
   •    A more complete description is needed for each effect and for the criteria used to provide
       data on that effect.
   •   Although the GIS data sets utilized for the GISST are existing data sets, better
       documentation is needed for the source data, for the  benefit of future users and to
       demonstrate understanding of the data set derivation.
          -  URLs should be provided for every data set.  The URLs should point to the
             original data sources and/or metadata that have been written for those data
             sources.
          -  For each data  layer there should be a description of the data by the source agency
             (e.g., the spatial aggregation of census data used in the GISST was by census
             block).
             If EPA Region 6  summarized the data by a geographic unit (e.g., HUC) after the
             data were obtained from the source agency, the procedure used to summarize the
             data should be documented.
             The time frame of the data (e.g., NPDES permits issued in 2004) should be
             provided.
   •    A description of scaling of data should be provided.
                                           B-3

-------