v>EPA
        United States ,    r;
        EnvironrnentarProtectioH:
        Agency      '
              National Environmental
              Supercomputing Center (NESC)
              135 Washington Avenue
              Bay City, Ml 48708-5845
EPA 208-K-93-001
December 31,1993
FY1993
Nationial
Environmental
Supercomputing
Centeit (NESC)
Annual Report
                                  Recycled/Recyclable
                                  printed with Soy/Canola Ink on paper that
                                  contains at least 50% recycled fiber

-------

-------
The NESC Mission:

The mission of the NESC is to
provide compute-intensive
processing for scientific appli-
cations that have a high prior-
ity within the EPA's overall
mission. These applications
will come from the EPA
researchers, regulatory staff,
and support personnel. In
addition, the NESC .will ser-
vice those universities, agen-
cies, and private companies
external to the EPA having
grants, cooperative agreements^ and memoranda of understand-
ing with the EPA, in which their &eieng$ cpalilies as compute-
intensive and has a high priority witfeia the EPA,
The computational services of thf NE^p include:
•  Management of a wide rang© of hardware!, software, and ser-
vices into an integrated BUp^rcomp^ting Africa
•  Centralized compufe-itolaiisiw processing facility.
•  Data commumeajppiis networks,
•  Consultation ^'fficol relating to the NESO*
A secondary mission of the NESC is to promote environmental
modeling goid. cbmputatieinal science within loeat schools, by
                                     UNTO) SOCKS EHVnomENnL PHOTECTOH AGENCY
                                     NATIONAL ENVRONIOmL SUPERCOMPUTHO CENTER
means
                               summer programs,
UNIX is a trademark of UNIX System Laboratories, Inc.
References to the X Window System are based on reference materials that are copyrighted © 1988 by the
Massachusetts Institute of Technology.
SPARCstation is a trademark of Sun Microsystems, Inc.
CRAY, CRAY Y-MP, UNICOS, and CP77 are trademarks of Cray Research, Inc.

-------

-------
 Contents
Contents ........................... . .................. . ......................... .... ...................... ill


Introduction            :

  Introduction to the NESC FY1993 Annual Report ......... ................................. 1

  Message from the NESC Director ...................................... ................................. 3


NESC: Operations and Activities

  The NESC - An Overview .1 ..................... . [[[ 9
                         i
                         i
  Great Lakes Environmental Visualization Conference
  Overview ............................. ,,; ..................... . ............................. ............................... 17

  EPA Computational Chemistry Workshop ...................... ............................... 19

  Queueing Theory Analysis of Service Level for NQS on the
  Cray Y-MP 81/232 ................ 1 ..................... . ............................. ............................... 23
                         i
  Optimization of Programs ;on a Supercomputer ............ ............................... 29

  EarthVision: EPA's Grand Challenge for High Schools .............................. 31
NESC Research Reports

  Estimation of Global Climate Change Impacts on Lake and
  Stream Environmental Conditions and Fishery Resources ...................... 43
Modeling Sediment and Contaminant Transport and Fate in
Aquatic Systems: ............... ,,; ..................... . ................................. •
                                                                47
  Development, Calibration and Evaluation of a User-Friendly
  Littoral Ecosystem Risk Assessment Model .................. ...... . .......................... 51

  Integration of Computational and Theoretical Chemistry
  with Mechanistically-Based Predictive Ecotoxicology
  Modeling ............................. .1 ..................... . ........................... ...... « .............. • ........... 55
The Supercomputer in Medicine: Interpretation of Gamma
Camera Composition ....... .. ..................... . ........................... ...... .
                                                                59
NESC Annual Report - FY1993

-------
  Contents
    Regional Oxidant Model Sensitivity Analysis	65

    Meteorological and Photochemical Grid Modeling in Urban
    Domains	j	71

    Regional Particulate Modeling			75

    Visualizing Environmental Data at the EPA	77

    Molecular Modeling on Supercomputers for Risk
    Evaluation	.„	85

    Regional Ozone Modeling to Support Clean Air Act
    Mandates	;	91

    Regional Acid Deposition Model (RADM) Evaluation	1	97

   Atmospheric Deposition of Nitrogen to Chesapeake Bay	:	99

   Projected Effects of the 1990 Clean Air Act on Acidic
   Deposition	^	 103

   The Role of the Zebra Mussel (Dreissena Plymorpha) fin the
   Uptake in Food Chain Transfer of Polychlorinated Biphenyls
   (PCBs)	 105

   Summary of Heavy Metal Distribution Data in the Saginaw
   Bay and a Computer Model to Interpret these Data	109
IV
                                                NESC Annual Report - FY1993

-------
                     to tlie ICBSC FY1B83 Ami u al Report
  The National Environmental Supercom-
puting Center (NESC) is the EPA's latest
investment to assure that science of the
highest quality is conducted to support
environmental protection.  The benefits of
the NESC to the public and EPA programs
center on fostering collaborative efforts
among scientists to bring the results of sci-
entific research to the decision maldng
process.
  To achieve success at the NESC, four
tightly integrated programs are main-
tained,                         j
• First, operation of a supercomputing
  resource to provide the maximum
  amount of computing time to research-
  ers is the backbone of the NESC mis-
  sion.                          ;
• Second, a strong computational science
  support effort for Agency scientists is
  essential to ensure the efficient use of
  resources and the improvement of math-
  ematical models.
• Third, an aggressive and innovative
  program in visualization and presenta-
  tion of scientific information is directed
  toward scientists, environmental deci-
  sion making officials, and the public.
• Fourth, collaborative efforts among all
  groups are strongly supported through a
  nationwide telecommunications net-
  work, workshops amd seminars, and
  educational programs such as EarthVi-
  sion: EPA's Grand Challenge for High
  Schools.
  In its first year of operation, the NESC
has become the central resource for carry-
ing out the research programs that are
vital to large-scale studies of ecological
and biological systems.  Its continued suc-
cess in supporting these efforts depends
upon the collaboration among scientists,
managers, and the public.
  The NESC remains dedicated to provid-
ing not only supercomputing resources,
but also to providing the collegia! environ-
ment necessary for that collaboration.
1 Walter M. Shackelford, Director of Scientific Computing, U.S. EPA, National Data Processing Division
    (NDPD), RTF, NC.
NESC Annual Report - FY1993

-------
It is our task in our time and in our
     generation to hand down
 undiminished to those who come
after us, as was handed dtawra to us
   by those who went before* the
natural wealth and beauty wliMi is
   hft l?itzgemM Kennedy $927-i983t Address at
           &&m$i&mes afihvNational Wildlife
                 Building [March 3, 1961]

-------
 •I' ' -V"     ..i \  '  •>-•••••• si '   •• ••    i '      % < v.% " "•
  Message from the MESC Director
 v.-,   %SV  % -. -."T^ •-''-.-.   -""A    ' ' <<•        \ , , w-  " '
Overview                     i
  The National Environmental Supercom-
puting Center (NESC) was established in
1992 by the United States Environmental
Protection Agency (EPA).  The NESC
focuses its resources on solving critical
problems in the environment by utilizing
the power of the most advanced supercom-
puters and visualization techniques avail-
able.                           |
  In 1992, an eighty year old building, in
Bay City, Michigan, was refurbished and
transformed into a high technology center.
On August 25,1992, the first EPA owned
supercomputer, a Cray Research Y-MP
81/232, was delivered.  The Cray Y-MP is
a two processor computer with 32 million
words of primary memory and 32  i
gigabytes (billions of characters) of jdisk
storage.                       ''.
  The NESC's computer runs Crayls stan-
dard operating system, UNICOS, which is
an acronym formed by the words U|NIX
Cray Operating System. As the first part
of the acronym suggests, it is a UNIX Sys-
tem V-based system with University of
California at Berkeley extensions.  The
second three letters indicate that the
UNIX internals have been heavily modi-
fied to make it viable for use on a super-
computer.
  To balance our computing facility, we
procured a StorageTek Automated Car-
tridge System (Silo) with one terabyte
(trillions of characters) of automated car-
tridge tape storage. Because of the,
storage demands made by large environ-
mental data files, a second silo was
procured giving the system two terabytes
of storage.                      ;
  The silos are integrated into the (Cray
system using Data Migration Facility
(DMF) software.  DMF manages the total
storage pool (disks and tapes) and coordi-
nates the movement of customer files
among the devices as needed.
  Our first year of operation (Fiscal Year
1993) was very successful. Immediately
upon installation, customers from around
the country began using the center. By
the end of calendar year 1993, the
resources were being fully utilized.

NESC Customers
  The NESC's customers include EPA sci-
entists and those agencies, universities, or
private companies having grants, coopera-
tive agreements, memoranda of under-
standing, or contracts with EPA.  The
models run at the NESC support the full
panoply of environmental work that
demands compute-intensive processing.
Examples of programs supported include
the Clean Air Act and the Clean Water
Act. There are also models dedicated to
particular areas, such as the Great Lakes,
Chesapeake Bay, and San Joaquin Valley.
Other areas of research include molecular
modeling of various types, such as CFC
studies.
  By the end of FY93, the NESC's super-
computer was being used by more than
200 active customers running over forty
projects.  Customers are scattered from
coast to coast. Representative customer
locations include: Grosse He, MI; Duluth,
MN; Research Triangle Park, NC; Las
Vegas, NV; Athens, GA; Ada, OK; Wash-
ington, DC; Seattle, WA; Kansas City,
KS; San Francisco, CA; University of
Michigan; University of Pennsylvania;
and State University of New York.

Computing at the NESC
  All scientific computers at the NESC use
a form of the UNIX operating system and
NESC Annual Report - FY1993

-------
 Message from the NESC Director
 the TCP/IP communication protocol. Vir-
 tually all supercomputer centers use this
 arrangement, since the user interface is
 the same regardless of whether one is
 using a Data General or Silicon Graphics
 workstation or a Cray supercomputer.
 This makes user access easy and mini-
 mizes the time required to learn a new
 operating system. Any scientific customer
 who knows UNIX can immediately use the
 NESC Cray.
   Since all customers are remote to the
 NESC, telecommunications play an essen-
 tial role in the NESC's effectiveness. One
 T3 (45 Mbps) line and one Tl (1.5 Mbps)
 line connect the NESC to the EPA's
 National Computing Center (NCC) in
 Research Triangle Park. Remote laborato-
 ries and regions connect to the NESC
 through the NCC. Plans are being formu-
 lated to have the remote sites connect
 directly to the NESC.
   The NESC also has an Internet connec-
 tion through a cooperative agreement with
 the Consortium for International Earth
 Science Information Network (CIESIN).
 Many universities and laboratories gaiin
 access to the NESC's supercomputer using
 the Internet.

 Visualization

  Visualization is an integral component
 of high-performance computing.  The
 NESC has established a visualization lab-
 oratory, which acts in conjunction with the
 visualization laboratory at RTP.  Virtually
 all customers of the NESC-use some form
 of visualization.
  Visualization training classes have been
held for customers.  A highly successful
visualization conference was held in July,
and featured speakers and attendees from
the EPA as well as external to the EPA,
(e.g., Jet Propulsion Lab).

Customer support

  The NESC's customers are supported by
 a special scientific .group dedicated to cus-
 tomer satisfaction. Staff members include
 scientists with advanced degrees in phys-
 ics, chemistry, and computer science. This
 group helps scientists port their codes to
 the supercomputer and optimize existing
 codes in order to meet customers' mis-
 sions. This group coordinated a computa-
 tional chemistry workshop in September,
 which had speakers and attendees from
 the EPA as well as external to the EPA
 (e.g., Dow Chemical and Midland Molecu-
 lar Institute).               \

 Future plans

   Because of the increasing demand for
 computation cycles, plans have been for-
 mulated for the replacement of the Cray
 Y-MP 81/232 with a more powerful Cray
 C94/264.  The C94 will have two proces-
 sors,  each two to three tunes more power-
 ful than our existing processors.  The
 primary memory, which has been a limit-
 ing factor on the Y-MP, will be doubled to
 64 million words. The increased memory
 will facilitate those jobs that require a
 large primary memory space. Examples
 are the Regional Oxidant Model (ROM)
 and the Regional Acid Deposition Model
 (RADM), which are required for
 compliance with the Clean Air Act.  In
 addition to primary memory, the NESC's
 high-speed disk storage will be almost tri-
 pled in size.

 EarthVision

  EarthVision, EPA's educational pro-
 gram, is administered through a coopera-
 tive agreement with a local university,
 Saginaw Valley State University.  This is a
 competitive program, whereby high
 schools submit proposals. If selected, the
 school's students take part in a tutorial
 program that takes place on Saturdays
 during the academic year. The high school
teams then submit another proposal and a
winner is selected for a three-week Sum-
mer Research Institute.
                                                        NESC Annual Report - FY1993

-------
                                                         Message from the NESC Director
  It is during this institute that the
schools work on their accepted projects.
They continue to work on then: projects
during the next academic year, and at the
end of the year produce a report oil the
research.  The winning proposal for FY93
was Uptake and Food Chain Transfer of
Polychlorinated Biphenyls (PCBs) in the
Zebra Mussel (Dreissenapolymorpha).
(Editors note:  Research articles from both
EarthVision teams appear on page 105
and page 109.)
1 Arthur G. Cullati, Director, National Environmental Supercomputing Center (NESC), 135 Washington
    Avenue, Bay City, MI 48708.      i
NESC Annual Report - FY1993

-------
  The NESC - An Overview
                                                         Communications links
                             Cray Y-MP/81232
                             2 processors
                             32 Megawords memory
       StorageTek STK 4400
          Tape Silos (2)
       2.4 Terabytes total
                                                                   DS60 disk drive units (16)
                                                                   31.5 Gigabytes total
                                 1600/6250 9 track tape drives (4)
                        Figure 1: Hardware Configuration Overview, National Environmental
                              Supercomputing Center (NESC) - September 1993.
 Mass data stnrag?

   In addition to disk drives, the NESC has
 two StorageTek 4400 robotic tape silos.
 Each silo contains about 6,'000 tape car-
 tridges and is capable of storing 1.2 tril-
 lion bytes of information. The two units
 combined provide a total storage capacity
 of 2.4 terabytes (2,400,000,000,000) of
 data. All tape handling is performed by
 robotic arms and is completely automatic
 and "transparent" to the user. Two 3 MB/
 second data lines connect the silos with
 the Cray.
  Traditional "round" tape facilities are
 also available by special request. Other
 data transfer media may be available.
 Contact the NESC for further information.
 Telecommunications
   In order to meet its mission, the NESC
 must serve customers throughout the
 United States. From its location in Bay
 City, the NESC uses a sophisticated tele-
 communications network to serve custom-
 ers at EPA sites around the country.
  The NESC's telecommunications net-
 work is illustrated in Figure 2 and Figure
 3 (page 11).  The network consists of both
 a Local Area Network (LAN) and a Wide
Area Network (WAN). Each is described
in greater detail in the following para-
graphs.
  The LAN, shown on Figure 4 (page 12),
is responsible for communications inside
the NESC. It consists of four Ethernet
10
                                                           NESC Annual Report - FY1993

-------
                                                                                     The NESC - An Overview
                               11 CIKCUir (!.£<;« Uli|:£>
                               HU ld-l:K Sllft li/iCKUCliE l.'LTXl'OUK
                               SU U-PS ?^H liAL'UHlJliiE liET\VUIi!i
                                iii t-iat'iii- aiicture
                     Figure 2: A sophisticated teliscommunications network connects the NESC witli
                                    EPA researchers throughout the United States.
                                                                                          COUUUNICATION5 UNKS

                                                                                                    T3-45Mb/S

                                                                                                    T1-1.5Mb/S
                     Figure 3: EPA Researchers -use the NESC's supercomputer from throughout the
                            United States via declicated and Internet communications lin
NESC Annual Report - FY1993
11

-------

-------
The NESC - An Overview
                                (HV1 SIA! E'aZB MOm3A/33NVyO
 12
                                                             NESC Annual Report - FY1993

-------
                                                               The NESC - An Overview
  backbones, capable of transmitting data at
  a rate of 10 million bits per second (MObS).
  In addition, there is a single Fiber Distrib-
  uted Data Interface (FDDI), which moves
  data at lOOMbS. These networks are
  responsible for moving data within the
  NESC.

    The WAN is responsible for moving data
  between the NESC and its customers. The
  WAN consists of one T3 transmission link
  and two Tl transmission links.  The T3
  link, which is capable of a peak transmis-
  sion rate of 45MbS, connects the NESC
  with the EPA's largest research facility in
  Research Triangle Park (RTF), North
  Carolina.
   One Tl link, which has a peak data
  transmission rate of 1.5MbS, also links
  the NESC with the EPA's RTF facilities.
  The second Tl link connects the NESC
  with MichNet which, in turn, connects the
  NESC with the NSF Net and
  the Internet.

   Telecommunications routing
  is handled through two NSC
  high speed routers. These rout-
  ers are fully redundant, with
  each router capable of manag-
 ing all telecommunications traf-
 fic.  Two 12MB/second data
 lines connect the routers to
 Sequoia.
  Through the use of TCP/IP
 protocols and the File Transfer
 Protocol (FTP), researchers can
 log into Sequoia, move data to
 and from Sequoia and their
 local network, and obtain the
 results of their research. The
 network's high speed and band-
 width offer the NESC's custom-
 ers the same speed and
 throughput as if Sequoia was in
 the same room.
        Facilities

         A significant amount of infrastructure is
        required to operate and maintain a super-
        computing center. As with an iceberg,
        much of this equipment lies below the
        surface.  The NESC is no exception. The
        two major supercornputing support areas
        are electrical power and equipment cool-
        ing.
        Electrical Power             i

         Supercomputers, such as the Cray Y-MP
        81, require considerable quantities of prop-
        erly conditioned electricity at the correct
       voltage. To support that need, the NESC
       has two 2,500 KVA utility feeds, each one
       of which is capable of supplying all neces-
       sary power. A motor/generator unit pro-
       vides both voltage transfer and
       conditioned power for essential systems
       and support equipment.
         In the event of a power failure, the
       NESC has two 750 KVA uninterruptable
       power supplies (UPS), which are capable
Table 1: Software Applications currently available on
        Sequoia (as of September 1993.)
Discipline
Chemistry
Mathematics /
Statistics
Data Exchange
Graphics /
"Visualization
Applit atioji j
Amber 4.0
AMSOL 3.0
CHARMm 22
DISCOVER 2.9.0
DMol 2.3 1
GAUSSIAN 92
MOPAC 6.0.2
BCSLIB/BCS-EXT Release 12
IMSL 2.0
NAGlib mark 15
LffiSCI
netCDF
AVS 5.0 i
NCAR Graphics 3. 1.3a
NESC Annual Report - FY1993
                                                                                13

-------
The NESC - An Overview
of operating the center for fifteen minutes,
sufficient time to permit an orderly shut-
down of the computer. The UPS are pow-
ered by a bank of 360 storage batteries.
Finally, in the event of a prolonged power
interruption, a 150 KVA natural gas-pow-
ered generator is available to support the
NESC's non-supercomputing functions.

Cooling                        :
  A by-product of supercomputing is the
generation of considerable heat by the
computer's densely-packed circuitry.
Without sufficient cooling, the Cray Y-MP
81 would be subject to extensive thermal
damage.
  Ib keep Sequoia functioning, the NESC
has three 110-ton chilling units operating
through two 175-ton cooling towers.  In
the event of a total failure of the chilling
units, a 2,000-gallon chilled water reser-
voir provides up to 15-minutes of emer-
gency cooling capacity.

Software
  To complement the NESC's supercom-
puting hardware, the NESC supports EPA
researchers and scientists with specialized
scientific application software packages.
Table 1, page 13, lists the software appli-
cations that are available to  researchers
as of September 1993.
  Sequoia uses an operating system based
on UNIX.  By using a UNIX-basecl operat-
ing system, researchers can easily move
their programs and applications between
their local environment and  that of the
NESC. Once a user becomes familiar with
UNIX, those skills are transferrable
across a number of hardware platforms,
including that of the Cray.       |
  Another advantage of UNIX is its adapt-
ability to Distributed Computingj Be it
through Massively Parallel Processing
(MPP) or some form of distributed
computing, UNIX permits the NESC to
easily embrace future trends in large-scale
scientific computing.

Visualization
  In addition to "crunching numbers", the
power and speed of a supercomputer sup-
ports the extensive use of graphical visual-
ization. EPA scientists can call upon
state-of-the-art graphical visualization
and computer-modeling capabilities to
augment their research.  These visualiza-
tion techniques permit the NESC's users
to "see the unseeable".
  Using graphically-based scientific work-
stations, environmental researchers
develop complex mathematical models of
air pollution, atmospheric conditions, the
chemical components of pollution, and
other Grand Challenges. The speed and
data handling capabilities of supercom-
puters allow environmental scientists to
model the interaction of the complex vari-
ables that, until now, could not be tested
in the laboratory.
  Another important aspect of the NESC's
visualization support is in the vital area of
Computational Chemistry. This rapidly
developing branch of chemistry permits
chemists to use a supercomputer in place
of their more traditional test tubes and
flasks.  Computational Chemistry experi-
ments are intuitive, fast, and cost-
effective.
  The NESC features a state-of-the-art
visualization laboratory staffed by experts
in scientific visualization. EPA research-
ers are encouraged to use the laboratory
and its staff to transform their research
data into strikingly meaningful graphical
images.
  The NESC's visualization group
includes six specialists located at RTF.
They are available to serve EPA's
researchers with personalized service and
advice.
 14
                                                          NB^SC Annual Report - FY1993

-------
                                                                The NESC - An Overview
  The NESC staff

    In addition to its hardware and soft-
  ware, a world-class supercomputing center
  requires considerable talent and expertise
  on the part of its staff. The NESC's staff
  includes experts in supercomputing opera-
  tions, planning, computational science,
  and related fields. The NESC staff is orga-
  nized into the following functional areas:
    • Management
    • Operations and Systems Support
    • Computational Science Support
    • Facilities
    • Visualization
    • Documentation
    • Telecommunications
    The NESC's staff is dedicated to sup-
  porting the users.  Staff expertise is avail-
  able to assist researchers with questions
  about computer systems, UNIX, code
 optimization, application porting, and doc-
 umentation. User contacts and inquiries
 are encouraged.

 NESC Usage

  Almost from its dedication in late 1992,
 EPA researchers have made extensive use
 of the NESC's computational resources.
 For the fiscal year, more than 99% of avail-
 able CPU hours were available to our
 users.  Figure 5 shows the CPU-hour utili-
 zation for the fiscal year.
  Plans are underway to increase the
 NESC's computing resources. With the
 addition of increased computational power
 scheduled for the spring of 1994^ the
NESC will continue to be a resource dedi-
cated to meeting our customers's compute-
intensive needs.
                              Sequoia CPU Utilization
                                     Fiscal Year 1993
                             Figure 5: FY1993 CPU utilization - Sequoia
NESC Annual Eeporfc - FY1993
                                                                                15

-------
Now 'tis the spring, and weeds are
           ,  ^t   ^-HS* f f f   ft f ff  f f

 shallow-rooted; Suffer them now

 and they'll o'ergrow
                  *
                           M$4~1616
             // , ,  Act: HI, Scene: 4 Line: 31
             ,"<,?? -

-------
      *ff|i iak&s Mi*vii?0iiiifcfciit£l Visualiz;ati<
   The United States Environmental Pro-
 tection Agency's (EPA) National Data Pro-
 cessing Division (NDPD) and the Great
 Lakes National Program Office (GLNPO)
 sponsored the Great Lakes Environmental
 Visualization Workshop, July 15-16,1993,
 in Cleveland's Marriott Society Center.
 Approximately 100 people, representing
 EPA research and policy-making activi-
 ties, other Federal agencies, the Interna-
 tional Joint Commission, the Great Lakes
 Commission, State environmental agen-
 cies, and local colleges, attended the two-
 day event.

 Day 1 - July 15

   Dr. Walter M. Shackelford, NDPD Direc-
 tor of Scientific Computing, set the stage
 for the workshop, and Dr. Arthur G. Cul-
 lati, Director of the EPA's National Envi-
 ronmental Supercomputing Center
 (NESC), provided an overview of high-per-
 formance computing and visualization
 technologies at the NESC.
   The workshop was designed to provide
 EPA and associated environmental
 researchers and policy analysts with an
 opportunity to learn about scientific visu-
 alization tools. The focus of the workshop
 was on the application of visualization and
 high-performance computing technologies
 to environmental research problems. With
 that goal in mind, three visualization
 experts shared their experiences and per-
 spectives on working with environmental
 data sets.
  Bill Hibbard, University of Wisconsin at
 Madison, gave a presentation on using
 visualization as a diagnostic tool for trou-
 bleshooting complex environmental
 models; Kevin Hussey, NASA Jet Propul-
 sion Laboratory, discussed creating high-
 end animation sequences for the environ-
 mental sciences; and, Lloyd Treinish, IBM
 Thomas J. Watson Research Center,
 talked about maintaining data integrity
 while transporting-environmental data
 sets into visualization toolkits.
   These visualization experts also con-
 ducted hands-on, 30-minute workshops
 on visualization tools their groups have
 developed. Hibbard provided an overview
 of Viz AD; Hussey presented a new soft-
 ware environment  (Surveyor) under devel-
 opment; and, Treinish demonstrated IBM
 Data Explorer.
   Martin Marietta Technical Services, Inc.
 staff taught sessions on visualizing envi-
 ronmental data with the Flow Analysis
 Software Toolkit (FAST) and the Applica-
 tion Visualization System (AVS)!.

 Day 2 - July 16

  An important component of visualiza-
 tion technology transfer includes com-
 puter graphics education programs. On
 the second day of the workshop, Dr. Acha
 Debela, chair of the Historically Black Col-
 leges computer graphics educatipn effort
 of the Association for Computing Machin-
 ery's Special Interest Group on Computer
 Graphics (ACM/Siggraph), and members
 of his committee examined education
 issues. Gloria Brown-Simmons, Jet Pro-
 pulsion Laboratory Visiting Scientist at
 Central State University in Ohio, gave a
 talk on establishing the Center for Scien-
 tific Visualization at Central State
 University.
  The day's technical program included
 presentations by EPA researchers using
visualization technology to solve environ-
mental science problems and by EPA
NESC Annual Report - FY1993
                                    17

-------
Great Lakes Environmental Visualization Conference Overview
managers responsible for implementing
high-performance computing technologies
within the Agency.               :
  James Giatinna, Deputy Director of
GLNPO, discussed the crucial function of
high-performance computing in under-
standing the Great Lakes ecosystem; Will-
iam Richardson, Station Chief of the
Large Lakes Research Station at Grosse
He, Michigan, discussed changing ithe
future of environmental research with the
use of scientific visualization at the Large
Lakes Research Station; and, Robin Den-
nis, Senior Program Manager at AJREAL
and Co-Chair of the Agency's HighLPerfor-
mance Computing and Communications
(HPCC) Program, highlighted the Agen-
cy's HPCC Program and components
related to multi-environmental media
visualization and collaborative computing.
  Pranas Pranckevicius, Chief of Data
Integration at GLNPO, concluded the
workshop with a presentation on
GLNPO's support of ecosystem protection
in the Great Lakes with visualization.
  Also on July 16., Bob Beltran from
GLNPO and Barry Bolka with the GIS
Management Office in Region 5 demon-
strated working with visualization tools in
the MS-DOS and 486 PC environment.
  NDPD and GLNPO look forward to
making the visualization workshop an
annual event for sharing high-perfor-
mance computing and visualization tech-
nology activities taking place within EPA.
 18
             NESC Annual Report - FY1993

-------

-------
             ^

  1B&A Computational
   The United States Environmental Pro-
 tection Agency's (EPA) National Data Pro-
 cessing Division (NDPD) and the  .
 Environmental Monitoring Systems Labo-
 ratory (EMSL) - Las Vegas jointly spon-
 sored a Computational Chemistry
 workshop. The workshop was held from
 September 27 through 29,1993 at the
 National Environmental Superconiputing
 Center (NESC) in Bay City, Michigan.
   This workshop was the first major event
 held at the NESC. Topics of discussion
 included some of the major work being
 done using the NESC's supercomputer.
 This was a unique opportunity to hear in-
 depth details about this new and exciting
 scientific discipline. The study of chal-
 lenging problems such as global warming,
 ozone depletion, air and water pollution,
 and acid rain has been tremendously
 accelerated as a result of the implementa-
 tion of computational chemistry   \
 approaches. Instead of conducting lexperi-
 ments with flasks and test tubes, the com-
 putational chemist works with computers,
 usually a workstation, mainframe, jand/or
 supercomputer. These experiments, called
 simulated experiments, are intuitive, fast,
 and cost-effective.                •
  Approximately 60 people, representing
 various EPA related institutions and oth-
 ers, attended the three-day workshpp.
 Prominent scientists from local institu-
 tions, such as Dow Chemical, Michigan
 Molecular Institute, the University; of
 Michigan, Saginaw Valley State Univer-
 sity, and Wayne State University, partici-
 pated in the scientific presentations and
 discussions. Other speakers inducted sci-
 entists from national and international
 universities specializing in the field| of
 computational chemistry.        ,|
   The main objective of this workshop was
 to survey the scientific objectives and
 achievements of the EPA in the field of
 computational chemistry. These objectives
 include studying problems such as: global
 warming, ozone depletion, air and water
 pollution, acid rain, radiation hazards,
 and chemical toxicity. Experts in areas
 such as Quantitative Structure Activity
 Relationship (QSAR) for chemical expo-
 sure and risk assessment, Computational
 Analytical Chemistry, Water and Atmo-
 spheric Chemistry Modeling, Toxicity Pre-
 diction, and Database Design, led the
 scientific presentations and subsequent
 discussions.
   Of particular interest was that, for the
 first tune, scientists from the regulatory
 wing and the research wing of the EPA
 met and exchanged ideas and discussed
 issues of mutual interest. Thus, this
 workshop at the NESC served as a melt-
 ing pot for the Agency's long term research
 and regulatory objectives in the field of
 Computational Chemistry.  The local press
 and scientific journals carried news items
 about the workshop. The coordinators of
 this workshop are pleased to note that the
 NESC workshop met and exceeded all
 expectations.

 Day One - September 27
  The opening session was chaired by Ben
 Bryan, Martin Marietta/NESC Manager.
 Dr. Walter M. Shackelford, NDPD Director
 of Scientific Computing, opened the work-
 shop by welcoming the participants. He
 specifically thanked the Environmental
Monitoring Systems Laboratory (EMSL) -
Las Vegas for jointly sponsoring this work-
shop. A welcome note from Wayne
Marchant, Director, Environmental
NESC Annual Report - FY1993
                                                                               19

-------
EPA Computational Chemistry Workshop
                         Figure 1: Computational Chemistry Workshop session
Monitoring Systems Laboratory (EMSL) -
Las Vegas, was subsequently read by Don
Betowski of EMSL - Las Vegas.
  Dr. Arthur G. Cullati, Director of the
NESC, then addressed the audience and
gave them a brief introduction and history
of the NESC. The next speaker was Dr.
George Delic, Martin Marietta/NESC
Computational Science Services Man-
ager.  Dr. Delic presented an overview of
his group's activities, including this work-
shop, as an outreach activity.  Dudley
Bromley, Martin Marietta/NESC Visual-
ization Support Manager, then presented
an overview of his group's activities.  Dr.
Don Betowski, one of the workshop coordi-
nators, delivered the vote of thanks.
  The keynote speaker was Gilles Klop-
man, Professor and Chairman, Depart-
ment of Chemistry, Case Western Reserve
University, Cleveland, Ohio. He spoke on
his years of research quantifying the activ-
ities of chemicals in terms of their struc-
tures. Klopman has been using computers
in his research since 1958, and he
acknowledged that "the structure-bioactiv-
ity challenge" still continues, whether it is
in toxicity prediction or in drug design.
  The next speaker was Dr. Steven Brad-
bury, Acting Associate Director for
Research Operations, U. S. Environmental
Research Laboratory, Duluth, MN. In his
talk, Dr. Bradbury stressed the impor-
tance of key experimental data for the
development of computer models for toxic-
ity prediction.  Dr. Stephen C. DeVito, Pro-
gram Manager, U. S. Environmental
Protection Agency, Headquarters, Wash-
ington, DC, gave an overview of various
problems experienced by the regulatory
wing of the EPA and stressed the impor-
tance of computational chemistry methods
in coping with those issues.
  The next session was titled, "Structure-
Activity:  Computational Methods." The
speakers for this session were:  Prof. Ber-
nard Schlegel of Wayne State University,
Prof. Donald Aue of the University of Cali-
fornia-Santa Barbara, Dr. Gilda Loew of
Molecular Research Institute, Dr. George
Famini of the U. S. .Army, Dr. Ann M.
20
             NESC Annual Kepqrt - FY1993

-------
                                                  EPA Computational Chemistry Workshop
 Richard of the U. S. EPA-RTP, Dr,,;Krish
 Namboodiri of Martin Marietta/NESC,
 and Dr. Scott DePriest of Tripos AJssoci-
 ates.  They spoke on various computa-
 tional chemistry methods used for| solving
 a number of environmentally-related prob-
 lems.                          ;
   In the evening, the panel session that
 followed the reception attracted almost all
 of the workshop's participants.  The main
 discussions centered around the strengths
 and weaknesses of various models 'and the
 sharing of information between thfe regu-
 latory and research wings of the EPA. Dr.
 Steven Bradbury of ERL-Duluth chaired
 this exciting session.

 Day Two - September 28

   The second day opened with the [session
 theme: "Molecular Basis of Toxicity."
 Prof. Roman Osman of Mt. Sinai Medical
 School, NY, presented a talk on the molec-
 ular changes/damages due to radiation
 effects on the most important biomplecule,
 deoxyribonucleic acid (DNA). Dr. J[ames
 R. Rabinowitz of the U. S. EPA-RTP and
 Prof. George Pack of the UIC College of
 Medicine presented their research pn
 explaining carcinogenesis based on1 the
 interaction of toxic molecules with DNA.
 Prof. Robert Pearlman of the University of
 Texas - Austin, Prof.  Robert Bach of
 Wayne State University, and Dr. Christo-
 pher Waller of Washington University pre-
 sented many approaches  to the stu dy of
 the basis of biological activity (e.g. toxic-
 ity) by extensively using workstations and
 supercomputers.                 '
  The next session was devoted to one of
 the most recent offshoots of computational
 chemistry, Computational Analytical
 Chemistry. Prof. Einar Uggerud of ithe
 University of Oslo, Norway, presented a
 method of calculating mass spectra of toxic
 molecules, a tool with plausible applica-
 tions in characterizing toxic waste dump
 sites.  Dr. Don Betowski, Prof. Don Aue,
 and Dr. Kathleen Robins presented their
 research on predicting the infrared spec-
 tra of toxic molecules, which has tremen-
 dous applications in remote sensing. The
 final session of the second day was "Com-
 putational Chemistry - Selected Topics."
 Bob Hunter of the NRRI-Duluth, talked on
 neural net approaches for classifying toxic
 chemicals. Dr. Robert Lipnick, of the U. S.
 EPA, Headquarters, Washington, DC,
 spoke of the necessity for using advanced
 computational chemistry tools for regula-
 tory purposes. Arid Dr. Sandy Sillman, of
 the University of Michigan, Ann Arbor,
 presented a talk on numerical methods
 used in the prediction of acid rain and
 ozone depletion processes.

 Day Three - September 29

  The last day of the workshop was spe-
 cially devoted to vendor presentations and
 hands-on sessions. The day started with
 four short presentations: Dudley Bromley,
 Martin Marietta/NESC, Visualization
 Support Manager, presented the video on
 the Center for Ecological Research and
 Training (CERT); Dr. Aileen Alvarado-
 Swaisgood presented various computa-
 tional chemistry software tools available
 from Biosym Technologies, Inc.; Dr. Eric
 Stahlberg, of Cray Research, Inc., pre-
 sented that firm's software, Unichem;
 Scott Hutton and Scott DePriest presented
various molecular modeling tools available
from Tripos Associates, Inc.; and Dr. Peter
Grant of Molecular Simulations,  Inc., pre-
sented computational chemistry tools
available from his company.  Following
those presentations, the vendors demon-
strated their software to the attendees on
an individual basis.
NESC Annual Report - FY1993
                                    21

-------
EPA Computational Chemistry Workshop
Workshop summation
  Using computers in chemical research
adds a new dimension to the traditional
experimental approach. For example, it is
difficult to comprehend the nature of
ozone depletion based on certain rare and
costly experimental results. By incorpo-
rating computers and graphic visualiza-
tion into this research, a whole hew world
of graphical and three-dimensional images
emerges. This not only improves compre-
hension, but also saves time and money,
while permitting more aggressive and
innovative scientific research.
22
            NESC Annual Report - FY1993

-------
                               alysis of Service Le«rel for 1SIQS on
Abstract                     ;
  The NQS complex on the National
Environmental Supercomputing Center's
Y-MP8I/232 has been analyzed to ;deter-
mine service levels as required by the U.S.
EPA's National Data Processing Division
under the existing contract with Martin
Marietta Technical Services (MMTS).  To
meet this contractual requirement,
elementary results from queueing theory
have been applied in the analysis of jobs
submitted to ten public and four private
queues on Sequoia over a nine month
period during the fiscal year 1993.  The
analysis shows that the probability
density function of queue wait times and
service (or wall-clock) times are hyperex-
ponential distributions and that process
rates and mean times are easily extracted
after a fit to the empirical data. The
queueing theory analysis has enabled the
installation of a program (qperf) on the
Cray that indicates the expected queue
wait time and service time for the queue
appropriate to the user specified CPU and
memory requirements. While past perfor-
mance is no guarantee of future results,
stability of the analysis is only suspect if
the character of the whole job population
changes drastically at some futurie time.
Clearly, the longer the time interval of the
sample the more stable the prediction and
therefore the analysis will be updated on a
quarterly basis at NESC.       |
                              i
                              i
Introduction
   It is a requirement of the U.S. EPA's
National Data Processing Division under
the existing contract with MMTS; that
users of the NESC have access to: a
quantitative measure of service levels for
jobs submitted to the Cray resource at the
NESC. At first sight, this seems a difficult
requirement to meet because of the non-
deterministic fashion in which resources
are allocated to jobs submitted to the Cray
under UNICOS. However, detailed infor-
mation on job processing is available from
Cray Standard Accounting (CSA) and, at
the NESC, locally written code extracts
job-level transaction data from the CSA
super-record on a daily basis. This process
tabulates (for each batch job) the CPU,
service (or wall-clock) and queue wait
times. In addition to these, the I/O activ-
ity (4K blocks moved) and memory inte-
gral (KW-minutes) is also recorded. This
sample has been accumulating since
inception and provides a valuable resource
for analysis with a view to determining
the quality of service that users enjoy at
NESC.

The NESC NQS System and Job
Level Data Collection
   Sequoia at NESC has ten public and
four private queues as given in Table 1,
page 24, which also shows the sample
sizes for the respective queues. The period
of the sample represents nine months of
throughput on Sequoia during which no
major system changes occurred that could
drastically affect throughput characteris-
tics. The sample represents a valuable
resource and this report summarizes how
the service and queue times are ana-
lyzed. While no detailed statistical analy-
 sis is present here, Table 1 does show
 some important basic results. In most
 cases, for both of these times, the sample
 standard deviation is larger than the
 mean which indicates the likelihood of a
 hyperexponential distribution. Also, it is
 NESC Annual Report - FY1993
                                                                                 23

-------
  Queueing Theory Analysis of Service Level for NQS on the Cray Y-MP 81/232
          O^'o'1,6 8tl£te*C8 for C,PU' servic?,aild 1ueue times for *e ten public and four private queues of the
             Queue limits are shown in million words (memory) and seconds (CPU time). The sample size for
                          each queue is shown in the column labelled *N'.               '•
>^ ^\;vV
^^tt>fc-"
small_8hort
medium_short
large_short
sma]l_normal
medium_normal
large_normal
smalljong
mediumjong
largejong
night
areall
area!2
areal_romdp
areaLradmp
Mem
,MW
4
12
18
4
12
18
4
12
24
24
10
16
6
4
<3»t?
: mvs
600
600
600
3600
3600
3600
100000
100000
999999
999999
999999
999999
1830
1830
N

1377
255
11
1440
1268
629
1197
1355 ,
494
42
570
505
388
22
CWti
Mean
1.1938
1.0492
2.4914
10.022
13.235
9.834
85.586
106.33
226.18
213.86
93.678
70.496
0.2947
0.7498
mafcbte)
ttdZfer
2.3719
2.036
3.7456
13.307
16.14
13.128
142.31
197.81
466.77
346.16
85.545
58.162
0.3353
0.4968
$e«iw1
Mgaa
21.175
6.4031
12.32
33.24
41.6
39.378
212.9
241.47
493.64
449.53
204.77
228.58
15.31
75.484
•HisKjE&n}
StdDw
164.93
11.065
17.61
52.43
54.57
131.83
420.59
452.62
1114.26
775.1
209.73
233.69
22.286
272.76
^tteiuet
Mean
74.93
8.8683
309.02
58.03
80.789
159.95
117.05
758.69
261.92
600.09
642.9
287.85
9.8729
4.4242
Jlws(jcdnJ
StdDsv
299.41
40.389
240.44
172.73
169.22
215.19
349.7
1777.53
598.46
616.1
960.74
626.94
26.0712
7.4717
 interesting to observe that the mean CPU
 times are invariably significantly smaller
 than the queue CPU time limits.

 Elements of queueing theory and
 analysis of NESC data

   Once a job has been dispatched to a par-
 ticular queue by NQS, it can be viewed as
 residing in a single server queue system.
 Queueing theory1 then applies and treats
 these times as random variables or observ-
 ables which do not have individually pre-
 dictable values but whose values show
 statistical regularity.  In particular, a ran-
 dom variable,-X", is completely described
 by a probability distribution function,
 F (t) =Prob { X£t}, or by the correspond-
 ing probability density
 f (t) =dF/dt. The latter is a frequency
 distribution and may have various possi-
 ble shapes depending on the details of the
 queueing system. However, in this analy-
 sis it is found that the distribution of
 queue and service times is dominated by
 exponential shapes such as
 f (t) = lie^, t>0, or combinations of
 exponentials  (hyperexponential). There-
 fore, the analysis2 requires the determina-
 tion of the rate parameter p, from which
 the probability distribution is computed as
 F (t) = 1 - e~^. The function F (t) is also
 known as the cumulative probability
 because it represents the "area" under the
 density curve f (t).  Since both job queue
wait and service times have been recorded
at NESC, each is analyzed in four simple
steps: (1) a sort into bins as is done in a
24
                                                         NESC Annual Report - FY1993

-------
                    Queueing Theory Analysis of Service Level for NQS on the Cray Y-MP 81/232
histogram plot, (2) the fitting of the result-
ing distribution with jie~*lt to determine \i,
(3) computation of the mean as 1 X|X (prop-
erty of the exponential distribution), (4)
computation of F (t).  Results of this
procedure are shown in Figure 1, page 26,
for the service time in the small_normal
queue. The distribution is characteristic
of a hyperexponential function where the
forward peak is described by an exponen-
tial with the service rate of 0.03218 job-
s/minute and the tail is described ;by
another exponential with a service rate of
0.004532 jobs/min. The mean service
times are simply the inverses of the
respective rates for each exponential dis-
tribution. The fact that there is a hyper-
exponential distribution shows that the
jobs in service are of (at least) two types.
If/? assigns a probability that the;job is of
either type, then Figure 2, page 27, shows
a sequence of probability densities corre-
sponding to different values of p. -The
large dots connected by the dotted line
represent the same data shown in
Figure 1, page 26.              ;
  Figure 3, page 27, shows the probability
distribution functions computed from the
two component exponential distributions
of Figure 2, page 27, for various values of
the probability, p. Each curve shown here
represents the accumulated area under
the corresponding probability density dis-
tribution of Figure 2.  The cumulative
probability distribution in Figure 3 shows
(on the vertical axis) the likelihood that a
given job has the predetermined total ser-
vice (or wall-clock) time chosen (on the
horizontal axis). While a table of jthe
mean rates (|x) and mean service times (a)
is shown for the tabulated choices of the
probability, p, the mean service time (a) is
not the best indicator of service level. A
job with a value p=0 has a 60% chance
(lowest curve) of receiving a service time of
200 minutes, whereas, a job with a p=l
value has a 60% chance of receiving a ser-
vice tune of approximately 25 minutes.
Stated in another way:  the expectation
that the same job would complete in 200
minutes changes from 60% to nearly 100%
if p moves closer to 1. Yet another way to
read the probability distribution of Figure
3 is to obtain the percentiles of the sample.
As an example, using the p=l curve for the
above-mentioned case, 60% (vertical axis)
of the sample of 1440 jobs had a service
time of approximately 25 minutes (hori-
zontal axis) or less.

Conclusions
  A nine month sample of job level NQS
data on Sequoia has been analyzed by a
simple single server queueing model.  The
resulting analysis demonstrates that
there  are at least two job classes. The first
job class shows shorter queue wait and
service times while the second job class
has longer times. The second job class
represents a small fraction of the mea-
sured sample and some reasons for large
values can be given. In the case of queue
wait time, users tend to submit multiple
jobs, but the system will only service two
jobs from the same user at any given
instant. Therefore, the queue wait time
increases for later jobs in a single batch
from the same user. In the case of service
(or wall-clock) time, large values can
result from several causes, such as: wait-
ing on migrated files, availability of
requested memory, heavy I/O activity, or
competition with interactive usage for
either memory or CPU time.
  A program called qperf has  been
installed on Sequoia and made available
to users of NESC. A simple command gen-
erates a tabular form of the probability
distribution function for both queue wait
and service times observed for the sample.
The user specifies the CPU time and mem-
ory requirement and the appropriate
queue is selected. Because no model exists
at present for estimating p in the range
0
-------
  Queueing Theory Analysis of Service Level for NQS on the Cray Y-MP 81/232


  characteristics a priori, qperf uses the
  results for the analysis of the first job class
  only since this class represents the major
  fraction of the sample.
   H. Kobayashi, Modeling and Analysis: an Introduction to System Performance Evaluation Methodology,
      Addison-Wesley Publishing Company, Reading, MA, 1978.
   J. Banks and J. S. Carson, Discrete-Event System Simulation, Prentice-Hall, Inc., Englewood Cliffs1, NJ, 1984.
                         Small_normal 4HW 3600  seconds
                   October  1992 to June  1993 -  1440 jobs
           &
1200
1000
800
600
200
0

'tMea
.IMe
I
I
I
*

i ser
n se


M<
M
*">
'ice r
vice


an si
an s
«,»
ime =


rvTce
rvice
E= 0.032
= 31.1 min


rate |i1 =
time =22
5/m



.OOJ
0.5 r
n

20

532,
iln


3ins

mln

                   O    100  200  300  400  500  60O  70O  800  900
                                 Service time in minutes
        Figure 1: Example of the probability density function for the service time of the small_normal queue which
                     shows the number of jobs versus the amount of service tune required.
26
                                                           NESC Annual Report - FY1993

-------
                  1200
                      Queueing Theory Analysis of Service Level for NQS on the Cray Y-MP 81/232
                            Small_normal  4MW 3600  seconds
                       October  1992 to June  1993 -  1440  jobs
                                 -Data fo
-  20  birs
                                       200     300     400     500
                                     Service time in minutes
                                600
          Figure 2 : Probability density functions computed from the two component exponential distributions of
         Figure 1 for various values of the probability, p, which measures if the job is likely to be in the tail (p=0) or
                                the peak (p=l) of the empirical distribution.
                   1.0
                            Small_normal 4MW  36OO  seconds
                       October  1992  to  June  1993  —  1440 jobs
                                                 I   (per £nin)
                                               1 .o^oioszTs
                                               0.8  0.02B&5
                                               0.6  0.02112
                                               0.4—0.01559
                                               0.2  0.01006
                                               O.O  0.004532
                                      200     300     4OO     500
                                     Service time in minutes
                               600
                                  i
       Figure 3: Probability distribution functions computed from the two component exponential distributions of
       Figure 1 for various values of the probability, p, which measures if the job is likely to be in the tail (p=0) or
       the peak (p=l) of the empirical distribution. Each curve shown here represents the accumulated area under
                        the corresponding probability density distribution of Figure 2.
NESC Annual Report - FY1993
                                                                                            27

-------
 All the mathematical sciences are
   founded on relations between
physical laws and laws of numbers,
so that the aim of exact science in to
 reduce the problems of nature to
 the determipatloipi of quantities by
     operations wlttt number
                          1831-1879
                       ofForce [1856]
* S', . " "••

-------
 Optimization of Programs on a Supercomputer
Introduction
  The NESC's Cray Y-MP 81 supercom-
puter (Sequoia) has been installed Eiiid
running for just over a year. During that
tune, many EPA programs have been exe-
cuted on it.  As we see an increasing num-
ber of different program types running on
Sequoia, we  are increasing our under-
standing of those program's performance.
  Many of the programs running on
Sequoia were ported from other types of
machines, such as Digital Equipment
VAX, and IBM 3090. Some users of the
Cray Y-MP have made the assumption
that their programs will execute with
supercomputer speed once they are run-
ning on the Cray Y-MP.  Unfortunately,
this assumption is simply not true, j
  Due to its unique design, the Cray Y-MP
can execute user programs over a large
range of performance levels. How effec-
tively these programs perform, depends on
their optimal use of the machine's special
capabilities. The difference in perfor-
mance between a well-optimized program,
and an unoptimized program, can easily
be an order of magnitude (as much as 448
tunes faster as described later in this arti-
cle.). That is, if the program, unchanged
after porting from another machine, takes
ten hours of execution (CPU) time to run,
a small effort at optimization could change
it to run in only one hour.  The example
that follows shows how dramatically per-
formance can be improved on a Cray Y-MP
machine.

 Why  make the effort to optimize?
   While optimization may require some
 effort, there are several major benefits
 that far outweigh that investment.
 NESC Annual Report - FY1993
• If researchers can get their model
  results back sooner, they will have
  more time to "do their science." Other-
  wise, they may be waiting for those
  results that will tell them what model-
  ing changes to try for the next run.
• Naturally, if a program runs faster, it
  will not deplete the scientist's time
  resource allocation on Sequoia as
  quickly as an unoptimized program.
• If the program can be made to run
  faster, the scientist using it may be able
  to run more iterations of the model.
  Faster execution may also lead the way
  to further refinements of the model,
  such as a smaller grid with more grid-
  points.
  It is definitely to the advantage of all
Sequoia users to optimize their programs
for improved Cray execution. The Compu-
tational Science Support (CSS) staff pro-
vides optimization help and advice.
    "Some users of the
      Cray Y-MP have
  made the assumption
   that their programs
     will execute with
      supercomputer
   speed once they are
   running on the Cray
             Y-MP.
    Unfortunately, this
        assumption is
      simply not true."
                                                                         29

-------
   Optimization of Programs on a Supercomputer
   The IDAHO4 Optimization Effort
     By making certain changes to the
   IDAHO4 program, NESC's CSS staff was
   able to speed it up by a factor of 448. This
   was accomplished using the Cray Y-MP
   machine at the National Environmental
   Supercomputing Center (NESC), located
   at Bay City, MI. The code was provided by
   Ross Kiester, a scientist at the Forest
   Research Station, which is associated with
   the Environmental Research Laboratory,
   EPA, Corvallis, OR (ERL-Corvallis).
     The program originally required 42,142
   seconds (11.7 hours) of computation (CPU)
   time to execute on the Cray Y-MP machine
   at the North Carolina Supercomputer
   Center (NCSC).  The program now
  requires only 94 seconds (1.6 minutes) of
  CPU time, running on the NESC machine.
  Thus, this scientist could perform 448
  runs of his model in the time previously
  required to execute only one run.
    It took about seven hours to achieve this
  speedup.  While this is a dramatic exam-
  ple of the benefits of optimization, even a
  small speedup in a long-running program
  can be of help  to the user of that program.

  Technical Issues

   The IDAHO4 program is written in the
  C language.  The problem this program
 solves is a combinatorial study. There are
 404 distinct sections (quads) of data, that
 must be compared in groups of four. Every
 possible combination of four must be tried.
 This causes the number of comparisons to
 "explode" to slightly more than one billion.
 Each section of data (quad) contains 359
 separate flags.  Each flag can take on a
 value of 1 (true) or 0 (false).
   The IDAHO4 program attacks this prob-
 lem using brute-force methods. Every one
 of the 359 flags is logically combined with
 every other flag in the group of four being
   checked. Thus, for each of the one billion
   combinations tested, 1436 boolean calcula-
   tions, and 359 boolean tests must be per-
   formed.                    :

    The first optimization issue was not one
   of vectorization, but rather the volume of
   work performed. The amount of work per-
   formed in the hardware was 64 times
   more than the program required, since
   each boolean flag occupied one Cray
   word. This fact led the optimizer to pack
  the flags into a 64-bit Cray word (64 flags
  per word), a better utilization of the Cray
  Y-MP hardware.
                             [
    Note that the Cray Y-MP machine can
  not only perform boolean operations'on 64
  bits at a time, but there are two parallel
  boolean functional units available in the
  machine's CPU.  Thus, this configuration
  can allow 128 simultaneous boolean oper-
  ations to take place in every machine cycle
  (six nanoseconds). Given in rate terms,
  this means that an ideal code cpuld per-
  form over 21 billion boolean operations per
  second per CPU.

   Once the amount of work was lessened,
 it was possible to continue the optimiza-
 tion process by improving the vectoriza-
 tion of the loops that the program
 performs.

   Note however, that vectorization alone
 cannot always improve a code's perfor-
 mance. In the case of the IDAHO4 pro-
 gram, it was necessary to first reduce the
 total amount of work performed by the
 vector functional units of the Cray Y-MP.
 The final execution time figure (94 sec-
 onds) represents a boolean rate  of 9.8 bil-
 lion operations per second (the peak
 theoretical rate of the Cray Y-MP is 21 bil-
 lion per second). This rate is 97 times
 higher than the boolean rate of the origi-
nal code, which was 101 million per
second.
30
                                                        NESC Annual Report - FY1993

-------
                  : MFA?s Grand Challenge for High Schools
What is Earth Vision?
  EarthVision is a professional develop-
ment and educational program for t0ams
of high school teachers and students.
Each team, is composed of two teachers
and four students. It is a joint venture
between, the United States Environmental
Protection Agency (EPA) and Saginaw Val-
ley State University (SVSU). EarthVi-
sion helps high schools develop   i
environmental research programs using
computational science and access to high
speed computers.
  EarthVision is the first computational
science education program to concentrate
solely on environmental issues, offer Sat-
urday tutorials, and provide multi-tiered
outreach activities.

EarthVision Components

Saturday Tutorials             !
  During the academic year, teams of high
school teachers and students learn the
specific skills                  i
required to conduct             |
environmental
research using
computational sci-
ence.  Each team
member is given
substantial hands-
on experience.
Each team is sup-
ported by mentors,
who are profes-
sional researchers
and/or specialists
in computational
  "The students are not only
excited, they are very com-
mitted!  They're going the
extra mile, they're taking
their spare time to learn as
much as i they can ... It's been
a surprise and excitement for
us all that they're far beyond
where we originally thought
they'd be at this time."
 science, and by a
 member of the
  Lynne Petterson, Project Officer, EarthVision: EPA's
   Grand Challenge for High Schools, National Data
      Processing Division (NDPD), U.S. EPA.
SVSU outreach support team. These ses-
sions are held at the National Environ-
mental Supercomputing Center (NESC) in
Bay City, Michigan. After the completion
of the Saturday tutorials, the teacher/stu-
dent teams write two proposals to apply
for admission to the Summer Research
Institute: (1) a research proposal and (2)
an education plan proposal to introduce
environmental research and computa-
tional science into their high school pro-
gram.
Summer Research Institute
  During the Summer, competitively
selected teams participate in a three-week
educational program at SVSU. Team
selection is based on proposals for a
research project to be conducted during
the academic year and an education plan.
Summer Research Institute  participants
receive instruction in conducting an envi-
ronmental research project of their own
design. Each team, is supported by men-
tors and by an SVSU outreach support
                team.  Each high
                school participating in
                the Summer Research
                Institute is supplied
                with a scientific work-
                station, and a telecom-
                munications link to
                the National Environ-
                mental Supercomput-
                ing Center (NESC) in
                Bay City. The Sum-
                 mer Research Institute
                 prepares the teams for
                 the next part of Earth-
                 Vision, conducting
                 environmental
                 research at their
                 school.
 NESC Annual Report - FY1993
                                                                             31

-------
   EarthVision: EPA's Grand Challenge for High Schools
                           at.HiPh Prhnnlfp)
    During the academic year, the partici-
  pants from the Summer Research Insti-
  tute conduct environmental research
  activities at their high schools.  They will
  use then- scientific workstation, and the
  supercomputer at the National Environ-
  mental Supercomputing Center in Bay
  City, to analyze data, conduct environmen-
  tal modeling and use scientific visualiza-
  tion to implement their research.  The
  teams will prepare a paper describing
  their research. The mentors advising each
  team will continue to provide support.
  The SVSU outreach team visits the
  schools which have participating teams
  and assists them in their research and in
  establishing an educational program in
  environmental research and computa-
  tional science.
 Benefits to the General Public

   The general public will benefit by expo-
 sure to student research projects and pre-
 sentations, enhanced awareness of new
 technologies, and new ways of Iconducting
 scientific research. Additionally, the gen-
 eral public will have opportunities to see
 the types of research conducted on a
 supercomputer.
   The EarthVision teams will present
 information about their environmental
 research, their experience and their cur-
 riculum plans at local and national confer-
 ences such as the IEEE SuperComputing
 1993 and the International Conference on
 Scientific Visualization. They have also
 given numerous local presentations to
 their colleagues in their own schools, other
 schools and community groupsJ  Addi-
tional presentations and publications are

      Figure 1: EarthVision Summer Research Institute 1993 Course Completion Ceremony, Bay City Central' Hirii
      School Tfeam. Fu-st row from left to right Marie Neal, teacher; Jill Bisel, teacher; Navid Mazloom, student- Kim
     Kukla student; Erin Gatza, student; Lynne Petterson, U.S. EPA; Second row from left to right Jon Whan, Super-
      intendent Bfa^6^CAIf); D°n ™f°Td> U'S-EPA;Jasoa Schroeder, student; George Charles, Principal; Wiffis
      Greenstreet, U.S. EPA; Joe Gonzales, Superintendent, Bay City Public Schools; Walter Shackelford, U.S. EPA;
                       The Honorable James Barcia, U.S. House of Representatives.
32
                                                            NESC Annual Report - PY1993

-------
                                    EarthVision: EPA's Grand Challenge for High Schools
being planned and pre-
pared. During the
Summer Research
Institute, the students
had several opportuni-
ties to present infor-
mation about their
research to various
constituencies.  It was
found that the experi-
ence of preparing for
presentations was a
valuable pedagogical
tool in assisting them
in focusing on specific
elements of their
research.

How has
EarthVision
developed beyond
its original intent?
Environmental
Research during the
Summer Research
Institute
  "The teams worked very
 hard and it was not a very
easy task for them because
    they had to combine
 everything they had been
  taught: the science, the
 computers, the operating
systems, the visualizations.
  All of this had to be put
together in such a way that
  they could go back and
  look: at the question they
     wanted to ask and
 determine whether or not
  they could actually ask
 that in the way proposed.
 Watching them do this and
 turn it into a visualization
      I was absolutely
      i  spectacular."
  Ken. Flurchick, Chief Scientist North Carolina
      i  Supercomputer Center
nical Services,
Inc., and Saginaw
Valley State Uni-
versity.
  Assistance from
experts in diverse
fields enabled the
EarthVision teams
to integrate compli-
cated views, meth-
ods and obtain
initial results
from their
research projects.
Initially, it was not
anticipated that
the diverse repre-
sentation of exper-
tise would promote
team building
among the partici-
pants, but it was
clear that the men-
tors, working
together to assist
the teams, served
as role models of
  Two of the three 1992-93 EarthVision
teams that finished the Saturday tutorials
also completed a three-week Summer
Research Institute. The original intent
was simply to prepare the teams to begin
conducting their research during the fol-
lowing academic year. Both teams actu-
ally began conducting their research
during the Summer Research Institute.
They made progress in formulating their
models, mathematically writing code to
represent those models, and preparing
initial visualizations of their data.
  During the Summer Research Institute,
lecture and hands-on sessions were lead
by experts in scientific research, modeling
and visualization. These experts were
from organizations such as: the North
Carolina Supercomputer Center, U.S. EPA
Visualization Lab, Martin Marietta (Ifech-
                                     team functioning.
                  Role differentiation on the part of the par-
                  ticipants was a key factor in facilitating
                  then- enhanced progress.  Each individual
                  had specific tasks to complete to contrib-
                  ute to the team's research effort.  The indi-
                  vidual achievement made possible the
                  team's success.
                  Cognitive Apprenticeships
                    In addition to expert lecturers, the
                  EarthVision teams were supported by
                  mentors working on environmentally-
                  related research at organizations such as:
                  the EPA's Large Lakes Research Station at
                  Grosse He, MI, Dow Chemical Company,
                  Alma College, HydroQual/Manhattan Col-
                  lege, and the EPA's Environmental
                  Research Laboratoiy, Duluth, MN. These
                  mentors donated their time to help the
                  teams refine and focus their approach, find
                  additional data, and construct computa-
NESC Annual Report - FY1993
                                                                           33

-------
  EarthVision: EPA's Grand Challenge for High Schools
  tional models. They will continue to work with
  the teams during the 1993-94 academic year
  until the teams complete their research
  projects. In addition to donating their time,
  mentors provide a more valuable commodity;
  they share their own thought processes, prob-
  lem solving methods and serve as role models.
  Thus, EarthVision teams learn environmental
  science in a cognitive apprenticeship with scien-
  tists solving current environmental challenges,
  very much like the traditional master-appren-
  tice learning format.

  Peer Reviews and Collaborations
   A multitude of currently available communi-
  cation methods help the teams keep in touch
  with other student-teacher teams with similar
 missions as well as with mentors. In addition
 to traditional face-to-face interactions and tele-
 phone conferences; electronic mail, videoconfer-
 ences, and collaborative visualization
   "These kids are very
surprising in a lot of ways
  and in some  cases very
   frightening. They are
extremely bright... Their
 enthusiasm continues to
 challenge anybody who
  works with them.  It's
 very difficult  at times to
   keep up with them.
 They're like a sponge ...
  they tend to  challenge
    and push anybody
    dealing with them
    because they're so
  interested in the way
   these things work."
Ken Flurchick, Chief Scientist North Carolina
       Supercomputer Center
          Figure 2: EarthVision Summer Research Institute 1993 Project Peer Review between EarthVision
               teams and NCSA SuperQuest participants, using compressed video teleconference.   :
34
                                                    NESC Annual Report - FY1993

-------
                                    EarthVision: EPA's Grand Challenge for High Schools
techniques are some of the
currently available means
of communication the
EarthVision teams can
employ.
  The NESC's videocon-
ference room has been
used to contact experts
and review research
projects with other stu-
dents nationwide.  One
such peer-review video-
conference took place dur-
ing the summer of 1993,
at which the teams shared
ideas with colleagues at
the National Center for
Supercomputing Applica-
tions (NCSA) SuperQuest
project. The teams, par-
ticularly the students, rec-
ognized the importance of
communication with oth-
ers doing similar work.
The teleconference with
NCSA was a means by
which the value of peer
review and collaboration
was  demonstrated. For
high school students, rec-
ognizing that they need
not operate in an isolated
environment was a
revelation. The students
believed that collabora-
tions and cognitive
apprenticeships were
highlights of the Summer
Research Institute.
  Collaborations among
the two current Earth-
Vision teams, mentors and
other students in national
projects, were motivating
for students and helped
them prepare for the
multidisciplinary-collabo-
 "You get to meet
  students from
   other schools
  with interests
 r
-------
  EarthVision: EPA's Grand Challenge for High Schools
    "We have just been
  thrilled by what's been
  reported to us about the
  students' reaction and
  their working 12 to 14 hour
  days. They're totally com-
  mitted to the project and
  the things  that they are
  doing and it's great to see."
   Don Fulford, Director, National Data Processing
         Division (NDPD), U.S. EPA
   "I was somewhat appre-
 hensive that we were offer-
 ing too much material for
 the teams to try to work on
 and perhaps too much
 work and not enough fun
 but I understand the work
 to them is the fun."
  Walter Shackelford, Director of Scientific Comput-
 ing,, National Data Processing Division (NDPD) U S
               EPA
  "We would hope that one
  or two of these kids might
      see fit to find their
   research careers in the
 environmental business.  I
  don't think we're going to
  run out of environmental
    problems in the near
   future.  So, it is a good
       future for them."
Willis Greenstreet, Director, Office of Administration
   and Resources Management/RTF, U.S. EPA
  enough funding to support one team during
  the 1993 Summer Research Institute. This
  support provided stipends for teachers and
  students and a scientific work-station and
  telecommunications link to the supercom-
  puter at NESC for their school. The Bay
  City team was chosen to receive this sup-
  port.  The EarthVision planning team
  invited the other Saturday Tutorial teams to
  participate in the Summer Research Insti-
  tute without financial support, since they
  worked so hard and had excellent proposals.
  The Saginaw Center for the Arts and Sci-
  ences team chose to participate in the Sum-
  mer Research Institute at their own
  expense. Both the Bay City team and SVSU
  volunteered the use of their workstations to
  the Saginaw team during the following aca-
  demic year. The Saginaw team participated
  with a great deal of enthusiasm and were
  significant contributors to the success of the
  summer experience.           '
 Experiencing the Rymtement of
 Environmental Science-        :
   Near the end of the Summer Research
 Institute, the EarthVision teams demon-
 strated their current accomplishments at an
 open house for parents, siblings, friends,
 and invited guests. This proved to be an
 excellent opportunity for parents and the
 general public to understand more about
 how research and computational science can
 contribute to the solution of some of today's
 most pressing environmental problems.
 Even at this stage in the program, before
 the teams' research is done, it appears that
 the experience is having an influence on the
 students career choices.

 What is Next for EarthVision?
 A New Annual (yip for EarthVision
  The EarthVision planning team is prepar-
 ing for a new group of participants to begin
 the Saturday tutorials in the 1993-94
 academic year. The enthusiasm and excite-
ment of the students and teachers partici-
36
                                                   NESC Annual .Report,- FY1993

-------
                                     EarthVision: EPA's Grand Challenge for High Schools
          Figure 3- EarthVision Summer Research Institute 1993 Course Completion Ceremony, Saginaw Cen-
          ter for the Arts and Sciences tearn. First row from left to right Chuck Rohde, student; Chris Stark, stu-
          dent: Natasha Sefcovic, student; Brian Weeden, student; Lynne Petterson, U.S. EPA. Second row from
          left to right Don Fulford, U.S. EPA; Willis Greenstreet, U.S. EPA; The Honorable James Barcia, U.S.
          House of Representatives; Dan Sealey, teacher; Gary Barker, teacher; Walter Shackelford, U.S. EPA;
                      Burris Smith, Assistant Superintendent Saginaw Public Schools.
pating in the 1993 Summer Research
Institute was demonstrated by thdir vol-
unteering to come              .,
back during the
1993-94 Saturday
tutorials to assist
the new teams.
  Invitations to
participate in the
EarthVision proj ect
have been sent to
high schools in the
State of Michigan.
Six new teams will
be selected this
year for the Satur-
day tutorials, and
                       and education proposals will be selected
                       for the Summer Research Institute. All
four schools with
the best research
  **Wheri I first heard the hours
of the Summer Research Insti-
tute ... 1 thought that's way too
much time to ask. Yet there
were tiitnes that it got to be 9
o'clock and I wasn't really ready
to go home.  There were things I
wanted to get done, and I'd go
home and log on my computer
and I'd work until I couldn't
stay awake anymore."
     Jill Bisk, teacher, Bay City Central High School
four EarthVision
teams will receive
a scientific work-
station to be used
for the environ-
mental research at
the teams'
schools. Participa-
tion in the Satur-
day tutorials is not
a prerequisite for
the Summer
Research Institute
and all research
and education pro-
posals will have
equal opportunity
in the competition.
 NESC Annual Report - FY1993
                                                                                37

-------
   EarthVision: EPA's Grand Challenge for High Schools
    "For the longest period of time,
     I have wanted to exploit the
        computers' power. This
    program not only gives me the
    opportunity, but when I heard
    about the supercomputer plus
    experts coming in and giving
   shape to some of the learning I
    had, it gave me an idea about
    how to bring it all together. It
      was an opportunity I was
            excited about."
  Dan Sealey, teacher, Saginaw Center for the Arts and Sciences
      "I was kind of undecided
      between Pediatrics and
    Environmental Engineering.
  This program has helped me get
  a better idea of what's involved
  in dealing with the environment
  and the technology behind it."
     Kim Kukla, student, Bay City Central High School
 "In a lot of ways I believe we are
  all born scientists.... We try to
 figure out how the world works.
 Scientists try to carry it a little
 further and these kids are great
 examples of that.  They want to
 know HOW, how everything fits
  together and how they can go
   about understanding that!"
 Ken Flurchick, Chief Scientist, North Carolina Supercomputer
                Center
   Intern and Team Role
   Development

    The current teams have surprised the
   planning team by their amount of con-
   tact and level of participation. After the
   Summer Research Institute course com-
   pletion ceremony, the project planning
   team assumed they would have a break
   after three intensive weeks of activity.
   However, on the following Monday
   morning, the teams were back at SVSU
   working as diligently as they had during
  the Summer Research Institute. This
  has continued into the academic year
  even after school has started. They have
  demonstrated a strong desire to continue
  and maintain a very close relationship
  with the EarthVision project, the
  instructors, the mentors and the plan-
  ning team. They have helped the plan-
  ning team to redefine the role of
  participants. They wish to be a part of
  the 1993-94 Saturday Tutorial Iprogram,
  and they have moved easily from the
  role of student to that of colleagues and
  life-long learners. We plan to use
  teacher and student interns from past
  EarthVision teams to interface with new
  EarthVision teams and provide demon-
  strations to the general public.
   Research Interns (graduate and
 undergraduate students) have been
 working as part of the EarthVision  plan-
 ning team to help with the implementa-
 tion of the outreach program, assisting
 school teams in their research and cur-
 riculum plans, providing technical sup-
 port, and networking with mentors.  The
 Research Interns will also conduct their
 own environmental-computational-sci-
 ence research projects.

 Projected program growth

  The original plan for EarthVision
identified a fourteen county area around
the NESC from which three teams would
be drawn for Saturday tutorials and one
team for the Summer Research Institute
38
                                                  NESC Annual Report r FY1993

-------
                                  EarthVision: EPA's Grand Challenge for High Schools
                       Figure 4:1 EarthVision students work on their project.
  "One of the things over the
  years that we've noticed is
that math and science is very,
  very competitive. If you're
   going to enter math and
  science as a female, you are
  going to be competing with
  people internationally who
have a serious attitude about
learning, maybe more serious
 than some of the Americans.
 ... This project has just been
   a wonderful opportunity."
   Kathlene Sefcovic, parent of EarthVision [student
for the 1992-93 project year.  Three teams
were invited to participate in the 1993
Summer Research Institute with two actu-
ally participating. During the 1992-93
project year, grants were received from the
Michigan Department of Natural
Resources and the Michigan Department
of Education to train high school teachers
to use the EPA's STORET dataset in envi-
ronmental research to be conducted with
their students at their schools. The ten
schools that participated were each given
Zenith 486 computers with modems. This
project was designed to build capacity at
the schools by providing professional
development for teachers, building a com-
puting infrastructure and introducing
environmental research. As such, it was
designed to prepare schools for participa-
tion in EarthVision.
  The proposed 1993-94 project year
included three teams for the Saturday
 NESC Annual Report - FY1993
                                                                         39

-------
   EarthVision: EPA's Grand Challenge for High Schools
   tutorials and four teams for the Summer
   Research Institute drawn from the entire
   state of Michigan. The planning team
   decided to invite six teams to participate
   in the 1993-94 Saturday Tutorial program.
   The projection for the 1994-95 project year
   called for three teams in the Saturday
   tutorials and eight teams participating in
   the Summer Research Institute, drawn
   from the eight Great Lakes States. It is
   expected that six teams will participate in
   the Saturday Tutorial phase and plans are
   being formulated to design an urban
   EarthVision initiative in Cleveland, Ohio,
   that would become a national model for
   implementation of EarthVision in large
   city environments.

   What Have the Teams
  Accomplished in their Research?
    EarthVision teams have studied two
  questions that have social implication
 today. The Bay City Central High School
 team has been researching the possible
 role of zebra mussels in the accumulation
 of contaminants in Great Lakes fish. The
 Saginaw Center for the Arts and Sciences
 team has been researching the ^transport
 of heavy metal ion contaminants in Sagi-
 naw Bay. The two research papers detail-
 ing the approach of the teams, current
 accomplishments and future plans are
 included elsewhere in this Annual Report
(the Bay City team's report starts on
page 105 and the Saginaw team's report
starts on page 109).
40
                                                        NESC Annual Report:- FY1993

-------

-------

-------
                  oOlobai Climate Change Imparts   it
Abstract

  Mathematical models have been devel-
oped for estimating the effects of climate
change (changed meteorological condi-
tions) on lake and stream thermal struc-
ture and dissolved oxygen concentrations
and on fishery resources. Regional impact
analyses require the development of lake
and stream classification systems to define
waterbody types, which in turn require
the availability of extensive regional data
bases for these resources. Fishery ;
resource response predictions require the
development of large field temperature
and fish distribution data bases from
which species and guild thermal require-
ments can be derived. Supercomputing
capabilities are being utilized in the devel-
opment and manipulation of the large
data bases to integrate the various? data
and program modules, and to make the
calculations required to perform regional
impact estimates.               i
 EPA Research Objectives
                               ; i
   According to a 1979 report by the U.S.
 National Academy of Sciences, and sup-
 ported by several general circulation mod-
 els of ocean atmosphere heat budgets,
 doubling atmospheric concentrations of
 CO2 could increase global mean air tem-
 peratures by 1.5° to 4.5° Celsius in: the
 next 100 years. This is likely to have
 many environmental consequences, such
 as changes in water temperature and dis-
 solved oxygen concentrations, which in
 turn are likely to affect fish populations.
 The fact that such changes are occurring
 many times faster than expected has
 resulted in requests for information on the
 causes, effects, and response options to the
projected climate changes. The Environ-
mental Research Laboratory - Duluth and
the University of Minnesota have initiated
a cooperative study,to determine the
impacts of global warming on lake and
stream environmental conditions and fish-
ery resources. In order to continue with
this study, fish thermal requirements need
to be estimated using a historical fish
presence/temperature record data base.


Background/Approach

  The Fish Temperature Distribution
Management System (FTDMS) is  a
national data base system that spatially
and temporally associates discrete fish
sample records with water temperature
data. Recent efforts have concentrated on
the expansion of data base content by
assembling information from a multitude
of sources, including federal agencies (e.g.
EPA/STORET and USGS) and private
museum and university collections.  The
assimilation of data from many sources
necessitates the need for automated spa-
tial and temporal matching of a fish record
with water temperature data. Prior ver-
sions of several program modules have
been converted from a PC data base plat-
form to C and have been run on the
NESC's supercomputer.  Program perfor-
mance has been closely monitored and
 several steps are being taken to increase
 performance.


 Scientific Accomplishments

   A modeling approach has been  devel-
 oped for estimating the effects of global
 warming on lake and stream environmen-
 tal conditions and fisheries resources.  The
  NESC Annual Report - FY1993
                                                                                43

-------
   Estimation of Global Climate Change Impacts on Lake and Stream Environmental Conditions and Fishery Resources
   initial phase of the work was partially
   supported by the EPA Office of Policy,
   Planning, and Evaluation, and the results
   are currently being applied to an economic
   impact analysis of global climate impacts
   on the United States.  One ongoing pro-
   gram is a component of the FCCSET Com-
   mittee on Earth and Environmental
   Sciences, Global Climate Research Pro-
   gram. At last count, more than ten publi-
   cations, most in the form of technical
  journal articles, had been produced by the
   project.


  Results

    The program module that calculates
  raw temperature data into weekly mean
  values has been run on the NESC's
  supercomputer for sixteen states. The
  results have been used to calculate the
  maximum (warmest throughout the year)
  95th percentile temperature where a fish
  species was collected for 32 species of
  North American freshwater fish. This
  temperature is used as an approximation
 of the lethal limit for that species and
 allows us to estimate the distribution of
 fish after global climate change.
   The speed by which the weekly mean
 temperatures are calculated on the super-
 computer makes it possible to perform the
 temporal matching in a number bf differ-
 ent ways. This has the potential for
 improving estimates of thermal require-
 ments for fish.  For example, the southern
 range of distribution of cool-water fish is
 generally near 40° latitude. Maximum
 weekly mean values from south of this
 parallel would be expected to provide a
 better estimate of thermal tolerances than
 values from all of North America.  Data
 manipulation on the basis of geographic
 regions is, therefore, desirable.! Temporal
 matching criteria can be restricted in
 other ways (fish and temperatures sam-
 pled in the same year, season, or month).
 Comparing "monthly" and "yearly"
 datasets provides a means of examining
the importance of the temporal relation-
ship  of temperature and fish records.
                                      Time (months)
                                                                              110
                             '16    20"   '24 ' ' ''as"' '  82"

                                     Time (weeks)

                       Figure 1: Graphic representation of raw temperature data
44
                                                         NESC Annual Report - FY1993

-------
Estimation of Global Climate Change Impacts on Lake and Stream Environmental Conditions and Fishery Resources
Future Objectives

  Currently, the weekly mean tempera-
ture is used to describe surface water con-
ditions for the week in which a fish sample
was taken. In the near future, the weekly
maximum temperature will be calculated
and daily means and daily maxima1 will be
matched to fish collections to re-calculate
the maximum 95th percentile tempera-
ture. These values will provide a unique
and valuable look at the relationship
between various expressions of a fish's
thermal regime and its geographic' distri-
bution. Most laboratory-derived mpasures
of thermal tolerance, the source of most
past temperature effects information,
have employed constant temperature
exposure conditions when short-tfsrm
peaks might be as much or more impor-
tant in nature. So far, studies of fish ther-
mal requirements and global climate
 effects have been focused on species in the
 central United States. A template has
 been developed for extending these analy-
 ses to other regions of the United States
 and beyond. The relationship of cpld tem-
 peratures to the distribution of warm-
 water fishes has only been examined
 superficially.  The data storage and manip-
 ulation requirements and modeling
 demands will be greatly increased, just for
 dealing with this single environmental
 component (fishes). Research is underway
to incorporate functional ecosystem
responses (e.g. system productivity) into
models projecting climate change
impacts. The area, of ecological processes
and effects research, only as related to
aquatic ecosystems, is obviously huge and
can benefit greatly from enhanced compu-
tational capabilities.


Publications

  Hokanson, K.E.F., B. Goodno, and J.G.
Eaton, Evaluation of field and laboratory
derived fish thermal requirements for glo-
bal climate warming impact assessment,
USEPA ORD "A" milestone report, p. 56,
 1990.
  Hondzo, M. and H.G. Stefan, Water tem-
perature characteristics of lakes subjected
 to climate change., University of Minne-
 sota, St. Anthony Falls Hydraulic Lab,
 Project Report No. 329, p. 156, 1992.
   Stefan, H.G., M. Hondzo, B. Sinokrot, X.
 Fang, J.G. Eaton, B.E. Goodno, K.E.F.
 Hokanson, J.H. McCormick, D.G.
 O'Brien, and J.A. Wisniewski, A methodol-
 ogy to estimate global climate change
 impacts on lake and stream environmental
 conditions and fishery resources with
 application to Minnesota, University of
 Minnesota, St. Anthony Falls Hydraulic
 Laboratory, Project Report No. 323, p. 222,
 2nd edition, March, 1992.
  i J G Eaton U.S. EPA, ERL-Duluth, 6201 Congdon Blvd., Duluth, MN 55804, 218-720-5557.
  2 H G Stefan St. Anthony Falls Hydraulic Lab, Dept. Civil & Mineral Eng., University of Minnesota,
     kssissippiR.& 3rd Ave.S.E., Minneapolis, MN 55414-2196, 612-627-4010
  3 D. G. O'Brien, Computer Sciences Corporation, ERL-Duluth, 6201 Congdon Blvd., Duluth, MN 55804, 218-
     720-5718.
  NESC Annual Report - FY1993
                                                                                    45

-------
For at least two million years men
   have been reproducing and
multiplying on * little automated
        Spaceship Egrttt,

-------
 Modeling Sediment mn& ContamteiaM ^lleaaspm*
           and Contaminant ^ransport Modeling, anal Fox Etaer / Green Bay
Overview                    ;
  Contamination of sediments by toxic
chemicals is a problem common to most of
the forty-two "Areas of Concern" identified
in the Great Lakes. Assessment a.hd
remediation of contaminated sediments
have become high priorities for applied
research within EPA, as State and Federal
agencies struggle with remediation deci-
sions involving great uncertainty, poten-
tial significant risks, and high cost.
Through involvement in the Assessment
and Remediation of Contaminated Sedi-
ments (ARCS) and Green Bay/Fox; River
Mass Balance Studies, LLRS and USCB
are developing and applying models for
understanding and predicting the trans-
port and fate of contaminated sediments.
These models are being applied to ;support
remediation decisions in several major
Great Lakes tributaries suffering exten-
sive sediment contamination. The models
will be used as tools to determine how var-
ious hydrologic events may affect contami-
nated sediments and to determine the
potential effects of proposed remediation
projects. The development of suet models
will offer predictive tools appropriate for
assessing the remediation of contaminated
sediments in the Great Lakes and nation-
wide.                         I
   Numerical models of the transport and
fate of sediments and hydrophobic con-
taminants in the Buffalo and Saginaw riv-
 ers and the Fox River/Green Bay system
have been developed to achieve this
 objective. The models are used tcrrun real
 time simulations of high flow events as
 well as low flow periods. The models pre-
 dict the movement of contaminants in the
river and the resuspension of in-place
pollutants from the sediment bed.  They
also predict the movement of pollutants
from the rivers into the Great Lakes. The
Buffalo, Fox and Saginaw Rivers are each
major Great Lakes tributaries; they were
chosen for modeling because of their dif-
ferences. The Buffalo River is a small,
narrow, winding river that has been
dredged from bank to bank to allow large
ship movement. This dredging has signifi-
cantly increased the cross sectional area of
the river in most places. Much of the time
the river is almost stagnant due to flow
rates less than 1 m3/s, with flushing of the
river dominated by seiche motion.  The
few high flow events contribute the major-
ity of sediment movement. The  Fox River
is controlled by a series of dams, which
stabilizes the flow response to hydrologic
events. Pools created above the dams also
function as reservoirs for contaminated
sediments. The Saginaw River is a much
larger, wider, more dynamic river.
Although dredged to allow shipping, this
dredging has not significantly increased
the cross sectional area of the river in
most areas. The median flow rates and
velocities are much greater on the
Saginaw.
   The area around each river is urbanized
and heavily industrialized. For many
years, industrial and municipal wastes,
including PCBs, PAHs, and heavy metals
have been dumped into these rivers.
These toxics have deposited in the river
bed and are now buried deep in the river
 sediments. Resuspension and transport of
 these contaminants is of primary con-
 cern. Many contaminants, such as PCBs,
 adhere to fine-griEiined sediments. There-
 NESC Annual Report - FY1993
                                                                                47

-------
   Modeling Sediment and Contaminant Transport and Fate in Aquatic Syste
                   ms:
   fore, prediction of the erosion, transport
   and deposition of sediments is essential
   for predicting the movement of contami-
   nants. Screening-level model calculations
   conducted in the Fox River indicate that
   bottom sediment resuspension is, in fact,
   the largest source of PCBs to the river
   during floods.

   Research Objectives

    The purpose of this study is to develop
   quantitative, predictive models describing
   the transport of fine-grained sediments
   and associated hydrophobic contaminants
  in the Buffalo and Saginaw rivers and the
  Fox River/Green Bay system. These mod-
  els provide the high spatial and temporal
  resolution necessary to predict the vari-
  ability and dynamics inherent to sedi-
  ment and associated contaminant
  transport.  The models also estimate the
  transport of sediments and contaminants
  from the rivers into the Great Lakes.
  Such estimates are essential to predict the
  impact of contaminated sediments and
  their remediation upon receiving water
  quality.

  Approach

   The SEDZL model, a numerical model of
  the resuspension, deposition, and trans-
 port of fine-grained cohesive sediments
 and associated contaminants, is used in
 this study.  The model consists of two-
 dimensional, vertically integrated, time
 dependent hydrodynamic and transport
 submodels coupled with a three-dimen-
 sional, time dependent submodel of the
 sediment bed and its properties. The
 three-dimensional sediment bed model is
 necessary to accurately model bed proper-
 ties which vary with age and depth as well
 as location beneath the river.  Contami-
 nant partition coefficients, sediment set-
 tling speeds, sediment resuspension
 parameters  and sediment bed properties
 needed in the model were determined from
 laboratory and field tests. The need to
   determine in situ sediment bed. properties
   has led to the development of a1 portable
   flume for measuring the resuspension
   properties at depth in cores collected from
   the sediment bed. Measurements of flow
   rates, suspended solids, and PCB concen-
   trations from each river were used to esti-
   mate input loadings for the model, while
   data from sediment core samples were
   used to develop initial sediment bed con-
   taminant concentrations.

  Accomplishments

    Model simulations for each river were
  performed to predict sediment and con-
  taminant transport over numerous time
  periods varying from three to six months
  in duration. The sediment deposition/ero-
  sion patterns were then compared to
  changes in bathymetry measured at
  transects across each river at the begin-
  ning and end of each time period. By com-
  paring the bathymetry measurements and
  model predictions for several time periods,
  as well as by comparing total suspended
  solids (TSS) predictions to data at the
  river mouth, we were able to calibrate and
  verify the models for sediment transport.
  Contaminant modeling predictions are
  also being analyzed and compared to exist-
 ing data, especially on the Fox River
 where extensive PCB water column mea-
 surements were conducted.
   Analysis of the transect measurements
 indicates that very little sediment trans-
 port occurs in the rivers during low flow
 periods and almost all erosion and deposi-
 tion occurs during high flow events.
 Therefore a method was developed for
 running long term simulations in which
 each event was modeled in detail ;and the
 low flow periods were modeled statisti-
 cally.  This greatly decreased the CPU
 time necessary for the three to six month
 simulations and allowed us to simulate
 the much longer periods that are relevant
 for persistent, in-place pollutantsJ Cur-
rently we are modeling 25 year scenarios
48
                                                        NESC Annual Report - FY1993

-------
                   Modeling Sediment and Contaminant Transport and Fate in Aquatic Systems:
to determine the long-term decrease in
load of contaminants from each river and
the erosion and burial of contaminants in
the sediment bed. These scenarios will
then be modified to test the effectiveness
of proposed remedial alternatives, such as
dredging or capping of contaminated sedi-
ments, or discontinuation of navigation
dredging.                      ;
                                I
                               >i
Cray Usage                   ;
  To accurately predict sediment iiad con-
taminant transport on a river requires
both a complex set of equations arid a fine
grid. Both the Saginaw and Buffalo rivers
require 900 grid elements to accurately
define the river; over 2000 grid elements
were required for the Fox River. This
makes the SEDZL code computationally
intensive. Workstations like the Sun
SPARCstation 2 require several CPU
hours to simulate one day. Therefore, the
only realistic way to do long term simula-
tions (even with statistical averaging of
low flow periods) is to use a vectorized
 code on a supercomputer. All long term
 simulations were done on the NESC Cray,
 with only short duration test rum=j being
 done on workstations.  The CPU time nec-
 essary for a 1-day simulation on, for exam-
 ple, the Buffalo River, is 0.3 CPU hours. A
 three-month simulation without averag-
 ing of low-flow conditions would require 30
 CPU hours. By using the statistical aver-
 aging procedure, the same simulation can
 be reduced to only 4-5 hours.  The 25-year
 forecasts, with statistical averaging,
 require about 70 CPU hours. This is a
 considerable quantity of supercomputer
 time; however, this is the relevant time
 scale for evaluating transport and fate of
 persistent toxic chemicals. Such computa-
 tions would be virtually impossible with-
out supercomputers.

Future Plans
  Our research team will continue to ver-
ify the sediment and contaminant trans-
port model predictions on the Buffalo and
Saginaw Rivers, and the Fox River/Green
Bay system, using existing data as well as
information gathered from additional
sampling. In the Fox River, the model is
being used to identify and prioritize sedi-
ment sampling efforts for Wisconsin
Department of Natural Resources.  Also,
long-term forecasts will be performed with
the models for specific remediation design
alternatives to assist decision makers.
  Our plans also include modeling sedi-
ments and contaminants in Green Bay
and Lake Michigan using a three-dimen-
sional version of the SEDZL model. The
ultimate goal of these efforts is to develop
and demonstrate a predictive model that
can be applied anywhere in the Great
Lakes, which can be run with minimal
field data from the area of study. The
model will be based on an accurate under-
standing and description of sediment and
contaminant transport and fate processes,
the former including particle settling,
 aggregation/disaggregation, and resuspen-
 sion. For PCBs, the contaminant pro-
 cesses include sorption, sediment/water
 and air/water transfer.

 Publications:
   Gailani, Joe, C. Kirk Ziegler, and Wil-
 bert Lick (1991) Transport of Suspended
 Solids in the Lower Fox River, Journal of
 Great Lakes Research. 17(4), pp.
 479-494.
  1 Douglas Endicott, U.S. EPA- Large Laikes Research Station, Grosse lie, MI.
  2 Mary Cardenas, Kirk Freeman, and Gilbert Lick, Department of Mechanical and Environmental Engineering,
    University of California, Santa Barbara, CA.                       ,„,.,-,    Ti n/rr
  3 Joseph Gailani and Mark Velleux, A S & I at U.S. EPA - Large Lakes Eesearch Station, Grosse He, MI.
  NESC Annual Report - FY1993
                                                                                   49

-------
                44
  The only solid pieee of scientific
  truth about which I feel totally
confident is that we are profoundly
  ignorant about nature. It is this
  sudden confrontation with the
depth and scope of ignorance that
  represents the most sfgnilieamt
contribution of twentSetli^OTtory
  science to the human Intellect.
           The Medusa and the Snail [1979]

               On Cloning a Human Being

-------
 	___ ^	f €ali&Fatiub-
stances (OPPTS) during the regis tration of
pesticides and industrial chemicals and by
risk managers for predicting changes in
ecological risk associated with watershed
management options.           ;    ;
  The NESC supercomputing facilities
will be used for three tasks:     |    ;
 1- Calibrate the parameters of the Littoral
   Ecosystem Risk Assessment Model
   (LERAM). This is done using an opti-
   mization algorithm to minimize a func-
   tion that measures the difference
   between the LERAM simulation of pop-
   ulation biomasses and biomass data
   from a littoral ecosystem.         \
 2- Create a user-friendly interface for
   LERAM that will allow the user to;
   make predictions of ecological risk to a
   littoral ecosystem from exposiire to
   specified stressors and display the
   results hi real tune.         ;
 3- Evaluate LERAM by simulating the
   effects of a number of concentrations of
   selected chemical stressors arid com-
   paring these simulations to field d£ta.
   The supercomputing environment will
  be used to run 500 (or more) Monte
  Carlo iterations of the model at each
  treatment concentration for the pur-
  pose of sensitivity and uncertainty
  analysis!

EPA Research Objectives
  At the present time, laboratory tests
and mathematical models form the basis
of the ecological risk assessment paradigm
used by the United States Environmental
Protection Agency (EPA). Until the early
1980s, single species tests were used
almost exclusively to provide hazard
assessments of chemicals.  At that time,
the National Academy of Sciences (1981)3
and others (Levin and Kimball, 1984> doc-
umented the need for supplementary
information from field tests, microcosm
experiments, and mathematical models to
better assess chemical hazards for differ-
ent geographic regions, seasons, applica-
tion methods, spatial scales, and levels of
biological organization.  Along with the
increased interest in using field tests,
microcosm experiments, and mathemati-
cal models to predict system responses to
perturbations, it became apparent that lit-
tle was known about the accuracy of pre-
dictions made by these techniques.  The
EPA's objectives for the research proposed
here include evaluating and refining one
ecological risk assessment technique using
field data from controlled experiments in
natural systems.

Background/Approach
   In order to reach the EPA's objectives,
 Lake Superior Research Institute (LSRI)
 and Environmental Research Laboratory -
 Duluth (ERL-D) researchers began the
 NESC Annual Report - FY1993
                                                                                51

-------
   Development, Calibration and Evaluation of a User-Friendly Littoral Ecosystem Risk Assessment
   development of the Littoral Ecosystem
   Risk Assessment Model (LERAM) in June,
   1989. LERAM is a bioenergetic ecosystem
   effects model that links single species tox-
   icity data to a bioenergetic model of the
   trophic structure of an ecosystem in order
   to simulate community and ecosystem
   level effects of chemical stressors. It uses
   Monte Carlo iterations of these simula-
   tions in order to calculate probabilistic
   ecological risk assessments of chemical
   stressors. To date, LSRI and ERL-D
   researchers have developed LERAM to the
   point where it models the unperturbed
   behavior of a littoral ecosystem (i.e., the
   "behavior" of control communities), and
   the response of that system to the insecti-
   cide chlorpyrifos, with a high degree of
  both accuracy and precision (Hanratty and
  Stay5, submitted to J. Appl. Ecol.; Han-
  ratty et al.8,1992). Current work using
  data from the 1988 esfenvalerate study
 (Lozano et al., 1989)6 appears even more
 promising. Further work is required, how-
 ever, to continue refinement of the model,
 to validate its output using data from
 other littoral enclosure studies, and to
 develop more accurate model predictions.
   The LERAM parameters haveibeen cali-
 brated to the control enclosures from two
 different littoral enclosure field experi-
 ments (Brazner et al., 19887; Lozano et al.,
 1989)6 using the North Carolina Super-
 computing Center's (NCSC) Cray. The
 simulations resulting from these;new
 parameter sets are much closer to the field
 data than the simulations earlier calibra-
 tion methods produced.  In many'cases,
 the model simulation overwrites the field
 data (figure 1, page 52).  The NCSC per-
sonnel worked on improving the yectoriza-
tion of the LERAM code by vectorizing
over the Monte Carlo iterations of the
                        Predaceous  Copepoda
                                  1986 Control
           0.2
             O  -
         -0.2
              160   170   180   190   2OO   210   220  230
                                  Julian Day
         i  ? f rfT    r6  °;amed by Calibratm&the LERAM parameters using field data. The solid line^epre-
   the field ThlafiST-SS/r Predaceou? C°PeP°da ™d the S9ua«* *ith ^or bars represent the biomass measured in
   the field. The field data 1S the mean predaceous Copepoda biomass in the control enclosures in the chlorpyrifos littoral
                              enclosure study (Brazner, et al., 1988).                       :
62
                                                           NESC Annual Report - FY1993

-------
   Development, Calibration and Evaluation of a User-Friendly Littoral Ecosystem Risk Assessment
                                                                            Model
model. In the future, we will work on
improving the efficiency of the calibration
algorithm and continue to improve the
vectorization of the LERAM code as much
as possible.                     |

Comparison with pre-Cray Results
  When the calibration program is run on
other computers, such as a DEC VAX or a
486 PC, it takes so long that the program
becomes impractical to use. The model
itself can be run on other computers, but
one loses the advantage of being able to sit
and wait about a minute for the results,
which allows the user to concentrate on
the ecological risk assessment or modeling
problem at hand.  When the simulations
are performed on other computers, the
model can take anywhere from 20 minutes
to three hours, depending on the computer
and the number of Monte Carlo iterations
performed.                     ',
                               \
Future Objectives           i
   The following tasks are proposed for
evaluating and refining LERAM using the
NESC supercomputing facilities..
   Fiscal year 1994:
 1- Port the LERAM code and input files  to
   the NESC supercomputer and establish
   communication with NESC personnel.
 2- Enhance the vectorization of the
   LERAM code in the supercomputing
   environment.                ;
 3- Investigate methods for improving the
   minimization algorithm that is used to
   calibrate the LERAM parameters.
 4- Use the above minimization algorithm
   to find the best parameter sets for
   representing the control experimental
    ecosystems measured in various field
    studies.                    ;
 5- Simulate the effect of a number of
    chemical stressors and evaluate the
    model's accuracy using field data from
    littoral enclosure and pond studies.
  Fiscal years 1995-96:
1- Create user-friendly interfaces for
  LERAM.
2- Calibrate the LERAM parameters
  using the control experimental units in
  future field studies.
3- Simulate the effect of the chemical
  stressors used in future field studies in
  order to further evaluate LERAM and
  define its domain of application.

Relevant Reports and Publications
  Bartell, S.M. 1987. Technical Reference
and User Manual for Ecosystem Uncer-
tainty Analysis (EUA):
1- The Pascal PC Demonstration Pro-
   gram.
2- The Standard Water Column Model
   (SWACOM).
3- The Comprehensive Aquatic System
   Model (CASM). U.S. EPA Office of
   Toxic Substances Report.
  Bowie, G.L., W:B. Mills, D.B. Porcella,
C.L. Campbell, J.R. Pagenkopf, G.L. Rupp,
K.M. Johnson, P.W.H. Chan, S.A. Gherini,
Tetra Tech, Inc. and C.E. Chamberlin,
1985. Rates, Constants, and Kinetics For-
mulations in Surface Water Quality Mod-
eling (Second Edition). EPA 600/3-85/
040. U.S. Environmental Protection
Agency, Athens, Georgia.
  Hanratty, M.P., RS. Stay, and S.J. Loz-
ano.  1992. Field Evaluation of LERAM:
The Littoral Ecosystem Risk Assessment
Model, Phase I.  U.S. EPA Environmental
Research Laboratory-Duluth, Minnesota.
  Hanratty, M.R and S.J. Lozano. In
 prep. Field evaluation of modeling and
 laboratory risk assessment methods.
   Press, W.H., B.P Flannery, S.A. Teukol-
 sky, W.T. Vetterling.  1989. Numerical
 Recipes in Pascal:  The Art of Scientific
 Computing. Cambridge University Press,
 New York.
  NESC Annual Report - FY1993
                                                                                 53

-------
   Development, Calibration and Evaluation of a User-Friendly Littoral Ecosystem Risk Assessment
                         WT          I               > University of Wisconsin-superior WI 54880. Phone:
                   .  Email (Internet): mhanratt@uuisuper.edu                                 '
                                          ReS6arCh Lab°ratory-Duluth> 6201 0««*» Blvd., Duluth, MN
                                                         ^^ on Ecosystems. National Academic

     L^'S'A-andK-D-Kimba11- MM- New Perspectives in Ecotoxicology. Environmental Management 8:375-

   5  ^h*^
   6  L0'Ha°lIoS;f " Jq°T' nra?f r> ™%*n?&> L'J- Heinis' KW- Sargent, D.K. Tanner, L.E. Anderson, S.L.
      PCSSrc««Sn  ^ A !n>  f^ ren/ E
-------
 Cli*fQjjitft? witlitteAaBlstieally
                            deling
Background                  :!
  In the field of environmental to^dcology,
and especially aquatic toxicology, first gen-
eration interpretive methods, such, as
Quantitative Structure Activity Relation-
ships (QSARs), have developed as^scientif-
ically credible tools for predicting the
ecological effect and fate of chemicals
when little or no empirical data are avail-
able. The proper application and contin-
ued acceptance of these techniques
requires that methods and models be
developed to systematically assign chemi-
cals, with an estimate of certainty, to the
appropriate QSARs. For example, failure
to use the correct model to predict the tox-
icity of a given compound can result in a
toxicity prediction error of 10 to 1000. In
addition to using quantitative models,
analog selection techniques are also
employed whereby data associated with
"structurally-similar" chemicals are used
to estimate risk levels of compounds for
which no data is available. Failure to
identify and select correct structiiral ana-
logs can also lead to order of magnitude
errors. The lack of sophisticated tech-
niques to select structural analogs and
 appropriate quantitative relationships
 represents a significant area of uncer-
 tainty in the ecological risk assessment of
 chemicals.                     :
                               l
   As the  use of first generation interpre-
 tive models have become well established
 in ecological hazard identifications, their
 use in ecological hazard and exposure
 assessments has increased. Coupled with
 these new application issues, it is neces-
 sary to address several areas of research
 to provide the scientific basis whereby sec-
 ond generation methods can be applied to
 these emerging needs.  Current QSAR
methods reliably estimate acute lethality
and chronic no-effect levels for 75 to 85
percent of the industrial chemicals; how-
ever, these methods underestimate the
risks of reactive chemicals by several
orders of magnitude. Consequently, QSAR
methods must be developed to accurately
predict the toxicity of electrophiles and
free radicals. In addition, the current
inability to quantify xenobiotic metabo-
lism can result in significant error when
predicting toxicity. For example, failure to
identify and quantify metabolic activation
can lead to toxicity predictions that under-
estimate hazards of xenobiotics to aquatic
organisms by several orders of magnitude.
Conversely, typical bioconcentration mod-
els, which assume no xenobiotic metabo-
lism, can result in estimates that are
overly conservative by several orders of
magnitude.

Research Objectives
   To reduce uncertainties in ecological
risk assessments; of chemical stressors, a
second generation of advanced predictive
modeling techniques is required.  The sec-
ond generation of models must be based on
fundamental principles of chemistry, bio-
 chemistry  and toxicology, and designed in
 such a way to efficiently assess the thou-
 sands of chemicals in commerce. Research
 must be directed towards establishing
 quantitative means of assessing chemical
 similarity  to reduce uncertainties in
 selecting chemical analogs and QSARs.
 Research must also be directed towards
 the development of mechanistically-based
 QSARs for reactive chemicals to improve
 toxicity predictions and estimates of
 metabolism.
  NESC Annual Report - FY1993
                                                                                   55

-------
                                                                          Predictive
   Approach

     Based on the areas of uncertainties pre-
   viously described, two major projects are
   being undertaken. The research within
   these projects is designed to resolve criti-
   cal scientific concepts that will result in
   the development of advanced interpretive
   methods for assessing the ecological risk of
   chemical stressors. The overall approach
   is designed to assess hypotheses by inte-
   grating computational and theoretical
   chemistry with toxicological principles.
   Through this integration, advanced
   QSARs for reactive toxicants  and xenobi-
   otic metabolism will be developed, as well
   as techniques to quantify chemical simi-
   larity in the context of toxicological prop-
   erties.

    For well established physico-chemical
  and ecotoxicological databases maintained
  within the U.S. EPA Environmental
  Research Laboratory-Duluth (ERL-
  Duluth) QSAR system, a variety of simi-
  larity metrics and artificial intelligence
  systems, including neural networks, are
  being evaluated. A critical component in
  this research effort will be to develop
  approaches whereby variation in a phys-
  ico-chemical or toxicological property can
  be related to variability in a given distance
  metric used in a multi-dimensional chemi-
  cal structure space. An important area of
 investigation has been to determine the
 means whereby chemicals with common
 modes of toxic action can be classified.
 Efforts to date have been addressing the
 use of neural networks as a complement to
 a currently used approach whereby a class-
 sic expert system, based on substructural
 rules, is used to assign chemicals to a
 given mode of toxic action for analog or
 QSAR selection.
   Concurrent with establishment of topo-
 logically-based chemical similarity
 approaches, efforts have been initiated  to
 generate global stereoelectronic parame-
 ters for chemicals within the industrial
   chemical inventory. Initial efforts have
   focused on specified subsets to assess com-
   putational needs to generate atomic and
   molecular parameters, which will be fol-
   lowed by an intensive computing exer-
   cise.  Generation of these data will be
   combined with topological indices
   described above to re-evaluate chemical
   similarity, especially in the context of
   assessing the probability of metabolic
   transformations and the toxicity iof reac-
   tive chemicals. Initial chemicals;selected
   for study have included those currently in
   the ERL-Duluth ecotoxicity database and
  additional compounds for which reliable
  metabolic maps and rate constants are
  available. Using these chemical datasets,
  exploratory QSARs are being established
  to evaluate the nature of global and local
  parameters that are needed to identify
  reactive centers for predicting mejtabolism
  and toxicity.

  Results and Future Objectives
   Neural Network^; During the past year
  a prototype PC-based neural network,
  designed to classify chemical structures by
  modes of toxic action, was installed on the
 National Environmental Supercomputer.
 Modifications to the software are currently
 underway to improve computing efficiency.
 After completion of these software modifi-
 cations, additional efforts will be under-
 taken to explore attenuation algorithms,
 to incorporate Monte-Carlo subsanipling
 techniques, and further develop training
 algorithms that are consistent with
 evolving chemical similarity metrics. As
 these research findings progress, neural
 networks will be employed to characterize
 and assess the industrial chemical inven-
 tory in the context of ecotoxicological end-
 points.                          ;

   Chemical Reactivity!  As previously dis-
 cussed, the broad objective of this effort is
to establish an efficient and mechanisti-
cally-relevant means  to incorporate stere-
56
                                                          NESC Annual Report - FY1993

-------
Integration of Computational and Theoretical Chemistry with Mechanistically-Based Predictive
Ecotoxicology Modeling             •
oelectronic parameters in the next'
generation of QSAR models.  This effort is,
in part, addressing the means whereby the
three dimensional structures of chemicals
currently in commerce (ca. 60,000 com-
pounds) can be quantified and stored for
subsequent modeling. During the past
year, a subset of 50 chemicals have been
used to examine techniques whereby
approximated three dimensional struc-
tures can be ported to the National Envi-
ronmental Supercomputer, for    ;
subsequent molecular structure optimiza-
tion using Gaussian-92. This effort is
establishing a procedure to generate
molecular structures for large files of com-
pounds and is providing estimates of the
resources needed for large scale   |
calculations.                    ;
     In related efforts, studies have been
undertaken to explore specific tosicologi-
cal processes and associated chemical
reactivity parameters to establish a mech-
anistically-based approach to screen com-
pounds and stereoelectronic indices and
thereby focus future three-dimensional
calculations.  Based on hypotheses con-
cerning toxic mechanisms and metabolic
activation pathways, several studies have
been undertaken to initially explore the
use of stereoelectronic descriptors to iden-
tify potentially reactive toxicants.
Descriptors of soft electrophilicity and one
electron reduction potentials have been
calculated for a diverse group of aromatic
compounds and used to discriminate the
narcosis mode(s) of toxic action from
mechanisms associated with covalent
binding to soft nvncleophiles in and oxida-
tive stress, respectively. These studies are
providing some insights into how a mecha-
nistically-based strategy may be devel-
oped for selecting and using electronic
indices in QSARs for biochemical and cel-
lular toxicity.
 i  Steven Bradbury and Oilman Veith, U.S. EPA, Environmental Research Laboratory - Duluth, 6201 Congdon
    Blvd., Duluth, MN 55804.         ;                                           .
 2  Ovanes Mekenyan, Lake Superior Research Institute, University ofWisconsin-Superior, Superior,Wl.
 3  Robert Hunter, Natural Resources Research Institute, University of Minnesota-Duluth, Duluth, MN.
 4  Eric Anderson, Computer Sciences Corporation, Duluth, MN.
  NESC Annual Report - FY1993
                                                                                      57

-------
  It is false dichotomy to think of
 nature and man*  Mankind is that
 factor in nature whieh exhibits in
its most intense form the
           \of nature*
                    : '-British philosopher

-------
                                n Jtifeffififtt IMterpretaiion of
             Camera Cbiiposltion **
Abstract
  The efficacies of inhaled pharmacologic
drugs in the prophylaxis and treatment of
airway diseases could be improved if parti-
cles were selectively directed to appropri-
ate sites. In the medical arena, planar
gamma scintillation cameras may; be
employed to study factors affecting such
particle deposition patterns withiii the
lung. However, the value and versatility
of such instruments are compromised by
the limited resolution of their images.
Specifically, it is not possible to determine
the composition of their central (0) or
large airway, intermediate (I), and periph-
eral  (P) or small airway zones. Herein, an
analytical model is presented to assist the
clinician in the systematic analysis and
interpretation of gamma camera images.
  Using a Cray Y-MP supercomputer,
human lung morphologies have been
mapped to function as templates to be
superimposed upon scans. The model is
intended to complement laboratory regi-
mens by providing a heretofore unavail-
able method to define the C, I, and P zones
of the human lung on an  airway genera-
tion-by-generation basis.  A quantitative
value can now be assigned to the degree of
overlapping that exists in the images. For
example, in the lung morphology Consist-
ing  of 16,777,215 airways (total), the C
zone itself may contain 1,608,24(5 airways
 of which 1,595,940, or more than|99%, are
 alveolated airways. By identifying compo-
 sition our intent is to integrate the model
 into future aerosol therapy protocols and
 thereby assist procedures for targeted
 delivery of airborne Pharmaceuticals.
Introduction
  Numerous clinical studies have been
performed in attempts to quantitate the
deposition patterns of radiolabeled test
aerosols and pharmacologic drugs as func-
tions of: (1) aerosol characteristics (e.g.,
particle size and density, the degree of
polydispersity) and (2) ventilatory param-
eters (e.g., tidal volume, breathing fre-
quency, breath-hold time).  The
experimental investigations (Barnes, et
al., 1978s, Newman et al., 19916, Ruffin et
al., 19787, and Smaldone et al., 19898)
have demonstrated that the efficacies of
inhaled pharmacologic drugs could be
enhanced when particles are  targeted to
regional areas of the lung ("large bronchial
airways", "small central airways", etc.).
  The benefits of such selective deposition
processes may be further improved with
the increased spatial resolution of deposi-
tion among the airways of the lung; that
is, if drug delivery were to be  attempted on
a more focused basis, perhaps even target-
ing airway bifurcations.9'10 Attempts to
relate deposition and response, indirectly,
have been made via pulmonary function
testing.  However, direct observations of
deposition patterns can be achieved via
gamma scintillation cameras, and PET
 and SPECT instrumentation.
    The clinical observations cited above
 may be explained, at least partially, in
 terms of the spatial dispersions of recep-
 tors and nerve endings within the lung.
 Their respective distributions could vary
 from being heterogeneous (e.g., localized)
 to  homogeneous (e.g., uniform). Let us
 address the former case. The roles of
 NESC Annual Report - FY1993
                                                                                 59

-------
   The Supercomputer in Medicine: Interpretation of Gamma Camera Composit:
                     ion
   afferent neural pathways were studied by
   Karlsson et al9. They determined, for
   example, that upper tracheobronchial (TB)
   bifurcations were very sensitive to cough
   stimulation. The results suggested that
   rapidly adapting stretch receptors (RAEs)
   could function as "cough receptors" com-
   mensurate with their relatively high con-
   centrations in the upper TB airways and
   the superficial locations of RARs within
   the mucosa. This would be consistent with
   the documented fast blocking effects of
   topically applied and aerosolized anaes-
   thetics.

    Let us now consider the second case.
  The first mappings of p-receptor distribu-
  tions within the lung were reported by
  Barnes et al.5 Using autoradiography
  techniques, they detected a dense labeling
  of smooth muscle that was greater in the
  smaller bronchioles than the larger (i.e.,
  cartilaginous) bronchi. The strong bron-
  chodilator effect of p-agonists, therefore,
  might be attributed to the rather wide-
  spread (but not uniformly so) dispersion of
  P-receptors throughout the lung.
    Gamma camera images cannot unam-
  biguously portray the morphology of the
  lung. That is, the respective central (C),
  intermediate (I), and peripheral (P) zones
  will be of mixed composition. For exam-
  ple, deposition within the alveolated
 region will be superimposed upon that
 within the larger bronchial airways. This
 inherent and indeterminate character of
 such instruments is an unresolved prob-
 lem which compromises their effectiveness
 and versatility.  Therefore, the mathemati-
 cal model presented herein was developed
 expressly to complement laboratory investi-
 gations by determining deposition on an
 airway generation-by-generation basis.
 The function of the model is twofold: (1) to
 aid the clinician in the analysis and inter-
 pretation of experimental data; and,
 thereby, (2) to assist in the design of future
 drug delivery tests.
   Methods

    In the medical arena, intersubject vari-
   abilities of patient lung morphologies and
   interlaboratory differences in experimen-
   tal techniques must be acknowledged. In
   our work we have focused upon matters
   related to morphology. The major factors
   that inherently affect and complicate the
   use and intercomparison of gamma cam-
   eras are the spatial discrimination of an
  image and the radioactive dissipation
  which occurs prior to reaching ah instru-
  ment's detector. These issues are
  addressed, albeit briefly, below.

  Human Lung Morphology

    The human adult lung morphology has
  been described by Soong et al.12 (see their
  Table 2). The branching angle of a daugh-
  ter airway is the deviation between the
  longitudinal axes of it and its parent.  The
  branching angles measured by Horsfield
  et al.13 have suggested a mean value of
  70° to be appropriate.          I
   In the human lung network described by
 Soong et al.12 there are 2n airways in each
 generation, n.  For instance, there are 223
 = 8,388,608 alveolar sacs in generation n =
 23. This numbering system is an impor-
 tant element in our later analyses.
   Our ultimate goal was to superimpose a
 well-defined format such as used in the
 clinical environment (see the next subsec-
 tion entitled Geometric Definition of a
 Scan, page 61) for the C, I, and P regions
 of gamma scans upon the lung. Therefore,
 it was incumbent upon us to develop an
 unambiguous technical representation of
 the lung's labyrinth of passages.  j
  A model for the structure of the lung
from Soong et al.12 was made using the
geometric parameters (lengths and
branching angles). To define each lung,
the pathway to each alveolar sac from the
trachea was first calculated. Then, the
60
                                                         NESC Annual Report - FY1993

-------
                  The Supercomputer in Medicine: Interpretation of Gamma Camera Composition
coordinates for the midpoint of each indi-
vidual airway in the scheme were  |
determined.  Therefore, the lung became,
literally, a matrix of points. Subsequently,
the gamma camera scan format selected
for use in our studies (i.e., that of McMas-
ter University) was superimposed, upon
the matrix and the composition of the C, I,
and P partitions determined directly by
sheer itemization. The problem was
thereby reduced to one of very extensive
bookkeeping.                    ;

Geometric Definition of a Scan
  To study aerosol deposition patterns
with gamma cameras, it is essential that
the boundaries of the lung be delineated.
Agnew et al.11 discussed two techniques
used to characterize regional lung deposi-
tion using planar images.  Newhouse et
al.14 and Smaldone et al.8 defined the outer
contours of the lung using xenon-127 and
xenon-133, respectively, whereas Aj^new et
al.15 used krypton-8 Im. The advantage of
using xenon rather than krypton gas is
that the former can be delivered to jthe
whole of the ventilated lung volume. More
recently, submicronic aerosols of  ;
99mTc-DTPA have been used successfully to
measure ventilation in both normal
patients and those with a variety of chest
diseases.  The distribution of the aerosol
in the lung is similar to xenon (Coates et
al., 198316.)  Alternatively, a perfusion
scan can be performed, using a sub-clinical
dose of 99mTc-MAA which will allow the
lung edge to be denned as well as provid-
ing a correction for chest wall and tissue
 attenuation for technetium-99m.  ',
   In Figure 1, the methodologies employed
 at St. Joseph's Hospital, McMaster Uni-
versity, to examine gamma camera; scans
 are displayed.  The protocols havelbeen
 described in other literature7-17>1S aind will
 be used herein to provide technical exam-
 ples of how our analytical model may be
 applied in the medical arena.  The
perimeter of the lung is initially deter-
mined from the xenon-127 gas ventilation
image.  This template is then transposed
to the aerosol deposition scan and the lung
regions denned as follows.  From the
outer contour of the lung, a one (1.0) inch
interval is offset to define the P zone.
Subsequently, another one (1.0) inch con-
centric segment is delineated to identify
the I zone. The balance of the image is
the C zone.   Radioactive markers placed
on the chest can be used to identify the
trachea.
   Figure 1: Diagrammatic outline of right lung illustrating
   the three-zone lung model. Squares 1.25 cm x 1.25 cm in
   original represent areas from which data were obtained.
   Dark squares represent the peripheral zone, light-shaded
   squares, intermediate zone; and white squares, perihilar
     zone; (reprinted with permission of Sanchis et al.)  7
 NESC Annual Report - FY1993
                                                                                     61

-------
   The Supercomputer in Medicine: Interpretation of Gamma Camera Composition
   Results and Discussion
    The mapping routine is quite simple in
   concept and theoretically straightforward,
   but requires numerous and repetitive com-
   putational procedures to track the posi-
   tions of all airways within the lung.  It is
   an ideal exercise, therefore, for the type of
  vector programming most efficient with a
  supercomputer.
    In Figure 2 (page 63), the tracheobron-
  chial tree of the adult human lung is pre-
  sented. The systematic branching
  Table 1: The airway generation-by-generation composition of the central
  (C), intermediate (I), and peripheral (P) zones of a gamma camera image.
   The description is based upon the dimensions of an adult human lung
                     (Soong, et. al.12).
-,;k$*f;-
(&me*atitm,
..>.... s^ v '**
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
Cumulative
% of Lung
'PaartMoia. Of Steam
Itegkme
0
4
8
16
4
4
12
24
42
90
182
372
754
1528
3054
6212
12440
25098
50412
100702
201338
402072
803878
1608246
9.59
Jtegkaal
f
£J
0
0
0
28
60
116
232
316
446
794
1372
2542
4964
9902
19530
38818
77148
154250
309416
619260
1240374
2483770
4963340
29.58
BfegfottF
0
C
0
0
0
0
0
0
154
488
1072
2352
4896
9892
19812
39794
79814
159898
319626
638458
1276554
2551858
5100960
10205628
60.83
 network of the 0 < n < 16 generations is
 clearly demonstrated. The airways mak-
 ing up the balance of the lung {airway
 generations 17 through 23, inclusive) are
 not depicted simply because of the extent
 of overlapping that occurs.  That is, those
 alveolated airways (if displayed) would be
 superimposed upon the currently visible
 structure and, as a result, no distinct air-
 way network could be discerned.  It should
 be noted that the TB network pf Figure 2
is drawn to a 0.75 scale; that is, the corre-
sponding generation-by-generation dimen-
sions  are systematically reduced by one
        quarter.
          In Table 1, the thidpiesses of
        the associated P and I concentric
        gamma camera zones were one
        (1.0) inch as prescribed in the
        subsection entitled Geometric
        Definition of a Scan (page 61).
        The data of Table 1 itemize the
        compositions of the C, ;I, and P
        components of the respective
        gamma camera scans., Specifi-
        cally, those partitions are explic-
        itly broken-down into their
       constituent elements. Tb our
       knowledge, our paper is the first
       to accomplish such an unambigu-
       ous spatial definition of the adult
       human lung.
         The tabulated information is
       quite self-explanatory. 'For brev-
       ity, we shall make only ja few com-
       ments to orient the reader.
       Regarding use of the results in
       the clinical laboratory, some criti-
       cal points are the following. The
       C partition contains a tiotal of
       1,608,246 airways of which
       1,595,940 are alveolated passages
       from generations 17 < n'< 23.
       Thus, more than 99% of the C air-
       ways are from the pulmonary
       compartment of the lung.
62
                                                           NESC Annual Report - FY1993

-------
                The Supercomputer in Medicine: Interpretation of Gamma Camera Composition
                                                                           S'E
NESC Annual Report - FY1993
63

-------

-------
The Supercomputer in Medicine: Interpretation of Gamma Camera Composition
  It may be prudent to view the data from
Table 1 from another perspective.  The
total number of all alveolated airways
(i.e., the entire pulmonary compartment of
the human lung) is 1,6646,144; of that
total, between 9-10% are in the C parti-
tion.
  In summary, regarding the difficulties
involved in standardizing such methodolo-
gies (i.e., the subsection entitled Geometric
Definition of a Scan, page 61.) between
laboratories, Agnew et al.11 submitted a
list of proposals to be incorporated into
future investigations to facilitate interpre-
tation of images and intercomparispn of
data. The resolution of such instruments,
however, will still not permit deposition to
be determined on an airway generation-
by-generation basis.  Therefore, our math-
ematical model was derived to comple-
ment such laboratory investigations.

Disclaimer
  Although the research described in this
article has been supported by the United
States Environmental Protection Agency,
it has not been subjected to Agency review
and, therefore, does not necessarily reflect
the views of the Agency, and no official
endorsement should be inferred. Mention
of trade names or commercial products
does not constitute endorsement or recom-
mendation for use.
1 T. B. Martonen - Health Effects Research Laboratory, U.S.  Environmental Protection Agency, Research
    Triangle Park, NC 27711 and Division of Pulmonary Diseases, Department of Medicine, University of North
    Carolina, Chapel Hill, NC 27599.
2 Y. Yang - Center for Environmental Medicine and Biology, University of North Carolina, Chapel Hill, NC
    27599.                           '
3 M.  Dolovich - Department of Nuclear Medicine, Health Sciences Centre - IP, McMaster University, Hamilton,
    Ontario L8N 3Z5.
4 Supported by funds provided by the U.S. Environmental Protection Agency under Collaborative Agreement
    CR817643 on Health Effects of Exposure to Air Pollutants through the Center for Environmental Medicine
    and Biology, University of North Carolina, Chapel Hill, NC.
5 Barnes P.J., Basbaum C.B., Nadel JA., I^oberts J.M. Localization of p-adrenoreceptors in mammalian lung by
    light microscopic autoradiography, Nature 1978, 299:444-447.
6 Newman S.P., Clark A.R., Talaee N., CliJrke S.W. Lung deposition of 5 mg Intal from a pressurised metered
    dose inhaler assessed by radiotracer technique. Inter. J. Pharm. 1991; 74:203-208.
7 Ruffin R.E., Dolovich M.B., Wolff R.K., Newhouse M.T.. The effects of preferential deposition of histamine in
    the human airway.  Am. Rev. Eespir. Dis. 1978; 117:485-492.
8 Smaldone G.C., Walser L., Perry R.J., Ilpwite J.S., Bennet W.D., Greco M.  Generation and administration of
    aerosols for medical and physiological research studies. J. Aerosol Med. 1989; 2:81-87.
9 Karlsson J.A., Sant'Ambrogio G., Widdicbmbe J. Afferent neural pathways in cough and reflex
    bronchoconstriction, J. Appl. Physiol 1988; 65:1007-1023.
10 Martonen T.B.  Aerosol therapy implications of particle deposition patterns in simulated human airways.  J.
    Aerosol Med.  1991; 4:25-40.         ;
11 Agnew J.E. Characterizing lung aerosol penetration.  J. Aerosol Med. 1991; 4:237-249.
12 Soong T.T., Nicolaides P., Yu C.P., Soong S.C. A statistical description of the human tracheobronchial tree
    geometry. Respir. Physio. 1979; 37:161-172.
13 HorsfieldK., Dart G., Olson D.E., FilleyiG.E, Gumming G.  Models of the human bronchial tree. J. Appl.
    Physiol. 1971; 31:207-217.           ;
14 Newhouse M.T., Wright F.J., Ingham G.K., Archer N.P., Hughes L.B., Hopkins O.L.  Use of scintillation
    camera and 135xenon for study of topographic pulmonary function, Respir. Physio.  1968; 4:141-153.
15 Agnew J.E., Lopez-Vidriero M.T., Pavia D., Clarke S.W. Functional small airways defense in symptomless
    cigarette smokers.  Thorax 1986; 41:524-530.
16 Coates G., Dolovich M., Newhouse M.T. ! A comparison of submicronic technetium aerosol with xenon-127 for
    ventilation studies. In: Raynaud C edl. Nuclear Medicine and Biology: Proceedings of Third World Congress
    of Nuclear Medicine and Biology, Vol 11. Elmsford, New York: Pergamon Press, 1983:91-96.
17 Sanchis J., Dolovich M., Chalmers R., Newhouse M. Quantitation of regional aerosol clearance in the normal
    lung.  J. Appl Physiol. 1972; 33:757-762.
18 Dolovich M.B.,  Sanchis J., Rossman C., Newhouse M.T. Aerosol penetrance: a sensitive index of peripheral
    airways obstruction. J.  Appl. PhysioL. 1976; 40:468-471.
64
              NESC Annual Report - FY1993

-------
                 Oxidant Model Sensitivity Analysis
  EPA Research Objectives

   The use of regional scale (~ 1000-2000
  km) air quality simulation models, such as
  BPA's Regional Oxidant Model (ROM), for
  analysis of regional air quality and for
  assessing potential future air quality con-
  trol strategies has become more common
  over the last decade. The specific objective
  of this work is to assess the sensitivity of
  ROM ozone predictions to key input vari-
  ables. The sensitivity case outlined here is
  that of mobile source (automobile) emis-
  sions.
 Overview of Project

   There are several classes of variables
 that are required to drive air quality mod-
 els such as the ROM. These classes
 include meteorological and emissions vari-
 ables, the principal forcing functions, as
 well as specification of initial and bound-
 ary conditions and terrain and land-use
 data.  There are uncertainties associated
 with the specification of every variable
 within each class; generally speaking, the
 emissions variables tend to be more uncer-
 tain than the others.  The concentration
 predictions from the air quality model will
 be sensitive, to one degree or another, to
 each of the input variables. We are most
 concerned with those variables which con-
 tain the greatest uncertainty in their spec-
 ification, and for which the model is also
 quite sensitive. To fully understand the
 model response to these variables, we
 must first characterize the degree of model
 sensitivity. The Atmospheric Research
 and Exposure Assessment Laboratory
 (AREAL) of EPA/ORD is attempting to
 characterize the ROM's sensitivity to a
 subset of chemical and physical input vari-
 ables to the model by direct simulation
 using perturbations of the input variables.
 We focus our discussion here on one of the
 sensitivity tests in our series.


 Background and Approach

   Photochemical oxidant pollution is a fre-
 quent and widespread phenomenon across
 the eastern U.S. in the warm-weather
 months of the year. Three-dimensional
 numerical grid models, such as.the ROM,
 are being developed, tested, applied, and
 evaluated to handle the air quality simu-
 lation needs of the pollution problem.
 These computationally demanding models
 simulate all of the necessary physical and
 chemical processes responsible for
 regional photochemical smog. Once evalu-
 ated, the models are used by EPA to assess
 potential future emissions control strategy
 options to reduce the levels of high ozone
 concentrations across the U.S.  ;
  As previously discussed, our confidence
 level in the results of the ROM is, in part,
 a function of how well we can characterize
 the model's sensitivity to key input vari-
 ables. As an illustration of this I process,
 we focus here on the ROM's sensitivity to
 the mobile source component of jthe source
 emissions inventory. Automobiles emit
 significant quantities of hydrocarbons and
 nitrogen oxides, precursor chemicals to
 the formation of ozone and photochemical
 smog. Subjectively, we expect the model to
 be responsive to this emissions compo-
 nent. To perform this test, we remove the
 automobile emissions from the inventory
 feeding the ROM and compare the results
 of an air quality model simulation with
 the altered inventory with that using a
 "base case" inventory. Specifically, for this
test all emissions from light duty cars and
trucks were set to zero, as well as associ-
ated emissions from gasoline marketing
and storage. The emissions changes were
NESC Annual Report - FY1993
                                    65

-------
Regional Oxidant Model Sensitivity Analysis
made in the U.S. portion of the modeling
domain only.


Scientific Results and Relevance
to EPA Mission

  The meteorological data used for these
simulations were taken from the period
July 2-10,1988, a time period character-
ized by high heat and air mass stagnation
over the eastern U.S. This entire lime
period was simulated with the ROM,
using the model's three vertical layers
extending through the  earth's boundary
and cloud layers, and with 18.5-km hori-
zontal resolution. We present here the
results of the model simulations using the
maximum hourly ozone concentration pre-
dicted by the ROM in each of the model's
lowest level grid cells over the July; 2-10
period. Figure 1, page  68, displays the
base case ozone results using the original
emissions inventory. Figure 2, page 69,
displays the results of the simulation
using the inventory without the mobile
source emissions, and Figure 3, page 70,
shows the concentration differences
between the simulations.         :
  Inspection of these figures reveals that
the removal of the mobile source emissions
component does indeed dramatically
reduce the ozone concentrations over the
eastern U.S. Regionally, concentrations
drop by 5 ppb to greater than 50 ppb over
their base case values, with the greatest
decreases occurring over and downwind of
the urban and industrialized portipns of
the model domain.  In Figure 3, note espe-
cially the concentration differences
throughout the eastern seaboard region,
industrial Midwest, and Great Lakes
region. Downwind influences of the emis-
sions change are seen over the Atlantic
Ocean and into interior New EngL^nd as
well. It is also worth noting that the
removal of the mobile source emissions did
not fully alleviate the ozone problem in the
model domain. Figure 2, page 69, reveals
estimated ozone concentration areas still
in excess of the 120 ppb Federal ambient
air quality standard to exist in portions of
the domain.
  Implications of the result of this study
indicate that the emissions reductions
incurred through programs such as the
Federal Motor Vehicle Control Program,
which aim to reduce mobile source emis-
sions of hydrocarbons and nitrogen oxides,
should serve to reduce ozone levels, but
the gains may be modest in many areas.
Other sources of ozone precursors, such as
naturally-emitted hydrocarbons and nitro-
gen oxides from industrial and power gen-
eration facilities can also combine to form
significant amounts of ozone.  Therefore,
while the motor vehicle emissions control
program is likely to be beneficial, other
source categories for targeted control
should also be considered to reduce high
ozone concentrations levels in the eastern
U.S.
Advantages of Using NESC Cray

  Earlier ROM simulations had been per-
formed on scalar mainframes, such as the
IBM-3090. Resource limitations on these
machines prevented our simulating the
entire Eastern U.S. with the ROM. Typi-
cally smaller domains were used, such as
the Northeastern U.S. only, about one-
fourth the size of the current model
domain depicted in Figures 1, 2, and 3.
With the availability of the NESC CRAY,
and the vector optimization of the ROM
code, we are now able to efficiently simu-
late the larger domain, providing a more
comprehensive picture of estimated air
quality throughout eastern North
America.
Future Plans

  We plan to continue our series of sensi-
tivity tests with the ROM, including more
emissions sensitivities as well as a series
of tests for meteorological sensitivities.
Results of these tests will help interpret
 66
             NESC Annual Keport - FY1993

-------
                                                      Regional Oxidant Model Sensitivity Analysis
  the significance of future air quality mod-
  eling analyses with ROM and other
  regional air quality models.


  Publications

    Roselle, S.J., T.E. Pierce, and K.L.
  Schere.  The Sensitivity of Regional Ozone
  Modeling to Biogenic Hydrocarbons.
Journal of Geophysical Research, Vol. 96,
pp. 7371-7394.
  Roselle, S.J., K.L. Schere, and S.H. Chu.
Estimates of Ozone Response to Various
Combinations ofNOx and VOC Emission
Reductions in the Eastern United States.
Proceedings of 1992 Quadrennial Ozone
Symposium, June  1992.
  1 Thomas E. Pierce and Kenneth L. Schere, U.S. EPA, Office of Research and Development, Atmospheric
      Research and Exposure Assessment Laboratory, Research Triangle Park, NC.             i
NESC Annual Report - FY1993
                                    67

-------
Regional Oxidant Model Sensitivity Analysis
                                      Figure 1: Matrix Base Cases
 68
                                                                  NESC Annual Report - FY1993

-------

-------
                                                                  Regional Oxidant Model Sensitivity Analysis
 0)
 o
 N
O
 I
 1
1
 I
O
DC
00
00
0)
  I
 CM
                                       Figure '2: Mobile Source Emissions Set to Zero
                                                                                                        00
                                                                                                        ONlO
OCD
OOO
                                                                                                           CD
                                                                                                        CD CO
                                                                                                         OJ
                                                                                                         o
                                                                                                         cr
                                                                                                         o
                                                                                                        CJ>
       NESC Annual Report - FY1993
                                                                                                      69

-------

-------
 Regional Oxidant Model Sensitivity Analysis
                             Figure 3: Matrix Base - Sensitivity Difference
70
                                                                                            I  o
                                                                                            A  A
                                                                                               I
                                                                                           LO

                                                                                           7 m
                                                                                            I  CN

                                                                                            II   II
                                                                                            A  A
                                                                                           LO
                                                                                           A A
                                                                                           LO

                                                                                           7-
                                                                                              II
                                                                                           V A
                                                                                           CL
                                                                                           Q.
                                                                                           0
                                                                                           !^
                                                                                           (— '
                                                                                           C
                                                                                           0)
                                                                                           u
                                                                                           C
                                                                                           o
                                                                                          o
                                                               NESC Annual Report - FY1993

-------

-------
   Meteorological and Photoe&emieal Grid Modeling in
   tMran Domains l
 Abstract

   Photochemical grid modeling is essen-
 tial to predicting the impacts of prescribed
 emissions changes on predicted ozone con-
 centrations and other secondary pollutant
 species. An intensive effort has been
 underway to develop improved meteoro-
 logical model algorithms and an updated
 version of the Urban Airshed Model
 (UAM) for use in urban-scale ozone| appli-
 cations.  Simulations with an efficient,
 computationally fast diagnostic meteoro-
 logical model and a computationally rigor-
 ous dynamic meteorological model, Which
 numerically solves the set of governing
 equations for atmospheric motion, t^mper-
 ature, and moisture, are being performed
 to generate meteorological data files for
 use by an updated UAM model. The
 dynamic model code also features a four-
 dimensional data assimilation technique
 which continuously adjusts modele
-------
Meteorological and Photochemical Grid Modeling in Urban Domains
meteorological parameter fields on pollut-
ant concentrations and their spatial pat-
terns, considerable effort has been
expended to design and to test updated
meteorological algorithms that can pro-
vide compatible input data files for the
UAM.  Consequently, both diagnostic and
dynamic meteorological modeling
approaches are being applied in order to
assess their strengths and weaknesses in
different urban environments. While  a
diagnostic model, relying on available
observations and interpolative technic[ues
to generate gridded data, is computation-
ally fast, complicated circulations on
smaller spatial and temporal scales can-
not be resolved.  In contrast, a dynamic
meteorological model is capable of simu-
lating physically-realistic land-sea breeze
circulations and terrain-induced flows.
However, the short computational time
step necessitates much longer CPU time
and results are not always expected to
reproduce observations.  Therefore, a four-
dimensional data assimilation (FDDA)
technique requiring little additional com-
putational time has been incorporated, into
the dynamic model code in order to pro-
vide more accurate winds at grids cells
near data sites and produce more realistic
flows in data-sparse areas.  Simulations
with and without the FDDA feature are
needed to assess key variables involved in
the approach for weighting the observa-
tions and modeled values.

Accomplishments and Preliminary
Results

  lest simulations with both meteorologi-
cal model codes are successfully being per-
formed on'the NESC Cray system in order
to generate meteorological data files for
use in simulations with the updated UAM
model code. Urban domains being
modeled include the greater New York
City metropolitan area with limited obser-
vational data and the Los Angeles basin
where numerous measurements were
obtained during an intensive experimental
field study.  An assessment of differences
in wind fields and other meteorological
parameters between these models and
analysis of the impact with different mete-
orological input data 011 simulated spatial
patterns and peak concentration, of ozone
are underway.
  Simulation results of both meteorologi-
cal models from the Los Angeles domain
reveal the existence of a sea breeze from
the Pacific Ocean developing during day-
time hours which transports pollutants
inland toward mountain ridges and
through gaps surrounding the urban
basin. A statistical evaluation of simu-
lated pollutant concentrations from the
updated UAM code versus measured con-
centrations at sites throughout the urban
domain is underway to assess model per-
formance using different meteorological
model inputs.
  A series of UAM sensitivity test simula-
tions are also underway for a case study
day in the New York City domain to inves-
tigate the impact on  peak ozone concentra-
tions from different meteorological inputs
and algorithm changes. Initial results
with a new horizontal advectioii scheme
designed to reduce numerical diffusion in
the updated UAM indicate a somewhat
narrower and more elongated urban ozone
plume downwind of the New York metro-
politan area than produced by the existing
horizontal advection technique in the
standard UAM code.          I


Future Plans              |
                            [
  Numerous simulations are being
planned with the updated UAM to
investigate differences in model results
due to prescribed changes in key input
parameters, and a series of model
simulations with selected variations in
emissions are anticipated in order to
assess the effect on ozone concentrations
and spatial patterns. Additional test
simulations with different wind fields,
72
                                                         NESC Annual Report - FY1993

-------
                                Meteorological and Photochemical Grid Modeling in Urban Domains
   mixing height fields, and vertical eddy
   diffusivity coefficients are expected to
   yield valuable information about the
sensitivity of modeled ozone
concentrations.
      Atmh11' P™dP*lInves%ator (JUG), U.S. EPA Office of Research and Development,
      Atmospheric Research and Exposure lAssessment Laboratory, Research IHangle Park, NC.
NESC Annual Report - FY1993
                                                                                       73

-------
Physics does not change ftie nature
   of the world it studies, and 110
science of behavior can change the
   essential native of man* «v«n
     though both sciences yield
 technologies with a vast power to
 manipulate their subject matters.
      •••.••,'.     •. , s    f^.-.
                      ic] Skinner [1904-J
   '-' pumttl&iimRemrd'lthird edition, 1972], ch. 5

-------
                Partiuuiate Modeling
Research Objective

  The objective of this project is to investi-
gate the behavior of atmospheric aerosol
particles on the regional scale. Such parti-
cles result from human activity as well as
natural processes. A major portion' of the
particles resulting from human activity
are acidic sulfates formed in the atmo-
sphere from the chemical transformation
of sulfur dioxide, a by-product of fossil fuel
combustion. Other anthropogenic particles
result from the chemical transformation of
reactive organic gases (ROG's) and oxides
of nitrogen.
     100
   I
   to
   3
   _E
   ~o
   to
   o
   <5
   cc

   I
10
         0
        12    24
 36
Hour
48    60    72
     Figure 1: Total aerosol mass (\igm~3)'for a high emission grid cell
                                      Approach

                                        A three-dimensional air quality model
                                      (RADM) has been extended to include the
                                      chemistry and dynamics of aerosol parti-
                                      cles. The model has a full photochemical
                                      mechanism which transforms emitted
                                      compounds from the 1985 NAPAP emis-
                                      sions inventory into atmospheric oxidants
                                      and particle forming species, such as sul-
                                      fates, nitrates and organics. The aerosol
                                      dynamics and chemistry codes explicitly
                                      recognize processes such as particle
                                      growth and coagulation, as well as the
                                      equilibration with atmospheric relative
                                      humidity. Particle size distributions are
                                                      treated in two ranges
                                                      or modes, both of
                                                      which are log normal.
                                                      The mean diameters
                                                      and geometric stan-
                                                      dard deviations of
                                                      these two modes are
                                                      time-dependent vari-
                                                      ables.
Accomplishment
Description

  The behavior of sul-
fate aerosol particles
has been studied for
cloud-free conditions.
Three sensitivity stud-
ies for a typical sum-
mer 72 hour
meteorological episode
have been conducted.
The first was a base
case using the full
inventory. The second
used the same inven-
tory but omitted
NESC Annual Report - FY1993
                                                                         75

-------
 Regional Paniculate Modeling
 ammonia. The third used the full inven-
 tory for all species except sulfur which was
 reduced by a factor of two. These studies
 indicated that the model behaved in a
 qualitatively correct manner. As expected
 the water  content of the particles is
 inversely related to the degree of neutral-
 ization of the acidity by ammonia. Thus in
 general, the more acidic the particles are,
 the higher the water content, and the
 larger the  particles become.


 Significance

  Aerosol particles are of great interest to
 EPA for several reasons. Major reasons
 are the health effect of inhaled particles,
 the role of these particles in acidic deposi-
 tion to sensitive ecosystems, and the
 reduction in visibility by the interaction of
the particles with light. Related'to this
last item is the role of aerosols in the glo-
bal climatic system. The model can be
used to investigate the impact of" proposed
control scenarios required under the 1990
Clean Air Act Amendments upon aerosol
particle behavior.              j


Future Plans               ;

  The model is constantly evolving as new
features are added. Current plans call for
the addition of cloud interactions with par-
ticles during FY1994. Semi-volatile
organic compounds will be added in
FY1995, thus allowing the model to be
used to study toxic species. Extensive
model evaluation is planned using existing
and future field study data sets.
1 Francis S. Binkowski, Principal Investigator - U.S. EPA, Office of Research and Development, Atmospheric
    Research and Exposure Assessment Laboratory, Research IHangle Park, NC.             '.
2 Uma Shankar, Co-Investigator - MCNC, Research Triangle Park, NC.                     :
76
            NESC Annual Report - FY1993

-------
                   Environmental Data at tta MPA
Overview

  Environmental visualization aims to
make visible the unseen. This helps
researchers and policymakers alike. Here
we examine the unique issues and prob-
lems involved.                   :
  Scientists at the U.S. Environmental
Protection Agency pursue a wide r:ange of
research interests.  One group develops
models for the transport and deposition of
airborne pollutants. Policy makers use
this information to develop control strate-
gies for managing air pollution, such as
the Clean Air Act. Another group evalu-
ates the positions of air quality monitoring
sites with respect to the distribution of
pollutants.  Other researchers collaborate
with investigators at NASA Langley and
Lawrence Livermore National Labs to
examine global climate change. Still oth-
ers study water quality and sedimentation
in the Great Lakes region (see Figure 1),
electrical properties of carcinogens (see
    Figure 1. Environmental researchers study water quality using visualiza-
   tion techniques. This image shows sediment Concentrations in Lake Erie for
    1940 storm data obtained from a collaboration between the University of
    California at Santa Barbara and U.S. EPA Large Lakes Research Station.
Figure 2, page 78), subsurface contamina-
tion of waste disposal sites, and the air
flow through and around buildings. In
this article we describe some of these
research efforts.

Environmental Data
  Environmental data sets are large, up to
4 Mbytes for a single time step. They
often contain many time steps, represent-
ing changing conditions. For example, one
data set consists of a set of ozone levels
across the world for each day of a 13-year
period. Another data set shows pollutant
concentrations for each hour of a 15-day
period. Still another data set describes
lake sedimentation levels over the course
of a storm.
  Data sets  arise from a wide range of
sources, including computationally inten-
sive models, atmospheric monitoring sta-
tions, satellites, laboratory experiments,
and meteorological, records. Data formats
also vary, compounding the handling prob-
               lem. Data can occupy
               either a 2D or 3D spatial
               domain, and we might
               need both surface and vol-
               ume representation tech-
               niques to handle them.
                A major problem afflicts
               environmental data: it is
               seldom regularly gridded,
               yet visualization software
               expects regular grids.  For
               example, water quality
               data might come on a cur-
               vilinear grid from moni-
               toring sites at points
               scattered irregularly
               across the nation.
NESC Annual Report - FY1993
                                     77

-------

-------
 Visualizing Environmental Data at the EPA
 Moreover, pollutant totals might be associ-
 ated with county-sized areas of space.

 Visualization Requirements
   Environmental visualization attempts
 to represent abstract environmental data
 using concrete visual metaphors, Sesveral
 factors influence the choice of representa-
 tion: type of data, relationships among dif-
 ferent parts of the data, placement of data
 in a spatial and temporal context, and
 interpretation of the data.
   Environmental processes can be com-
 plex, involving many chemical species,
 atmospheric conditions, and geographical
 factors. Scientists frequently want to see
 multiple data sets or data variables (dis-
 played in the same visualization. For
 example, the Regional Qxidant Model
 (described below) calculates airborne pol-
 lutant transport at three different atmo-
 spheric levels. Sometimes scientists want
 to see those three layers displayed
 together.
   To validate the model, other visualiza-
 tions display data produced by environ-
 mental models and data collected by ;
 monitoring stations together. Data sets
 displayed together can have different
 sources, different data
 types, and different
 internal coordinate sys-
 tems. Sometimes, we
 must first register the
 various data sets so
 that we can represent
 them in a single display
 coordinate system.
 Unfortunately, these
 time-intensive and
 error-prone procedures
 often dominate the real
 applications.
  Adding information to
 place the data in its spa-
 tial and temporal con-
 text can make
                visualization of environmental data more
                valuable. Tb add spatial context, for
                example, we can take pollutant levels
                across the U.S. and overlay them with a
                map showing state borders (see Figure  3,
                page 80). We can show fluid flow through
                and around a building in conjunction with
                a simple representation of the building.We
                can show the electrical charge distribution
                of a molecule relative to the atoms and
                bonds of the molecule. These spatial land-
                marks, while not pairt of the data itself,
                provide valuable cues for interpretation.
                We can provide temporal context by show-
                ing the progression of a 'distribution over a
                time series. Showing a time series, rather
                than a single frame, emphasizes the devel-
                opment and progression of environmental
                processes.  Labeling a visualization with
                its date and time also provides temporal
                context.

                Configuration, General Strategies
                 The EEA uses a heterogeneous computer
                architecture for scientific computing. It
                supports a range of platforms and tools,
                including workstations and networked
                supercomputers.
Figure 2: U.S. EPA environmental researchers evaluate molecular properties
of carcinogens from pollution sources. This images shows electric field vectors
  for the benzopyrene molecule. This is based on data from the ILS. EPA
    i          Health Effects Research Laboratory.
78
                                                          NESC Animal Eeport - FY1993

-------

-------
                                                Visualizing Environmental Data at the EPA
   The Visualization Centers software con-
 figuration is similarly heterogeneous,
 using a wide range of tools and packages
 on the various platforms.  Despite the
 desirability of agency-wide standardiza-
 tion on a single visualization environment,
 none of the tools we use provides all the
 functionality we need.  For example, while
 one tool provides photorealistic rendering
 and specialized functionality for generic
 gridded data sets,  our environmental
 researchers find its script-driven environ-
 ment unnatural. Another package offers
 interactive control and a variety of fea-
 tures useful for representing computa-
 tional fluid dynamics data, but it runs only
 on one platform and requires researchers
 to convert data into a specific format for
 visualization.
  The visual programming and multiple-
 platform capabilities of tool kits are
 attractive for the EPA's heterogeneous
 hardware configuration. However, for a
 tool kit to provide the required functional-
 ities, the EPA visualization team needs to
 develop, support, and inventory custom-
 ized modules that support environmental
 research requirements.
  Through necessity or by design we have
 adopted a few general strategies for visu-
 alization. First, we introduce scientists to
 the possibilities of visualization, and then
 we let their needs and vision drive the
 research product. Second, we push soft-
 ware packages beyond their intended lim-
 its, then supplement and extend them
 with customized modules.  Third, we seek
 a balance between image quality and
 interactive control.  We can interactively
 manipulate small-  to medium-sized data
 sets. Fourth, we support the EPA's!
 nationwide heterogeneous hardware and
 software environment.            ;

 Regional Oxidant Model
  The Regional Oxidant Model is an air
 quality model  designed to examine the
 transport and  deposition of airborne
pollutants. EPA scientists use the model
to develop control strategies for the Clean
Air Act.  The Regional Oxidant Model sim-
ulates most of the significant chemical and
physical processes responsible for photo-
chemical production of ozone over a 100-
kilometer domain for episodes lasting 15
days. These processes involve
•  horizontal transport.
•  atmospheric chemistry and subgrid-
   scale chemical processes.
•  nighttime wind shear and turbulence
   associated with the low-level nocturnal
   jet.
•  effects of cumulus clouds on vertical
   mass transport and photochemical
   reaction rates.
•  mesoscale vertical motions included by
   terrain and the large-scaleflow.
•  terrain effects on advection, diffusion,
   and deposition.
•  emissions of natural and anthropogenic
   ozone precursors.
•  dry deposition.
  Three separate atmospheric layers are
associated with every Regional Oxidant
Model run, and the output provides the
geographic distribution of three chemical
species (ozone, NOX, and ROG). A typical
geographic domain of the model is the
northeast corridor of the United States
(see Figure 4, page 80).
  We use visualization techniques to
examine the inputs generated for the
Regional Oxidant Model, such as
sequences of wind inputs (see Figure 5,
page 82). We animate wind vector repre-
sentations over time. This visualization
improves researchers understanding of
the wind inputs and results in modifica-
tions to the respective model algorithms.
  We have also applied volume visualiza-
tion techniques to Regional Oxidant Model
outputs.  Using a splatter technique origi-
nally developed by Westover,4 we gener-
ated a simultaneous visualization of the
three atmospheric layers of the model (see
NESC Annual Report - FY1993
                                     79

-------

-------
Visualizing Environmental Data at the EPA
                                      Oseone  9/2/88  O  GMT
             95
          ppb
    Figure 3. Cutting planes visualization for ozone concentration values of the Regional Arid Dej toaitiou Model, which.
     is a nonlinear air quality model used for assessing acid rain, impacts. The U.S. EPA Atmospheric Research and
                          Exposure Assessment Laboratory supplied the data.       ':
Figure 6, page 82). The graphical repre-
sentation displays cloud-like structures
for each of the chemical species. This pro-
vides a new way for atmospheric research-
ers to view their data.  Interestingly  I
enough, our scientists did not find the vol-
ume visualization effort insightful.  The
researchers had trouble comprehending
the interaction between the three layers
depicted. They requested that we return
to the surface modeling approach to visu-
alize the individual                  '
Regional  Oxidant
Model layers for the
respective chemical
species.
  We developed
movie sequences of
15-day episodes of
Regional  Oxidant
Model ozone concen-
trations, incorporat-
ing discrete color
mapping  display
and titling with time
and data  stamping.
We also examined
output with stan-
dard visualization
tool kits and
                      extended these tool kits with customized
                      modules to support environmental
                      researchers visualization requirements.

                      Total Ozone Mapping Spectrometer
                        In a cooperative project between the
                      EPA and NASA Langiey Research Center,
                      the EPA Visualization. Center has visual-
                      ized data from the Total Ozone Mapping
                      Spectrometer (TOMS) satellite.  The
Figure 4. texture (left side) and transparency (right side) mapping for ozone concen-
trations of the IRegional Oxidant Model, northeastern domain.. Researchers use this
model to develop control strategies for the Clean Air Act. Data came from, the VS.
      EPA Atmospheric Research and Exposure Assessment Laboratory.
80
                                                              NESC Annual Report - FY1993

-------

-------
                                                Visualizing Environmental Data at the EPA
 TOMS instrument measures total atmo-
 spheric ozone by analysis of back-scat-
 tered solar radiation across several
 ultraviolet bands.  The instrument is
 aboard a south-north, sun-synchrdnous,
 polar-orbiting satellite and performs more
 than 200,000 measurements each day over
 the entire globe. The measurements are
 then processed into 1- degree latitude by
 1.25-degree longitude cells (see Figure 7,
 page 82).  The purpose of the EPA portion
 of the study is to compare the results from
 the agency's nonlinear models (like the
 Regional Oxidant Model) with the mea-
 sured data available from TOMS.  '.
   Logistic problems included format con-
 versions and projections to various! coordi-
 nate systems. The EPA Visualization
 Center staff worked extensively on writing
 code to support converting the TOMS data
 into a 3D visual display. We also encoun-
 tered issues associated with handling
 missing data values. Large-scale dropouts
 in the TOMS data resulted from substa-
 tion failures during the satellite tracking
 and data collection process. The computa-
 tional process for handling the smoothing
 of the data resulted in potential artifacts,
 which environmental researchers
 reviewed and critiqued. Our visualization
 efforts focused on developing an adaptive
 filtering process to handle variations in
 missing data regions, data confidence, and
 sampling density issues (see Figure 8,
 page 83).
   The large size of the TOMS data set
 poses other concerns. Interactive \iewing
 of the daily change of global ozone for 365
 days requires computationally intensive
 hardware currently not available on stan-
 dard desktop visualization workstations.
 This concern highlights our need to
 address metacomputing for environmental
 data sets.                        :

 Unsolved Problems
  There remain a number of unsolved sys-
 tems problems with the visualization of
 environmental data.  Our current work-
 stations provide insufficient computa-
 tional power to interactively step through
 the time steps of large 3D data sets. Large
 data sets likewise present storage and
 access difficulties. The need to support
 visualization on multiple platforms has
 not yet been completely addressed.  Simi-
 larly, we have just begun to address the
 issues associated with remote access.
 Remote access issues appear in our collab-
 orations with scientists at remote sites
 and in our use of the National Environ-
 mental Supercomputing Center (NESC) in
 Bay City, Michigan.6
   A number of software limitations  and
 challenges also remain. Foremost is the
 need for a general-purpose data import
 tool for visualizing multiple data formats.
 Tool kit data conversion utilities currently
 available begin to address the problem,
 but stop short of a complete solution. For
 example, our scientists often run models
 on a mainframe or supercomputer and
 wish to view the model output.  Data con-
 version routines that run on workstations
 often cannot handle large binary files gen-
 erated on a supercomputer.
   Data collected from atmospheric moni-
 toring sites causes other problems. These
 sites might be scattered irregularly across
 the area of interest. The  data collected
 does not exhibit the grid structure (either
 regular or irregular) expected by standard
 tools (see Figure 9, page 83). Finally, some
 data associates values with irregularly
 shaped areas (such as counties) and
 requires the functionality found in GIS
 packages. Until the tool kit data conver-
 sion utilities develop sufficient power to
 handle all the data used by scientists, we
 will continue writing conversion pro-
 grams for each new type of data.
  As the EPA Visualization Center contin-
 ues to support agency visualization needs,
we will develop new approaches  and
tools. The installation of our first high-
performance computer at  the NESC has
NESC Annual Report - FY1993
                                    81

-------

-------
Visualizing Environmental Data at the EPA
                                                                Figure 5. Visualization of
                                                                wind vector inputs to the
                                                                Regional Oxidant Model.
                                                                Meteorologists at the U .S.
                                                                 EPA use visualization to
                                                                  examine inputs to th@
                                                                 Regional Oxidant Model
                                                                 before executing it on a
                                                                tiigTi-parfarmflnrtf* comput-
                                                                 ing platform. Data came
                                                                from the U.S. EPA Atmo-
                                                                  spheric Research and
                                                                  Exposure Assessment
                                                                       Laboratory.
                     Figure 6. This image shows three layers of the Regional Oxidant Model using a splatter volume-
                      visualization technique. Because of the complexiiy associated with the simultaneous display of
                     the £hree layers, environmental researchers requested that we use standard surface renderings
                       for each layer. We obtained the data from the U.S. EPA Atmospheric Research and Exposure
                                             i        Assessment Laboratory.
      Figure 7. Visualization
      of "Ibtal Ozone Mapping
       Spectrometer data.
        This visualisation
      projects the TOMS data
      onto a spherical object
      in three views, helping
      EPA and NASA Langley
       researchers examine
      total global ozone distri-
      bution.  NASAAGod-
      dard supplied the data
 82
NESC Annual Report - FY1993

-------

-------
                                                     Visualizing Environmental Data at the EPA
 resulted in the need to address high-speed
 data transfer, storage, and compression
 requirements for visualizing environmen-
 tal data at other EEA. research siteis. As
 part of our participation in the U.S. Fed-
 eral High Performance Computing and
 Communications (HPCC) effort, EPA will
 develop techniques to transfer high- per-
 formance tools to key state, federal, and
 industrial users with decision making
 responsibility. These future techniques
 are indeed grand challenges.       ',
Acknowledgments

  We thank the many environmental
researchers throughout the EPA who have
brought us challenging visualization
projects. Special thanks are also in order
to our colleagues at the North Carolina
Supercomputer Center, University of Cali-
fornia at Santa Barbara, NASALangley
Research Center, Numerical Design, Com-
puter Sciences, Sterling Software, and
Unisys who have provided support and
suggestions on our various projects.
                                                             S,ne.  Sprinted, with permission,
    trom 1&H.1L Computer Graphics and Applications, Vol. 13, No. 2, March 1993
   TheresaEhyne, MarkBolstad, and Penny Rheingans, Martin Marietta Technical Services, Visualization
    Support, Research Triangle Park, NC.
 3 L   6 PetterS°n and Walter **• shacteUprd, U.S. Environmental Protection Agency, Eesearch Triangle Park,
        ™' F°otprint Evaluati°n for Volume Rendering, Computer Graphics (Proe. Siggraph), VoL 24, No. 4
    Aug. 1990, pp. 367-376-
 5 A. Cullati, E. Idaszak, and T. Bhyne, Scientific Visualization Efforts at the U.S. Environmental Protection
    Agency, Landscape and Urban Planning, Special Issue on Data Visualization Techniques in Environmental
    Management, Vol. 21, No. 4, May 1992, pp. 323-326.
          Figure 8. -Visualization of Ibtal Ozone Mapping Spectrometer data near the Antarctic legion.
          The amage is a polar orthographic projection of data from NASA Goddard, -with missing data
                                valuias shown as square holes.
        Figure 9, Abstract repreaentation of ozone concentrations from the Regional Oxidant Model for the north-
       eastern ILS. The geographic map of the domain distorts mth the height values of the ozone concentrations
         Weottamed the data from the U.S. EPA Atmospheric Eesearch and Exposure Assessment Laboratory
NESC Annual Report - FY1993
                                                                                         83

-------

-------
The machine does not isolate
man from the great problems
  of nature but plunges him
   more deeply mto them*
                French aviator* writer

-------

-------
  Molecular H0d&lin,g 
-------
Molecular Modeling on Supercomputers for Risk Evaluation
interaction of chemicals in this class with
DNA and other electron rich biological
molecules (See Figure 1, page 88). This
interaction does not require metabolism.
Other competing pathways are metabolic
epoxidation of the double bond and enzy-
matic de-esterification (See Figure 2,
page 88). Epoxidation could provide a
pathway for binding to biologically rele-
vant molecules, and also provide a mecha-
nism for elimination.  Using ab initio
quantum mechanical techniques, a three
dimensional structure and a charge distri-
bution were obtained for both acrylates
and methacrylates. From these charge
distributions, the molecular electrostatic
potential (MEP) was computed for both of
these moieties (See Figure 3, page 89).
From the MEP it was observed that a neg-
atively charged species,  or the negatively
charged part of a biopolymer, could
approach more closely to the relevant
reactive center of the acrylate moiety than
the similar reactive center in a methacry-
late. This strongly implies that acrylates
are more likely to bind to biopolymers
than the corresponding methacrylates.
Again, ab initio quantum mechanical tech-
niques were used to model Michael addi-
tion between these chemicals and small
electron donor targets. The results of
these calculations indicate that Michael
addition to a reasonable mechanism for
activity. From these calculations, it is pos-
sible to predict that a simple acrylate is
likely to be more biologically active than
the corresponding methacrylate in sys-
tems with minimal metabolism and that
the addition of metabolic enzymes
decreases the acrylate activity but may
increase the methacrylate activity. Both
of these predictions were confirmed by
genetic bioassays which also provide a
way to compare the two  different path-
ways in Figure 2 (page 88).  Even with the
addition of metabolic enzymes, the acry-
late is much more active than the corre-
sponding methacrylate further indicating
the importance of the Michael addition
pathway. The implication of these studies
is that positive data for the activity of a
methacrylate can be used to imply activity
for the corresponding acrylate, and nega-
tive data for an acrylate can be used to
imply that the corresponding metjhacry-
late is inactive.  However, the reverse of
these paradigms is not true.  Thejconfir-
mation of Michael addition as a mecha-
nism for biological activity, using j
molecular modeling techniques,  allows
the chemical class to be expanded to
include all chemicals that contain! an elec-
tron withdrawing group adjacent jto a dou-
ble bond.  More realistic biomolecular
targets have been used to enhancle the
insight that this  model provides. !
  Polycyclic Aromatic Hydrocarbons
(PAHs) are a class of pervasive environ-
mental chemicals produced by the incom-
plete combustion and pyrolysis of [fossil
fuels and other organic materials j  Mole-
cules within this class show considerable
variation in toxicity. Some class members
are powerful mutagens and animal carcin-
ogens, while other molecules show no such
activity after considerable testing. The
activity of many  class members falls
between these two extremes. While the
details of the molecular mechanism of
action depend on the specific PAH and the
particular test system used, therd is
always at least one metabolic oxidation
step  to a reactive epoxide.  Experimental
scientists in HERL have been studying the
biological activity of PAHs and thpir spe-
cific mechanisms of action. Cyclopenta-
PAHs (cPAHs) (PAHs that contain a five
membered ring in addition to six mem-
bered rings) form one of the relevant sub-
classes.  The carbon-carbon bond that
closes the five membered ring is o|ften the
bond that is metabolically activated to an
epoxide. The direction of ring opening for
that  epoxide indicates the atom that will
be bound to the nucleophilic site in DNA
(See  Figure 4, page 89). In conjuiiction
with the experimental studies underway
in HERL, computational methods were
used to predict the direction of the epoxide
86
             NESC Annual Report! - FY1993

-------
                                    Molecular Modeling on Supercomputers for Risk Evaluation.
  ring opening for a series of cPAHs. In the
  initial studies the semi-empirical quan-
  tum mechanical method AMI was used to
  obtain three dimensional structures of the
  epoxides and carbocations and ab initio
  quantum mechanical methods were used
  to obtain energies and charge distribu-
  tions for those structures. The differences
  in energy between carbocation pairs
  resulting from the same epoxide was used
  to predict the direction of ring opening and
  the charge distribution used to provide
  information about the reactivity of the car-
  bocation. It was found that the cPAHs
  being considered divided into two groups,
  one for which the energy difference
  between the possible carbocatiorns was
 large (>7.5 kcal/mol) and those for which
 this difference was small (< 4.0 kcal/mol).
 It was possible to enunciate structural dif-
 ferences between the classes. Available
 experimental data partially confirmed this
 result. They suggest that the trends
 observed by the energy differences
 between carbocations are correct, but the
 computational methods may  be overesti-
 mating this difference.
   In later studies, as better computational
 resources became available, ab initio
 quantum mechanical methods and a semi-
 empirical method that includes the solvent
 water (AM1/SM2) were used to improve
 these results. We found that the improved
 molecular geometries changed individual
 carbocation energies significantly but the
 difference in energy between  carbocation
 pairs was not changed significantly. How-
 ever, the energy differences between car-
 bocation pairs was lowered significantly
 by the inclusion of solvent and the results
 improved. The implication of these results
 is that, for charged molecules or reactive
 intermediates at least, a greater impact on
 the results will be made by methods that
 include solvent.
   In most of the studies where molecular
 modeling techniques are applied to prob-
 lems of environmental interest, the envi-
 ronmental agent (that is, the acrylate or
 the PAH or other molecule) is modeled and
 its reactivities or interacts with a small
 surrogate target studied to predict how it
 will act in a biological system. However,
 more and more experimental information
 is becoming available about the actual bio-
 molecular targets for these agents (recep-
 tors, enzymes and DNA, for instance). As
 algorithms and computational facilities
 improve, it becomes possible to model the
 interaction of the agent or the ultimate
 toxicant with its actual target. In order to
 obtain experience for this eventuality,  a
 specific DNA sequence has been modeled
 (the sequence was chosen because it is a
 known target sequence for an environmen-
 tal agent being studied.) using the molecu-
 lar mechanics features in the program
 Discover.  Figure 5, page 90, shows the
 preliminary results of these computer
 experiments. The importance of including
 the solvent molecules is clear from these
 results.

 Future Plans
  Molecular modeling methods will be
 applied to additional classes of importance
 to the Agency. Methods that include sol-
 vent in both quantum mechanical and
 molecular dynamics and mechanics will be
 employed. Biomolecular targets will be
included in the models for activity as the
specific targets are identified experimen-
tally.
  James R. Rabinowitz, PhJD., CMB/GTD/HERL, U.S. EPA, Research Triangle Park, NO.
NESC Annual Report - FY1993
                                                                                 87

-------

-------
Molecular Modeling on Supercomputers for Risk Evaluation
                          Kgurel: Electron Rich Bidogieal Molecule
                                                          NESC Annual Report - FY1993

-------

-------
                                       Molecular Modeling on Supercomputers for Risk Evaluate
                       KgureS: Molecular Electrostatic Potential (MEP) Con.putation !
                 AGEANTHRYLENE 1,2-EPOXIPE AMD HYOROXY CATIONS
                        ,Kgure4: Example of Nucleophffic Site Bonding
NESC Annual Report - FY1993
                                                                                   89

-------

-------
    Molecular Modeling on Supercomputers for Risk Evaluati
                                                     ion
                                  Figured: DNA Sequence Model
90
                                                            NESC Annual Report - FY1993

-------

-------
   Regional Ozone Modeling to Support oWn Air Act
                                                        '"       *
  Abstract

    Attainment of the ozone National Ambi-
  ent Air Quality Standards (NAAQS) has
  been a particularly difficult enviironmental
  problem to solve. In fact, nearly 100 areas
  of the United States are still not attaining
  the NAAQS despite the efforts of the U. S.
  Environmental Protection Agency (EPA),
  states, and industry over the last 15 years.
  In 1990 the Congress reauthorized the
  Clean Air Act (CAA) in part to mandate
  programs and milestones for attainment of
  the ozone NAAQS. In order to support
  states in meeting these requirements and
  schedules, the EPA has committed to a
  regional scale modeling program; using the
  Regional Oxidant Model (ROM)., The pur-
  pose of the proposed work is to use ROM
  simulations to  support development of
  credible control strategies to solve the
  ozone problem  for many of the most seri-
 ously polluted areas of the Eastern U. S.
 Because of the  computational complexity
 of models such as ROM, the space and
 time scales of the applications, and the
 number of simulations required, it is nec-
 essary to run this model in a supercomput-
 ing environment.
   The key objectives of this project are to
 1) assess the relative benefits of alterna-
 tive emission control strategies for reduc-
 ing regional ozone concentrations and 2)
 provide the data bases containing the
 regional model predictions to states for
 use in estimating the impacts of controls
 on future levels of pollutant transport.
  Results from completed applications
 have been used by the EPA for developing
 policies on the relative effectiveness of
 alternative emissions reduction scenar-
 ios. Analyses for the Ozone Transport
  Commission have been used to assess the
  benefits of emissions reductions from man-
  dated CAA control programs.

  Background

   In November 1990 the United States
  Congress amended the Clean Air Act
  (CAA) to provide for attainment and main-
  tenance of the National Ambient Air Qual-
  ity Standards (NAAQS) to protect public
  health and welfare against the harmful
  effects of air pollutants. Major portions of
  the Clean Air Act Amendments (CAAA)
  contain provisions oriented toward attain-
  ment of the ozone NAAQS through pro-
  grams that reduce ozone precursor
  emissions of volatile  organic compounds
  (VOC), nitrogen oxides (NOX), and carbon
 monoxide (CO).  Section 182  and guidance
 issued by the U. S. Environmental Protec-
 tion Agency (EPA)3 require that urban
 scale photochemical grid modeling be used
 by states to demonstrate attainment in
 areas classified as having moderate (inter-
 state), serious, severe, and extreme ozone
 problems. Most of these urban modeling
 efforts involve the Urban Airshed Model
 (UAM)4.  In addition,  Section 184 of the
 CAAA establishes an Ozone Transport
 Commission (OT'C) to address the specific
 regulatory and technical issues associated
 with modeling and implementing regional
 control strategies needed to achieve
 attainment in the Northeast.  The forma-
 tion of the OTC is in recognition of the
 complexity of the ozone problem in this
 region where summer meteorological con-
 ditions, combined with the spatial distri-
 bution and large magnitude of ozone
precursor emissions, contribute to long
range and interurban  transport across
political boundaries over several days.
NESC Annual Keport - FY1993
                                                                               91

-------

-------
 Regional Ozone Modeling to Support Clean Air Act Mandates
 Overview and Objectives
  A key component of the modeling by
 states is the estimation of expected future
 year urban boundary conditions which
 reflect changes in upwind emissions
 between current baseline and future
 attainment dates required by the CAAA.
 The EPA has developed a methodology for
 deriving these boundary conditions from
 predictions of the Regional Oxidant  I
 Model5-6 (ROM). The EPA is committed to
 providing the ROM applications that cou-
 ple with the episodes and emissions sce-
 narios to be simulated with UAM by
 states.  The objectives of this project are to
 use the capabilities of the National Envi-
 ronmental Supercomputing Center
 CNESC) to perform the ROM simulations
 necessary to support the states. Because
 of the computing requirements of ROM for
 the spatial and temporal scales needed for
these applications, using the NESC is the
only available means to achieve results in
time to meet the scheduJ.es required in the
CAAA.
  Also, states in the OTC are examining
regional controls beyond those in the
CAAA that may be needed for attainment
in this region. The EiPA has completed
several ROM analyses for the OTC to
examine the relative effectiveness of possi-
ble regional strategies. Regional strate-
gies selected by the OTC will be used by
individual states as an initial baseline set
of controls in then? attainment demonstra-
tion modeling. This report presents the
results of ROM simulations performed for
the OTC.

Accomplishments

  The OTC modeling analysis has focused
on developing an initial regional
         Mgurel:
          Map of
         Northeast
          ROM
          domain
          and the
          Ozone
         Transport
          Region.
92
                                                         NESC Annual Report - FY1993

-------

-------
                                   Regional Ozone Modeling to Support Clean Air Act Mandates
   assessment of the impacts on Northeast
   ozone levels due to the control programs
   specified in the CAAA. The results have
   not only been valuable to the OTC, but
   also to the EPA since they provide the only
   integrated assessment of the likely ozone
   benefits of controls mandated by the
   CAAA across a region containing over 25
   nonattainment areas.

   Results from Completed Analyses

    The OTC assessment included ROM
   simulations for emissions scenarios
   described below.  The location of the
   Transport Region in the ROM Northeast
   simulation domain are shown in Figure 1
  page 92.

  Base Case Scenario

    The Base Case Scenario was largely cre-
  ated from the 1985 National Acid Precipi-
  tation Assessment Program (NAPAP)7
  emissions inventory.  In the Transport
  Region, area sources and highway vehicles
  account for the bulk of man-made VOC
  emissions. However, biogenic emissions
  make up half of the total VOCs. Although
  they are highest in rural areas, biogenics
  contribute about 30% to total VOC emis-
  sions in many of the Northeast urban
  areas.  An exception is New York City
  where biogenics are only about 1Q% of the
  total. In contrast to VOC, NOX emissions
  are dominated by highway vehicles and
 point sources, with very low natxiral emis-
 sions.

   Of importance for designing regional
 strategies is the spatial distribution of
 emissions. In the Transport Region, 84%
 of the man-made VOC and 78% of the NOX
 emissions are from sources in nonattain-
 ment areas. In addition, 72% of the VOC
 and 61% of the NOX in the region are from
 sources along the Northeast Corridor from
   Washington, DC to southern Maine. In
   rural areas, biogenics dominate VOC
   emissions, but the largest contribution to
   NOX comes from electric generation by
   utilities.

   2005 Scenario

    The Base Case emissions were projected
   to 2005 considering the effects of growth,
   existing controls and the new stationary'
   and mobile source control programs in the
   1990 CAAA. The year 2005 was selected
  since it is near or beyond the dates when
  most of the areas in the Transport Region
  are required to be in attainment.
  Although the focus of this analysis is on
  the Transport Region, the appropriate
  CAAA control was also applied to emis-
  sions in other nonattainment areas.

  Impact on Emissions of CAAA
  Controls

    The result of applying the above control
  programs and estimates of emissions
  growth (where applicable) was a net
  reduction in man-made VOC emissions in
 the Transport Region of 46% between the
 Base Case and the 2005 Scenario. When
 biogenic emissions are factored in, the
 overall reduction in VOC is only 24%. For
 NOX, the reduction in emissions is 36%
 (recall that natural NOX emissions, as cur-
 rently quantified, are relatively insignifi-
 cant.)

 ROM Ozone Predictions

   The ozone predictions from ROM for the
 two scenarios simulated are presented and
 compared in terms of 1) spatial patterns
 and the percent change hi episode maxi-
 mum concentrations and 2) changes in
 daily maximum ozone. Ozone predictions
from the lowest layer in ROM are used in
the comparisons.
NESC Annual Eeport - FY1993
                                                                               93

-------

-------
    Regional Ozone Modeling to Support Clean Air Act Mandates
Cb
Y.
s
0)
94
                                                           NESC Annual iReport - FY1993

-------

-------
                                     Regional Ozone Modeling to Support Clean Air Act M
    Comparisons for Episode
    Maximum Ozone
      The spatial distribution of episode maxi-
    mum ozone for the Base Case Scenario is
    shown in Figure 2, page 94. The episode
    maximum values represent the highest
    1-hour ozone concentrations predicted in
    each grid all over the entire episode. The
    figure indicates ozone concentrations
    above the level of the NAAQS stretching
    along the entire Northeast Corridor with
   values exceeding 180 to 200 ppb near
   Washington, DC/Baltimore and Philadel-
   phia, and from New York City across Con-
   necticut. High ozone levels are also
   predicted near and downwind of other cit-
   ies (e.g. Chicago, Cleveland, Pittsburgh
   and Toronto) as well as across rural sec-
   tions of several states (e.g., Virginia, West
   Virginia, Pennsylvania, and New York).
    Predictions for the 2005 Scenario in Fig-
   ure 3, page 94, indicate that the CAAA
   controls result in a substantial reduction
   m peak values as well as the aerial cover-
   age of concentrations above 120 ppb  The
  number of grids with ozone above 120 ppb
  m the Transport Region declined by 56%
  Peak values along the Northeast Corridor
  are reduced by 15 to over 20%. Elsewhere
  m the region, peak ozone levels declined
  by 10 to 20% across Maryland and Penn-
  sylvania with lesser reductions across
  upstate New York. Despite these reduc-
  tions, levels are still predicted to remain
  above 120 ppb in many of areas of the
  region.


 Comparisons for Daily Maximum
 Ozone

   The 95th percentile of daily maximum
 ozone concentrations for selected areas in
 the Transport Region indicates reductions
 in ozone of approximately 20% from the
 New York City area northeastward along
                                   andates
    the Corridor with slightly less reduction
    Ub to 18%) southwest of New York  In
    Pittsburgh, daily maximum ozone '
    declined by 13%. These reductions are
    similar m magnitude to those revealed by
    the episode maximum concentrations
    above.  The 95th percentile values follow-
   ing the implementation of CAAA control
   measures decline to near or below the
   level of the NAAQS in most areas The
   exception is New York City and southern
   New England, which is immediately down-
   wind of New York on most days in this epi-
   SOQG.


   Summary and Conclusions

    Preliminary applications of the ROM
  have been made to assess the expected
  benefits of emissions reductions specified
  in the 1990 CAAA on  ozone in the North-
  east. This analysis Is part of a larger
  effort to develop a set of regional and/or
  subregional strategies that, when com-
  bined with urban, area-specific control pro-
  grams, will provide for attainment of the
  ozone NAAQS  in the Northeast. The ROM
  applications include simulations for a
  Base Case Scenario and a 2005 Scenario
  that includes the net effects on emissions
  of growth together with controls in the
  CAAA. The results indicate that the
 OAAA controls may provide significant
 reductions in ozone levels in this region
 However, even  with these controls, ozone
 may remain close to or above the NAAQS
 by the year 2005 in portions of this region
 under meteorological conditions of the
 type simulated. Thus, optional regional
 control programs are being considered to
 provide the additional emissions reduc-
 tions which may be needed for attainment
 The expected effectiveness of these pro-
 grams will be assessed as part of upcom-
ing regional and urban model
applications.
NESC Annual Report - FY1993
                                                                                95

-------
Regional Ozone Modeling to Support Clean Air Act Mandates
3 JSS"^™^^










 '
     Park, NC, 1989, 692 pp.                                                       I
                                                              NESC Annual Report - FY1993

-------
OMI Aeid Iposition Model
                                                                 Myaluaf ion
    EPA Research Objectives
      Regional air quality models are needed
    and used to extrapolate outside current
    conditions, therefore, these advanced mod-
    els are developed with parameterization
    and physical and chemical mathematical
    descriptions as close to first principles as
    possible. The purpose of the evaluation is
    to test the science incorporated into the
    advanced models. Evaluation is diagnos-
   tic, to explore quality of predictions and
   develop an appraisal of model strengths
   and weaknesses.  The data used were spe-
   cially collected for the RADM evaluation
   as part of the National Acid Precipitation
   Assessment Program (NAPAP) and a bi-
   national effort, the Eulerian Model Evalu-
   ation Field Study (EMEFS). The data
   were collected over a two-year period with
   special, several-week intensives that used
  very advanced instruments to collect air
  concentrations to provide data that would
  support the most diagnostic testing.

  Overview of Project

    Early evaluation research concentrated
  on examining the predictions for the sul-
  fur cycle.  Significant improvements to the
  KADM were accomplished (see refer-
  ences). Current research is  investigating
  the nitrogen cycle, which is much more
  complex. This investigation  focuses on the
  testing the ability of the model to accu-
 rately replicate in time and space the con-
 7TSST (°r oxidation) of nitrogen oxides
 (NOX) to their oxidized products, PAN and
 nitrates (particulate nitrate, NO3-, and
 nitric acid, HNO3). Measurements taken
 by aircraft carrying sophisticated instru-
 ments to measure air quality in the
 EMEFS's 1988 aircraft intensive are used
 tor the diagnostic testing.
                                   Background and Approach

                                     The observations were developed from
                                   measurements taken during 12 aircraft
                                   flights over a 35 -day period. The standard
                                   80-km. version of RADM2.6 was used to
                                   simulate the 35-day period and a data-
                                   probe was "flown" through the model
                                   These "data" from the model are compared
                                   to equivalent data from the aircraft mea-
                                  .surements. An early sensitivity study on a
                                   3-day period using a smaUer grid resolu-
                                   tion indicated that grid size may be affect-
                                  ing the rate of conversion of NOX to PAN
                                  and HN03. Thus a full set of runs was
                                  ?ae?foeod f°r the August 25 to September
                                  ^9,1988 period that includes the aircraft
                                  2?™ ^md Using a ^-^meter version of
                                  KADM) the High Resolution RADM (HR-
                                  RADM), using a subdomain of the 80-km.
                                  KADM that encompassed the aircraft
                                  flights. The meteorology was interpolated
                                  from 80 kilometers to 20 kilometers to
                                  maintain consistency in scales that are
                                  resolved, and because the more advanced
                                  meteorological modeling capability at 20
                                  km., requiring a new convective parame-
                                  terization designed for scales less than 25
                                  km., was not yet ready. The HR-RADM
                                  was run as a one-way nest within the full
                                 RADM domain.

                                 Accomplishments Using the NESC's
                                 Cray

                                   The initial 80-krn. RADM runs for the
                                 August 25 to September 29, 1988 period
                                 required approximately 48 hours on a sin-
                                 gle-processor Cray Y-MP. The HR-RADM
                                 runs for this same period required
                                 approximately 180 hours on a single pro-
                                 cessor of the NESC Cray Y-MP. It
                                required about five weeks to prepare the
                                emissions. The model runs on the Cray
                                were able to be completed in about two
                                weeks.
NESC Annual Report - FY1993
                                                                               97

-------
Kegional Acid Deposition Model (KADM) Evaluation


Scientific Results and Relevance to
EPA Mission
  The full set of comparisons showed that
reducing the grid size from 80 km. (typical
for regional acid deposition models) to 20
km. (typical for regional oxidant models)
had little effect on the rate of conversion of
NOX to total-nitrate. The conversion to
PAN increased.  The manner of the
changes in the model predictions indicated
that changing grid size does not have the
effect one expects. It was very important
to be able to test the entire period at the
higher grid resolution to avoid the possi-
bility that a single case may not be repre-
 sentative of behavior over the entire
 period.
   This study enhances our understanding
 of the working of regional model photo-
 chemistry for rural ambient concentra-
 tions conditions. Proper computation of
 the photochemistry for rural conditions is
 important to the ability of models to sup-
 port exploration and establishment of
 appropriate emissions controls to reduce
 and eliminate violations of the ozone
 health standard. Examination of nitrogen
 chemistry is important because it is a cen-
 tral part of the oxidation process forming
 ozone and because rural oxidant produc-
 tion is generally believed to be NOX-
 limited.

  Future Objectives and Plans
    The evaluation will continue with addi-
  tional sensitivity studies directed at
understanding meteorological infhjences,
especially temperature, on the RADM
chemistry. Preliminary indications are,
that to improve temperature predictions
from the mesoscale meteorological|model,
its vertical resolution will need to be
increased from 15 to 25 layers and param-
eters affecting the surface heat flux, such
as a soil moisture, updated. Once the
newly adapted meteorological model has
been tested, roughly 60 Cray Y-MP hours
will be required to regenerate new, meteo-
rology for the September 1988 evaluation
period, 80 Cray Y-MP hours to generate
new 80-km. RADM results, and 300 Cray
Y-MP hours to generate new HR-^ADM
 results for the next round of diagnostic
 testing of the chemistry in RADM;.

 Publications and Reports   j
   Dennis, R. L., W. R. Barchet, T. L. Clark,
 S. K. Seilkop, and P. M. Roth, 1990: Eval-
 uation of regional acid deposition; models
 (Part 1), NAPAP SOS/T Report 5.; In:
 National Acid Precipitation Assessment
 Program: State of Science and technology,
 Volume I. National Acid Precipitation
 Assessment Program, 722  Jackson Place
 NW, Washington, D.C.         j
    Dennis, R. L.,  J. N. McHenry, W. R. Bar-
 chet, F. S. Binkowski, and D. W. Byun,
  1993: Correcting RADAM's sulfate under-
 prediction: discovery and correction of
  model errors and testing the corrections
  through comparisons against field data.
 Atmospheric Environment 27A, 975-997.
      Exposure Assessment Laboratory, Research Triangle Park, NO
                                   >  Cohn, William Hwang, and Daewon Byun, Regional Acid
                                         of Research and Development, Atmospheric Research and
   98
                                                            NESC Annual Report - FY1993

-------
    EPA Research Objectives
      Nitrogen is the primary cause of
    eutrophication in Chesapeake Bay. Nitro-
    gen input from the atmosphere represents
    fo^f^r3^ SOUrce of nitr°gen to the Bay
    (25-35% of the nitrogen loading). Water
    quality models have incorporated atmo-
    spheric nitrogen, but in a very simple
    manner. One objective of this research is
    to provide more accurate estimates of the
    quantity and the pattern of nitrogen load-
   ing from the atmosphere to the Chesa-
   peake Bay watershed and the Bay itself.
   These estimates will be provided as inputs
   to ^e water quality models for the water-
   shed (the HSPF model adapted by the
   Chesapeake Bay Program Office) and the
   Bay (the 3-D Bay Water Quality model
   developed by the Army Corp of Engineers)
   Another objective of this research is to
   determine the extent of the airshed that is
   primarily responsible for the atmospheric
   nitrogen affecting the Bay watershed. The
  airshed will be larger than the water-
  shed. The overall purpose is to develop an
  understanding of which controls of NOX
  emissions to the atmosphere will have the
  greatest benefit on reducing the nitrogen
  loading to coastal estuaries. This work is
  important to the Chesapeake Bay Pro-
  gram Office's efforts to achieve a 40%
  reduction in controllable nitrogen loading
  to the Bay by the year 2000, and to the
  upcoming 1996 Agency decision on the
  amount of Phase 2 NOX controls required
 Overview of Project

  Development of more accurate spatial
 fields of nitrogen loading estimates
 involves estimation of annual average
 nitrogen deposition to coastal areas using
 the Regional Acid Deposition Model
    (RADM).  These deposition estimates are
    made for 1985 emissions and representa-
    tive meteorology.  They are also made for
    ^005 emissions projections representing
    estimates of changes stemming from
    growth combined with emissions reduc-
    tions called for in the 1990 Clean Air Act
   Amendments with the same meteorology.
   Development of an understanding of the
   airshed influencing the Chesapeake Bay
   watershed involves using RADM as a labo-
   ratory of the real world to carry out sensi-
   tivity studies that elucidate the
   contributions of different emissions
   sources to the Bay  watershed. This
   source-receptor understanding is very dif-
   ficult and nearly impossible to develop
   from empirical data and requires the
   designing of sensitivity studies that will
   extract that information from a mathe-
  matical model.

  Background and Approach

    Because the RADM is very computation-
  ally intensive, it is not feasible, with
  today's computing power, to simulate an
  entire year's worth of meteorology to
  develop annual average estimates of depo-
  sition loading. Instead, annual averages
  are developed from a weighted average of
  a statistical sample of 30 5-day model
 runs.  The average is representative of
 meteorology for the 1982 to 1985 period
 which has a rainfall pattern very close to a
 30-year average.  Meteorological events
 (synoptic progressions of high and low
 pressure systems) with similar 850-mb
 wind-flow patterns were grouped or classi-
 faed by applying cluster analysis to them.
 ihis resulted in 19 sampling groups or
 strata. Meteorological cases were ran-
 domly selected from each stratum, based
 on the number of wind-flow patterns in
that stratum and on the number in each of
NESC Annual Report - FY1993
                                                                                99

-------
Atmospheric Deposition of Nitrogen to Chesapeake Bay
the other strata. This procedure approxi-
mates proportionate sampling. The num-
ber of cases, 30, was set after carrying out
a sampling-error analysis on wet sulfur
deposition and taking into consideration
computer resource limitations. These are
termed the aggregation cases.
   Development of a source-receptor under-
standing on an annual basis requires an
experimental design that will extract this
information from sensitivity studies with
RADM. Because NOX emissions contrib-
ute to oxidant production and there is a
 dynamic interplay between the production
 of ozone and nitric acid, the dominant
 form by which nitrogen is deposited to the
 Earth's surface, the modeling of nitrogen
 must incorporate full photochemistry as is
 done in RADM. As a first approximation to
 the source-receptor relations implicit in
 the model calculations, emissions sources
 of interest are subtracted from the emis-
 sions fields. The 30 aggregation cases Eire
 run and the results subtracted from
 results obtained with unperturbed emis-
 sions fields. For this study, the objective
 was to develop an understanding of the
 difference in the range of influence of
 ground-level NOX emissions (such as auto-
 mobiles) from upper-level NOX emissions
 (such as power plants) with regard to
 nitrogen deposition.

  Accomplishments Using the NESC's
  Cray
    The basic 1985 and 2005 RADM model
  runs were the same as those produced hi
  the 1990 CAAA Projections project
  (requiring 240 hours on a single-processor
   Cray Y-MP). The two sensitivity runs that
   were part of this study required 240 hours
   on a single-processor Cray Y-MP.

   Scientific Results and Relevance to
   EPA Mission
     The comparison of predicted wet and
   dry nitrogen deposition from RADM
indicated that dry deposition of nitrogen
appears to be less than wet. Up td now,
many researchers were assuming that dry
deposition equaled wet deposition.; The
model result was confirmed by comparing
the model estimates with Nationaj Dry
Deposition Network data. Thus, the
model results are helping to provide more
accurate estimates of the magnitude of
nitrogen deposition from the atmqsphere
to the coastal estuaries. The model results
have also identified the size of the gradi-
ents in the pattern of total nitrogen depo-
sition to the watershed. This information
will now be input to the water quality
models.                       |
   The new, projected reductions in total
 nitrogen deposition expected to result
 from the 1990 CAAA are estimated to be
 50% larger than earlier estimates used in
 sensitivity studies with the watei- quality
 models. Thus, reductions in NOX; emis-
 sions due to the 1990 CAAA could play a
 more beneficial role for Chesapeake Bay
 than previously thought.  As a result, con-
 tinued study of atmospheric reductions
 appears to be important to Chesapeake
 Bay work.                     !
    The sensitivity study on the range of
 influence of NOX emissions on nitrogen
 deposition depending on the height of the
 emissions (surface or tall-stack) produced
 a somewhat unexpected result. The range
 of influence of surface emissions; appears
 to be roughly 70% that of emissions from
  tall stacks. This is different than the "con-
  ventional" wisdom which would |"predict
  that the range of influence from fall stacks
  would be much greater. It is possible that
  conventional wisdom has been influenced
  by study of the sulfur system where the
  primary specie, SO2, plays a significant
  role in the total deposition. In the nitro-
  gen system, nitrogen deposition; is almost^
  entirely due to the secondary specie nitric
  acid, HNO3 (ignoring ammonia p>r the
  moment).  The distance over which NOX
   100
                                                            NESC Annual Report - FY1993

-------
     enussions from the western Pennsylvania
     region appear to noticeably influence the
     f^iff deP°sition is approximately 400
     to 500 kilometers. This means that the
     airshed affecting Chesapeake Bay is sig-
     nificantly larger than the watershed and
    is expected to include many sources along
    the upper Ohio River.

    Future Objectives and Plans
      Future plans call for from two to three
    additional sensitivity runs to help better
    define the airshed affecting the Bay. Plus
    the relative contribution from different   '
    economic sectors to the nitrogen deposi-
    tion are important to establish where the
    effort should be placed in the study of
   future control options. National trends
   mask regional differences that are impor-
   tant to regions such as the Chesapeake
   tfay. The additional sensitivity runs are
   expected to require 240-360 Cray Y-MP
   hours and the economic sector study on
Atmospheric Deposition of Nitrogen to Chesapeake Bay


       1985 and 2005 emissions to require 720
       Cray Y-MP hours. The exploration of con-
       trol options is also planned and is expected
       to require 240 to 360 Cray Y-MP hours In
       addition, the degree of error in the tech-
       nique used to approximate the nitrogen
       source-receptor relations needs to be
       established. Such an experiment would be
       expected to require the order of 240 Cray
       Y-MP hours.  The Chesapeake Bay Pro-
      gram would also like to develop an esti-
      mate of the portion of the nitrogen
      deposition that is coming from several
      .major urban areas within the watershed.
      Tins will require using the High-Resolu-
      tion RADM to better resolve the urban
      areas. The base runs plus the urban sen-
      sitivities would be expected to require
      1,800 Cray Y-MP hours.

     Publications and Reports

       In the process of being written.
                                                NC
                                                                               ,nd
NESC Annual Report - FY1993
                                                                                101

-------
   Pollution is nothing but tiie
 resources we are not harvesting.
We allow them to disperse because
we've been igiioratit of their value.
               ••' "


               ,*
-------
                                         ""WCtea* Air Act
                        •>   ,
    EPA Research Objectives
      Developing accurate estimates of the
    impact of the 1990 Clean Air Act Amend-
    ments (CAAA) on acidic deposition and
    atmospheric sulfate (key to visibility deg-
    radation in the Eastern United States) are
    important to the Agency. The amount of
    reduction in sulfur deposition to be antici-
    pated by 2005 or 2010 due to implementa-
    tion of Title IV Phases I and II sets an
   important baseline for understanding how
   much mitigation in deposition we are
   expected to achieve and how much farther
   we might need to go to provide protection
   to ecological resources.  The reduction in
   deposition loading in Canada that is likely
   coming from the United States and vice
   versa 1S important to the U.S. Canada Air
   Quality Accord. These estimates are
   important to the Canadians for them to
   project whether they will achieve their
  Inf ° o^f SUlfur deP°siti°n being below
  20kg-S042-/ha. As well, the European com-
  munity is interested in estimates of the
  long-range transport across the Atlantic of
  sulfur-related acidic deposition that could
  be affecting them. The objective is to
  develop best estimates from evaluated and
  well-characterized models of changes in
  acidic deposition loading, visibility impact-
  ing pollutants, and oxidants. Also, the
  cross-program effects from the different
  Titles of the 1990 CAAA need to be charac-
  terized.

 Overview of Project
   Development of estimates of future dep-
 osition involves creation of estimates of
 future emissions that account for popula-
 tion and economic growth plus the incor-
 poration of emissions controls called for bv
 the 1990 CAAA. The new emissions are
 input to RADM simulations to estimate
    the new deposition, assuming the same
    meteorology as today's. A difficult element
    tor the projection of future emissions is the
    estimation of power-plant retirements
    and the installation of new generating
    capacity to make up the difference in on-
   hne capacity and projected demand  Of
   special difficulty is locating or siting
   potential future plants for the modeling.
   These projections are generated by experts
   in the field of emissions estimation and
  - projection.

  Background and Approach

     Because the RADM is very computation-
  ally intensive, it is not feasible, with
  today's computing power to simulate an
  entire years worth of meteorology to
  develop annual average estimates of depo-
  sition loading. Instead, annual averages
  are developed from a weighted average of
  a statistical sample of thirty 5-day model
  runs. The average is representative of
  meteorology for the 1982 to 1985 period
  which has a rainfall pattern very close to a
  30-year average. Meteorological events
  (synoptic progressions of high and low
  pressure systems) with similar 850-mb
  wind-flow patterns were grouped or classi-
  fied by applying cluster analysis to them.
 This resulted in 19 sampling groups or
 strata.  Meteorological cases were ran-
 domly selected from each stratum, based
 on the number of wind-flow patterns in
 that stratum and on the number in each of
 the other strata.  This procedure approxi-
 mates proportionate sampling. The num-
 ber of cases, 30, was set after carrying out
 a sampling-error analysis on wet sulfur
 deposition and taking into consideration
 computer resource limitations.  These are
 termed the aggregation cases.

  Two emissions cases were developed for
the year 2005. The first emissions
NESC Annual Report - FY1993
                                                                                103

-------
Projected Effects of the 1990 Clean Air Act on Acidic Deposition
projection case considered Title II (auto-
motive NOX emissions) and Title IV, Phase
2 (acid rain-related utility emissions of
SO2 and NOX) emissions reductions. The
second emissions projection case consid-
ered the addition of Title I (ozone State
Implementation Plan requirements on
NOX emissions) to the other two CAAA
titles. The projections represented full
implementation of the CAAA titles. Thus,
emission reductions for SO2 included all
provisions of Phase 2, which are more rep-
resentative of 2010 emissions of SO2.

 Accomplishments Using the NESC's
 Cray
   The basic 1985  and the Titles-II and IV
 only 2005 RADM model runs required 240
 hours on a single-processor Cray Y-MP.
 The "sensitivity" run for 2005 that incor-
 porated Titles I, H, and IV required an
 additional 120 hours on a single-processor
 CrayY-MP.

 Scientific Results and Relevance to
 EPA Mission
    The reductions of total sulfur deposition
  due to the Title IV acid rain controls pro-
  duced by this study were similar to those
  projected for the National Acid Precipita-
  tion Assessment Program's 1990
  assessment. Comparison of the two cases
  indicates that the reduction in nitrogen
  deposition across the northeast (affecting
the Chesapeake Bay watershed in particu-
lar) is expected to be a factor of 1.2 larger
due to consideration of Title I State Imple-
mentation Plan requirements in addition
to the controls mandated by Titles II and
IV. Thus, it is important to incorporate
the impacts of interprogram (inter-pollut-
ant) air quality mandates of the 1990 ;
Clean Air Act on estimates of future air
quality and deposition loading. This is
consistent with the new emphasis in EPA
on considering the full range of multi,
media and multi-program effects and tak-
ing a more holistic perspective towards
pollution control.                  !

 Future Objectives and Plans   ,
   Future plans call for updating the 2005
 projections, in coordination with the |
 Regional Oxidant Model runs by Office of
 Air Quality Planning and Standards,
 using the new 1990 emissions inventory as
 a baseline, more current economic grpwth
 projections and the latest mobile source
 emissions model from EPA-Ann Arbor
 Plans are also to assess the potential for
 differences in sulfur deposition at regions
 of particular interest across the eastfern
 United States that might occur through
  the trading of SO2 emission allocations.
                                  i
  Publications and Reports      j
    In the process of being written.
     Robin L.
       S EPA, Office of Research and Development,
                     '  ' NC.      :
   104
                                                            NESC Annual Report •! FY1993

-------
     %w ^ -f •• ^  •? v  *. _^  ,•









               s. ..    .*•    - .      w           •. _.        s     ....                     **        *°
    Abstract

      In this research project we are investi-
    gating the possible role of the zebra mus-
    sel (Dreissena Polymorpha) in the
    increase of Polychlorinated Biphenyls
    (PCBs) in Great Lakes fish. A computa-
    tional approach using high performance
    computers, such as a Silicon Graphics
   (bGI) workstation and a Cray supercom-
   puter, to run code and visualize output of
   the model was used to deal with the vol-
   ume of data involved in the research of
   this problem.

     Background preparation for the investi-
   gation began during Saturday Tutorial
   sessions held at the National Environmen-
   tal bupercomputing Center (NESC) This
   continued during the Summer Research
   Institute at Saginaw Valley State Univer-
   sity by researching literature and contact-
  ing scientists doing work on zebra mussels
  and food chain modeling. A food chain
  model developed by John Connolly (Con-
  nolly  etal.) was then applied to a simple
  food chain, in Western Lake Erie, generat-
  ing preliminary data. The output data was
  then visualized on a Silicon Graphics sci-
  entific workstation using the Explorer
  visualization software.

   Future plans include refinement of the
 model, locating accurate data for input
 variables and improvement of the current
 visualization. Along with continuing the
 research project, a course in environmen-
 tal research and computational science
 will be developed and integrated into the
 curriculum in Bay City Central High
 School for the 1994-95 school year.
   Research Objectives
     Recent studies have shown that PCB
   levels have increased in Lake Erie fish
   over the last few years. The purpose of this
   research project is to explore whether
   zebra mussels are contributing to this
   increase. Determination of this role will be
   of importance to researchers, fisheries and
   recreationalists who use or examine any
   fish that have been or will be colonized by
   the zebra mussel.

  Approach

    The first step is to research current lit-
  erature to identify a specific zone and food
  chain m western. Lake Erie into which the
  zebra mussel has been inserted.  Next
  identify PCB concentrations in feces '
  pseudofeces and zebra mussel tissue in the
  food chain and examine how PCBs may be
  biomagnified. The next steps are to iden-
  tify feeding rates at each trophic level, and
 mechanisms controlling uptake and loss of
   Based on the above steps we will con-
 struct a mathematical model of this food
 chain, and translate this mathematical
 model into a high level language such as
 FORTRAN. This code will™ on a Silicon
 Graphics workstation or if necessary on
 the Cray Y-MP 81 located at the National
 Environmental Supercomputing Center
 (NESC) in Bay City, Michigan. The output
 of the computational model will be visual-
ized using the Explorer software package
on an SGI workstation.
NESC Annual Report - FY1993
                                                                                105

-------
The Role of the Zebra Mussel (Dreissena Plymorpha)
Polychlorinated Biphenyls (PCBs)

Accomplishments to date
  A typical food chain involving the zebra
mussel in Lake Erie has been identified.
The box diagram on the right hand side ot
Figure 1 shows a direct feeding relation-
ship between yellow perch and zebra mus-
sel as reported by Fisher. (Fisher, 1993).
We used a mathematical model for bioac-
cumulation of contaminants by John Con-
nolly (Connolly, 1992) to represent this
relationship (Figure 2, page 107).
   Data on zebra mussel PCB tissue con-
 centrations have been collected from
 research literature and personal contacts
 with various researchers in the field. The
 tissue concentration used for the computai-
 tional model was taken from Russ Kreiss
 (Kreiss, 1993). Published data on PCB
 uptake and loss mechanisms have not yet
 been identified.
   The mass balance equation has been put
 into FORTRAN code to generate synthetic
 data of PCB levels over a 90 day time
 period. The output of this code was trans-
 ferred into a 2D array data format
in the Uptake in Food Chain Transfer of j
                               j

 readable by the Explorer visualization
 software and visualized using a Siblcon
 Graphics workstation. The visualization
 in Figure 3, page 107, shows how the
 PCBs concentrations in prey and th0
 water column relate to PCB concentra-
 tions in fish tissue (as represented by
 varying colors).                 j
   The visualization output has been ana-
  lyzed and  possible scenarios have been
  explored for variables which contribute to
  the increase of contaminants (e.g., PCB
  concentrations in the water column and in
  prey.) Although our visualization indicates
  that PCB  concentrations in the perf h are
  highly dependent on contaminant concen-
  trations in the zebra mussel, at this point,
  due to lack of substantial data, no Reliable
  conclusions can be drawn.       J
                               I
                               I
  Future plans                j
    The most pertinent need is to continue
  literature review and conversations with
  mentor scientists to obtain more accurate
  and current data as it becomes available.
                 Food  Cliain  Routes
                    Figure 1. Feeding relationship between zebra mussel and yellow perch.
    106
                                                          NESC Annual Report - FY1993

-------
          The Role of the Zebra Mussel (Dreissena Pl^orpha) in the Uptake in Pood Chain OVansfer of
                                                       Polychlorinated Biphenyls (PCBs)
                          Mass Balance Equation Representing
                           Uptake and Loss of Contaminants
         dv/dt = KuCd
                                                -  (K + G)v
v =
        concentration of contaminants (PCB's) in the animal (perch)

K
G
       - concentration of contaminants (PCB 's) in food
        assimilation efficiency of contaminants (PCB's) in food
        consumption rate of food
        excretion rate
        growth rate of the animal
                 KuCd = uptake of contaminants (PCB's) through gills
                 T^1^ contaminant (pCB> assimilation through feeding
                 (K + G)v = loss of contaminants (PCB's) through excretion and growth
               Figure 2.
                                                            '. Conolly)
             Figure 3. Visualization of PCB concentration in prey, water column and fish ti
                                                       tissue.
NESC Annual Report - FY1993
                                                                           107

-------
The Role of the Zebra Mussel (Dreissena Plymorpha) in the Uptake in Food Chain Transfer of!
Polychlorinated Biphenyls (PCBs)

The mathematical model will be refined to
more accurately represent bioaccumula-
tion and gain/loss of PCBs. A graph of the
actual numbers which are represented by
the model will also be added to our
visualization.
  A second food chain route (represented
 on the left hand side of Figure 1
 (page 106), will also be identified by deter-
 mining the intermediate food chain links
 from a specific species of gammarid to the
 upper trophic levels. The FORTRAN code
 will be refined to represent the new model
 and include place holders for other food
 sources.
   As an extension of the EarthVision pro-
 gram, environmental research and compu-
 tational science will be incorporated into
 the existing curriculum at Bay City Cen-
 tral High School using an integrated cur-
 riculum approach. Teachers in disciplines
 such as math, science, computer science
 and art have agreed to work together to
 develop units in the courses which will
 become part of a multidisciplinary
 environmental-computational-science  cur-
riculum. A Silicon Graphics workstation
with ongoing access to the Cray is to be
installed, and academic activities which
will utilize the equipment will be deyel-
oped.                           i
                                I
Literature cited               j
  Connolly, John P., Thomas F. Parkerton,
James D. Quadrini, Scott T Taylor, j
Andrew J. Thumann, Development and
Application of a Model of PCBs in the^
Green Bay, Lake Michigan Walleye and
Brown Trout and Their Food Webs, Envi-
ronmental Engineering & Science ?ro-
 gram, Manhattan College, October,! 1992.
   Fischer, Susan W., The Role of Zebra
 Mussels in Contaminant Cycling, Zebra
 Mussel Information Clearinghouse! Vol-
 ume 4, Number 1, January/February
 1993.                          |
   Ereiss, Russell, Telephone interview on
 September 9,1993 regarding measured
 PCB concentrations in whole zebra mus-
 sels taken from Sterling State Park on
 Lake Erie.                     j
                                                                             i
  i  M.Neal.J.Bisel, N. Masloom, K. Kukla, E. Gatza, and J. Schroeder, Bay City Central High School ,

      EarthVision Team.
    108
                                                            NESC Annual Report - FY1993

-------
                                   IfetaJ Bfstrflmttoit Data In the
                                ««omp«ter Model t« latent
                                   ' '        "•
    Abstract

      The Great Lakes make up a substantial
    portion of the world's fresh water supply
    and the Saginaw Bay watershed is one of
    the largest and most heavily populated
    and industrialized segments feeding Lake
    Huron. Contamination of the Saginaw Bay
    is interesting both locally and nationally.
    Industrial and municipal waste treatment
    includes sophisticated removal tech-
    niques for most organic contaminants but
    removal of certain inorganic contami-
    nants, particularly heavy metals, has
    proven to  be difficult.

     This project initially focuses on nickel
   ion species because many industrial pipes
   are lined with nickel alloys and this metal
   has recently come under closer scrutiny
   as a possible health hazard. It is hoped
   however, that this research can be applied
 to a broad spectrum of heavy metal con-
 tamination in the Saginaw Bay.

 Research Objectives

   The project objectives include the collec-
 tion and visualization of historical heavy
 metal data for the Saginaw Bay using
 computer graphics. A mathematical model
 of how the ions interact with the water,
 sediments and living organisms will then
 be developed, visualized and tested by
 comparing it to the actual historical data.

 Approach

  Historical data on contamination in the
 Saginaw Bay will be collected from sources
 such as the Saginaw Bay Watershed Ini-
 tiative located on the campus of Saginaw
 Valley State University and also from
STORET. A grid will be placed over a map
          Figure 1:
        Contamination
        isosurface and
     transverse orthogonal
       slice rendering of
     historical data (1978) on
     nickel ion distribution
     in the Saginaw Bay. Red
      represents highest
     concentration. Total
     depth of slice is 50 cm.
     Red outer ring is an
     outline of Saginaw Bay.
      Bay City is at the
     bottom of the photo.
NESC Annual Report - FY1993
                                                                                109

-------
Summary of Heavy
these Data
Metal Distribution Data in the Saginaw Bay and a Computer Model to interpret
of the Bay and each sampling site will be
given a set of coordinates. The sampling
data will then be entered into a three
dimensional array to include concentra-
tions at various depths in the sediments
and stored as a data file. Using the visual-
ization software package called Explorer
on a Silicon Graphics workstation, the
data will be visualized in various ways
including, contour mapping, isosurfacing,
orthogonal slicing and volume rendering
of ion concentrations. Data sets separated
in time by several years will be visualized.
   For the purpose of modeling, the Bay
 will be divided into three dimensional
 'cells' (3.5km X 3.5km x 50 cm) centered on
 grid coordinate  intersections. Mathemati-
 cal expressions  of the more obvious inter-
 actions between the ions and the
 environment within a 'cell' and each cell's
 influence on surrounding cells will be
 written as FORTRAN code. The code will
 be executed on the workstation or, if neces-
 sary, on a CRAY-YMP at NESC in Bay
 City, Michigan. The results will be visual-
 ized as before and compared to the histori-
 cal data. By using an earlier historical
 data set as initial conditions and execut-
 ing the program forward in tune to a later
 historical data set, the model can be tested
  and modified.

  Accomplishments to Date
    Some historical data on heavy metal
                            contamination in the Saginaw Bay (J. A.
                            Robbins) have been encoded and visual-
                            ized as described. These data values were
                            interpolated for sampling sites where none
                            existed for some visualizations.  ,
                              The mathematical model is stip in the
                            early stages of development. In general,
                            Saginaw Bay is a very complex system to
                            model and there is a danger that, due to
                            its shallow depth, any detailed jnodel
                            might be overwhelmed by storm; activity.
                            Nevertheless, nickel ion'distribution in
                             the Saginaw Bay does seem to be associ-
                             ated with high clay content in the sedi-
                             ment.                        |
                                                          i
                                                          |
                              Future Plans               !
                               The model will be fine tuned and tested
                             against other data sets. Several questions
                             that will be investigated include: Does
                             point source contamination affect distribu-
                             tion differently than multiple site contam-
                             ination? How do instantaneous emissions
                             compare with long term seepage in affect-
                             ing heavy metal distribution? Jfhat, if
                              any contributions to heavy metal contam-
                              ination in the Saginaw Bay come from the
                              main body of Lake Huron itself?
                                Since this EarthVision project is part of
                              the EPA's Grand Challenge forjHigh
                              School Students, the research will be used
                              to teach students the fundamentals of
                              computational science and the j interdisci-
                              plinary nature of scientific investigations.

      EarthVision Team.

-------
     Summary of Heavy Metal Distribute, Data in the Saginaw Bay and a Computer Model to Interpret
                                                                                       these Data

                                                                             Figure 3. Volume
                                                                              rendering of
                                                                             same data. Pur-
                                                                              ple is highest
                                                                             concentrations
                                                                             (about 60 mg/
                                                                             kg.) Bay City is
                                                                              at lower left.
NESC Annual Report - FY1993
                                                                                          111

-------
Summary of Heavy Metal Distribution Data in the Saginaw Bay and a Computer Model toj Interpret
these Data
      Figure 4.
     Combination
     multiple level
     contours and
     longitudinal
      orthogonal
     slice. Red is
     highest con-
      centration.
     Bay City is at
      lower left.
   112
                                                                    NESC Annual Report - FY1993

-------

-------
"O
   5'
  - 
-------