T10 68 United States Office of I ivr.wu 1986 Environmental Protection Toxic Substances Agency - Washington DC 20460 TSCA Assistance Office vxEPA Expjaining Environmental Risk 230R86900 ------- ------- Explaining Environmental Risk Some Notes on Environmental Risk Communication by Peter M. Sandman November 1986 U S. Environmental Protection Agency Region 5, Library (PL-12J) 77 West Jackson Boulevard, 12th Ftooi Ctiicago, IL 60604-3590 ------- Contents "Important if True" 1 Dealing with the Media 4 1. Environmental risk is not a big story 4 2. Politics is more newsworthy than science 5 3. Reporters cover viewpoints, not "truths" 6 4. The risk story is simplified to a dichotomy 8 5. Reporters try to personalize the risk story 9 6. Claims of risk are usually more newsworthy than claims of safety 10 7. Reporters do their jobs with limited expertise and time 12 Dealing with the Public 14 1. Risk perception is a lot more than mortality statistics 14 2. Moral categories mean more than risk data 16 3. Policy decisions are seen as either risky 01 safe 17 4. Equity and control issues underlie most risk controversies 18 5 Risk decisions are better when the public shares the power 20 6. Explaining risk information is difficult hut not impossible if the motivation is there 21 7. Risk communication is easiei when emotions are seen as legitimate 2.'! Selected Bibliography 26 ------- 'Important If True' In colonial times newspaper "correspondents" were nothing more than acquaintances of the publisher, writing home from their travels. Unable to confirm or disconfirm their reports, cautious publishers often printed them under the headline "Important If True." "Explaining Environmental Risk" should be read in the spirit of this caution. While I have leaned heavily on the risk communication research literature where I could, many questions haven't been thoroughly studied, and here I have relied on my experience, my sense of other people's experience, and, frankly, my biases. If your experience and biases suggest different answers, try them. If you want to stick more closely to research findings, check the sources listed at the end. Why are so many risk assessment and risk management people beginning to take an interest in risk communication? There are two answers, I think, one entirely admirable and the other more open to question. The good news is that experts and managers are coming to recognize that how people perceive a risk determines how they respond to it, which in turn sets the context for public policy. It is hard to have decent policies when the public ignores serious risks and recoils in terror from less serious ones. The task of risk communication, then, isn't just conveying information, though that alone is a challenge; it is to alert people when they ought to be alerted and reassure them when they ought to be reassured. If your job is directing the cleanup at chemical spills, or running a right-to-know program, or siting new waste facilities—in fact, if your job has anything to do with setting or administering or following environmental regulations—explaining environmental risk is an important piece of your job. And it's probably a piece for which you have had little training. The more questionable reason for the growing interest in risk communication is the hope in some quarters that communicating about the environment can somehow replace managing it or regulating it aggressively. This is a common dilemma for communication specialists—advocates of bad policies sometimes imagine that they can get away with anything if they sell it cleverly enough, while advocates of good policies sometimes imagine that they don't have to sell at all. At a January 1986 national conference on risk communication (co-sponsored by the Conservation Foundation, the National Science Foundation, the ------- Environmental Protection Agency, and other organizations), the sessions on how to alert people to serious risks were sparsely attended, while overflow crowds pondered ways of calming people down. People sometimes need to be calmed down—but the ultimate goal of risk communication should be rational alertness, not passive trust. If a public that views risk with rational alertness strikes you as a desirable outcome, "Explaining Environmental Risk" should help. This is neither a theoretical treatise nor a nitty-gritty cookbook; along with the practical suggestions for effective communication, I have tried to explain why some strategies work and others fail, so that you can build on this understanding to design your own strategies. Though I hate to admit it, risk communication is a simpler field than risk assessment or risk management. It just isn't that hard to understand how journalists and nontechnical publics think about risk. But it is crucial to understand, and not mastering the rudiments of risk communication has led a lot of smart people to make a lot of foolish mistakes. With apologies to busy readers, I have therefore resisted the urge to produce an executive summary or a list of recommendations. Technicians can get by on cookbooks, perhaps, but decision-makers need to understand. Much depends, in fact, on whether you think risk communication is a job that can safely be left to "technicians" (public relations staff, community affairs officers) or whether—as I am convinced—you believe it must become an integral part of risk management. Although I hope public information people will find some value in what I have to say, my main goal is for environmenlal protection commissioners and plant managers to read it ... not merely pass it along to the public information office. The temptation to pass it along to the public information office—and then forget it—is almost overwhelming, I know. It's not just that decision-makers are busy people. It's not even that decision-makers don't realize how greatly their success depends on dealing effectively with the media and the public. It's more that they wish it weren't so, that dealing with the media and the public seems in so many ways the least pleasant, least controllable, least fair part of their work. Most risk managers, I suspect, spend a good deal of time hoping the media and the public will go away and leave them to do their jobs in peace. But since they won't, the next best thing is to understand better why they won't, how they are likely to react to what you have to say, and what you might want to say differently next time. I hope "Explaining Environmental Risk" will help. ------- Four on-going research projects have added greatly to my understanding of risk communication. They are: (1) "Environmental Risk Reporting" and "Risk Communication for Environmental News Sources" (with David B. Sachsman, Michael Greenberg, Audrey R. Gotsch, Mayme Jurkat, and Michael Gochfeld), both funded by the National Science Foundation Industry/University Cooperative Center for Research on Hazardous and Toxic Substances; (2) "Getting to Maybe: Building Toward Community-Developer Negotiations on New Hazardous Waste Facilities" (with Jim Lanard and Emilie Schmeidler), funded by the Fund for New Jersey; (3) "Manual and Conference for DEP Risk Communication" (with Caron Chess and B.J. Hance), funded by the New Jersey Spill Fund, New Jersey Department of Environmental Protection; and (4) "Radon Risk Communication Symposium and Recommendations" and "Radon Knowledge, Attitudes, and Behavior in New Jersey" (with Neil Weinstein), both funded by the New Jersey Department of Environmental Protection. Of course my colleagues and funders on these projects are not responsible for my speculations in this report. Several organizations have invited me to address them on strategies of risk communication, providing an opportunity to develop the ideas expressed in this report and test them on thoughtful and experienced audiences. I am grateful especially to the National Governors' Association, the New Jersey Hazardous Waste Facilities Siting Commission, the Council of Scientific Society Presidents, the Institute for Environmental Studies of the University of North Carolina, and the Air Pollution Control Association. Peter M. Sandman is Professor of Environmental Journalism at Cc-ok College, Rutgers University, New Brunswick, NJ, and Director of the Environmental Communication Research Program of the New Jersey Agricultural Experiment Station. Preparation of this report was funded by the Office of Toxic Substances of the United States Environmental Protection Agency as part of the Agency's effort to obtain diverse views on risk communication. Publication of this document does not signify that the contents necessarily reflect the views and policies of the Agency. ------- Dealing With The Media 1. Environmental risk is not a big story. The mass media are not especially interested in environmental risk. Reporters do care whether or not an environmental situation is risky; that's what makes it newsworthy. But once the possibility of hazard is established—that is, once someone asserts the risk on the record—the focus turns to other matters: how did the problem happen, who is responsible for cleaning it up, how much will it cost, etc. Assessing the extent of the risk strikes most journalists as an academic exorcise. The reporter's job is news, not education; events, not issues or principles. And the news is the risky thing that has happened, not the difficult determination of how risky it actually is. In an emergency, of course, the extent of the acute risk is the core of the story; radio reporters in particular want to know first and foremost whether to tell listeners to stay indoors, to evacuate, not to drink the water, etc. But the media don't especially want to know the ins-and-outs of risk assessment, the details of how great the risk is likely to be, how sure the experts are, or how they found out. If the story is important enough, these technical details merit a follow-up, a sidebar on the third or fourth day—but few stories are important enough. The typical news story on environmental risk, in other words, touches on risk itself, while it dwells on more newsworthy matters. In 1985 newspaper editors in New Jersey were asked to submit examples of their best reporting on environmental risk, and the articles were analyzed paragraph by paragraph. Only 32 percent of the paragraphs dealt at all with risk. Nearly half of the risk paragraphs, moreover, focused on whether a substance assumed to be risky was or was not present (e.g. is there dioxin in the landfill), leaving only 17 percent of the paragraphs that dealt directly with riskiness itself (e.g. how hazardous is dioxin). In a parallel study, reporters were asked to specify which information they would need most urgently in covering an environmental risk emergency. Most reporters chose the basic risk information, saving the details for a possible second-day story. What happened, how it happened, who's to blame, and what the authorities are doing about it all command more journalistic attention than toxicity during an environmental crisis. ------- The nature ot the crisis determines how nuK.h stress the media put on risk as opposed to other issues. Reporters know, tor example, that a chemical spill is a risk story, and at the scene ot a spill they will keep asking about toxic effects even alter they art: told the chemical is benign and inert. A tire story, on the other hand, automatically raises questions about how the tire started, how much damage was done, who turned in the alarm, and the like; many reporters won't realize unless told that a lire in a battery factory or a supermarket warehouse is a toxic event. But even when reporters understand that environmental risk is a ke\ element of the crisis, their appetite tor risk information is strong but easily sated; they want to know badly, but they don't want to know much. And when there is no crisis? The extent ot a chronic risk is newsworthy only when events make it so—for example. when a court battle or a regulatory action hinges on a disputed risk assessment. Sources wishing to "sell" a chronic risk story to the media must therefore work to make it newsworthy. Give it a news peg—that is, make something happen that reporters can cover. Make it interesting. Build the case tor its importance. Provide a prop worth focusing a camera on. But expect only partial success; reporters tlock to the scene of a crisis, but they have to be seduced into covering chronic risk. Among the greatest environmental risks in New Jersey is indoor radon contamination Because it is new and serious, it received considerable media attention in 1985 and early 1986. Then the coverage began to slip. The easy news pegs were over: the discovery of the problem, the first home in the state with a super-high reading, the passage of radon legislation. With no "radon industry" to tight back, the conflict that journalism teeds on has been conspicuously missing from the radon story. Radon is more a health problem and a housing problem than an environmental controversy, and its coverage is correspondingly muted. And radon at least has the "advantage" of cancer, the disease we love to hate. Imagine its low visibility it it gave people emphysema instead. 2. Politics is more newsworthy than science. The media's reluctance to focus on risk tor more than a paragraph or two might be less of a problem if that paragraph or two were a careful summary of the scientific evidence. It seldom is. In fact, the media are especially disinclined to cover the science ot risk. Most of the paragraphs devoted to risk in the New Jersey study consisted of unsupported opinion—someone asserting or denying the risk without documentation. Only 4.2 percent of the paragraphs (24 percent ot the risk paragraphs) took an intermediate or mixed or tentative ------- position on the extent of the risk And only a handful of the articles told readers what standard (if anv) existed tor the hazard in question, much less the status of research and technical debate surrounding the standard. The media's focus on the politics of risk rather than the science of risk is most visible in the sources relied upon in risk coverage. In the New Jersey study. 57 percent of the sources cited were government, with state government (22 percent) leading the pack. Industry captured 15 percent of the paragraphs; individual citizens and advocacy groups were cited in 7 percent each. Uninvolved experts such as academics—those least likely to have an axe to grind, most likely to have an intermediate opinion and a technical basis for it—were cited in only 6 percent of the paragraphs. Of course sources from government, industry, and environmental groups may also have scientific rationales for their judgments, and "experts" are not always neutral. Still, it is important that the media get their risk information from people who are directly involved in the news event; only occasionally do they seek out uninvolved experts for guidance on the extent of the risk. Trying to interest journalists in the abstract issues of environmental risk assessment is even tougher than trying to get them to cover chronic risk: abstract issues are not the meat of journalism. Yet the public needs to understand abstractions like the uncertainty of risk assessments, the impossibility of zero risk, the debatable assumptions underlying dose-response curves and animal tests. Where possible, it helps to embed some of these concepts in your comments on hot breaking stories—though reporters and editors will do their best to weed them out. When there is no breaking story, try to sell your favorite reporter on a feature on the fight over how conservative risk assessment ought to be. Emphasize that the problem underlies many of the stories he or she is covering. But understand why you will have only partial success, why the science of risk is inevitably less newsworthy than the politics of risk. 3. Reporters cover viewpoints, not "truths." Journalism, like science, attempts to be objective, but the two fields define the term very differently. For science, objectivity is tentativeness and adherence to evidence in the search for truth. For journalism, on the other hand, objectivity is balance. In the epistemology of journalism, there is no truth (or at least no way to determine truth): there are only conflicting claims, to be covered as fairly as possible, thus tossing the hot potato of truth into the lap of the audience. Imagine a scale from 0 to 10 ot all possible positions on an issue. Typically, reporters give short shrift to 0. 1. 9. and 10: ------- these \ievvs art: too extreme to be credible, and are covered as "oddball" if they are covered at all. (You may think some pretty extreme viewpoints get respectful media attention— but you haven't met the people reporters decide not to quote.) Reporters also pay relatively little attention to 4, 5, and B. These positions are too wishy-washy to make good copy; how do you build a story out of "further research is needed?" And sources with intermediate positions are unlikely to be heavily involved in the issue, certainly unlikely to seek media attention. Most ot the news, then, consists of 2's and 3's and 7's and 8's, in alternating paragraphs if the issue is hot, otherwise in separate stories as each side creates and dominates its own news events. Objectivity to the journalist thus means giving both sides their chance, and reporting accurately what they had to say. It does not mean tilling in the uninteresting middle, and it certainly does not mean figuring out who is right. Journalists who insist on trying to figure out who is right are encouraged to become columnists ... or to leave. If a risk story is developing and you have a perspective that you feel has not been well covered, don't wait to be called. You won't be. And you don't need to wait. Reporters are busy chasing after the sources they have to talk to, and listening to the sources who want to talk to them. If you're in the former category—if you're safety manager at a plant that just experienced an uncontrolled release, for example— reporters will find their way to you, like it or not. Otherwise, rather than suffer in silence, become one of the relatively few experts who keep newsroom telephone numbers in their rolodex. You will find reporters amazingly willing to listen, to put you in their rolodexes, to cover your point of view along with all the others. Insofar as you can, try to be a 3 or a 7—that is, a credible exponent of an identifiable viewpoint. Don't let yourself be pushed to a position that is not yours, of course, but recognize that journalism doesn't trust O's and 10's, and has little use for 5's. In deciding whether to brave the considerable risks of media exposure, bear in mind that the story will be covered. whether or not you arrange to be included. News items are allotted media attention to the extent that journalists see them as important and interesting. Then the search begins for information to fill the vacuum—preferably new, solid, comprehensible information that reflects an identifiable point of view, but if there's not enough of that to fill the time or space that the story "deserves." reporters will scrounge tor angles to make up the difference. The result can be an enlightening feature on the problems of technical prediction. but it's more likelv to be a "color storv"—the fears ot ------- bystanders, the views of ideologues, the speculations of spokespeople, the history of mismanagement. Environmental risk stories often turn into political stories in part because political content is more readily available than technical content. Experienced sources work at filling the vacuum. Although journalists tend not to believe in Truth-with-a-capital-T, they believe fervently in facts. Never lie to a reporter. Never guess. If you don't know, say you don't know. (But expect reporters to ask why you don't know.] If you don't know but can find out later, do so, and get back to the reporter as soon as possible, remembering that journalistic deadlines are measured in minutes, not months. If you know but can't tell, say you can't tell, and explain why. If you know but can't manage to say it in English, find someone who can. Reporters do not expect you to be neutral; in fact, they assume that you probably have an axe to grind, and prefer that you grind it visibly. They do expect you to grind it with integrity. 4. The risk story is simplified to a dichotomy. The media see environmental risk as a dichotomy; either the situation is hazardous or it is safe. This is in part because journalism dichotomizes all issues into sides to be balanced. But there are other reasons for dichotomizing risk. (1) It is difficult to find space for complex, nuanced, intermediate positions in a typical news story, say 40 seconds on television or 15 short paragraphs in a newspaper. (2) Virtually everyone outside his or her own field prefers simplicity to complexity, precision to approximation, and certainty to tentativeness. As Senator Edmund Muskie complained to an aide when the experts kept qualifying their testimony "on the other hand": "Eind me an expert with one hand." (3) Most of the "bottom lines" of journalism are dichotomies— the chemical release is either legal or illegal, people either evacuate or stay, the incinerator is either built or not built. Like risk managers, the general public is usually asked to make yes-or-no decisions, and journalists are not wrong to want to offer information in that form. Reporters are accustomed to the fact that technical sources invariably hedge, that nothing is ever "proved." They see this as a kind of slipperiness. Someone can always be found to advocate a discredited position (the tobacco industry has plenty of experts); no one wants to go too far out on a limb in case new evidence points in a different direction; researchers in particular like to leave the issue open so they can justify more research. Pinning down evasive sources is a finely honed journalistic skill. In terms of our O-to-10 scale, reporters spend a fair amount of time trying to get 5-ish sources to make clear-cut 3 or 7 statements. ------- Sources, especially technical sources, greatly resent the pressure from journalists to dichotomize and simplify. The dichotomization of risk distorts the reality that nothing is absolutely safe or absolutely dangerous, and polarizes "more-or-less" disagreements into "yes-or-no" conflicts. And oversimplification of any sort can mislead the audience and damage the reputation of the source. But recognize that journalists must simplify what they cover. If you refuse to simplify what you say, the reporter will try to do the job for you (at great risk to accuracy) or will turn to a more cooperative source. The most qualified person to simplify your views is you. Decide in advance what your main points are, and stress them consistently and repetitively, even if you have to hook them onto your answers to irrelevant questions. Leave out the technical qualifiers that your colleagues might insist on but the general public doesn't need to know (but leave in the qualifiers that really affect the bottom line). Stay away from jargon, and explain the technical terms you can't avoid. Check to make sure the reporter understands what you are saying; if the reporter looks glassy-eyed or starts frantically taking down every word, back up and start over. When you explain the significance of a toxic substance to reporters, try to avoid the "is it there or not" dichotomy, which can so easily alarm people about tiny concentrations. On the other hand, don't expect reporters to sit still for a dissertation on uncertainty in dose-response curves. Your best bet, when you can, is to specify the amount involved, then set it against some standard of comparison, ideally a government exposure standard. This is still a dichotomy, of course; it leaves the misimpression that exposures just under the standard are perfectly safe while exposures just over are deadly. But as dichotomies go, "over or under" is preferable to "there or not." If you want to fight the journalistic tendency to dichotomize risk, fight it explicitly, asserting that the issue is not "risky or not" but "how risky." Recognizing that intermediate positions on risk are intrinsically less dramatic and more complex than extreme positions, work especially hard to come up with simple, clear, interesting ways to express the middle view. Even so, expect reporters to insist on knowing "which side" you come down on with respect to the underlying policy dichotomy. 5. Reporters try to personalize the risk story. Perhaps nothing about media coverage of environmental risk so irritates technical sources as the media's tendency to personalize. "Have you stopped drinking it yourself?" "Would you let your family live there?" Such questions fly in ------- the i'ace of the soiuce's technical training to keep oneself out of one's research, and they confuse the evidentiary requirements of policy decisions with the looser ones of personal choices. But for reporters, questions that personalize are the best questions. They do what editors are constantly asking reporters to do: bring dead issues to life, make the abstract concrete, focus on real people facing real decisions. Personalizing also forces the source to dichotomize, to make the same "yea" or "nay" decision the reader or viewer must make. In a sense, experts and policy-makers work at a different level of analysis than reporters and the public. As an EPA study on the ethelyne dibromide controversy noted, the agency wanted to talk about "macro-risk" (how many deaths will result from EDB contamination), while reporters kept asking about "micro-risk" (is it okay to eat the cake mix). The connections between macro-risk and micro-risk are difficult to draw. But for the individual citizen (faced with a cake mix, not a regulatory proposal), micro-risk is the issue, and reporters are not off-base in pushing technical sources to trace the connections. This is what personalizing questions are designed to do. Knowing that reporters will inevitably ask personalizing questions, be prepared with answers. It is often possible to answer with both one's personal views and one's policy recommendations, and then to explain the difference if there is one. Or come with colleagues whose personal views are different, thus dramatizing the uncertainty of the data. If you are not willing (or not permitted) to acknowledge your own views, plan out some other way to personalize the risk, such as anecdotes, metaphors, or specific advice on the individual micro-risk level. 6. Claims of risk are usually more newsworthy than claims of safety. On our O-to-10 scale of risk assertions, the 3's and 7's share the bulk of the coverage, but they don't share it equally. Risk assertions receive considerably more media attention than risk denials. Sometimes, in fact, the denials get even less coverage than the intermediate position, and reporters wind up "balancing" strong assertions of risk with bland statements that the degree of risk is unknown. In the New Jersey study, the proportions were 58 percent "risky," 18 percent "not risky," and 24 percent mixed or intermediate. This is not bias, at least not as journalism understands bias. It is built into the concept of newsworthiness. If there were no allegation of risk, there would be no story. That something here might be risky is thus the core of the story; having covered it, the media give rather less attention to the counterbalancing notion that it might not be risky. 10 ------- Other factors contribute to the tilt toward alarming news. One is the reporter's desire to "build" the story, to come back with something that editors will want to showcase. (Reporters are much more interested in selling stories than in "selling newspapers.") Another factor is the journalist's preference for simple, graphic language, for "dump" rather than "land emplacement." Risks sound riskier in simple language than in technical jargon. The factor closest to outright bias—but still distinguishable in the minds of journalists—is the media's traditional skepticism toward those in authority. Most news is about powerful people, but along with the advantage of access government and industry must endure the disadvantage of suspicion. Environmental groups, by contrast, receive li-ss attention from the media, but the attention is more consistently friendly. On the other hand, the media are often and justly criticized for being too slow to alert the public to new environmental hazards. Considering that we rely largely on journalism as an "early warning system" for social problems on the horizon, this is a serious criticism. To gain a journalistic hearing, the first source to assert a particular risk must be reasonably credible, highly committed, and very lucky or very skilled. Almost invariably, new technologies start out with sweetheart coverage. The environmental controversy comes later, and only after the controversy is on the media agenda (and the technology is perhaps too deeply embedded to be dislodged) does the risky side of the argument catch up and pull ahead. This may be the worst of all possible patterns: to fail to warn us about risks when it's early enough to make a societal go/no-go decision, then to frighten us deeply about risks after the decision has been made. The principal exception to this pattern is emergencies. On a chronic risk story, the risk is the story. But a genuine emergency is by definition a big story; freed from the need to build the story, the reporter—especially the local reporter— may try to prevent panic instead. The President's Commission on the Accident at Three Mile Island conducted a content analysis of network, wire service, and major newspaper coverage during the first week of the 1979 accident. The Commission's expectations of sensationalism were not confirmed. Of media passages that were clearly either alarming or reassuring in thrust, 60 percent were reassuring. If you stick to the technical issues, eliminating passages about inadequate flow of information and general expressions of tearfulness from local citizens, the preponderance of reassuring over alarming statements becomes 73 percent to 27 percent. It didn't seem that way at the time, of course. The information that something previously assumed to be safe may or may not be hazardous naturally strikes people as 11 ------- alarming, almost regardless ot the amount of attention paid to the two sides; imagine reading this evening that scientists disagree over whether your favorite food is carcinogenic;. Thus, sociologist Allan Mazur has found that public, fearfulness about risky new technologies is proportional to the amount of coverage, not to its character. Media coverage of environmental risk alerts the public to risks it was otherwise unaware of, and thus increases the level ot alarm even when it is balanced. None of this is a rationale for avoiding the media. Even balanced media coverage may not reliably lead to balanced public opinion, but balanced coverage is preferable to unbalanced coverage. And the coverage is most likely to be balanced when sources on all sides are actively trying to get covered. People with knowledge and opinions to share perform a public service when they share them. What can you do to alert people to the risks of a new technology before it is too late? What can you do to redress the alarming imbalance once the media have begun to overdramatixe the risks? Energetic public relations will help with both tasks. though in both cases you will be working against the grain 7. Reporters do their jobs with limited expertise and time. At all but the largest media, reporters covering environmental risk are not likely to have any special preparation tor the assignment. Specialized environmental reporters are more the exception than the rule. Reporters covering an environmental emergency, for example, are mostly general-assignment reporters or police reporters, sent to the scene (or the phones) without time to scan the morgue, much less a technical handbook. And reporters tend to be science-phobic in the first place; the typical college journalism major takes only two science courses, and chooses those two carefully in an effort to avoid rigor. Though there are many exceptions, the average reporter approaches a technical story with trepidation (often hidden by professional bravado), expecting not to understand. It doesn't help that the average reporter covers and writes two to three stories a day. Here too there are exceptions, hut most journalists are in a great hurrv most ot the time. They must make deadline not just on this stor\. but quite often on the story they will be covering after this one. Their goal. reasonably, is not to find out all that is known, but just to find out enough to write the story Even it thev knew more, they would not have the space or airtime to report more, nor do they believe their readers or viewers would have the interest or patience to absorb more. Note also that irrespective of what journalistic superstars earn, the average reporter at a small daily newspaper takes home perhaps S13.000-S18.000 a year Considering their ------- incomes, journalists are shockingly competent and dedicated. but there are limits to hcnv much competence and dedication a salary in the teens can purchase. If the idea appeals to you. by all means otter to teach local journalists the basics of your iield—but don't expect general assignment reporters to find much time (or much stomach) for technical training they will use only a tew times a year. A beat reporter who covers your issue full-time (it you are lucky enough to have one) is a much better candidate for technical training. Better still, train yourselt (and \our colleages and stati) in dealing with the media. Hiring effective public information specialists also helps, but reporters much prefer to talk to the people in charge and the people in the know Especially during an emergency, press calls often go to the boss and the expert instead of the press office, so the boss and the expert should know how to talk to reporters 'The annals of risk communication are full of stories of corporate managers and agency bureaucrats who shot themselves in the toot—and permanently damaged their organizations—because they hadn't the least idea of how to deal with the media. Even the best communication skills can't rescue a technical disaster, of course; who wants to handle the PR at Chernobyl or Bhopal? But inadequate communication skills can create a disaster that needn't have been. And adequate communication skills are not so hart! to develop. All it takes is a little understanding of how the media work, a little training in dealing with reporters, and a little experience to smooth out the rough edges. \Yhv. then. do so many managers, bureaucrats, and technical experts avoid all contact with the media7 Because it's risky. Reporters don't always understand what you're telling them: they don't always share your goals and values. the\ don't always handle their jobs the way you want them to. In all these ways and mam others, reporters may be different from the people you usually work with And so working with reporters may sound like something less than an unalloyed pleasure. Pleasure or not. the risks ot due,king the media are far greater than the risks ot working with them. Every news story about environmental risk is a collaboration between the journalists working on the story and the sources they talk to. There's not too much you can do to change the nature ot journalism or the performance of journalists But you can understand them and figure out hou to deal with them B\ improving your own performance as a source, you < (in bring about a real improvement in media coverage of environmental risk ------- Dealing With The Public 1. Risk perception is a lot more than mortality statistics. If death rates are the only thing vou care about, then the public is afraid of the wrong risks. Thai is. public fears are not well correlated with expert assessments or mortality statistics. This is often seen as a perceptual distortion on the part of the public, but a more useful wax to see? it is as an oversim- plification on the part of many experts and policy-makers. In other words, the concept of "risk" means a lot more than mortality statistics. Virtually everyone would rather dri\-e home from a party on the highway than walk home on deserted streets. Even if we do not miscalculate the relatix'e statistical likelihood of a fatal mugging versus a fatal car crash, the possibility of getting mugged strikes us as an outrage, xvhile xve accept the possibility of an auto accident as x'oluntary and largely controllable through good driving. (Eighty-five percent of all drivers consider themselves better than average.) Similarly, a household product, however carcinogenic., seems a lot less risky than a high-tech ha/ardous waste treatment facility—the former is familiar and under one's oxvn control, while the latter is exotic and controlled by others. Risk perception experts (especially psychologists Paul Slovic, Sarah Lichtenstein, and Baruch Eischhoff) have spent years studying hoxv people interpret risk. The folloxving list identifies some of the characteristics other than mortality that factor into our working definitions of risk. Remember, these are not distortions of risk; they are part of xvhat we mean by the term. Less Risky Voluntary Familiar Controllable Controlled by self Fair Not memorable Not dread Chronic Diffuse in time and space Not fatal Immediate Natural Individual mitigation possible Detectable More Risky Involuntary Unfamiliar Uncontrollable Controlled by others Unfair Memorable Dread Acute Focused in time and space Fatal Delayed Artificial Individual mitigation impossible Undetectable 14 ------- The very same risk—as experts see these things—will be understood quite differently by the lay public depending on where it stands on the dimensions listed above. Some thirty percent of the homes in northern New Jersey, for example, have enough radon seeping into their basements to pose more than a one-in-a-hundred lifetime risk of lung cancer, according to estimates by the U S. Environmental Protection Agency and the State Departments of Health and Environmental Protection. But despite considerable media attention (at least in the beginning), only five percent of North Jersey homeowners have arranged to monitor their homes for radon, and even among these few the level of distress is modest—compared, say. to the reaction when dioxin is discovered in a landfill, objectively a much smaller health risk. State officials were initially concerned about a radon panic, but apathy has turned out to be the bigger problem. The source of the radon in New Jersey homes is geological uranium; it has been there since time immemorial, and no one is to blame. But three New Jersey communities— Montclair, Glen Ridge, and West Orange—have faced a different radon problem: landfill that incorporated radioactive industrial wastes. Though their home readings were no higher than in many homes on natural hotspots, citizens in the three communities were outraged and fearful, and they successfully demanded that the government spend hundreds of thousands of dollars per home to clean up the landfill. The state's proposal to dilute the soil nearly to background levels and then dispose of it in an abandoned quarry in the rural community of Vernon has provoked New Jersey's largest environmental demonstrations in years, with thousands of residents swearing civil disobedience sooner than let the trucks go through. In nearby communities threatened by naturally occurring radon, meanwhile, the concern is minimal. It doesn't help to wish that people would confine their definitions of risk to the mortality statistics. They won't. Mortality statistics are important, of course, and policy- makers understandably prefer to focus on the risks that are really killing people, rather than the risks that are frightening or angering people because they are involuntary, unfamiliar, uncontrollable, etc. But successful risk commuication begins with the realization that risk perception is predictable, that the public overreacts to certain sorts of risks and ignores others, that you can know in advance whether the communication problem will be panic, or apathy And since these differences between risks are real and relevant, it helps to put them on the table. Merely acknowledging that a risk seems especially tearful because it 13 ------- is unfamiliar or unfair will help. Doing something to remedy the unfamiliarity or unfairness will help even more. Just to make things more complicated, risk perception is not linear, not for anybody. That is, you can't just multiply how probable a risk is by how harmful it is to get how badly people want to prevent it. (It you could, there would be no insurance industry and no gambling industry.) In general, people will pay more to protect against low-probability loss than to pursue low-probability gain—but if the price is low enough to be dismissed as negligible, even an infinitesimal chance at a big payoff looks good. Risk judgments are also very responsive to verbal cues. Doctors, for example, are much more likely to prescribe a new medication that saves 30 percent of its patients than one that loses 70 percent of them. A pollutant or an accident that will eventually give cancer to 10,000 people sounds very serious, but one that will add less than one tenth of one percent to the national cancer rate sounds almost negligible. There is in fact no "neutral" way to present risk data, only ways that are alarming or reassuring in varying degrees. Finally, people's perception of risk is greatly influenced by the social context. Our responses to new risks, in fact, are largely predictable based on our enduring values and social relationships. Do we like or dislike, trust or distrust the people or institutions whose decisions are putting us at risk? Do our friends and neighbors consider the risks tolerable or intolerable? Are they enduring higher risks than ours, or escaping with lower ones? All these factors, though they are irrelevant to the mortality statistics, are intrinsic parts of what we mean by risk. 2. Moral categories mean more than risk data. The public is far from sure that risk is the real issue in the first place. Over the past several decades our society has reached near- consensus that pollution is morally wrong—not just harmful or dangerous, not just worth preventing where practical, but wrong. To many ears it now sounds callous, if not immoral, to assert that cleaning up a river or catching a midnight dumper isn't worth the expense, that the cost outweighs the risk, that there are cheaper ways to save lives. The police do not always catch child molesters, but they know not to argue that an occasional molested child is an "acceptable risk." Government agencies build their own traps when they promulgate policy (and public relations) in the language of morality, depicting food additives or chemical wastes or polluted water as evils against which they vow to protect the innocent public. It is not at all obvious which environmental "insults" (another term with moral overtones) a society should reject on moral grounds and which it should assess 16 ------- strictly in terms of impact. But an agency that presents itself and its mission in moral terms should expect to he held to its stance. And an agency that wishes to deal with environmental risk in terms of costs-and-benefits instead of good-and-evil should proceed gently and cautiously, aware that it is tramping on holy ground. Nor is morality the only principled basis for questioning the costs-and-benefits premises of risk assessment. Just as the moralist challenges the Tightness of trading oil certain risks against costs or benefits, the humanist challenges the coherence of the tradeoffs. How, the humanist asks, can anyone make sense of a standard that tries to put a cash value on human life? Or, indeed, of a standard that assumes that a hundred widely scattered deaths per year are equivalent to a one-in-a-hundred chance of obliterating a community of 10,000? Similarly, the political critique of the premises of risk assessment begins by noting that "the greatest good for the greatest number" has always been a convenient rationale for the oppression of minorities. Democratic theory asserts that individuals and groups should be free to bargain for their own interests, and should be protected from the tyranny of the majority. There is nothing unreasonable about the suggestion that equitable distribution of risks and benefits—and of the power to allocate risks and benefits—is often more important than the minimization of total risk or the maximization of total benefit. It may be efficient to dump every environmental indignity on the same already degraded community, but it is not fair. 3. Policy decisions are seen as either risky or safe. Like the media, the public tends to dichotomize risk. Either the risk is seen as very frightening, in which case the response is some mix of fear, anger, panic, and paralysis; or the risk is dismissed as trivial, in which case the response is apathy. In their personal lives, people do not necessarily dichoto- mize risk. Most of us are quite capable of understanding that the picnic might or might not be rained out, that the boss might or might not get angry, even that smoking might or might not give us lung cancer. Of course quantified probabilistic statements are genuinely hard to understand, especially when the probabilities are small, the units are unfamiliar, and the experts disagree. But beyond these perplexities lies another issue of enormous importance to risk communication. While people may (with difficulty) master a probabilistic risk statement that concerns what they should do to protect themselves, they are bound to resist probabilistic risk statements that concern what others (government, say) should do to protect them. On my own 17 ------- behalf, I may choose to tolerate a risk or to piotect against it, but lor you to deude that mv risk is tolerable is itselt intolerable. Quantitative risk assessments, risk-beneiit calculations, risk-cost ratios, and risk-risk comparisons are all hard to hear when we bear the risk and someone else makes the decision. 4. Equity and control issues underlie most risk controversies. Trust and credibility are often cited as the key problems ot risk communication. Certainly few people trust government and industry to protect them from environmental risk. This is just as true ot the passive, apparently apathetic public as it is of the activist, visibly angry public. The former is simply more fatalistic, more prone to denial, more completely drowned in undiscriminating chemophobia. The activist public, in other words, distrusts others to protect its interests and thus chooses to protect its own. The far larger passive public is passive not because it believes others will protect its interests, but because it doubts it can protect its own. Both publics listen to the reassurances of government and industry—if they listen at all—with considerable suspicion. But to say that trust is the problem here is to assume that the goal is a passive public that doesn't mind being passive. If the goal is an actively concerned public, then the problem isn't that people are distrustful, but rather that government and industry demand to be trusted. Translate the question of trust into the underlying issue of control. Who decides what is to be done? Any environmental risk controversy has two levels. The substantive issue is what to do; the process issue is who decides. So long as people feel disempowered on the process issue, they are understandably unbending on the substantive issue, in much the same way as a child forced to go to bed protests the injustice of bedtime coercion without considering whether he or she is sleepy. It isn't just that people oppose any decision they view as involuntary and unfair, regardless of its wisdom; because the equity and control issues come first, people typically never even ask themselves whether they agree on the merits. Outraged at the coercion, they simply dig in their heels. It is hardly coincidental that risks the public tends to overestimate generally raise serious issues of equity and control, while most of the widely underestimated risks (smoking, fat in the diet, insufficient exercise, driving without a seatbelt) are individual choices. Specialists in negotiation and conflict resolution have long understood this relationship between substantive issues and the process issues ot equity and control. Consider for 18 ------- example a community chosen by the state government to "host" a hazardous waste incinerator. Justly offended at this infringement of local autonomy, the community prepares to litigate, frantically collecting ammunition on the unacceptability ot the site. Both their anger and the legal process itself encourage community members to overestimate the risk of the proposed facility, to resist any argument that some package of mitigation, compensation, and incentives might actually yield a net gain in the community's health and safety, as well as its prosperity. In interviews with community members faced with such a situation, the control issue tends to overshadow the risk assessment, But when citizens are asked to hypothesize a de facto community veto and envision a negotiation with the site developer, they become quite creative in designing an agreement they might want to sign: emissions offsets. stipulated penalties, bonding against a decline in property values, etc. It is still too early to tell whether a negotiated hazardous waste treatment facility is feasible. But thinking about such a negotiation becomes possible for community members only when they feel empowered—that is, when the issue of outside coercion has been satisfactorily addressed. On this dimension people's response to information is not much different from their response to persuasion. We tend to learn for a reason—either we're curious, or we're committed to a point of view and looking for ammunition, or we're faced with a pending decision and looking for guidance. These three motivations account for most information-seeking and most learning—and none of them exerts much influence when an individual citizen is ottered information about, say, a Superfuncl clean-up plan. A few stalwart souls will read out of curiosity, though it won't take much technical detail to put a stop to that. Activists will scour the plan for evidence to support their position or for evidence that their position wasn't properly considered. (Activists know what they think and believe they can make a difference.) And those charged with litigating, funding, or implementing the plan study it in order to do their jobs. And the general public? Why learn if you feel powerless do anything about what you have learned? On the other hand, when the public has felt it was exercising real influence on a decision—the ASARCO smelter in Tacoma comes to mind—it has shown a surprising ability to master the technical details, including risk assessment details. Not that every citizen wants to play a pivotal role in environmental decision. U'e have our own lives to lead, and we would prefer to trust the authorities. It the issue is unimportant enough we often decide to trust the authorities despite our reservations; it the crisis is urgent enough we 19 ------- may feel we have no choice but to trust the authorities, again despite our reservations. The gravest problems of risk communication tend to arise \\lien citizens determine that the issue is important, that the authorities cannot be trusted. and that they themselves are powerless. Then comes the backlash of outrage. 5. Risk decisions are better when the public shares the power. People learn more and assess what they learn more carefully if they exercise some real control oxer the ultimate decision. But this sort oi power-sharing is, of course. enormously difficult for policy-makers, for a wide range of political, legal, professional, and psychological reasons. Interestingly, corporate officials may sometimes find power-sharing less unpalatable than government officials. Corporations have a bottom line to nurture, and when all else fails they may see the wisdom of sharing power in the interests of profit. But government officials have no profit to compensate for the loss of power, so they may find it harder to share. "Public participation," as usually practiced, is not a satisfactory substitute for power-sharing. To be sure, telling the public what you're doing is better than not telling the public what you're doing. Seeking "input" and "feedback" is better still. But most public participation is too little too late: "After years of effort, summarized in this 300-page report, we have reached the following conclusions... Now what do you folks think?" At this point it is hard enough for the agency to take the input seriously, and harder still for the public to believe it will be taken seriously. There is little power- sharing in the ''decide-announce-defend" tradition of public participation. The solution is obvious, though difficult to implement. Consultations with the public on risk management should begin early in the process and continue throughout. This means an agency must be willing to tell the public about a risk before it has done its homework—before the experts have assessed the risk thoroughly, before all the policy options have been articulated, way before the policy decisions have been made. There are dangers to this strategy: people will ask the agency what it proposes to do about the problem, and the agency will have to say it isn't sure yet. But on balance an agency is better off explaining why it doesn't yet have all the answers than explaining why it didn't share them years ago. In fact, not having all the answers can be made into an asset, a demonstration of real openness to public input. The goal, after all, is to enlist the rationality of the citizenry, so that citizens and experts are working together to figure out how great the risk is and what to do about it. 20 ------- Of course no responsible agency will go public without any answers. What's important is to propose options X, Y, and Z tentatively, with genuine openness to V and W, and to community comments that may eliminate Z. A list of options and alternatives—and a fair and open procedure for comparing them and adding new ones—is far more conducive to real power-sharing than a "draft" decision. This sort of genuine public participation is the moral right of the citizenry. It is also sound policy. Undeterred by conventional wisdom, lay people often have good ideas that experts can adapt to the situation at hand; at a minimum, lay people are the experts on what frightens them and what would reassure them. When citizens participate in a risk management decision, moreover, they are far more likely to accept it, for at least three reasons: (1) They have instituted changes that make it objectively more acceptable; (2) They have got past the process issue of control and mastered the technical data on risk; that is, they have learned why the experts consider it acceptable; and (3) They have been heard and not excluded, and so can appreciate the legitimacy of the decision even if they continue to dislike the decision itself. 6. Explaining risk information is difficult but not impossible, if the motivation is there. High school teachers have long marveled that a student who couldn't make sense of Dickens's A Tale of Two Cities had no trouble with Hot Rod's far more complex instructions on how to adjust one's sparkplugs for a fast start on a rainy day. Motivation makes the difference. When people have a reason to learn, they learn. It is still possible for communicators to make the learning easier or harder—and scientists and bureaucrats have acquired a fairly consistent reputation for making it harder. At Three Mile Island, for example, the level of technical jargon was actually higher when the experts were talking to the public and the news media than when they were talking to each other. The transcripts of urgent telephone conversations between nuclear engineers were usually simpler to understand than the transcripts of news conferences. To be sure, jargon is a genuine tool of professional communication, conveying meaning (to those with the requisite training) precisely and concisely. But it also serves as a tool to avoid communication with outsiders, and as a sort of membership badge, a sign of the status difference between the professional and everyone else. Like any piece of professional socialization, the tendency to mystify outsiders becomes automatic, habitual more than malevolent. It's hard for a layperson to get a straight answer from an expert even when nothing much is at stake. When a potentially serious risk is at stake, when people are 21 ------- frightened or angry or exhausted, when the experts aren't sure what the answers are. when the search for a scapegoat is at hand, effective communication is a lot to expect. In many risk communication interactions, in short, the public doesn't really want to understand (because it feels powerless and resentful) and the experts don't really want to be understood (because they prefer to hold onto their information monopoly). The public finds it convenient to blame the experts for obfuscation, and the experts find it convenient to blame the public for obtuseness. These motivational issues are probably more important than the traditional concerns of clarity in determining whether real knowledge will pass from expert to public. Within the traditional concerns of clarity, the major issue is simplification. Even assuming a public that wants to understand and an expert who wants to be understood, risk information must still be simplified. Insofar as possible, of course, it is wise to simplify language rather than content. That is, take the extra words to make hard ideas clear. Unfortunately, neither the expert source nor the lay audience is usually willing to dedicate the time needed to convey complex information a step at a time. So inevitably simplification becomes a matter of deciding what information to leave out. Experts are famous for their conviction that no information may be left out; unable to tell all, ihey often wind up telling nothing. In fact, there are three standard rules of thumb for popularizing technical content. (1) Tell people what you have determined they ought to know—the answers to the questions they are asking, the instructions for coping with the crisis, whatever. This requires thinking through your information goals and your audience's information needs, then resolutely keeping the stress where you have decided it should be. (2) Add what people must know in order to understand and feel that they understand the information— whatever context or background is needed to prevent confusion or misunderstanding. The key here is to imagine where the audience is likely to go off-track, then provide the information that will prevent the error. (3) Add enough qualifiers and structural guidelines to prepare people for what you are not telling them, so additional information later will not leave them feeling unprepared or misled. Partly this is just a matter of sounding tentative; partly it is constructing a scaffolding of basic points on which people can hang the new details as they come in. Applying these three rules isn't easy, but it is a lot easier than trying to tell everything you know. The hardest part of simplifying risk information is explain- ing the risk itself. This is hard no! only because risk assess- 22 ------- merits are intrinsically complex and uncertain, but also because audiences cling tenaciously to their safe-or- dangerous dichotomy. One path out of dichotomous thinking is the tradeoff: especially risk benefit, but also risk-cost or risk-risk. But there is solid evidence that lay people resist this way of thinking; trading risks against benefits is especially offensive when the risks raise moral issues and the "victims" are not the ones making the choice. Another alternative to dichotomy is the risk comparison: X is more dangerous than Y and less dangerous than Z. But as we have already noted, risk means a lot more than mortality statistics, and comparing an involuntary risk like nuclear power to a voluntary one like smoking invariably irritates more than it enlightens—as does any risk comparison that ignores the distinctions listed at the start of this section. The final option to dichotomy is to provide the actual data on deaths or illnesses or probability of occurrence or whatever. This must be done carefully, with explicit acknowledgement of uncertainty, of moral issues, and of non-statistical factors like voluntariness that profoundly affect our sense of risk. Graphs and charts will help; people understand pictorial representations of probability far better than quantitative ones. Don't expect too much. People can understand risk tradeoffs, risk comparisons, and risk probabilities when they are carefully explained. But usually people don't really want to understand. Those who are frightened, angry, and powerless will resist the information that their risk is modest; those who are optimistic and overconfident will resist the information that their risk is substantial. Over the long haul, risk communication has more to do with fear, anger, powerlessness, optimism and overconfidence than with finding ways to simplify complex information. 7. Risk communication is easier when emotions are seen as legitimate. It follows from what we have been saying that an important aspect of risk communication is finding ways to address the feelings of the audience. Unfortunately, experts and bureaucrats find this difficult to do. Many have spent years learning to ignore feelings, their own and everyone else's; whether they are scientists interpreting data or managers setting policy, they are deeply committed to doing their jobs without emotion. At an even deeper level, scientists and bureaucrats have had to learn to ignore the individual, to recognize that good science and good policy must deal in averages and probabilities. This becomes most obvious when a few people feel threatened by a generally desirable action, such as the siting of a hazardous waste facility. Experts who are confident that the risk is small and the facility needed may ------- welt try to sympathize with the target community—but their training tells them playing the odds is a good bet, somebody has to take the risk, the decision is rational, and that's the end of the matter. Thus the most common sources of risk information are people who are professionally inclined to ignore feelings. And how do people respond when their feelings are ignored? They escalate—yell louder, cry harder, listen less—which in turn stiffens the experts, which further provokes the audience. The inevitable result is the classic drama of stereotypes in conflict: the cold scientist or bureaucrat versus the hysterical citizen. Breaking this self-defeating cycle is mostly a matter of explicitly acknowledging the feeling (and the legitimacy of the feeling) before trying to explain anything substantive— because any effort to explain substance first will be experienced by people as just another way of not noticing how they feel. The trick, in other words, is to separate the feeling from the substance, and respond to the feeling first. "I can tell you're angry about this" won't eliminate the anger— nor should it—but it will eliminate the need to insist on the anger, and will thus free energy to focus on the issue instead. "A lot of people would be angry about this" and "in your position I would be angry about this" are even more empathic remarks, legitimating the anger without labeling the citizen. All three responses are far more useful than pretending that the anger isn't there or, worse yet, demanding that it disappear. Techniques of this sort are standard practice in many professional contexts, from police crisis intervention to family counseling. Training is available; risk communicators need not reinvent the wheel. It helps to realize that experts and bureaucrats—their preferences notwithstanding—have feelings too. In a public controversy over risk, they are likely to have very strong feelings indeed. After all, they consider themselves moral people, yet they may be accused of "selling out" community health or safety or environmental protection. They consider themselves competent professionals, yet they may be accused of egregious technical errors. They very likely pride themselves on putting science or public service ahead of personal ambition, yet they may be accused of not caring. They chose their careers expecting if not gratitude at least a calm working environment and the trust and respect of the community. Instead they are at the center of a maelstrom of community distrust, perhaps even community hatred. It hurts. 24 ------- The pain can easily transform into a kind of icy paternal- ism, an "I'm-going-to-help-you-even-if-you-don't-know-what's- good-for-you" attitude. This of course triggers even more distrust, even stronger displays of anger and fear. Risk communication stands a better chance of working when both sets of feelings—the expert's and the community's—are on the table. Feelings are not usually the core issue in risk communica- tion controversies. The core issue is usually control, and the way control affects how people define risk and how they approach information about risk. But the stereotypical conflict between the icy expert and the hysterical citizen is nonetheless emblematic of the overall problem. The expert has most of the "rational" resources—expertise, of course; stature; formal control of the ultimate decision. Neither a direct beneficiary nor a potential victim, the expert can afford to assess the situation coldly. Indeed, the expert dare not assess the situation in any other way. The concerned citizen, meanwhile, has mainly the resources of passion— genuine outrage; depth of commitment; willingness to endure personal sacrifice; community solidarity; informal political power. To generate the energy needed to stop the technical juggernaut, the citizen must assess the situation hotly. A fundamental premise of "Explaining Environmental Risk" is that risk understanding and risk decision-making will improve when control is democratized. We will know this is happening when citizens begin approaching risk issues more coolly, and experts more warmly. 25 ------- Selected Bibliography Covello, Vincent T., "The Perception of Technological Risks: A Literature Review," Technological Forecasting and Social Change, 1983, pp. 285-287. Covello, Vincent T., Detlof von Winterfeldt, and Paul Slovic, "Communicating Scientific Information about Health and Environmental Risks: Problems and Opportunities from a Social and Behavioral Perspective," in V. Covello, A. Moghissi, and V.R.R. Uppuluri, Uncertainties in Risk Assessment and Risk Management (New York: Plenum Press, 1986), in press. Fischhoff, Baruch, "Protocols for Environmental Reporting: What to Ask the Experts," The Journalist (Foundation for American Communications), Winter 1985, pp. 11-15. Klaidman, Stephen, "Health Risk Reporting," Institute for Health Policy Analysis, Georgetown University Medical Center, Washington, DC, 1985. Mazur, Allan, "Media Coverage and Public Opinion on Scientific Controversies," journal of Communication, 1981, pp. 106-115. Mazur, Allan, "Bias in Risk-Benefit Analysis," Technology in Society, 1985, pp. 25-30. Nelkin, Dorothy, Science in the Streets (New York: Twentieth Century Fund, 1984). President's Commission on the Accident at Three Mile Island, Report of the Public's Right to Information Task Force (Washington, DC: U.S. Government Printing Office, 1979). RuckeJshaus, William, "Risk in a Free Society," Risk Analysis, September 1984, pp. 157-163. Sandman, Peter M., "Getting to Maybe: Some Communications Aspects of Hazardous Waste Facility Siting," Seton Hall Legislative Journal, Spring 1986. 26 ------- Sandman, Peter M., David B. Sachsman, Michael Greenberg, Mayme Jurkat, Audrey R. Gotsch, and Michael Gochfeld, "Environmental Risk Reporting in New Jersey Newspapers," Environmental Risk Reporting Project, Department of Journalism and Mass Media, Rutgers University, January 1986. Sharlin, Harold I., "EDB: A Case Study in the Communication of Health Risk," Office of Policy Analysis, U.S. Environmental Protection Agency, January 1985. Slovic, Paul, "Informing and Educating the Public About Risk," Decision Research Report 85-5, November 1984. Slovic, Paul, Baruch Fischhoff, and Sarah Lichtenstein, "Facts and Fears: Understanding Perceived Risk," in R.C. Schwing and W. Al Albers, eds., Societal Risk Assessment: How Safe Is Safe Enough? (New York: Plenum, 1980), pp. 181-216. Weinstein, Neil D., and Peter M. Sandman, "Recommendations for a Radon Risk Communication Program," Office of Science and Research, New Jersey Department of Environmental Protection, November 1985. ------- ------- |