Marie-Christine Therrien, Professor
École nationale d'administration publique
marie-christine.therrien@enap.ca
With the collaboration of Anaïs Valiquette-L'Heureux
A crisis has been defined as “a situation that threatens the high-priority goals of the decision making unit, restricts the amount of time available before the decision is transformed, and surprises the members of the decision making unit by its occurrence” (Hermann, 1972, p. 13). Crisis and risk management is not simply a matter of public relations, but instead calls for strategic and proactive measures, a culture of prevention, well-tested preparedness and mechanisms for detecting and catching “weak” warning signals in time. Another, somewhat fuller definition of the word crisis is given by Pauchant and Mitroff (1988): “a disruption that physically affects the system as a whole and threatens its basic assumptions, its subjective sense of self, its existential core.” The authors go on to say that a crisis “can threaten the legitimacy of an entire industry… reverse the strategic mission of an organization” and “disturb people's subjective world: the way they perceive the world and themselves; their inner sense of self-worth, power and identity; their inner cohesion”.
The word crisis comes from the Greek krisis, designating the action of distinguishing, separating, choosing or judging. The earliest use of the word in English dates to the 16th century, in reference to a turning point, either for better or for worse, in the course of an illness; by the 17th century, the meaning of “crisis” had extended to designate an important or vital stage in the progress of anything. Today the term is applied especially to times of difficulty, insecurity and suspense in politics or commerce. (Oxford English Dictionary; see also Bolzinger, 1982, pp. 475-480; Béjin and Morin, 1876, pp. 1-2).
The field of risk and crisis management developed as advances were made in the discipline of risk analysis (or risk assessment), which is concerned with industrial safety and the anticipation of post-accident crises. Researchers in this field introduced the notion of major technological risk (Lagadec, 1979 and 1981). Knowledge about natural and human-related disasters, risk assessment and crisis management has grown as a result of work by sociologists, anthropologists, geographers, psychologists, international relations specialists and, more recently, researchers in administration and management science (Lagadec, 1991, pp. 13-15).
It is sometimes tempting to reduce crisis management to managing media coverage of a crisis or disaster. While this aspect is important, crisis management encompasses much more. It is an ongoing process that should be an integral part of an organization's management and culture. Crisis management requires considerable organizational and management skills. Crucially, those in charge must have the ability to think “far outside of the boxes” and imagine the unimaginable, as well as being able to gain support from every unit of the organization (Mitroff, 2001, p. 121 and p. 151).
The issues of safety and crisis vulnerability in high-technology systems are viewed differently by two schools of thought (Sagan, 1993, p. 46). Researchers belonging to the “high reliability” school hold that a certain number of preventive measures can be put in place to avoid accidents. According to this approach, the reliability of an organization can be reinforced through wise design and management techniques, redundancy in employees' duties and the information they receive, decentralized decision making and a culture that places high priority on safety, along with trial-and-error learning, supplemented by anticipation and simulation (Sagan, 1993, p. 27, citing La Porte, 1982; Weick, 1987; Rochlin, La Porte and Roberts, 1987; Roberts, 1989 and 1990; La Porte and Consolini, 1991).
The other school of thought espouses the “normal accident” theory, according to which accidents are inevitable in highly interactive, “tightly coupled” systems, since the goals of safety and prevention must compete with other organizational objectives like continuity, productivity and profits (Perrow, 1984). According to Perrow, it is impossible for organizations to “train for unimagined, highly dangerous or politically unpalatable operations.” In addition, an inherent organizational contradiction arises from the fact that “decentralization is needed for complexity, but centralization is needed for tightly coupled systems” (Perrow, 1984, cited by Sagan, 1993, p. 46; Rasmussen, 1990; Reason, 1990, cited by Smith, 2000, p. 66).
Unlike private organizations, public organizations do not have to struggle with the same choice between productivity and safety, since they are more concerned with the social sphere than with economic profits. However, they have certain characteristics that lead to silo management in times of crisis, rather than fluid, effective, integrated horizontal management. Dupuy (1999, p. 5) identifies these characteristics as the high formalization of tasks and procedures, legal obligations that hinder decisional flexibility, and a culture of bureaucratic protection.
Crisis management can be viewed as involving three phases (Pauchant and Mitroff, 1982, p. 163). The first phase consists in preparing for crises, taking preventive measures and being aware of early warning signals to watch for. According to Mitroff (2001), organizations should be ready to deal with seven kinds of crises. These are economic, informational, physical and reputational crises, as well as those related to human resources, psychopathic acts and natural disasters. Organizations should encourage an awareness of prevention and develop a system that alerts them to any “trail of early warning signals.” Professional ability (including stress management and media readiness) and organizational skills (such as flexibility, maintaining communication flow, rapid resource mobilization and response capacity) can be tested and improved. This initial phase in crisis management can be summed up as corresponding to two mechanisms – anticipating and sensing (Mitroff, 2001, p. 40).
The second phase corresponds to the set of actions adopted by an organization in order to recover. During this phase, one of the most important aspects of crisis management – damage containment – is undertaken to ensure that the crisis does not spread to other sectors of the organization or network. This aspect takes on particular strategic importance in the highly interdependent relationships that characterize today's systems (the organizations themselves), subsystems (divisions or units) and supra-systems (broad networks and infrastructures like the energy or telecommunications industry). Decision quality depends on “the quality of information inputs into the decision process… the fidelity of objective articulation and trade-off articulation… and cognitive abilities of the decision group” (Smart and Vertinsky, 1977, p. 640; Cooley 1994, cited by Smith, 2000). The measures taken during this second phase correspond to the actions of reacting to the crisis and containing it (Mitroff, 2001).
Finally, in the third phase, an organization that has come through a crisis should draw lessons from it and recognize the weaknesses that led to it or hindered its effective management. However, it can be difficult to take this approach in the aftermath of a crisis, since organizations may be tempted to cover up their errors rather than to learn from them (Sagan, 1993, p. 46). Such is the power wielded by experts that organizations may again make decisions based on scientific rationales that induce an erroneous belief in the invulnerability of systems, strategies and structures (Pauchant and Mitroff, 1988; Smith, 2000). According to Mitroff (2001), blaming individuals should be avoided except in cases of criminal malfeasance or negligence. The focus should instead be on lessons learned, as well as on redesigning systems and mechanisms to reduce the possibility of future crises and mitigate those that occur. This third phase may be thought of as corresponding to the actions of learning from a crisis and redesigning crisis management procedures. (Mitroff, 2001, p. 41).
Bibliography
Béjin, A. and E. Morin (1976). “Introduction,” Communications, no. 25, pp. 1-3.
Bolzinger, A. (1982). “Le concept clinique de crise,” Bulletin de psychologie, vol. 35, no. 355, pp. 475-480.
Compact Edition of the Oxford English Dictionary (1979). Oxford, Oxford University Press.
Cooley, E. (1994). “Training an Interdisciplinary Team in Communication and Decision-Making,” Small Group Research, vol. 25, no. 1, pp. 5-25.
Dupuy, F. (1999). Why Is It So Difficult to Reform Public Administration?, Symposium Paris, Paris, OECD, September 14 and 15, www.oecd.org/officialdocuments/publicdisplaydocumentpdf/?cote=PUMA/SGF(99)7&docLanguage=En] (last retrieved in June 2010).
Gilgun, J. F. (1988). “Decision-Making in Interdisciplinary Treatment Teams,” Child Abuse and Neglect, vol. 12, no. 2, pp. 231-239.
Hermann, C. F. (1972). “Some Issues in the Study of International Crisis,” in C. F. Hermann (ed.), International Crises: Insights from Behavioral Research, New York, Free Press, pp. 3-17.
La Porte, T. (1982). “On the Design of Nearly Error-Free Organizational Control Systems,” in D. L. Sills, C. P. Wolf and V. B. Shelanski (eds.), Accident at Three Mile Island: The Human Dimensions, Boulder (CO), Westview Press, pp. 185-200.
La Porte, T. and P. M. Consolini (1991). “Working in Practice but Not in Theory: Theoretical Challenges of High Reliability Organizations,” Journal of Public Administration Research and Theory, vol. 1, no. 1, pp. 19-47.
Lagadec, P. (1991). La gestion des crises, www.patricklagadec.net/fr/pdf/integral_livre1.pdf (last retrieved in July 2010).
Lagadec, P. (1981). Le risque technologique majeur: Politique, risque et processus de développement, Paris, Pergamon Press (collection: Futuribles) [English translation: Major Technological Risk: An Assessment of Industrial Disasters, Oxford, Pergamon Press (Translated from the French by H. Ostwald), 1982].
Lagadec, P. (1979). “Le défi du risque technologique majeur,” Futuribles, no. 28, pp. 11-34.
Mitroff, I. (2001). Managing Crises Before They Happen: What Every Executive and Manager Needs to Know About Crisis Management, New York, Amacom Editions.
Pauchant, T. C. and I. Mitroff (1992). Transforming the Crisis-Prone Organization: Preventing Individual, Organizational and Environmental Tragedies, San Francisco, Jossey-Bass.
Pauchant, T. C. and I. Mitroff (1988). “Crisis Prone Versus Crisis Avoiding Organizations: Is Your Company's Culture Its Own Worst Enemy in Creating Crises?,” Industrial Crisis Quarterly, vol. 2, no. 1.
Perrow, C. (1984). Normal Accidents: Living with High-Risk Technologies, New York, Basic Books.
Rasmussen, J. (1990). “Human Error and the Problem of Causality in Analysis of Accidents,” Philosophical Transactions of the Royal Society of London, Series B., vol. 327, no. 1241, pp. 449-462.
Reason, J. (1990). Human Error, Cambridge, Cambridge University Press.
Roberts, K. H. (1989). “New Challenges in Organization Research: High Reliability Organizations,” Industrial Crisis Quarterly, vol. 3, no. 2, pp. 111-125.
Roberts, K. H. (1990). “Some Characteristics of One Type of High Reliability Organization,” Organization Science, vol. 1, no. 2, pp. 160-176.
Rochlin, G. I., T. R. La Porte and K. H. Roberts (1987). “The Self-Designing High-Reliability Organization: Aircraft Carrier Flight Operations at Sea,” Naval War College Review, pp. 76-90.
Sagan, S. D. (1993). The Limits of Safety: Organization, Accidents and Nuclear Weapons, Princeton, Princeton University Press.
Shrivastava, P. (1992). Bhopal: Anatomy of a Crisis, 2nd ed., London, Paul Chapman Publishing.
Smart, C. and I. Vertinsky (1977). “Designs for Crisis Decision Units,” Administrative Science Quarterly, vol. 22, no. 4, pp. 640-657.
Smith, D. (2000). “Crisis Management Teams: Issues in the Management of Operational Crises,” Risk Management, vol. 2, no. 3, pp. 61-78.
Weick, K. E. (1987). “Organizational Culture as a Source of High Reliability,” Californian Management Review, vol. 29, no. 2, pp. 112-127.
_________________________
Reproduction
Reproduction in whole or part of the definitions contained in the Encyclopedic Dictionary of Public Administration is authorized, provided the source is acknowledged.
How to cite
Therrien, M.-C. with the collaboration of A. Valiquette-L'Heureux (2012). “Crisis Management,” in L. Côté and J.-F. Savard (eds.), Encyclopedic Dictionary of Public Administration, [online], www.dictionnaire.enap.ca
Legal deposit
Library and Archives Canada, 2012 | ISBN 978-2-923008-70-7 (Online)