Selecting Improvement Projects
David Birnbaum
Epidemiology, as a process for logical inquiry, has much in common with systems analysis or industrial engineering (also known as management engineering) (1). Similar perspectives and complementary methods shared by these disciplines make them ideal for managing healthcare quality improvement (2). However, to succeed, these disciplines must be applied in a supportive setting and on worthwhile quality improvement projects. There are underlying principles and precedents of both successes and failures; these can serve as important guides to anyone contemplating extension of epidemiologic skills from familiar areas of infection control to less familiar areas of quality improvement. Healthcare as a business sector has lagged far behind the cutting edge of other industries in advancing its methods to assure and improve service quality. Healthcare organizations have generally failed to use the full potential of epidemiology in discerning alternative strategies and informing consensus on best practices; exploring the natural course of conditions; performing cost-benefit and effectiveness analyses; surveying patient preferences; measuring organizational effectiveness; establishing indicators, criteria, and other measures; and designing and evaluating surveillance systems (3,4). Although it is noteworthy and unfortunate that epidemiology is not listed among team leadership in seminal reference books and motivating reports (5,6), this reflects the simple fact that relatively few healthcare epidemiologists rose to embrace challenging new opportunities.
Although some of the language in general underlying principles for selecting improvement projects might introduce foreign concepts, the principles are not complicated. Mozena and Anderson (7) list the following essential criteria to consider:
Impact on patient care or external customer
Impact on favorable patient outcomes
Magnitude of potential cost savings
Cost of implementation
Difficulty of implementation
Ability to measure performance of process
Potential benefits outweigh cost of the project
Deals with key business issue
High error rate
Availability of data
Impact on profitability
Potential for success
Impact on ongoing quality
Ability to quantify results
High visibility to customers or patients
Elimination of rework
High risk to patients or employees
A National Demonstration Project on Quality Improvement in Health Care reported that nomination and selection of projects often are run by steering committees (a quality council) but that the best ideas come from “listening to the voice of the customer” (in which external customers are patients who receive services and internal customers are staff members who collectively provide and support service delivery) (8). Surveying customer opinion is an active way to listen; design and conduct of surveys are familiar grounds in epidemiology. Epidemiologists also may be more aware than most about the distinction between measuring patient satisfaction, patient safety, and service quality of healthcare (9). Relating service attributes to customer expectations may involve less familiar but still simple techniques such as Quality Function Deployment matrices (a simple two-dimensional matrix in which the strength of association between specific items and categories of customer expectations is summarized) (10,11). However, all of these criteria and methods are disjointed considerations. What is needed to bring efficiency and acceptance is an effective system for their implementation.
GUIDANCE FROM HISTORICAL PRECEDENTS
Three precedents bear consideration as effective systems to select improvement projects. Although two were successful quality improvement systems, they failed to persist and become today’s North American gold standard models. Williamson’s Achievable Benefits Not Achieved (ABNA) system to identify and prioritize potential projects (12) has a remarkable track record among alternatives (13). Similarly, the so-called Denver Connection of the same era is a story of successful amalgamation and reorganization of two hospitals in a way that put quality improvement supported by real-time performance data
analysis as the centerpiece of medical staff departmental meetings and continuing medical education (encouraged by board-level involvement while the usual array of advisory committees was eliminated) (14). Finally, the Institute of Medicine’s (IOM’s) Model Process for semiquantitatively ranking alternative projects is instructive for its mathematical approach (15).
analysis as the centerpiece of medical staff departmental meetings and continuing medical education (encouraged by board-level involvement while the usual array of advisory committees was eliminated) (14). Finally, the Institute of Medicine’s (IOM’s) Model Process for semiquantitatively ranking alternative projects is instructive for its mathematical approach (15).
Dr. John W. Williamson developed systems for health accounting and ABNA during an impressive body of work that spans decades on the faculty of the Department of Health Services Administration at the Johns Hopkins University School of Hygiene and Public Health, Medicine and Medical Informatics at the University of Utah School of Medicine, Regional Medical Education Center of the Salt Lake Veterans Affairs Medical Center, and service on government commissions. Health accounting, conceptualized in the early 1960s, is “a management model to integrate continuing education and patient care research into an ongoing cyclic function to systematically improve the quality of medical care” (16). It is an evidence-based outcome-focused approach that selects project priorities through ABNA, a formal, efficient process refined in the 1970s. Although proven effective and cost-effective in a wide range of applied research and demonstration projects at the American National Institutes of Health, Veterans Administration, and elsewhere, Williamson acknowledges that the most successful application of his system to enhance national quality is in the Netherlands (16). The fact that Williamson’s name and work are unfamiliar to so many throughout North American hospitals and healthcare leadership is hauntingly reminiscent of the history of W. Edwards Deming. Deming’s influence took decades to return to North America, heeded by American manufacturers only after Japan capitalized on Deming’s leadership to outperform their American counterparts (and later heeded by health service organizations decades after that!) (17). Williamson stresses the following:
The importance of applying principles of epidemiology, sampling, and simple statistical testing to QA-focused reviews
The necessity of a multidisciplinary team approach to QA, in which the consumer was the most important member of the team
The need to use structured group judgment methods for establishing priorities, criteria, and standards as well as QA action decisions under the usual conditions of factual uncertainty
The need for a unique set of statistical methods for QA that allowed comparison of measured results against consensus standards that reflected reasonably achievable projected outcomes
Consistent with concurrent surveillance methods that have become a mainstay of contemporary infection surveillance programs, Williamson recognized long ago that chart-based audits as a basis for quality assurance may be misleading and severely limit the potential impact of programs to improve quality (16). His ABNA process consists of selecting a team (ideally 7-11 persons, including “at least four knowledgeable and respected staff physicians” and representatives of other functional areas and the lay public) and then supporting that team through two 2-hour meetings 3 to 4 weeks apart. The first meeting is a training session simulating the later priority-setting session. A master list of potential topics developed during the training session, together with team recommendations on additional “data, literature, or consultation” required, provides an indication of support materials that will be needed at the second session. Ideally, these are gathered during the 3- to 4-week hiatus. There are seven tasks in the prioritysetting meeting:
Introductory remarks by the moderator clarify meeting purpose, tasks, and timing of the 2-hour session and review the ABNA framework (5 to 10 minutes).
A simple four-column form is distributed so that individual team members can each list as many topics as they wish, listing along one row for each topic:
exactly who (what group) will benefit
b. for what health problem
from what action(s)
by which provider(s)
A cue sheet is provided to give examples of various patient characteristics, health problem characteristics, provider characteristics, and (inter)action types that might be considered. Time is given to work individually (10 minutes).
Each team member, sequentially, is then asked by the coordinator to nominate one problem from their list. The coordinator develops a summary chart; in addition to the four columns identified in step two, when acting as coordinator, I found it helpful to list in two additional columns an indication of whether the intervention is known to work (nature of evidence for efficacy or effectiveness) and whether it is feasible (information on cost, cost-effectiveness, case study, etc.). Discussion is limited to clarification at this stage, and the process repeats until all the most promising ideas from each member’s list are presented (30 minutes).
Individuals then vote in an “initial weights” column on their form to assign a priority rank (high to low on a fivepoint scale) for each project nominated (5 minutes).
The coordinator then collects the votes, anonymously recording both initial individual weights and their sum for each nominated project. This information, superimposed on the summary chart, is projected back to the group (10 minutes).
Discussion of results, one topic at a time, then examines whether priority ranking is tightly or widely dispersed, the strength of evidence, and other detailed considerations. On completion of comprehensive discussion, members are then asked to vote again on every topic, in a “revised weights” column on their form. The coordinator again collects and records votes anonymously (50 minutes).
The highest-ranked ABNA topics are then forwarded as recommendations from the team, along with any further recommendations for additional data or evidence required for any of the topics. The meeting is adjourned, the team thanked for completing its work, and special teams of qualified individuals then take responsibility for moving approved projects forward.
Meanwhile, also in the 1970s, radical changes under the amalgamation of the medical staffs at Denver’s Swedish
Medical Center and Porter Memorial Hospital occurred. Radical change was needed because as Dr. William Robinson, Director of Medical Education, noted, “In spite of the hundreds of physician hours devoted to medical staff activities, little actually was accomplished. It was almost impossible to demonstrate that quality of care was in any way influenced by the physicians’ repetitive, duplicative, unrewarding medical staff activities.” The usual litany of committees was reduced to just three (executive, professional activities, and credentials). A reduced number of subcommittees composed of small numbers of individuals, the bulk of whom were not physicians, served these committees, and much of the quality-related work was shifted from committees to medical staff departments supported by the work of subcommittees or research and education department employees. Medical staff members were strongly encouraged to ask questions at their departmental meetings about quality issues, and then make policy decisions based on evidence delivered soon thereafter (answers supplied through real-time research capacity in their own institution) (14).
Medical Center and Porter Memorial Hospital occurred. Radical change was needed because as Dr. William Robinson, Director of Medical Education, noted, “In spite of the hundreds of physician hours devoted to medical staff activities, little actually was accomplished. It was almost impossible to demonstrate that quality of care was in any way influenced by the physicians’ repetitive, duplicative, unrewarding medical staff activities.” The usual litany of committees was reduced to just three (executive, professional activities, and credentials). A reduced number of subcommittees composed of small numbers of individuals, the bulk of whom were not physicians, served these committees, and much of the quality-related work was shifted from committees to medical staff departments supported by the work of subcommittees or research and education department employees. Medical staff members were strongly encouraged to ask questions at their departmental meetings about quality issues, and then make policy decisions based on evidence delivered soon thereafter (answers supplied through real-time research capacity in their own institution) (14).
Although successful into the 1980s, by the turn of the 21st century, the Denver Connection was so far dismantled that it no longer even existed in the institutional memory of Porter Hospital’s present administration! Porter and Swedish, partners of the so-called Denver Connection formed in 1972, went their separate ways in 1992. These two hospitals serving health needs of southeast Denver had remained separate corporate entities, yet collaborated successfully for many years following formal merger of their medical staff organizations. That merger was initiated by the doctors, not by the administrators, for the purpose of improving quality of care. Administrative support grew following demonstrated successes, and the hospitals cooperated in division of complementary health services, instead of duplication of what the other offered solely to compete. That was before big business entered sickness. Ultimately, poor quality of management in emerging, large, managed-care corporations led to unexpected deficits in profitability of operations, and corporate vision shifted to preoccupation with that debt. A grass-roots anticompetitive way of serving the community’s health needs could not sustain itself in the face of powerful market forces and growing business empires awash in corporate debt. Personal ideologies in administrative leadership compounded the difficulty of making effective alliances, and mistrust grew where open communication once thrived. Key participants later interviewed conveyed a sense of loss and regret, a realization they participated in something very unique and beneficial that was lost for illogical reasons (“nobody had any appreciation what we were doing was special or unique … only in retrospect we came to appreciate the specialness of what we were doing.”). The influence of accreditation programs in this saga was noteworthy only for its lack of influence (18). It is tempting to speculate that this visionary effort thrived because it followed characteristic principles that seem to distinguish great companies from others (19) and that the Denver Connection ultimately fell when it strayed from core values and these fundamental principles.
The IOM Model Process to set priorities in health technology assessment addresses similar dimensions as ABNA, but in a more quasinumeric than nominal group consensus manner. The IOM priority score for each technology, instead of being assigned by consensus ranking, is calculated as ΣW1lnS1 + W2lnS2 + … + W7lnS7, where Wi represents the criterion weight and Si the criterion score for (i = 1-7).
Prevalence (e.g., cases per 1,000 persons)
Burden of illness (e.g., difference in quality-adjusted life years of individuals with vs. without the condition under consideration)
Cost (total direct and indirect costs per person with the condition)
Variation in rates of use of the technology (coefficient of variation)
Potential to change health outcome (subjective assessment on a five-point scale)
Potential to change costs (subjective assessment on a five-point scale)
Potential of assessment result to inform ethical, legal, or social issues (subjective assessment on a five-point scale)
The first three criteria are objective measures; the remaining four are subjective and are addressed by one or more expert panels. Criterion weighting values are arbitrary choices; the process described has an expert panel select one criterion as least important, which then is assigned weight of one. Mean weights given by panel members for each of the remaining criteria, relative to the least important one, then determine the other six weights. After discussion of results to resolve any wide disagreement, results of a second vote are final. Pilot test results with small conventional and mailed response panels are examined in the IOM report, which gives the following values: W1 = 1.6, W2 = 2.25, W3 = 1.5, W4 = 1.2, W5 = 2.0, W6 = 1.5, and W7 = 1.0. Logarithms of criterion scores are used to make the model multiplicative rather than additive, thus responsive to relative rather than absolute differences in scores (algebraically, the formula can be restated as ΣWilnSi, which equals the product ΠSii for i = 1-7). Subjective item scales, therefore, run from a value of one for least likely to five for most likely.
Health accounting provided a philosophy, the Denver Connection a forum, and ABNA a method all consistent with today’s emphasis on evidence-based practice and continuous quality improvement (CQI). What lessons should one take from these all-but-forgotten precedents? Williamson reflects on lessons learned from 25 years of experience (16), naming three premises on which quality assurance or improvement is based and five principles that evolve from it: