sábado, 18 de agosto de 2018

World Health Organization Methodology to Prioritize Emerging Infectious Diseases in Need of Research and Development - Volume 24, Number 9—September 2018 - Emerging Infectious Diseases journal - CDC

World Health Organization Methodology to Prioritize Emerging Infectious Diseases in Need of Research and Development - Volume 24, Number 9—September 2018 - Emerging Infectious Diseases journal - CDC





Volume 24, Number 9—September 2018

Online Report

World Health Organization Methodology to Prioritize Emerging Infectious Diseases in Need of Research and Development

Massinissa Si MehandComments to Author , Piers Millett, Farah Al-Shorbaji, Cathy Roth, Marie Paule Kieny, and Bernadette Murgue
Author affiliations: World Health Organization, Geneva, Switzerland (M. Si Mehand, P. Millett, F. Al-Shorbaji, M.P. Kieny, B. Murgue)Department for International Development, London, UK (C. Roth)

Abstract

The World Health Organization R&D Blueprint aims to accelerate the availability of medical technologies during epidemics by focusing on a list of prioritized emerging diseases for which medical countermeasures are insufficient or nonexistent. The prioritization process has 3 components: a Delphi process to narrow down a list of potential priority diseases, a multicriteria decision analysis to rank the short list of diseases, and a final Delphi round to arrive at a final list of 10 diseases. A group of international experts applied this process in January 2017, resulting in a list of 10 priority diseases. The robustness of the list was tested by performing a sensitivity analysis. The new process corrected major shortcomings in the pre–R&D Blueprint approach to disease prioritization and increased confidence in the results.
Recent outbreaks of Ebola virus disease, Middle East respiratory syndrome, and Zika virus disease illustrate that emerging infectious diseases will continue to cause major public health emergencies. Further work is needed to strengthen defenses with medical countermeasures (MCMs) and other protective interventions. Building on recent experiences and at the request of the World Health Assembly in May 2015 (1), the World Health Organization (WHO) launched the R&D Blueprint for action to prevent epidemics. This global strategy and preparedness plan is designed to ensure that targeted research and development (R&D) will strengthen emergency response by accelerating availability of biomedical technologies to populations and patients during epidemics (2). The R&D Blueprint focuses on severe emerging diseases that pose a major risk for causing a public health emergency and for which MCMs or substantial R&D initiatives and pipelines are insufficient or nonexistent (3).
Experts compiled an initial list of relevant diseases at an informal consultation in December 2015 (4). A more robust methodology was needed, one that could be standardized and repeated regularly for reviewing and, if necessary, updating the list in the light of successful development of new interventions or the emergence of new disease threats.
WHO settled on a 3-pronged approach: 1) a methodology development and review process; 2) an annual review of a list of prioritized diseases; and 3) a decision instrument to guide decision-making on a novel disease (Technical Appendix 1). All 3 processes use a common set of weighted criteria and subcriteria, such as the human-to-human transmissibility of the disease or its potential societal impact (Technical Appendix 2). This process is inherently expert-driven because the R&D Blueprint addresses pathogens that are yet to be fully characterized and for which an understanding of how to diagnose, prevent, and treat the resulting diseases is incomplete. Further, these pathogens might behave differently on different occasions because of variation in the biologic, cultural, or environmental context. Decisions have to be made on the basis of partial information supplemented by expert opinion. Any methodology will be prone to biases (3).
This article assesses the application of this methodology for the 2017 annual review of the WHO R&D Blueprint priority list of diseases. We consider its effectiveness and assess the degree of confidence that can be placed in the list produced.

Developing a Prioritization Process

WHO developed a comprehensive methodology (3) to ensure the list of the R&D Blueprint prioritized diseases best reflects targeted global health needs and focuses on the most pressing threats. The approach taken drew heavily on established best practice (57) and is based on practical national and regional experiences in compiling similar lists (814). This approach also specifically addressed criticism of pre–R&D Blueprint attempts by WHO to prioritize diseases by developing tools for assessing confidence in the results generated and addressing potential biases (5).
Disease prioritization is not a straightforward task and requires a defined set of criteria on which to base prioritization (7). These criteria can be qualitative, intangible, or subjective, changing for different stakeholders (15). The criteria can also be interdependent, complicating separate assessment (16). For instance, the case-fatality rate of a disease has a social effect, which in turn has an economic effect. Given the complexity and the challenges of disease prioritization, ensuring the process is transparent and reproducible is important (5,7,17).
Recent disease prioritization methods were summarized in a 2015 review by the European Centre for Disease Prevention and Control (ECDC) (5), which extrapolated a series of best practices. Several subsequent studies were also identified (Technical Appendix 3). Past disease prioritization studies have been conducted for different purposes, such as communicable diseases surveillance (18,19), biosecurity (20), and resource allocation (21,22), and have covered disease in humans, livestock, or wildlife. Many studies were conducted primarily at national (810,19,2136) and regional (1114,16,20,3743) levels (e.g., Europe and North America) but rarely at a global level (44,45). None of the disease prioritization exercises matched the aims of the R&D Blueprint, its public health focus, and its global reach; thus, WHO needed to develop its own methodology.
Several different disease prioritization methods exist (5), including Delphi processes (38,40), multicriteria decision analysis (MCDA) (14,26,28,36,46), H-index (42,43), questionnaires (11,13,22), and qualitative algorithms (47,48). Each method has its own strengths, weaknesses, and context-dependent utility, but 3 methods most closely matched the requirements of the R&D Blueprint (5): 1) a semiquantitative Delphi process to narrow the list of diseases under consideration; 2) MCDA to rank the remaining diseases (Technical Appendix 4); and 3) questionnaires in the form of online survey tools to standardize information gathering from participating experts.
 Top

Methods and Tools

The resulting methodology was developed over a year-long process, involving informal consultations, internal and external expertise, and guidance from the R&D Blueprint Scientific Advisory Group (4). Methods and tools were subsequently reviewed and validated by an external group of experts (49) and used in the review of the list of priority diseases in January 2017 (50).
Prioritization Committee
Selecting the right group of experts is critical for ensuring an outcome as accurate as possible (39,51). Gathering a diverse field of expertise with a broad geographic distribution, including an in-depth knowledge of the diseases and pathogens being considered, is important. The multidisciplinary committee convened for the 2017 annual review included 24 experts drawn from Africa, Asia, Europe, North America, and South America (Technical Appendix 2). The persons present at the meeting covered all 7 areas of expertise detailed in the methodology (Technical Appendix 1) (3). To ensure the process was as transparent as possible, representatives from several additional organizations were present, including the World Organisation for Animal Health (OIE), which helped ensure a One Health approach was followed, as well as the Coalition for Epidemic and Preparedness Innovations and the Global Research Collaboration for Infectious Disease Preparedness, which facilitated cooperation, coordination, and the sharing of experiences outside of WHO. To minimize bias related to expert opinions, the prioritization committee is changed yearly (7).
Triage of the Diseases
To narrow the list of potential priority diseases, a 2-step semiquantitative Delphi technique was adapted from established environmental horizon scanning methods (52). Each proposed disease was scored from 0 to 1,000, where 1,000 represented a perfect fit for the R&D Blueprint and 0 represented diseases with no epidemic potential, diseases for which effective and commercially available MCMs exist, or both.
Disease Scoring
Thumbnail of Example screenshot of tool developed in R Shiny (https://shiny.rstudio.com) using the slide bar function to compare candidate diseases for each criteria and subcriteria considered in the development of the World Health Organization R&D Blueprint to prioritize emerging infectious diseases in need of research and development. Experts were requested to compare candidate diseases to each other for each criteria, placing them in ranked order according to their knowledge. The World He
Figure 1. Example screenshot of tool developed in R Shiny (https://shiny.rstudio.com) using the slide bar function to compare candidate diseases for each criteria and subcriteria considered in the development of the World...
To rank the short list of diseases, an online survey tool was designed by using the slide-bar function of R Shiny (https://shiny.rstudio.com) (Figure 1). Because absolute scoring scales require broadly accepted standards (53) and these standards are not evident in the context of emerging infectious disease, for which many characteristic remain unclear or unknown, the WHO tool makes use of a relative scale that compares values between diseases rather than against absolute values (i.e., the impact of scoring diseases A and B at 3 and 5 is the same as scoring the same diseases at 7 and 9).
The data collected were processed by an in-house program implemented in R Studio. A custom analytic hierarchy process (AHP) implementation was used to calculate disease scores for each subcriterion (Technical Appendix 5). This process included normalization and weighting procedures. Comparison matrices were built from data provided by each expert and then averaged by using the geometric average (54). Disease scores for each subcriterion were computed, and an overall multicriteria score for each disease was ultimately computed by using the disease scores and criteria weights. Following best practices, the criteria definition and weighting steps were separated from the disease scoring (5,7,14,36,38). The criteria were defined in 2015 by a group of experts (4) and were then reviewed, validated, and weighted by another group (49).
Sensitivity Analysis and Confidence Estimation
Past prioritization processes have used sensitivity analysis, commonly with lower and upper 95% CIs (31). Other processes included modifying the weight of criteria used (9,55) and removing them one at a time (55). We describe a series of sensitivity analyses, including setting all the criteria at the same weight, removing 1 criterion at a time, increasing the weight of each criterion by 20%, and doubling the weight of a criterion. This approach to sensitivity analysis enables assessment of the impact of different scenarios on the final disease ranking and provides important insights into the robustness of the ranking and the impact of potential biases (9,55).
As a confidence indicator, differences among expert opinions were considered. The arithmetic average scores and the corresponding SDs for each disease were calculated and tracked through the process by using an error propagation technique (Technical Appendix 5).
 Top

Results

Compiling and Reviewing a Long List of Diseases
The long list of diseases was drawn from diseases identified as requiring urgent R&D support in the 2015 priority list, diseases recommended but not included by the 2015 consultation, and diseases suggested by participants in the 2017 review. As a result, 8 diseases on the original 2015 list were supplemented by another 10 diseases. In 2017, no additional disease was selected by the decision instrument.
Thumbnail of Process for compiling the short list of diseases for inclusion in the World Health Organization R&D Blueprint to prioritize emerging infectious diseases in need of research and development.
Figure 2. Process for compiling the short list of diseases for inclusion in the World Health Organization R&D Blueprint to prioritize emerging infectious diseases in need of research and development.
Each of these 18 diseases was then considered in turn. Two experts introduced each of the diseases on the 2015 list and those selected at the 2015 consultation. A single expert introduced each of the diseases proposed by the 2017 committee. Consensus was rapidly reached that diseases on the 2015 list should be reassessed by using the MCDA tool. A triage of the remaining 10 diseases was then carried out. The results were discussed in detail, and a further 5 diseases were added to the short list (Figure 2). Additional considerations of those diseases not incorporated into the list were also discussed (Technical Appendix 6), such as the importance of continuing relevant R&D (50).
Ranking the Short List of Diseases
Thumbnail of Multicriteria scores of diseases considered in the 2017 prioritization exercise for the development of the World Health Organization R&D Blueprint to prioritize emerging infectious diseases in need of research and development. A) Disease final ranking using the geometric average of the comparison matrices. B) Disease final ranking using the arithmetic average of the raw data. Error bars correspond to SD, indicating disagreement among experts. C) Disease final ranking using the S
Figure 3. Multicriteria scores of diseases considered in the 2017 prioritization exercise for the development of the World Health Organization R&D Blueprint to prioritize emerging infectious diseases in need of research and development....
The MCDA tool was used to generate 1) scores for each disease against each subcriterion; 2) aggregated scores for each criterion for each disease; and 3) multicriteria scores for each disease. The aggregated scores for each criterion for each disease were considered in more depth (50). The multicriteria scores (Figure 3, panel A) were then used to rank diseases on the short list. Six diseases (P1, P2, P3, P4, P8, and P9) were highly ranked; a group of 3 diseases (P5, P6, and P7) were ranked next, and the final 4 diseases (P10, P11, P12, and P13) had the lowest ranking.
A consensus was quickly reached that the group of 6 top-ranking diseases (P1, P2, P3, P4, P8, and P9) should be on the 2017 priority list. An uncertainty analysis revealed overlapping results for the remaining pathogens, particularly noticeable for the lower 2 tiers (Figure 3, panel B). As a result, the multicriteria scores alone were insufficient to differentiate between the remaining 8 diseases. An additional round of the Delphi technique enabled the committee to compile a final list: arenaviral hemorrhagic fevers (including Lassa fever); Crimean-Congo hemorrhagic fever; filoviral diseases (including Ebola and Marburg virus infections); Middle East respiratory syndrome coronavirus infection; Nipah virus infection and related henipaviral diseases; other highly diseases coronaviral diseases (such as severe acute respiratory syndrome); Rift Valley fever; severe fever with thrombocytopenia syndrome; and Zika virus infection. A more detailed discussion as to how the list was compiled can be found in the WHO report on the 2017 prioritization exercise (50).
Assessing Confidence in the Results
The multiscenario sensitivity analysis detailed the influence of each criterion on the final ranking. When all the criteria were set at the same weight as those used in a similar exercise conducted in Kenya (9), the multicriteria scores were affected, but the overall ranking remained largely the same, with 2 diseases (P5 and P7) switching positions.
When highly weighted criteria, such as human-to-human transmissibility, were suppressed, major changes in the multicriteria scores were observed, but a much smaller impact was evident on the overall disease ranking. A notable exception was when the MCMs criterion was suppressed, after which no notable impact on the multicriteria scores or the final ranking was observed.
When the weight of each criterion was increased by 20%, no notable changes in ranking were observed. Doubling the weights of highly weighted criteria resulted in changes in the overall multi criteria scores but had a minimal impact on the overall ranking of diseases. Once again, doubling the weight of the MCMs criterion had minimal impact on the multicriteria scores and the overall disease ranking.
To further validate the 2017 priority list, the same data were analyzed by using the SMART Vaccines prioritization tool (56). Unlike the methodology discussed in this article, the SMART Vaccines tool makes use of absolute rather than relative values. This feature precludes a direct comparison of specific results but helps explore the reproducibility of the list as a whole. The results from the SMART Vaccines prioritization tool also grouped the same diseases together in the same 3 tiers (Figure 3, panel C).
 Top

Discussion

The 2017 annual review resulted in a list of diseases that pose a risk for a public health emergency and for which an urgent need for R&D exists (50). The earlier ECDC review highlighted numerous general weaknesses among published prioritization processes (5). It identified shortcomings in WHO approaches toward disease prioritization before the R&D Blueprint, including insufficient detail in reporting; a lack of transparency, in particular as to how the prioritization criteria were developed; a need for greater consideration on sources of bias; a better discussion of implementation challenges; methodologic anomalies, such as the use of only a single round of the Delphi technique; and a lack of external review of the methodology and subsequent publications.
The methodology developed by WHO for the R&D Blueprint explicitly addresses these shortcomings (e.g., mitigation of numerous sources of bias). Past methodologic anomalies have also been addressed (e.g., a 2-step semiquantitative Delphi technique is now being used). The reporting process has been strengthened; the methodology has been published in full (3), as has a detailed report of its use during the 2017 annual review (50).
The new approach is also much more transparent, with all publications being more detailed and openly available. These publications explain the reasoning behind why certain diseases ultimately were (or were not) included on the list. The process by which the prioritization criteria were developed (Technical Appendix 2) is also well documented in meeting reports (4,49).
Shortcomings in the review process have been addressed, in part, by separate committees to develop and implement the methodology because these committees effectively review each other’s work. The methodology itself was validated through a dedicated expert consultation, improving the review procedures further. Finally, publication of this article further expands opportunities to review the approach and its implementation.
Several challenges to implementation exist. Other prioritization studies have invested extensive resources into identifying potentially relevant diseases. For example, Cox et al. conducted a bibliometric analysis of >3,000 infectious organisms in North America to identify the 651 pathogens relevant to their study (43). In Belgium, Cardoen et al. complimented a literature review with expert consultations (36). In the Netherlands, Havelaar et al. went a step further, supplementing their literature review with consultations with international, regional, and national experts to identify the relevant subset of pathogens (26).
At present, the community proposing additional diseases to be considered in an annual review of the R&D Blueprint is limited. To address this issue in future prioritization exercises and to better reflect regional factors in the long list of diseases, the WHO regional offices will be more actively involved in the inclusion of a wider range of experts.
The next methodology review should look at the reproducibility of these tools. In the interim, further improvement of the MCDA model might include reviewing the pertinence of the MCMs criterion given that it had little effect on the multicriteria scores of the diseases, reweighting the criteria, drawing on a wider community of relevant expertise and a larger sample size, and reviewing and simplifying the specific wording of the subcriteria.
The R&D Blueprint methodology was developed to mitigate numerous sources of bias, including flaws in study design, selection bias, interviewer bias, chronology bias, and recall bias (3). These efforts were largely successful; however, further work might be necessary to mitigate selection and recall bias.
The selection of experts to participate in the MCDA is important for mitigating selection biases. WHO’s policies on geographic and gender representation go some way to address selection bias. Considerable resources were also expended to create a committee with the diverse range of expertise required, with experts from microbiology and virology, clinical management of severe infections, epidemiology and outbreak investigation and response, public health policy, animal health, mathematical modeling of disease, environmental and social science, nongovernmental organizations, and the security sector. This diversity is consistent with and exceeds the range of participants found in other studies, allowing for some variation based on their specific purposes (8,9,11,13,26,30). Ensuring that future reviews also have a sufficient range of expertise will be important.
The number of experts participating in the annual review meeting also deserves careful consideration. Larger groups increase the likelihood of reproducibility and decrease the risk for certain biases (57). Smaller groups can simplify the consensus-building process (15,58). Group size can also impact group dynamics. Too large a group can make face-to-face consultations impractical, complicating efforts to review and discuss the results, correct eventual inconsistencies, reach consensus, and avoid misunderstanding (8). Although exploring ways that a greater number of experts might be involved with developing an initial long list of diseases to be considered might be useful, a more limited group will probably need to continue to analyze the short list in the years to come.
Additional efforts are also needed to address recall bias. Discussions during the 2017 review highlighted that the diseases that enjoyed the greatest levels of support for inclusion in the revised priority list had all caused recent major outbreaks. An annual “landscape review” (in which each disease on the long list of proposed diseases is independently reviewed, considering factors such as the current knowledge regarding prioritization criteria, risk for emergence, and availability of countermeasures, regardless of recent events) should contribute to avoiding a disproportionate emphasis based on recent events. In the short term, participants in the next annual review should be briefed on, and discuss the impact of, recall bias before undertaking the MCDA scoring exercise. In the longer term, options for weighting against recent public health emergencies might be explored, perhaps through the development of a calibration curve, which has been used to mitigate recall bias in other types of processes (59).
This methodology is expert-driven, and despite all efforts to minimize biases related to their efforts, biases still occur. To address this problem, WHO should 1) change the composition of the prioritization committees yearly and expand the geographic range of the experts involved and 2) review the methodology separately with different experts.
The similarity between the WHO list of prioritized diseases and those found in other studies suggests a degree of consistency with previous findings (811,13,14,35,41). The results of the sensitivity analysis demonstrate that the R&D Blueprint ranking is robust, corresponding with earlier observations that the analytic hierarchy process is not sensitive to minor changes in criteria weights (55). Even when major changes on the weight of criteria were applied, the final ranking remained largely stable. Throughout all the scenarios used in this sensitivity analysis, the same 3 groupings of diseases remained consistent. In some scenarios, the ranking of diseases within the group changed, but this observation is consistent with the findings of other prioritization exercises (9). Being able to produce a similar 3-tiered group ranking using another model, the SMART Vaccines prioritization tool, also suggests that the approach employed for the R&D Blueprint is producing valid results.
However, the impact of the MCMs criterion needs further consideration. The sensitivity analysis showed that the contribution of this criterion to the final ranking is limited despite its high weight. This observation is probably explained by the objectives of the R&D Blueprint itself, which focuses on diseases for which few or no MCMs exist, meaning that all the diseases considered score equally in this regard. Ensuring that sufficient attention is paid to this issue when selecting diseases for inclusion on the long-list will be useful as a screening process. Ensuring that distinct R&D gaps are a prerequisite for inclusion could result in this criterion being dropped from the MCDA.
In conclusion, the R&D Blueprint fills a considerable gap in public health preparedness by supporting R&D on highly infectious diseases for which few or no countermeasures exist. To translate this objective into effective action, WHO had to determine the diseases that most urgently required the commencement of work. For each of these priority diseases, WHO is developing roadmaps; target product profiles for vaccines, therapeutics, and diagnostics (60); and generic protocols for vaccine and therapeutic clinical trials. The R&D Blueprint is also enabling cross-cutting support activities, such as data and sample sharing norms, regulatory preparedness aspects, and overall research coordination (61). Aware of the shortcomings of past efforts to develop similar lists, WHO explored lessons learned and best practices for developing a new approach. The challenge was in balancing competing needs for a standardized, robust methodology that can be repeated on a regular basis, with a reliance on expert opinion. Because this methodology and its supporting tools will be subjected to a full review within 2 years, WHO hopes that the lessons learned through the R&D Blueprint’s repeated use, including those we have identified, will be used to improve it further.
 Top
Dr. Si Mehand is a technical officer at the World Health Organization. His research interests are applied mathematics, outbreak preparedness and response, priority setting, decision aid, and risk analysis.
 Top

Acknowledgments

We thank the Prioritization and Methodology Review Committees for their inputs and Farah Al-Shorbaji for editing the revised manuscript.
This work has been partially supported by the United Kingdom Department for International Development.
 Top

References

  1. Word Health Organization. Sixty-ninth World Health Assembly resolution WHA69.23: follow-up of the report of the Consultative Expert Working Group on Research and Development: financing and coordination [cited 2016 May 28]. http://apps.who.int/gb/ebwha/pdf_files/WHA69/A69_R23-en.pdf
  2. Word Health Organization. An R and D blueprint for action to prevent epidemics—plan of action [cited 2016 May 30]. http://www.who.int/blueprint/about/r_d_blueprint_plan_of_action.pdf?ua=1
  3. Word Health Organization. Methodology for prioritizing severe emerging diseases for research and development [cited 2017 Feb 19]. http://www.who.int/blueprint/priority-diseases/RDBlueprint-PrioritizationTool.pdf?ua=1
  4. World Health Organization. The 2015 WHO R and D Blueprint priority diseases [cited 2015 Dec 10]. http://www.who.int/medicines/ebola-treatment/WHO-list-of-top-emerging-diseases/en/
  5. European Centre for Disease Prevention and Control. Best practices in ranking emerging infectious disease threats: a literature review [cited 2015 Feb 1]. https://ecdc.europa.eu/sites/portal/files/media/en/publications/Publications/emerging-infectious-disease-threats-best-practices-ranking.pdf
  6. Rist  CLArriola  CSRubin  CPrioritizing zoonoses: a proposed one health tool for collaborative decision-making. PLoS One2014;9:e109986DOIPubMed
  7. Brookes  VJDel Rio Vilas  VJWard  MPDisease prioritization: what is the state of the art? Epidemiol Infect2015;143:291122DOIPubMed
  8. Kadohira  MHill  GYoshizaki  ROta  SYoshikawa  YStakeholder prioritization of zoonoses in Japan with analytic hierarchy process method. Epidemiol Infect2015;143:147785DOIPubMed
  9. Munyua  PBitek  AOsoro  EPieracci  EGMuema  JMwatondo  Aet al. Prioritization of zoonotic diseases in Kenya, 2015. PLoS One2016;11:e0161576DOIPubMed
  10. Garner  MJCarson  CLingohr  EJFazil  AEdge  VLTrumble Waddell  JAn assessment of antimicrobial resistant disease threats in Canada. PLoS One2015;10:e0125155DOIPubMed
  11. Ng  VSargeant  JMA quantitative and novel approach to the prioritization of zoonotic diseases in North America: a public perspective. PLoS One2012;7:e48519DOIPubMed
  12. Ng  VSargeant  JMA stakeholder-informed approach to the identification of criteria for the prioritization of zoonoses in Canada. PLoS One2012;7:e29752DOIPubMed
  13. Ng  VSargeant  JMA quantitative approach to the prioritization of zoonotic diseases in North America: a health professionals’ perspective. PLoS One2013;8:e72172DOIPubMed
  14. Humblet  M-FVandeputte  SAlbert  AGosset  CKirschvink  NHaubruge  Eet al. Multidisciplinary and evidence-based method for prioritizing diseases of food-producing animals and zoonoses. Emerg Infect Dis2012;18:e1DOIPubMed
  15. Wilson  S-JWard  MPGarner  MGA framework for assessing the intangible impacts of emergency animal disease. Prev Vet Med2013;111:1949DOIPubMed
  16. Ng  VSargeant  JPrioritizing zoonotic diseases: differences in perspectives between human and animal health professionals in North America. Zoonoses Public Health2015.PubMed
  17. Giesecke  JChoosing diseases for surveillance. Lancet1999;353:344DOIPubMed
  18. Carter  ANational Advisory Committee on Epidemiology Subcommittee*Establishing goals, techniques and priorities for national communicable disease surveillance. Can J Infect Dis1991;2:3740DOIPubMed
  19. Doherty  J-AEstablishing priorities for national communicable disease surveillance. Can J Infect Dis2000;11:214DOIPubMed
  20. Brioudes  AWarner  JHedlefs  RGummow  BDiseases of livestock in the Pacific Islands region: setting priorities for food animal biosecurity. Acta Trop2015;143:6676DOIPubMed
  21. Rushdy  AO’Mahony  MPHLS overview of communicable diseases 1997: results of a priority setting exercise. Commun Dis Rep CDR Suppl1998;8:S112.PubMed
  22. Horby  PRushdy  AGraham  CO’Mahony  MPHLS Overview of Communicable Diseases CommitteePHLS overview of communicable diseases 1999. Commun Dis Public Health2001;4:817.PubMed
  23. Lapid  RKing  RYakobson  BShalom  UMoran-Gilad  JWildlife Pathogen Surveillance in Israel to Inform Human and Animal Infectious Disease Control: a Prioritization Exercise. Isr J Vet Med2016;71:3341.
  24. McKenzie  JSimpson  HLangstaff  IDevelopment of methodology to prioritise wildlife pathogens for surveillance. Prev Vet Med2007;81:194210DOIPubMed
  25. Maino  MPérez  POviedo  PSotomayor  GAbalos  PThe analytic hierarchy process in decisionmaking for caprine health programmes. Rev Sci Tech2012;31:88997DOIPubMed
  26. Havelaar  AHvan Rosse  FBucura  CToetenel  MAHaagsma  JAKurowicka  Det al. Prioritizing emerging zoonoses in the Netherlands. PLoS One2010;5:e13965DOIPubMed
  27. Kurowicka  DBucura  CCooke  RHavelaar  AProbabilistic inversion in priority setting of emerging zoonoses. Risk Anal2010;30:71523DOIPubMed
  28. Cox  RSanchez  JRevie  CWMulti-criteria decision analysis tools for prioritising emerging or re-emerging infectious diseases associated with climate change in Canada. PLoS One2013;8:e68338DOIPubMed
  29. Hongoh  VMichel  PGosselin  PSamoura  KRavel  ACampagna  Cet al. Multi-stakeholder decision aid for improved prioritization of the public health impact of climate sensitive infectious diseases. Int J Environ Res Public Health2016;13:419DOIPubMed
  30. Stebler  NSchuepbach-Regula  GBraam  PFalzon  LCUse of a modified Delphi panel to identify and weight criteria for prioritization of zoonotic diseases in Switzerland. Prev Vet Med2015;121:1659DOIPubMed
  31. Stebler  NSchuepbach-Regula  GBraam  PFalzon  LCWeighting of criteria for disease prioritization using conjoint analysis and based on health professional and student opinion. PLoS One2016;11:e0151394DOIPubMed
  32. Brookes  VJHernández-Jover  MCowled  BHolyoake  PKWard  MPBuilding a picture: Prioritisation of exotic diseases for the pig industry in Australia using multi-criteria decision analysis.Prev Vet Med2014;113:10317DOIPubMed
  33. Del Rio Vilas  VJVoller  FMontibeller  GFranco  LASribhashyam  SWatson  Eet al. An integrated process and management tools for ranking multiple emerging threats to animal health. Prev Vet Med2013;108:94102DOIPubMed
  34. Ruzante  JMDavidson  VJCaswell  JFazil  ACranfield  JAHenson  SJet al. A multifactorial risk prioritization framework for foodborne pathogens. Risk Anal2010;30:72442DOIPubMed
  35. Dahl  VTegnell  AWallensten  ACommunicable diseases prioritized according to their public health relevance, Sweden, 2013. PLoS One2015;10:e0136353DOIPubMed
  36. Cardoen  SVan Huffel  XBerkvens  DQuoilin  SDucoffre  GSaegerman  Cet al. Evidence-based semiquantitative methodology for prioritization of foodborne zoonoses. Foodborne Pathog Dis2009;6:108396DOIPubMed
  37. Bouwknegt  MHavelaar  ANeslo  Rde Roda Husman  AMHogerwerf  Lvan Steenbergen  Jet al. Ranking infectious disease risks to support preparedness prioritization in the European Union. Eur J Public Health. 2015;25(Suppl 3):ckv167.036-ckv167.036.
  38. Balabanova  YGilsdorf  ABuda  SBurger  REckmanns  TGärtner  Bet al. Communicable diseases prioritized for surveillance and epidemiological research: results of a standardized prioritization procedure in Germany, 2011. PLoS One2011;6:e25691DOIPubMed
  39. Gilsdorf  AKrause  GPrioritisation of infectious diseases in public health: feedback on the prioritisation methodology, 15 July 2008 to 15 January 2009. Euro Surveill2011;16:19861.PubMed
  40. Krause  GWorking Group on Prioritization at Robert Koch InstituteHow can infectious diseases be prioritized in public health? A standardized prioritization scheme for discussion. EMBO Rep2008;9(Suppl 1):S227DOIPubMed
  41. Ciliberti  AGavier-Widén  DYon  LHutchings  MRArtois  MPrioritisation of wildlife pathogens to be targeted in European surveillance programmes: Expert-based risk analysis focus on ruminants. Prev Vet Med2015;118:27184DOIPubMed
  42. McIntyre  KMSetzkorn  CHepworth  PJMorand  SMorse  APBaylis  MA quantitative prioritisation of human and domestic animal pathogens in Europe. PLoS One2014;9:e103529DOIPubMed
  43. Cox  RMcIntyre  KMSanchez  JSetzkorn  CBaylis  MRevie  CWComparison of the H‐index scores among pathogens identified as emerging hazards in North America. Transbound Emerg Dis2016;63:7991DOIPubMed
  44. Remme  JHFBlas  EChitsulo  LDesjeux  PMPEngers  HDKanyok  TPet al. Strategic emphases for tropical diseases research: a TDR perspective. Trends Parasitol2002;18:4216DOIPubMed
  45. World Health Organization. The Dubrovnik pledge on surveillance and prioritization of infectious diseases [cited 2002 Nov 23]. http://www.euro.who.int/__data/assets/pdf_file/0005/128039/E78888.pdf
  46. Cox  RRevie  CWSanchez  JThe use of expert opinion to assess the risk of emergence or re-emergence of infectious diseases in Canada associated with climate change. PLoS One2012;7:e41590DOIPubMed
  47. Morgan  DKirkbride  HHewitt  KSaid  BWalsh  ALAssessing the risk from emerging infections. Epidemiol Infect2009;137:152130DOIPubMed
  48. Palmer  SBrown  DMorgan  DEarly qualitative risk assessment of the emerging zoonotic potential of animal diseases. BMJ2005;331:125660DOIPubMed
  49. World Health Organization. R&D Blueprint: progress in prioritizing diseases with epidemic potential [cited 2016 Nov 18]. http://www.who.int/blueprint/what/research-development/prioritizing_diseases_progress/en
  50. World Health Organization. The 2017 annual review of diseases prioritized under the Research and Development Blueprint informal consultation 24–25 January 2017, Geneva, Switzerland [cited 2017 Mar 30]. http://www.who.int/blueprint/what/research-development/2017-Prioritization-Long-Report.pdf?ua=1
  51. Toloie-Eshlaghy  ANazari Farokhi  EMeasuring the importance and the weight of decision makers in the criteria weighting activities of group decision making process. American Journal of Scientific Research. 2011;24:612.
  52. Sutherland  WJBroad  SCaine  JClout  MDicks  LVDoran  Het al. A horizon scan of global conservation issues for 2016. Trends Ecol Evol2016;31:4453DOIPubMed
  53. Saaty  TLHow to make a decision: the analytic hierarchy process. Eur J Oper Res1990;48:926DOI
  54. Saaty  TLDecision making with the analytic hierarchy process. Int J Serv Sci2008;1:8398DOI
  55. Saito  EKShea  SJones  ARamos  GPitesky  MA cooperative approach to animal disease response activities: Analytical hierarchy process (AHP) and vvIBD in California poultry. Prev Vet Med2015;121:12331DOIPubMed
  56. Madhavan  GPhelps  CSangha  KLevin  SRappuoli  RBridging the gap: need for a data repository to support vaccine prioritization efforts. Vaccine2015;33(Suppl 2):B349DOIPubMed
  57. Pannucci  CJWilkins  EGIdentifying and avoiding bias in research. Plast Reconstr Surg2010;126:61925DOIPubMed
  58. Brookes  VJHernández-Jover  MNeslo  RCowled  BHolyoake  PWard  MPIdentifying and measuring stakeholder preferences for disease prioritisation: A case study of the pig industry in Australia. Prev Vet Med2014;113:11831DOIPubMed
  59. McCormick  THZheng  T. Adjusting for recall bias in “How many X’s do you know?” surveys [cited 2017 Apr 10]. http://www.stat.columbia.edu/~tzheng/files/McCormick_jsm07.pdf
  60. World Health Organization. The R&D Blueprint—accelerating R&D processes [cited 2018 Apr 5]. http://www.who.int/blueprint/what/research-development/en
  61. World Health Organization. The R&D Blueprint—developing new norms and standards tailored to the epidemic context [cited 2018 Apr 5]. http://www.who.int/blueprint/what/norms-standards/en
 Top

Figures

 Top
Suggested citation for this article: Si Mehand M, Millett P, Al-Shorbaji F, Roth C, Kieny MP, Murgue B. World Health Organization methodology to prioritize emerging infectious diseases in need of research and development. Emerg Infect Dis. 2018 Sep [date cited]. https://doi.org/10.3201/eid2409.171427
DOI: 10.3201/eid2409.171427


Original Publication Date: 7/31/2018

No hay comentarios:

Publicar un comentario