Grading the Quality of Information and Synthesis of mHealth Evidence MPH Capstone Project Dr. Jaime Lee Abstract Background: Despite the growing mHealth evidence base, it comprises of literature with varying levels of methodological rigor due to the rapid pace of technology and the multi-‐disciplinary nature of mHealth. As such, a grading tool to assess the quality of information will help researchers improve the completeness, rigor and transparency of their research reports for mHealth interventions for the purpose of guidance development. Objective: To propose a grading tool to rate the quality of information in mHealth, and for synthesis of the available high quality information about a particular mHealth intervention. Methods: We performed a comprehensive search for published checklists used to assess quantitative and qualitative studies, evaluation of complex interventions in Medline, and author or reviewer guidelines of major medical journals including those specific for mHealth or eHealth. 85 items from 7 checklists were compiled into a comprehensive list and we recorded the frequency of each item across all checklists. Duplicate items and ambiguous items were removed. The grading tool was subjected to an extensive iterative process of feedback and revision. A preliminary validation study to assess inter-‐rater reliability and clarity of item descriptions was conducted. We tested the use of the tool on 2 papers, a peer-‐reviewed article and a grey literature article with 8 graduate students. Results: Items most frequently included in the checklists were identified. All items were grouped into two domains: 1) Reporting and methodology and 2) Essential mHealth criteria. Preliminary testing of the mHealth grading tool showed moderate agreement between the rater for scoring of items with overall kappa statistic of 0.48 for the grey literature piece and 0.43 for the peer-‐reviewed article. Conclusions: The mHealth grading tool was developed to improve the quality of information of mHealth studies. Dr. Jaime Lee MPH Candidate, 2013 Acknowledgments I would firstly like to thank the support of my advisor Dr. Alain Labrique for his constant support and guidance, and kindness for giving me the opportunity to work on this World Health Organization mHealth Technical Advisory Group project. It has been a fantastic experience and without him this capstone project would not be possible. To Dr. Smisha Agarwal and Dr. Amnesty Lefevre, it has been truly been wonderful to work with such a brilliant team – thank you, my friends. I must also thank the other members of the Johns Hopkins WHO mTAG team who gave invaluable advice and reviewed the multiple drafts: Dr. Larissa Jennings and Michelle Colder Carras. To my friends, Estefania, Hamish, Madelyn, Mariam, Melissa, Sam, Shaymaa and Sneha who tested the mHealth grading tool, thank you for taking the time out of your busy schedules to help me complete this capstone project. Finally, it is important to acknowledge that the WHO mTAG Quality of Information Task Force have given multiple rounds of feedback that have been critical to the development of the grading tool. Disclosure Statement This Capstone is based on work that I am currently doing as a part of the Johns Hopkins WHO mTAG team. 2 Dr. Jaime Lee MPH Candidate, 2013 Table of Contents Abstract ............................................................................................................................................. 1 Acknowledgments ......................................................................................................................... 2 Disclosure Statement .................................................................................................................... 2 1. Background and current mHealth evidence base .......................................................... 4 1.1 Current tools for grading quality of information ................................................................... 7 2. Objectives ..................................................................................................................................... 8 3. A new grading tool for mHealth research ......................................................................... 8 3.1 Grading Quality of Information .................................................................................................... 9 3.1.1 Methodology: Development of grading criteria .............................................................................. 9 3.2 Using the mHealth grading tool to assess quality of Information .................................. 11 3.3 Calculation of Quality Score and Quality of Information rating ...................................... 18 3.4 Synthesis of evidence ..................................................................................................................... 19 3.5 Convene an expert review panel ................................................................................................ 20 4. Preliminary validation of the mHealth grading tool and inter-‐rater reliability 21 4.1 Objectives .......................................................................................................................................... 21 4.2 Methodology ..................................................................................................................................... 21 4.3 Results ................................................................................................................................................. 22 5. Discussion .................................................................................................................................. 23 6. References ................................................................................................................................. 25 7. Appendices ................................................................................................................................ 27 8. Reflection on the Capstone Project ................................................................................... 39 3 Dr. Jaime Lee MPH Candidate, 2013 1. Background and current mHealth evidence base Mobile health, or mHealth, is defined by the World Health Organization (WHO) as “medical and public health practice supported by mobile devices, such as mobile phones, patient monitoring devices, personal digital assistants (PDAs), and other wireless devices. mHealth involves the use and capitalization on a mobile phone’s core utility of voice and short messaging service (SMS) as well as more complex functionalities and applications including general packet radio service (GPRS), third and fourth generation mobile telecommunications (3G and 4G systems), global positioning system (GPS), and Bluetooth technology” (1). Mobile technologies offer an effective means of delivering healthcare services to underserved populations. With the overall improvements in telecommunications, there has been increasing enthusiasm in the use of mobile technologies for health from multiple sectors such as health, computer science, engineering, and telecommunications, to capitalize on the rapid uptake of mobile communication technologies. Whilst mHealth is still a nascent field, there are indications that it has shown promise to revolutionize health systems through (2): 1) Increased access to healthcare and health-related information, particularly for hard-toreach populations 2) Improved ability to diagnose and track diseases 3) Timely more actionable public health information 4) Increased access to ongoing medical education and training for health workers 5) Increased efficiency and lower cost of service delivery There has been a growing body of literature documenting mHealth studies and initiatives. However, a number of literature reviews have noted the lack of rigorous, high quality evidence in the mHealth domain (3-6). The varying levels of rigor found in the current 4 Dr. Jaime Lee MPH Candidate, 2013 mHealth evidence base are attributable to two major factors: first, the multi-disciplinary nature of mHealth, which combines the health and technology worlds, and second, the rapid pace of development of technology. The first factor refers to how the health industry and the technology industry use different methodology to assess an intervention, with different speed and ways of dissemination of findings. In the technology space, prototypes are usually assessed by proof-of-concept or demonstration studies with fast turn-around time for modification. Then these results are generally disseminated quickly in the grey literature, including white papers, conference papers, presentations and blogs. In contrast, the health field moves at a slower pace. In general, more emphasis is placed on methodological rigor and the timeframe for a study may be longer than in the technology industry. The majority of results in the field of health and public health are disseminated through peer-reviewed journals and conference papers, and a smaller proportion in the grey literature. So this leads into the second issue. The time it takes for a study of high methodological rigor to be completed and then published in a peer-reviewed journal could take over a year, even up to a few years, for the findings to be disseminated. However with the rapid pace at which technology changes, a newer model or new technology may be available in a significantly shorter timeframe and hence the study results may potentially be less relevant to the mHealth field. Consequently, the current mHealth evidence base varies in the quality of information that is disseminated in multiple forms from peer-reviewed literature to white papers, theses, reports, presentations and blogs. The World Bank reported that there are more than 500 mHealth 5 Dr. Jaime Lee MPH Candidate, 2013 studies in 2011 (7). For the purpose of guidance development, the varying quality of information will offer different levels of value to stakeholders. As such, the present mHealth evidence base is not sufficient to inform governments and industry partners to invest resources in nationally scaled mHealth interventions (3, 6). The Grading of Recommendations Assessment, Development and Evaluation (GRADE) approach is one method used to develop guidelines, and is being increasingly used by international organizations such as the World Health Organization (WHO) and the Cochrane Collaboration (8). The GRADE system has brought greater transparency and a systematic approach to rating the quality of evidence and grading the strength of recommendations (8). It is tool that was developed for systematic reviews of evidence on effectiveness. There are also other tools that have been developed to assess the reporting of systematic reviews and metaanalyses such as the PRISMA checklist (9), and to assess their methodological quality or reliability such as the SUPPORT tools (10) and AMSTAR (11). However, synthesis of the mHealth evidence base, as to what works and what does not work, has yet to be rigorously assessed and established (3). Such information would provide a valuable contribution to guidance development. There have been a number of efforts to review and synthesize the mHealth evidence base using the grading tools previously mentioned (12-15). Free et al. conducted two recent systematic reviews (13, 14). In one systematic review of 42 controlled trials, the authors concluded that mobile technology interventions showed modest benefits for supporting diagnosis and patient management outcomes (14). They also reported in this study that none of the trials were of high quality and the majority of studies were conducted in high-income countries. In the other systematic review of 59 trials that investigated the use of mHealth interventions to improve disease 6 Dr. Jaime Lee MPH Candidate, 2013 management, and 26 trials examined their use to change behavior, Free et al. found that there was mixed evidence regarding the benefits of interventions (13). Text messaging interventions were shown to increase adherence to anti-retroviral medication in a low-income setting, and increased smoking cessation was demonstrated in high-income settings (13). While in other areas, the evidence suggested potential benefits of mHealth interventions. Using the results of these systematic reviews for guidance development are limited due to the lack of high-quality trials with adequate power. Hence there has been a call of high quality evidence and a set of standards that identify the optimal strategies for delivering and informing scale-up of mHealth interventions (16). 1.1 Current tools for grading quality of information Systematic and transparent approaches to grading the quality of mHealth information are particularly important given the complexity of mHealth interventions and the need for adequate integration with the existing health system of a particular country. One challenge to grading public health interventions is that typically randomized controlled trials (RCTs) have been held as providing the highest quality of evidence that a particular strategy can yield a specific outcome or result, and non-‐randomized designs are often perceived as less useful to the synthesis of evidence. This may particularly affect grading of emerging and complex public health interventions, where RCTs may be infeasible or otherwise inappropriate (17). Evidence reporting and synthesis approaches such as MOOSE (18), TREND (19) and STROBE (20) provide suggestions for improving quality of reporting in observational studies, but do not provide a framework to grading strength of evidence, when data sources are varied and depend on mixed methods. 7 Dr. Jaime Lee MPH Candidate, 2013 The current tools used to evaluate the quality of quantitative or qualitative studies are very specific to the study design. For example CONSORT is specific to RCTs (21), STROBE is specific to observational studies (20), TREND is for non-randomized studies particularly for behavioral and public health interventions (19), and COREQ (22) is for qualitative studies. There is no grading tool that assesses the quality of information of mHealth research with specific mHealth criteria that can help develop recommendations. 2. Objectives The objective of this capstone is to propose a grading tool to rate the quality of information in mHealth, as part of a two-stage process: first, to identify the highest quality information generated by a range of methodologies (from qualitative to quantitative), reported according to the best standards for that methodology, and second, to provide the raw materials for a synthesis of the available, high quality information about a particular mHealth strategy. 3. A new grading tool for mHealth research The grading tool that we propose has been developed because there is a need to examine the Quality of Information (QoI) in mHealth studies. The majority of publications lack clarity, transparency and rigor in the conduct of the research, and there is a tendency, given the rapid pace of this emerging field, to report formative and even research findings in the non-peerreviewed literature (4, 5, 23). Hence, it is important to develop a grading tool to help researchers improve the completeness, rigor and transparency of their research reports and to facilitate the more efficient use of research findings for those seeking to select and implement 8 Dr. Jaime Lee MPH Candidate, 2013 mHealth interventions, potential funders of evaluation studies, and policymakers (23). Figure 1 shows an overview of mHealth guidance development. The evaluation of mHealth research requires a unique approach using a combination of quantitative and qualitative evaluation methods that take into account the context and setting of the mHealth intervention. This proposed approach is specific to mHealth research but has been designed to be easily understood and applied by anyone interested in assess evidence to strengthen health systems. 3.1 Grading Quality of Information 3.1.1 Methodology: Development of grading criteria The mHealth grading tool development process was designed to produce consensus among a broad constituency of experts and users on both the content and format of guideline items. We first performed a comprehensive search for published checklists used to assess the methodology of quantitative and qualitative studies, the evaluation of complex interventions, guidelines for reporting quantitative and qualitative studies in Medline, and author or reviewer guidelines of major medical journals including those specific for mHealth or eHealth. We extracted all criteria for assessing quantitative or qualitative studies from each of the included publications. Duplicated items were excluded. 85 items from 7 checklists (19-22, 24-26) were compiled into a comprehensive list for reporting and methodology criteria. We recorded the frequency of each item across all the publications (see Appendix 1). For the mHealth criteria, we generated additional items through literature searches. We subjected consecutive drafts to an extensive iterative process of consultation. 9 Dr. Jaime Lee MPH Candidate, 2013 We grouped all items into two domains: 1) Reporting and methodology criteria, and 2) Essential mHealth criteria (Box 1). Within each domain we reviewed all relevant items and simplified the criteria by rephrasing items for clarity and removing duplicates and items with ambiguous definitions. We drafted a provisional list of items deemed important to be included in the checklist. This draft checklist was used to facilitate discussion at a December 2012, there was a 3-day WHO mHealth Technical Advisory Group meeting of 18 global representatives. All participants discussed the draft checklist and it was subjected to intensive analysis, comment and recommendations for change. Furthermore, five members of the Quality of Information Task Force reviewed the first draft in depth and applied the checklist to various pieces of literature. Then they provided additional feedback. Box 1. Overview of mHealth Grading tool Domain 1: Reporting and Methodology Criteria A. Essential criteria for ALL studies B. Essential criteria based on type of study (choose at least 1 of the following:) i. Quantitative ii. Qualitative iii. Economic evaluation Domain 2: Essential mHealth Criteria After the meeting we subsequently revised the checklist. During this process, the coordinating group (ie. the authors of the present paper) met on six occasions and held several telephone conferences to revise the checklist. We obtained further feedback of the checklist from more than 10 people. 10 Dr. Jaime Lee MPH Candidate, 2013 Figure 1: Overview of proposed mHealth guidance development Articulation of mHealth intervention for review Systematic access of relevant information in database Use mHealth grading tool to rate quality of information for each study based on Methodology and Reporting criteriia (Domain 1) and mHealth criteria (Domain 2) Synthesis of evidence -‐ summarize the quality of information for every study across both Domains Convene expert panel to assess the overall quality of information, develop recommendations and identify evidence gaps Consensus statement on mHealth intervention based on the quality of information, and direction, consistency and magnitude of evidence 3.2 Using the mHealth grading tool to assess quality of Information The mHealth grading tool for assessing the QoI is a flexible approach that allows the grading of reporting and methodology for varied study designs (Table 1) As indicated in Box 1, all evidence under consideration must be scored against the “Essential criteria”. After that, the evidence can be classified as qualitative or quantitative or economic evaluation based on the 11 Dr. Jaime Lee MPH Candidate, 2013 methodology employed for the study. A detailed description of the steps in grading quality of information is presented in Box 2 and an example of using the mHealth tool to grade an article is shown in Appendix 2. Box 2: How to use the mHealth grading tool Step 1: In Domain 1 Part A, apply the criteria to all studies. Step 2: For Domain 1 Part B, you can choose to apply 1 or more of the following criteria, as appropriate to the mHealth study: i. Quantitative ii. Qualitative iii. Economic evaluation Step 3: In Domain 2, apply all essential mHealth criteria to all studies. Step 4: Record the scores in the Scoring Summary Grid (Table 2) Step 5: Calculate the Quality Score for Domains 1 and 2 (Quality score = # points / maximum score* X 100%) Step 6: Based on the calculated Quality Score, you can determine the Quality of information for each domain separately as Weak <50%, Moderate 50-75% or Strong >75%. Step 7: Steps 1 to 6 can be repeated for every study identified for a particular mHealth intervention. * The maximum score for Domain 1 will depend on which set/s of criteria were applied in Part B i.e. If it is a quantitative study, the maximum score for Domain 1 is 38. But if it is a study with quantitative and qualitative methods, then the maximum score for Domain 1 is 41. 12 Dr. Jaime Lee MPH Candidate, 2013 Table 1. Grading criteria of assessing Quality of Information from mHealth studies Domain 1: Reporting and Methodology Criteria A. Essential criteria for all studies Criteria Introduction Rationale/ scientific background Objectives/ hypotheses Intervention model and theoretical considerations Item no. Description 1 Scientific background and explanation of rationale 2 Specific objectives or hypotheses 3 Description of development and piloting of intervention and any theoretical support used to design the intervention (how the intervention is intended to bring about change in the intended outcomes) 4 Clear description and justification of chosen study design, especially if the design has been chosen based on a compromise between internal validity and the complexity and constraints of the research setting or research question 5 Clearly defined primary and secondary outcome measures to meet study objectives 6 Description of data collection methods, including training and level of data collection staff 7 Eligibility criteria for participants 8 Method of participant recruitment (eg. Referral, selfselection), including the sampling method if a systemic sampling plan was implemented 9 Method of participant allocation is clearly described 10 Information presents a clear amount of sampling strategy 11 Justification for sample size is reported Setting and locations Comparator 12 Settings and locations where the data were collected 13 Describes use of a comparison group from similar population with regard to socio-demographics or adjusts for confounding Data sources/ measurement 14 Describes the source of data for each variable of interest and detailed measurement criteria for the data 15 Enrollment: the numbers of participants screened for Methodology Study design Participants Sampling Results Participants Score as: 1 – Found/ Met 0 – Not found/ not met 13 Dr. Jaime Lee MPH Candidate, 2013 eligibility, found to be eligible or not eligible, declined to be enrolled, and enrolled in the study 16 Assignment: the number of participants assigned to a study condition and the number of participants who received each intervention 17 Analysis: the number of participants included in, or excluded from, the main analysis, by study condition Recruitment 18 Dates defining the periods of recruitment and follow-up Baseline data 19 Baseline demographic and clinical characteristics of participants in each study cohort Fidelity 20 Degree to which the intervention is implemented as planned with a description of adherence, exposure, quality of delivery, participant responsiveness and program differentiation Context 21 Description of the organizational, social, economic and political context in which the intervention is developed and operated Attribution 22 The link between the intervention and outcome is reported Bias 23 The risk of biases is reported 24 The risk of confounding is reported 25 Ethical and distributional issues are discussed Discussion Summary of evidence 26 General interpretation of the results in the context of current evidence and current theory Limitations 27 Discussion of study limitations, addressing sources of potential bias, imprecision, and, if relevant, multiplicity of analyses Generalizability 28 Generalizability (external validity) of the study findings, taking into account the study population, the characteristics of the intervention, length of follow-up, incentives, compliance rates, and specific settings involved in the study and other contextual issues Conclusions/ interpretation 29 Interpretation of the results, taking into account study hypotheses, sources of potential bias, imprecision of measures, and other limitations or weaknesses of the study 30 Discussion of the success of, and barriers to, scaling up Ethical considerations 14 Dr. Jaime Lee MPH Candidate, 2013 the intervention Other Funding Competing interests 31 Discussion of research, programmatic or policy implications 32 33 Sources of funding and role of funders Relation of the study team towards the intervention being evaluated i.e. developers/sponsors of the intervention Subtotal of Quality Points for Essential criteria for all studies (out of 33): B. Essential criteria based on type of study - Must choose at least 1 of the following criteria to apply as appropriate: i. Quantitative ii. Qualitative iii. Economic evaluation Criteria i. Quantitative Statistical methods Outcomes and estimation Item no. Description 34 Statistical methods used to compare groups for primary and secondary outcomes 35 Methods for additional analyses, such as subgroup analyses and adjusted analyses 36 Methods of imputing or dealing with missing data 37 For each primary and secondary outcome, study findings are presented with for each study cohort, and the estimated effect size and confidence interval to indicate the precision 38 Estimate for random data variability and outliers are clearly stated Score as: 1 – Found/ Met 0 – Not found/ not met Subtotal for Quantitative study design (out of 5) ii. Qualitative N/A Analytical methods Use of verification methods to demonstrate credibility Reflexivity of account provided 39 40 41 Analytical methods clearly described (In-depth description of analysis process, how categories/themes were derived) Discusses use of triangulation, member checking (respondent validation), search for negative cases, or other procedures Relationship of researcher/study participant has been discussed, examining the researcher’s role, bias, or potential influence Subtotal for Qualitative study design (out of 3): 15 Dr. Jaime Lee MPH Candidate, 2013 iii. Economic evaluation 42 Competing alternatives clearly described (e.g. costeffectiveness of zinc and ORS for treatment of diarrhea versus standard treatment with ORS alone) 43 The chosen analytic time horizon is reported 44 The perspective / viewpoints (e.g. societal, program, provider, user, etc.) of the analysis is clearly described 45 The alternatives being compared are clearly described 46 The sources of effectiveness estimates are clearly stated 47 Details of the design and results of the effectiveness study and/or methods for effect estimation are clearly stated 48 Methods for estimation of quantities and unit costs are described 49 Details of currency of price adjustments for inflation or currency conversion are given 50 Currency and price data are recorded 51 The choice of model used and the key parameters on which it is based are reported 52 The discount rate(s) are reported 53 Sensitivity analyses are reported 54 Incremental analyses are reported 55 Major outcomes are presented in a disaggregated as well as aggregated form Subtotal for Economic Evaluation (out of 14): Domain 2: Essential mHealth Criteria for all studies Criteria Item no. Infrastructure 56 Technology architecture Intervention 57 58 Description Score as: 1 – Found/ Met 0 – Not found/ not met Clearly presents the availability or kind of infrastructure to support technology operations (eg. electricity, access to power, connectivity) Describes the technology architecture including the software and hardware mHealth intervention is clearly described with frequency and mode of delivery of intervention (i.e. SMS, face-toface, interactive voice response) for replication 16 Dr. Jaime Lee MPH Candidate, 2013 59 Details of the content of the intervention are clearly described or link is presented and content is publically available Usability 60 Clearly describes the ability of different user groups to successfully use the technology in a given context eg. literacy, computer/Internet literacy, ability to use device User feedback 61 Describes user feedback about the intervention Identifies constraints 62 mHealth solution states one or more constraints in the delivery of current service, intervention, process or product Access and affordability 63 Presents data on the access and affordability of the mHealth solution from varying user perspectives Cost assessment 64 Presents basic costs assessment of the mHealth intervention from varying perspectives Training inputs 65 Clearly describes the training inputs for the adoption of the mHealth solution Strengths and limitations 66 Clearly presents mHealth solution considerations, both strength and limitations, for delivery at scale Language adaptability 67 Describes the adaptation, or not, of the solution to the local language Replicability 68 Clearly presents the source code/screenshots/flowcharts of the algorithms/ examples of messages to ensure replicability Data security 69 Describes the data security procedures/ confidentiality protocols Subtotal for mHealth criteria (out of 14): 17 Dr. Jaime Lee MPH Candidate, 2013 3.3 Calculation of Quality Score and Quality of Information rating After using the grading tool to assess an mHealth study, record the scores into the Scoring Summary Grid (Table 2) to calculate the Quality Score for Domains 1 and 2. The quality of information is defined under 2 areas: 1) Domain 1: Reporting and Methodology – This is indicative of the quality of methodological rigor employed by the studies under consideration, as well as the reporting standards that have been adhered to. 2) Domain 2: Essential mHealth criteria – Classifies the studies under consideration based on the quality of information presented about the mHealth intervention. The Quality Score for each domain is calculated using the formula: Quality score = !"#$%& !" !"#$%& !"#$%& × 100% !"#$%&% !"#$% !"# !"#$%& For Domain 1, the maximum will depend on which set/s of criteria were applied in Part B. That is, if it is a quantitative study, the maximum score for Domain 1 is 38. But if it is a study with quantitative and qualitative methods, then the maximum score for Domain 1 is 41. So then the quality score will be calculated accordingly. Domain 2 is more straightforward as the maximum score is set at 14 quality points. Then based on the Quality Score, you can determine the Quality of Information rating for each domain as Strong (>75%), Moderate (50% to 75%) or Weak (<50%). 18 Dr. Jaime Lee MPH Candidate, 2013 Table 2. Scoring summary grid Number of Quality points Maximum score (If criteria were not applicable, circle N/A) Domain 1: Reporting and Methodology A. Essential criteria for all studies 33 B. Essential criteria based on type of study (choose at least 1 of the following options) i. Quantitative 5 or N/A ii. Qualitative 3 or N/A iii. Economic evaluation 14 or N/A Total number of Quality points for Reporting and Methodology Quality score (# divided by Maximum total score x 100%) Quality of Information (<50% Weak, 50-75% Moderate, >75% Strong) Domain 2: Essential mHealth Criteria Total number of Quality points for mHealth Quality score (# divided by 14 x 100%) Quality of Information (<50% Weak, 50-75% Moderate, >75% Strong) 14 3.4 Synthesis of evidence The mHealth tool can be used to grade all available studies on a specific mHealth intervention and get a QoI rating for Domains 1 and 2 for each paper. Then the evidence can be synthesized by cross-tabulating the domain QoI ratings for each paper in a matrix. Table 3 shows a hypothetical example of recording the domain QoI ratings of 30 papers for an mHealth intervention. 19 Dr. Jaime Lee MPH Candidate, 2013 Table 3: Synthesis of Evidence Matrix (including a hypothetical example of 30 papers for a specific mHealth intervention, with the number of papers for each option) Domain 2: Quality of Information Rating for mHealth criteria Domain 1: Quality of Information Rating for Reporting and Methodology criteria Weak Moderate Strong Weak 3 10 2 Moderate 0 1 4 Strong 5 2 3 3.5 Convene an expert review panel Once the QoI for all available studies are graded, an expert panel can convene to review the synthesis of the evidence and provide guidance. We propose that the expert panel reviews papers that are rated at least as “Moderate” in both domains. They can look at the direction, consistency and magnitude of the evidence to help develop recommendations. Furthermore this method of synthesis will allow reviewers to see where the “gaps” and areas for improvement are for a specific mHealth intervention. For example, a paper rated as “Strong” in Domain 1, may be Weak or Moderate in reporting the mHealth considerations important for scale-up. Hence the review panel can provide guidance by recommending that the investigators aim to increase their SOE rating in Domain 2 in order to improve the potential sustainability of the mHealth intervention. 20 Dr. Jaime Lee MPH Candidate, 2013 4. Preliminary validation of the mHealth grading tool and inter-‐ rater reliability 4.1 Objectives To assess the inter-rater reliability of the grading tool and the clarity of the item descriptions, we held a workshop in April, 2013 and tested the use of the tool on 2 papers, a peer-reviewed article and a grey literature article with 8 graduate students who are in the Master of Public Health program at Johns Hopkins Bloomberg School of Public Health. 4.2 Methodology The eight participants were selected through the author’s network and approached to participate in this workshop. All students are currently completing the MPH program, have completed at least 3 units of the Statistical Methods in Public Health (Course number 140.621-623), Principles of Epidemiology (340.601), as well as Observational Epidemiology (340.608) or Epidemiological Methods (340.751-753). All students received a meal as compensation for their time to complete this exercise. Each participant was asked to grade the QoI for two papers selected by the author. One paper is a peer-reviewed article (27) and the other is a piece of grey literature (28). The author was present during the grading exercise to answer any questions about use of the tool or to clarify the items in the checklist. After the grading exercise, participants were asked to give feedback on use of the tool. The major themes were noted. 21 Dr. Jaime Lee MPH Candidate, 2013 The percentage of overall agreement and the Kappa statistic was calculated for each paper using Stata Version 12. Further analysis was conducted to assess the percentage of overall agreement by each item. 4.3 Results For the article by Lester and Kariri, the percentage overall agreement between the raters for scoring the QoI for both domains (as Strong, Moderate or Weak) was 73.2%. If we breakdown the analysis to each domain separately, Domain 1 had 100% agreement between raters, and Domain 2 had 62.5% agreement. Further sub-analysis by the scoring of each item showed moderate agreement overall with kappa-statistic of 0.48. For the peer-reviewed article by Mbuagbaw et al., the percentage overall agreement for QoI for both domains was lower than the grey literature article with a result of 42.9%. Subanalysis by scoring of each item also showed moderate agreement with kappa-statistic of 0.43 We also analyzed the percentage overall agreement by each individual item so that we could see which items had less than 75% agreement between the raters. For the Lester and Kariri paper, there were 7 items that had 62.5% overall agreement between the raters and 1 item that had 50% overall agreement. For the Mbaugbaw paper, there were 9 items that had 62.5% overall agreement between the raters and 3 items that had 50% overall agreement. Overall, the raters found the tool “easy to use,” as one rater put it. There were a few items that they had identified required further clarification and these correlated with the items that had less than 75% overall agreement. 22 Dr. Jaime Lee MPH Candidate, 2013 5. Discussion The mHealth grading tool was developed to promote comprehensive reporting of mHealth studies with the goal of improving the completeness, rigor and transparency of mHealth research. We developed the checklist through an open process, taking into account the experience gained with previous published checklists. It allows the assessment of the quality of information in 2 domains, methodological rigor and mHealth contextual and technical considerations. The checklist consists of items specific to mHealth interventions as well as generic reporting and methodology criteria that are applicable to quantitative, qualitative, economic evaluation or mixed methods study designs. The process of grading the quality of information in the 2 distinct domains allows reviewers to develop recommendations for a particular mHealth intervention in 2 dimensions and to see the gaps in the evidence. For a specific intervention, reviewers could analyze papers that have rated as at least “Moderate” in Domain 1 and Domain 2, and then assess the direction, consistency and magnitude of the evidence to develop further recommendations. The grading tool should not be interpreted as an attempt to dictate the reporting of mHealth research in a rigid format. The checklist items should be addressed with clarity and in sufficient detail somewhere in an article, but the order and format for presenting the information depends on the authors’ preferences and publication style. We conducted a small preliminary validation test to assess the inter-rater reliability and to identify which items in the checklist would require reviewing or clarification. Whilst both papers showed moderate agreement between the raters, the peer-reviewed paper showed lower percentage overall agreement than the grey literature paper. This was most likely due 23 Dr. Jaime Lee MPH Candidate, 2013 to the paper being a study protocol. Consequently there was confusion between the raters. Some raters awarded 1 point if the authors mentioned what they would do in the study while others raters awarded 0 points as there are technically no results as this was a study protocol. So for future use of this tool we could recommend that it should not be used to grade study protocols. Analyzing the percentage overall agreement by each item was valuable to identify which items required further clarification or review. The feedback from the raters about which items needed further explanation at the end of the workshop also supported those findings. So this will help us revise those specific checklist items. It should be acknowledged that the mHealth grading tool is currently limited to threes main types of studies – quantitative, qualitative and economic evaluation study designs. It is not designed for stakeholder analysis, policy analysis or equity analysis. As mentioned earlier, study protocols may also not be appropriate for the grading tool. Another limitation was that while no one was excluded from the process, existing professional networks influenced the composition of the group of contributors. We emphasize that the mHealth grading tool should be viewed as an evolving document that requires ongoing assessment and refinement. We will revise the checklist in the future taking into account new evidence, comments, criticism and experience from its use. 24 Dr. Jaime Lee MPH Candidate, 2013 6. References 1. World Health Organization. mHealth: New horizons for health through mobile technologies. WHO Global Observatory for eHealth series volume 3. Geneva, Switzerland: 2011. 2. Vital Wave Consulting. mHealth for development: The opportunity of mobile technology for healthcare in the developing world. Washington, D.C. and Berkshire, UK: UN Foundation-‐Vodaphone Foundation Partnership, 2009. 3. Mechael P, Batavia H, Kaonga N, Searle S, Kwan A, Goldberger A, et al. Barriers and gaps affecting mHealth in low and middle income countried: Policy white paper. Center for Global Health and Economic Development, Earth Institute, Columbia University, May 2010. 4. Philbrick WC. mHealth and MNCH: State of the evidence. mHealth Alliance, 2012. 5. Tamrat T, Kachnowski S. Special Delivery: An Analysis of mHealth in Maternal and Newborn Health Programs and Their Outcomes Around the World. Matern Child Health J. 2012;16(5):1092-‐101. 6. AT Kearney, GSMA. Improving the evidence for mobile health. 2011. 7. Qiang CZ, Yamamichi M, Hausman V, Altman DG. Mobile applications for the health sector. Washington DC: World Bank, 2011. 8. Guyatt G, Oxman AD, Akl EA, Kunz R, Vist G, Brozek J, et al. GRADE guidelines: 1. Introduction—GRADE evidence profiles and summary of findings tables. Journal of Clinical Epidemiology. 2011;64(4):383-‐94. 9. Liberati A, Altman DG, Tetzlaff J, Mulrow C, Gøtzsche PC, Ioannidis JPA, et al. The PRISMA statement for reporting systematic reviews and meta-‐analyses of studies that evaluate health care interventions: explanation and elaboration. Journal of Clinical Epidemiology. 2009;62(10):e1-‐e34. 10. Lewin S, Oxman A, Lavis J, Fretheim A. SUPPORT Tools for evidence-‐informed health Policymaking (STP) 8: Deciding how much confidence to place in a systematic review. Health Research Policy and Systems. 2009;7(Suppl 1):S8. 11. Shea B, Grimshaw J, Wells G, Boers M, Andersson N, Hamel C, et al. Development of AMSTAR: a measurement tool to assess the methodological quality of systematic reviews. BMC Med Res Methodol. 2007;7(1):10. 12. Car J, Gurol-‐Urganci I, Vodopivec-‐Jamsek V, Atun R. Mobile phone messaging reminders for attendance at healthcare appointments. Cochrane Database of Systematic Reviews. 2012(7). 13. Free C, Phillips G, Galli L, Watson L, Felix L, Edwards P, et al. The Effectiveness of Mobile-‐Health Technology-‐Based Health Behaviour Change or Disease Management Interventions for Health Care Consumers: A Systematic Review. PLoS Med. 2013;10(1):e1001362. 14. Free C, Phillips G, Watson L, Galli L, Felix L, Edwards P, et al. The Effectiveness of Mobile-‐Health Technologies to Improve Health Care Service Delivery Processes: A Systematic Review and Meta-‐Analysis. PLoS Med. 2013;10(1):e1001363. 15. Krishna S, Boren SA, Balas EA. Healthcare via cell phones: a systematic review. Telemedicine and e-‐Health. 2009;15(3):231-‐40. 16. Tomlinson M, Rotheram-‐Borus MJ, Swartz L, Tsai AC. Scaling up mHealth: Where is the evidence? PLoS Med. 2013;10(2):e1001382. 25 Dr. Jaime Lee MPH Candidate, 2013 17. Armstrong R, Waters E, Moore L, Riggs E, Cuervo LG, Lumbiganon P, et al. Improving the reporting of public health intervention research: advancing TREND and CONSORT. Journal of Public Health. 2008;30(1):103-‐9. 18. Stroup DF, Berlin JA, Morton SC, et al. Meta-‐analysis of observational studies in epidemiology: A proposal for reporting. JAMA. 2000;283(15):2008-‐12. 19. Des Jarlais DC, Lyles C, Crepaz N, the TREND Group. Improving the reporting qualtiy of nonrandomized evaluations of behavioural and public health interventions. Am J Public Health. 2004;94:361-‐6. 20. von Elm E, Altman DG, Egger M, Pocock SJ, Gotzsche PC, Vandenbroucke JP. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: Guidelines for reporting observational studies. PLoS Med. 2007;4(10):e296. 21. Schulz KF, Altman DG, Moher D, CONSORT Group. CONSORT 2010 Statement: updated guidelines for reporting parallel group randomized trials. BMC Medicine. 2010;8:18. 22. Tong A, Sainsbury, Craig J. Consolidated criteria for reporting qualitative research (COREQ): a 32-‐item checklist for interviews and focus groups. International Journal for Quality in Health Care. 2007;19(6):349-‐57. 23. Moher D, Schulz KF, Simera I, Altman DG. Guidance for Developers of Health Research Reporting Guidelines. PLoS Med. 2010;7(2):e1000217. 24. Bossuyt PM, Reitsma JB, Bruns DE, al. e. The STARD statement for reporting studies of diagnostic accuracy: explanation and elaboration. Ann Intern Med. 2003;138:W1-‐W12. 25. Eysenbach G, CONSORT-‐EHEALTH Group. CONSORT-‐EHEALTH: Improving and standardizing evaluation reports of web-‐based and mobile health interventions. J Med Internet Res [Internet]. 2011;13(4):e126. 26. Davidoff F, Batalden P, Stevens D, Ogrinc G, Mooney SE. Publication guidelines for quality improvement studies in health care: evolution of the SQUIRE project. BMJ. 2009;338:402-‐4. 27. Mbuagbaw L, Thabane L, Ongolo-‐Zogo P, Lester R, Mills EJ, Smieja M, et al. The Cameroon Mobile Phone SMS (CAMPS) trial: a randomized trial of test messaging versus usual care for adherence to antiretroviral therapy. PLoS Med. 2012;7(12). 28. Lester R, Kariri A. Mobilizing cell phones to improve antiretroviral adherence and follow-‐up in Kenya. World Health Organization: Essential Medicines Monitor. 2009;2:1-‐3. 26 Dr. Jaime Lee MPH Candidate, 2013 7. Appendices Appendix 1: Items included in 7 published checklists CONSORT-‐ EHEALTH (25) Consolidated Consolidated Standards of Standards of Reporting Reporting Trials (for Trials (for RCTs) mHealth or eHealth trials) CONSORT (21) STROBE (20) Strengthening the Reporting of Observational studies in Epidemiology STARD (24) STAndard for the Reporting of Diagnostic Accuracy Title and abstract x x x x x Introduction Rationale/scientific background x x x x x Objectives/hypotheses x x x x x x Methodology Description of study/trial design x x x x x x x x x Eligibility criteria for participants Methods for participant recruitment Consolidated criteria for Reporting Qualitative research x x x x x x x x x x x x x x x x x x x x x Presence of non-‐participants Description of sample Describe participant sampling method Standards for Quality Improvement reporting excellence COREQ (22) x Setting and location of data collection SQUIRE (26) x Theoretical framework TREND (19) Transparent Reporting of Evaluations with Non-‐ randomized Designs 27 Dr. Jaime Lee MPH Candidate, 2013 Description of Interventions Outcomes -‐ define primary and secondary outcome measures, how and when they were assessed x x x x Exposure quantity and duration Sample size/study size x x Randomisation -‐ method Sequence generalation Allocation -‐ concealment, mechansism Implemention x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x x Unit of analysis x x Data sources/measurement x Methods for additional analyses (subgroup analyses and adjusted analyses) Describe any sensitivity analyses x x x Methods for imputing missing data, if used x x Statistical methods Data collection process x Assignment method Blinding x Interim analysis and stopping guidelines x Define exposures, predictors, potential confounders and effect modifiers Unit of delivery/content/deliverer x x x 28 Dr. Jaime Lee MPH Candidate, 2013 Bias x Summary measures Ethics and informed consent x Activities to increase compliance or adherence (incentives) x x x x Derivation of themes Test methods (for diagnostic tests -‐ reference standards and rationale, technical specifications of materials and methods, units, cutoffs; number, training and expertise of persons executing and reading the index tests and reference standards Computer/internet literacy x Open vs closed, web-‐based vs. face-‐ to-‐face assessments Developers/owners/sponsors of software Development process of application and previous formative evaluations x x x x x x Revisions and updating Source code/screenshots/flowcharts of the algorithms Digital preservation Access Mode of delivery, features/functionality/ components of the intervention and comparator x x x Use parameters x Level of human involvement x 29 Dr. Jaime Lee MPH Candidate, 2013 Prompts/reminders used Description of co-‐interventions eg. Training If outcomes were obtained through online questionnaire Personal characteristics of the research team x x x x x x Repeat interviews x Duration of interviews x Field notes x Results Participants -‐ numbers at each stage of study: potentially eligible, examined for eligibility, included in the study, assigned to which group, completing follow-‐up, and analysed x x x x x Reasons for non-‐participation at each stage x x x x Recruitment -‐ dates defining periods of recruitment and follow-‐up x x x x x x Research team's relationship with participants Interview guide (i.e. questions, prompts) Baseline data x x x Information on exposures and potential confounders x x 30 Dr. Jaime Lee MPH Candidate, 2013 Summary of missing data for each variable or interest Diagnostic test results Report demographics associated with digital divide issues eg. Age, education, gender, SES, computer/internet literacy Baseline characteristics for each study condition relevant to specific disease prevention research Baseline comparisons of those lost to follow-‐up and those retained x x x x x x x Numbers analysed x x x x x Outcomes and estimation x x x x x x x Ancillary analysis x x x Comparison between study population at baseline and target population of interest Harms Privacy breach/technical problems x x x x x x x Qualitative feedback from participants or observations from staff/researchers x x Quotations presented x Clarity of major themes x Clarity of minor themes x Discussion Summary of evidence x x x x Limitations x x x x x 31 Dr. Jaime Lee MPH Candidate, 2013 Generalizability x x x Conclusions/interpretation x x x x x x x x Other x x Funding x x Registration x x Protocol x x x Competing interests 32 Dr. Jaime Lee MPH Candidate, 2013 Appendix 2: Sample of grading an mHealth study Paper: Mbuagbaw L, Thabane L, Ongolo-‐Zolo P, Lester RT, Mills EJ et al. (2011) The Cameroon Mobile Phone SMS (CAMPS) Trials: A randomized trial of text messaging versus usual care for adherence to antiretroviral therapy. PLoS ONE 7(12): e46909. Domain 1: Reporting and Methodology Criteria A. Essential criteria for all studies Criteria Introduction Rationale/ scientific background Objectives/ hypotheses Intervention model and theoretical considerations Methodology Study design Participants Sampling Setting and locations Comparator Item no. Description Score as: 1 – Found/ Met 0 – Not found/ not met 1 Scientific background and explanation of rationale 1 2 Specific objectives or hypotheses 1 3 Description of development and piloting of intervention and any theoretical support used to design the intervention (how the intervention is intended to bring about change in the intended outcomes) 1 4 Clear description and justification of chosen study design, especially if the design has been chosen based on a compromise between internal validity and the complexity and constraints of the research setting or research question 1 5 Clearly defined primary and secondary outcome measures to meet study objectives 1 6 Description of data collection methods, including training and level of data collection staff 1 7 Eligibility criteria for participants 1 8 Method of participant recruitment (eg. Referral, selfselection), including the sampling method if a systemic sampling plan was implemented 1 9 Method of participant allocation is clearly described 1 10 Information presents a clear amount of sampling strategy 1 11 Justification for sample size is reported 1 12 Settings and locations where the data were collected 1 13 Describes use of a comparison group from similar 1 33 Dr. Jaime Lee MPH Candidate, 2013 population with regard to socio-demographics or adjusts for confounding Data sources/ measurement 14 Describes the source of data for each variable of interest and detailed measurement criteria for the data 1 15 Enrollment: the numbers of participants screened for eligibility, found to be eligible or not eligible, declined to be enrolled, and enrolled in the study 1 16 Assignment: the number of participants assigned to a study condition and the number of participants who received each intervention 1 17 Analysis: the number of participants included in, or excluded from, the main analysis, by study condition 1 Recruitment 18 Dates defining the periods of recruitment and follow-up 1 Baseline data 19 Baseline demographic and clinical characteristics of participants in each study cohort 1 Fidelity 20 Degree to which the intervention is implemented as planned with a description of adherence, exposure, quality of delivery, participant responsiveness and program differentiation 1 Context 21 Description of the organizational, social, economic and political context in which the intervention is developed and operated 0 Attribution 22 The link between the intervention and outcome is reported 0 Bias 23 The risk of biases is reported 1 24 The risk of confounding is reported 0 25 Ethical and distributional issues are discussed 1 Discussion Summary of evidence 26 General interpretation of the results in the context of current evidence and current theory 1 Limitations 27 Discussion of study limitations, addressing sources of potential bias, imprecision, and, if relevant, multiplicity of analyses 1 Generalizability 28 Generalizability (external validity) of the study findings, taking into account the study population, the characteristics of the intervention, length of follow-up, incentives, compliance rates, and specific settings involved in the study and other contextual issues 1 Results Participants Ethical considerations 34 Dr. Jaime Lee MPH Candidate, 2013 Conclusions/ interpretation 29 Interpretation of the results, taking into account study hypotheses, sources of potential bias, imprecision of measures, and other limitations or weaknesses of the study 1 30 Discussion of the success of, and barriers to, scaling up the intervention 1 31 Discussion of research, programmatic or policy implications 1 32 33 Sources of funding and role of funders Relation of the study team towards the intervention being evaluated i.e. developers/sponsors of the intervention 1 0 Subtotal of Quality Points for Essential criteria for all studies (out of 33): 30 Other Funding Competing interests B. Essential criteria based on type of study - Must choose at least 1 of the following criteria to apply as appropriate: i. Quantitative ii. Qualitative iii. Economic evaluation Criteria Item no. i. Quantitative Statistical methods Outcomes and estimation Description 34 Statistical methods used to compare groups for primary and secondary outcomes 1 35 Methods for additional analyses, such as subgroup analyses and adjusted analyses 1 36 Methods of imputing or dealing with missing data 1 37 For each primary and secondary outcome, study findings are presented with for each study cohort, and the estimated effect size and confidence interval to indicate the precision 1 38 Estimate for random data variability and outliers are clearly stated 1 Subtotal for Quantitative study design (out of 5) ii. Qualitative Analytical methods Use of verification methods to demonstrate credibility Reflexivity of Score as: 1 – Found/ Met 0 – Not found/ not met 5 N/A 39 40 41 Analytical methods clearly described (In-depth description of analysis process, how categories/themes were derived) Discusses use of triangulation, member checking (respondent validation), search for negative cases, or other procedures Relationship of researcher/study participant has been 35 Dr. Jaime Lee MPH Candidate, 2013 account provided discussed, examining the researcher’s role, bias, or potential influence Subtotal for Qualitative study design (out of 3): iii. Economic evaluation N/A 42 Competing alternatives clearly described (e.g. costeffectiveness of zinc and ORS for treatment of diarrhea versus standard treatment with ORS alone) 43 The chosen analytic time horizon is reported 44 The perspective / viewpoints (e.g. societal, program, provider, user, etc.) of the analysis is clearly described 45 The alternatives being compared are clearly described 46 The sources of effectiveness estimates are clearly stated 47 Details of the design and results of the effectiveness study and/or methods for effect estimation are clearly stated 48 Methods for estimation of quantities and unit costs are described 49 Details of currency of price adjustments for inflation or currency conversion are given 50 Currency and price data are recorded 51 The choice of model used and the key parameters on which it is based are reported 52 The discount rate(s) are reported 53 Sensitivity analyses are reported 54 Incremental analyses are reported 55 Major outcomes are presented in a disaggregated as well as aggregated form Subtotal for Economic Evaluation (out of 14): Domain 2: Essential mHealth Criteria for all studies Criteria Item no. Infrastructure 56 Technology 57 Description Clearly presents the availability or kind of infrastructure to support technology operations (eg. electricity, access to power, connectivity) Describes the technology architecture including the Score as: 1 – Found/ Met 0 – Not found/ not met 0 0 36 Dr. Jaime Lee MPH Candidate, 2013 architecture Intervention 58 software and hardware mHealth intervention is clearly described with frequency and mode of delivery of intervention (i.e. SMS, face-toface, interactive voice response) for replication 1 59 Details of the content of the intervention are clearly described or link is presented and content is publically available 1 Usability 60 Clearly describes the ability of different user groups to successfully use the technology in a given context eg. literacy, computer/Internet literacy, ability to use device 1 User feedback 61 Describes user feedback about the intervention 1 Identifies constraints 62 mHealth solution states one or more constraints in the delivery of current service, intervention, process or product 1 Access and affordability 63 Presents data on the access and affordability of the mHealth solution from varying user perspectives 0 Cost assessment 64 Presents basic costs assessment of the mHealth intervention from varying perspectives 0 Training inputs 65 Clearly describes the training inputs for the adoption of the mHealth solution 0 Strengths and limitations 66 Clearly presents mHealth solution considerations, both strength and limitations, for delivery at scale 1 Language adaptability 67 Describes the adaptation, or not, of the solution to the local language 1 Replicability 68 Clearly presents the source code/screenshots/flowcharts of the algorithms/ examples of messages to ensure replicability 1 Data security 69 Describes the data security procedures/ confidentiality protocols Subtotal for mHealth criteria (out of 14): 0 8 37 Dr. Jaime Lee MPH Candidate, 2013 Scoring summary grid Maximum score Number of Quality points (If criteria were not applicable, circle N/A) 30 33 i. Quantitative 5 5 or N/A ii. Qualitative N/A 3 or N/A iii. Economic evaluation N/A 14 or N/A 35 38 (33+5) Domain 1: Reporting and Methodology A. Essential criteria for all studies B. Essential criteria based on type of study (choose at least 1 of the following options) Total number of Quality points for Reporting and methodology Quality score (# divided by Maximum total score x 100%) Strength of Evidence (<50% Weak, 50-75% Moderate, >75% Strong) 92% Strong Domain 2: Essential mHealth Criteria Total number of Quality points for Reporting and methodology Quality score (# divided by 13 x 100%) Strength of Evidence (<50% Weak, 50-75% Moderate, >75% Strong) 8 14 57% Moderate Synthesis of evidence Domain 2: Strength of evidence for mHealth criteria Weak Domain 1: Strength of evidence for Reporting and Methodology criteria Strong Weak Moderate Strong Moderate 1 38 Dr. Jaime Lee MPH Candidate, 2013 8. Reflection on the Capstone Project The experience that I gained from completing my Capstone project has been invaluable. This Capstone was also part of the work that I am doing with the JHU WHO mTAG team. Through this opportunity I have learnt about process and impact evaluation of mHealth interventions to guide national scale-up of mHealth strategies. Developing the grading tool taught me skills about grading the quality of information, and methodological rigor in study designs. It has also taught me to think critically about the principles required for an mHealth intervention to be sustainable. In my initial goals analysis, I had also wanted to gain skills in statistical analysis, economic evaluation and health finance and management. Whilst I did not learn or apply these skills in my Capstone, I completed those courses during my MPH year. Overall, the skills and experience that I gained from completing my Capstone and working with the JHU WHO mTAG team have exceeded my expectations and MPH goals, and I have thoroughly enjoyed this experience. 39
© Copyright 2026 Paperzz