Plan International UK M&E Framework DRAFT 30.05.2013 DRAFT M&E Framework Plan International UK Table of Content Table of Content ............................................................. 1 Introduction .................................................................... 2 Section 2: Evaluation Plan............................................. 32 Section 3: Monitoring Strategy ..................................... 48 Appendices.................................................................... 54 1 DRAFT M&E Framework Plan International UK Introduction Plan International UK has been asked by DFID and PWC to put together an M&E Framework specific to our GEC funded project and to incorporate the data required for this in to our reporting schedule. Our M&E Framework provides an overview of our approach and methodology for monitoring and evaluating our project Supporting Marginalised Girls in Sierra Leone to Complete Basic Education with Improved Learning Outcomes. This framework is presently in draft form and has been put together by Plan International UK and Plan Sierra Leone with support from the GEC consortium partners and our independent baseline research team. 2 DRAFT M&E Framework Plan International UK Section 1. Baseline Plan Plan International UK has contracted GK Consulting LLC to undertake independent and thorough baseline survey research ahead of the GEC funded project – Supporting Marginalised Girls in Sierra Leone to complete Basic Education with improved Learning Outcomes. The plan for undertaking this piece of work will be finalized after their pilot phase during April 8-12th. However to date, based on a) both GK and Plan’s experience of carrying out similar pieces of work b) the information provided by PWC to guide this process and c) the discussions we have been having with our local team and consortium partners, we have drafted the following plan: Background Plan has approved the following team structure for the purposes of the baseline survey. (CVs and experience summaries of the below can be provided on request, as can proof of our competitive process and rationale behind choosing GK Consulting): Gwendolyn Heaner, PhD: (GK Consulting) Baseline Team Leader Sybil Bailor, Chief of Party (Plan Sierra Leone) Plan Sierra Leone Team Lead Maggie Korde, GEC Programme Manager, (Plan UK) Plan UK Team Lead Nils Gardek (GK Consulting)– Quantitative Research Lead Samuel Byrne, Head Evaluation and Research Coordinator, (Plan Sierra Leone) Plan Technical Review Team Lead Jake Phelan, Learning, Impact and Assessment Officer (Plan UK) Plan Technical Review Support Stan Peabody, PhD (EEA)– Senior Technical Reviewer Carmen Aldinger, PhD (GK Consulting) – Child Education Specialist Sharron Kelliher, MA (GK Consulting)- Field Coordinator Marissa Van, MPhil (EEA)– Qualitative Research Analyst Our proposal to conduct a baseline study on girls’ education in Sierra Leone is developed with focus on our 4 main project objectives: Increasing access and retention in grades 5 and 6 of primary school and grades 7,8.9 in junior secondary school, and of children with disabilities in primary school Strengthening girls’ learning in grades 5 and 6 of primary school and junior secondary school Ensuring that girls are learning in an inclusive environment and protected from harm Ensuring girls’ voices and needs are listened to and responded to and that girls participate in decision-making concerning their education Education in Sierra Leone Sierra Leone faces great challenges with respect to education. One consequence of the war was destruction of schools: over 70% of all schools (1270 primary schools) had been destroyed (IMF, 2001); this led to almost 70% of school-age children being out of school. In 2001 Sierra Leone 3 DRAFT M&E Framework Plan International UK started implementing a policy of providing free primary education (FPE) for all pupils in governmentowned and government-assisted schools. During 2001-2005, the government committed almost 20% of its budget towards education; almost half was devoted to primary education (UNDP 2007); by 2005 there were at least 4300 primary schools. Each district received new funding for education and the number of students doubled from 0.6 million to 1.3 million between 2001 and 2004; in 2004 Sierra Leone had the highest primary school enrolment rates amongst all Sub-Saharan countries (World Bank 2007). However, statistics also show that girls, particularly those at JSS level in rural areas, have much lower rates of enrolment than boys; in addition, the disabled have strikingly lower rates of enrolment. Analysing data from before and after Sierra Leone’s 2001 implementation of free and compulsory primary education, Cannonier and Mocan (2012) found that “an exogenous increase in education triggers a change in attitudes that is empowering for women”, indicating that not only does education contribute to technical skills for girls’ to compete in the workforce, but it also gives them the confidence to do so. However, gaining access to and continuing education is not necessarily easy for girls. Maclure and Denov (2009) highlight the huge importance of educating girls, and call for more targeted interventions to affect girls’ education in Sierra Leone: “In post-war contexts, education is widely regarded as essential not only for civic reconciliation, but also as a key force for gender equity. In Sierra Leone, however, despite enhanced educational opportunities for girls, much of the emphasis on post-war educational reconstruction is unlikely to rectify gender inequities that remain entrenched within mainstream schooling and in the broader social context”. Given this, the need for the GEC project is clear and from that, the need for a solid baseline in order to measure the projects outcomes. Baseline Study Approach A. Sampling Strategy Qualitative and quantitative data will inform all facets of the project (i.e. not just data on targeted population [girls] to measure the program’s effects over time, but also data from those around them regarding their perceptions of targeted population in order to better understand the reasons behind girls’ low retention), but it is key to select representative communities/schools/respondents in order to obtain an accurate baseline. A good cluster sample will ensure that the size of the sample (n) that is randomly chosen from the sampling frame (or population) is closely representative of the population as a whole. Thus the baseline team will utilize 3-stages of cluster sampling with randomized selection and stratification, and ensuring some quotas are met, throughout the five districts. The first cluster’s sampling frame will be made up of villages in which there are intervention schools stratified by district; the second cluster’s sampling frame will be a list of the communities (village/town) that lie within the catchment area of those schools; the third cluster will be made up of the households within that particular community. The total number of communities and number of households 4 DRAFT M&E Framework Plan International UK surveyed in each depends upon the total population size of the intervention group, as detailed below. Sample size This sampling strategy starts by looking at the ‘ideal case’ for the sample size. This gives a total sample size of 5,280 girls. In this ‘ideal’ case, the treatment and control groups would be equal, 2,640 in each, and individuals would be randomly assigned. Calculations for this ideal case are set out in Appendix 1. There are a number of reasons why this ‘ideal case’ cannot be achieved: the project design does not allow for randomisation; it will not be possible to identify a substantial and appropriate control group; the project has multiple different interventions within it at different levels (school, individual and community based) and is too complex to break down into separate treatment arms; project resources are not sufficient for a survey on this scale and there are more efficient ways of approaching the issue. Accordingly, the proposed approach outlined below will split data collection into two separate datacollection exercises. A household level survey will be administered to measure changes in attitudes and perceptions at the household level. A school survey level will administer learning tests to girls to track learning of study subjects over time and provide data to gauge impact. Since expected changes in learning performance are significantly smaller than changes in attitudes and perceptions, separating the learning tests from the household survey allows a significant reduction of the latter sample. Instead of 5,280 households, the proposed case therefore suggests a household level sample of 1,440. This helps to significantly reduce survey costs while maintaining a minimum level of statistical power to assess the programme performance. Power calculations have been undertaken to inform the sample size of both the school and the household survey. Calculations for the MDES have been made for a power of .80 and a significance level of 0.5 using the following formula: MDES = (t(1-k)+ta) * sqrt(C/(P(1-P)J))*sqrt(1-C/(P(1-P)NJ)) Whereby: (t(1-k)+ta) = 2.8 C = Intra cluster correlation P = Portion of girls allocated to treatment J = Total number of clusters N = Data points within each cluster The household survey will be administered in a total of 1,440 households in 144 villages. This includes a control group of 240 households in 24 villages and 120 intervention villages. This control is not ideal but is constrained by a number factors (see section below on control villages). However, on certain outcome variables it will be effectively increased by additional control schools, which are significantly less problematic to identify (see below). 5 DRAFT M&E Framework Plan International UK Table 1.1 Household Survey Power Calculations Household survey effect size intra cluster correlation Multiplier Allocation ratio Total # clusters Intra-Cluster Sample Size Response rate sample size Treatment cluster control cluster MDES 0.3 0.2 2.8 0.83 144 9 90% 1440 120 24 0.34 The household survey will be administered in a total of 1,440 households in 144 villages. This includes a control group of 240 households in 24 villages and 120 intervention villages. For the household survey the parameters used are specified in the table below. The household sample size given allows the measure impact on the household level of 34%. The allocation between treatment and control does not follow a 50/50 split. The treatment sample comprises 90% of the total number of clusters. Following Bloom (1995) one can argue, however, that substantial deviation from the 50/50 treatment/control group mix does not significantly reduce the statistical power of an experiment. Nonetheless, the MDES is slightly higher than absolutely desirable given an expected effect size of around 30%. However, this is the best that can be achieved given constraints on control villages. The table above specifies the parameters chosen for calculating the survey sample size. Table 1.2 School Survey Power Calculations Primary School JSS effect size 0.2 effect size intra cluster 0.2 intra cluster correlation correlation Multiplier 2.8 Multiplier Allocation ratio 0.7 Allocation ratio Total # clusters 200 Total # clusters N 21 n Response rate 90% Response rate N 24 N sample size 4800 sample size Treatment cluster 140 Treatment cluster control cluster 60 control cluster MDES 0.210819 MDES 0.2 0.2 2.8 0.714286 84 21 90% 24 2016 60 24 0.329983 Combined effect size intra cluster correlation Multiplier Allocation ratio Total # clusters n Response rate N sample size Treatment cluster control cluster MDES 0.2 0.3 2.8 0.704225 284 21 90% 24 6816 200 84 0.210185 They key outcome indicator with regards to the school survey is learning. It is more efficient to administer the learning tests separately from the household survey, so this will be analyzed at the school and individual level. Learning assessments at school level will be carried out with a randomly selected number of girls per target grade (at P5, P6 and JSS1 for the baseline). These girls will form a 6 DRAFT M&E Framework Plan International UK cohort that will be tracked throughout the project. Re-contact details will be taken at baseline, and at subsequent waves information on girls exposure to specific elements of the project (such as bursaries, involvement in study groups, children’s clubs etc.) will be recorded. At Primary level, 140 treatment schools will be randomly selected and 60 control schools. Within each school, 12 girls per sampled grade (P5 and P6) will be tested, giving a total sample size of 4,800. At JSS, 60 treatment schools will be sampled and 24 control schools with 24 girls tested in the first target grade of the cohort (JSS1), giving 2,016 girls in total. Details of sample size calculations are given in the table 1.2. Oversampling within clusters is used in expectation of attrition within the cohort. The minimum detectable effect size at JSS is not ideal, but the gains from increasing the number of treatment clusters when it is not possible to increase the number of control clusters is minimal. Within each class, girls will be randomly selected by counting off the nth girl (where n equals the number of girls in the class divided by 12 (or 24 in JSS) and rounded down), starting at the girl nearest the door and working along each bench in turn. In cases where there are not enough girls in a grade in a sampled school, the team leader will keep track of the shortfall and randomly sample a larger number of girls in the next available school with more than 12 (or 24 girls for JSS) available to ensure the total sample size is met within the same number of clusters. In addition, all disabled children benefitting from CBR activities will be administered the test, a total of 2,125 girls and boys. The table below summarizes the number of girls who will be assessed at each wave in control and intervention groups. Table 1.3: Overall Sample Composition Intervention HH questionnaire 1,200 School based testing 4,800 Beneficiary testing 2,125 TOTAL 8,125 Control 240 2,016 0 2,256 Total 1,440 6,816 2,125 10,381 Using ASER tests will provide internal validity within the project (and control) over time, but not external validity. The additionally of learning outcomes will be measured through changes pre to post-test within control and intervention groups. At mid-term and end-line the learning assessment will be re-administered to the same project beneficiaries and the same girls interviewed through the household survey. Within control schools the same cohort of individual girls will also be followed throughout. Re-contact details will be recorded at baseline. If girls in the cohort drop out of school their replacements will be selected randomly (the sample allows for a 10% attrition rate). Total sample sizes for learning assessments will obviously be increased by the 1,440 girls assessed through the household survey but this is not factored into these calculations. The priority in this sampling strategy is put on measuring learning outcomes. The school level control will not provide a larger counterfactual for the household questionnaire administered at the village level. However, this is counter-balanced to some extent as key indicators measured through the household survey (such as changes in attitudes and behaviours, some based on binary ‘yes’ or ‘no’ questions) expect to see a larger effect size. 7 DRAFT M&E Framework Plan International UK Enrolment and retention will be measured in 100% of intervention schools and the 80 control schools. Attendance will be measured in a randomly selected sub-sample of intervention schools and the 80 control schools. This strategy is designed to measure one treatment arm only, that is, the cumulative effect of all project activities (see Evaluation strategy below). Given that it is not possible to achieve the ‘ideal case’ of treating disabled children as a sub-group (or separate treatment arm), an alternative approach will be taken. Enrolment of children with disabilities can be measured directly though school surveys for all control and intervention schools. A quota sample in the household survey will ensure a small but insignificant number of disabled girls are surveyed in control and intervention groups. To supplement these two measures, all children registered through CBR activities (2,125 boys and girls) will be administered the learning test and a very short questionnaire (questions from which will be the same as those used in the household survey that relate directly to key indicators). Although there will not be a robust control group for children with disabilities, changes in their learning, attitudes and perceptions, disaggregated by sex, will be measured pre and post-test and compared internally to girls in control and intervention groups. Table 1.4 summarises the main sources of data and sample sizes for key outcome variables. 8 DRAFT M&E Framework Plan International UK Table 1.4: Summary of Data Sources and Sample Composition Indicator Survey School Project beneficiaries Notes Intervention Control Intervention Control 1200 240 720 80 785 Intervention Schools comprised of 180 JSS and 540 PS (100% of target schools). Disaggregated for PS/JSS and disability Retention 720 80 785 Attendance 120 80 177 Intervention Schools comprised of 180 JSS and 540 PS (100% of target schools). Disaggregated for PS/JSS Spot checks in one school closest to the randomly sampled village (as per household survey) 4800 2016 8256 Enrolment Learning 1200 240 Individual attitudes and behaviours i.e. % of marginalised girls who feel their parents and community members support their education 1200 240 Disabled children (attitudes/perceptions of and learning outcomes) Unknown - Est. 10% of surveyed = 120 Unknown Est. 10% of surveyed = 24 Intervention Total 2,125 Control - Data collection consists of both household and school survey 1,440 Questions relating specifically to the attitudes and perceptions of children with disabilities (i.e. % of CwD who feel their parents and community members support their education) will also be asked alongside learning assessments during registration of CwD 2,269 Survey data will relate to girls only; project beneficiaries will be boys and girls 9 DRAFT M&E Framework Plan International UK Sample selection This sample selection process is based on the practicalities of the sample frame and village level data available in Sierra Leone (see section below) and the need to ensure all villages within the catchment areas of target schools have an equal chance of being randomly selected. In summary, sampling of intervention communities will be based on a multi-stage process as follows: 1. Randomly select x number of villages (where x is the total number of sample points required, and where all villages in the first stage are also the location of a target schools, PS or JSS), stratified by the number of schools per district and by PS and JSS within each district 2. Per randomly selected village, randomly select one village per school from a list of all villages that feed into the school, including the village where the school is located 3. Per randomly selected village, randomly select x number of households (where x is the sample size within one cluster) as outlined below Selection of sites in which to collect data (stage one cluster) A total of 144 sample points will be selected at the first stage in the target areas, which is a sufficient number to use in order to represent each strata (each level of school) according to the actual target population, as detailed above. Figures have been rounded up/down to the nearest whole number to obtain the given numbers in Table 1.2. Table 1.2 indicates the number of schools that will be targeted after the sample has been stratified relative to a) school level (JSS or PS) and b) the total GEC project intervention population of each targeted district by school level, so that the sample is representative of the GEC project intervention population as a whole (i.e. so that each district has an equal chance of being included in the sample, even though their project intervention populations may be greater/far less than others). Ensuring representation from all districts is important because it will allow for various features that may be different across districts such as ethnicity, languages spoken, different leadership (country and district-based), geography and relative socioeconomic status (all of which may play a role in the outcomes of GEC) to be represented in the sample and thus in the collected data. Table 1.5 below indicates the total number of communities that should be sampled based upon the above criteria. The total number of JSS intervention schools is 180, or 25% of the total. The numbers can be calculated for the PS once we have the total number selected in each district. 10 DRAFT M&E Framework Plan International UK Table 1.5: District-level stratification District # of intervent ion schools (JSS) # of interventio n schools (PS) % of interventio n schools per district total # of villages to be sample d per district # of treatme nt villages to be sampled # of control PS to be sample d per district # of control JSS to be sample d per district 31 # of control villages to be sampled per district* * 6 Port Loko 47 141 0.26 37 16 5 Kenema 27 81 0.15 22 18 4 9 3 Kono 29 87 Kailahun 34 102 0.16 23 19 4 10 3 0.19 27 23 5 11 4 Moyamba 43 129 0.24 35 29 6 14 5 Total 180 540 1 144 120 24 60 20 The number of control school sampling points per district may need to vary depending on actual number of available control villages per district but the total number will remain the same i.e. as there are no eligible control JSS in Moyamba more will be sampled from Port Loko instead (see selection of control villages and control schools below). Community (town/village) cluster – (Stage 2 cluster) After village/school clusters are selected at the first stage, numbered lists of the communities within the catchment area of that school will be used to randomly select one community to visit. The list of villages using is given as the second stage tab within the sampling frame (columns O to W list the villages to be randomly sampled at the second stage according to which row (first stage village/school has been sampled). Households/girls cluster – (Stage 3 cluster) Within each of the communities that have been selected, the adults residing in 10 households will be administered a questionnaire first. Households will be randomly chosen by going to the town center, spinning a bottle on the ground and walking in the direction of the bottle, then choosing every Xth house (X determined by total households in town/10) while walking in a straight line. A degree of quota sampling will be employed in this method in order to ensure that socio-economic variations within the communities are represented. In this case it may be necessary to over-sample particularly poor or particularly wealthy households in order to ensure they are represented. At a minimum, one household per community should be relatively better off than other households. Within each household in which eligible girls reside (ages 9-16) as targeted by the Plan Sierra Leone GEC project logframe, adult heads (the primary caregiver in each household) will be interviewed based upon convenience sampling (i.e. whether they are there and willing to talk when the team arrives). Quota sampling will ensure that there is an even split between male 30.05.2013 11 DRAFT M&E Framework Plan International UK and female caregiver interviews (i.e. if after 5 houses, only men have been contacted, the team will continue to every nth household until a female is contacted). In total we will obtain data from a minimum of 1440 households (10 in each community). We will also collect quantitative data from a total of 1,440 targeted girls aged 9-16 who reside in each randomly selected house that also has an eligible girl; thus, 10 girls in each of the communities will be surveyed. Also important is that girls’ with disabilities are represented in the sample, again obtained through quota sampling. Where possible, quotas should be met in each community according to the following table: Table 1.6 – Quotas to fulfill in quantitative surveys Persons to be Rural SL % Female SL interviewed M/F (assume rural and females urban % fewer) Female Caregiver Girls with disabilities 2.7 Child-headed households Quota of meet (per community) 2 2 1* Quotas within households will only relate to specific sub-groups (female caregivers, girls with disabilities, and child-headed households) and not related to socio-economic characteristics or educational/learning attainment. If there are no girls in a household who fulfil the quota requirements or the quota has already been fulfilled in that community, but there is more than one eligible girl, then the girl to be interviewed will be selected randomly. Pre-prepared numbered tokens numbered 1 to 9 will be pulled out of a bag (as a more obviously transparent random process for household members than a Kish grid). If a quota cannot be fulfilled in a village the process will carry on regardless and the quota will not be rolled over to any subsequent community. If there are no girls in the household who match the target age and/or do not fulfil quota requirements, the next Xth household will be chosen until twenty targeted girls are reached. If a household is reached and there are no primary caregivers available, then the next house should be visited (i.e. even if children are available, they must not be interviewed without contacting primary caregiver first). If the village boundary is reached in attempts to fulfil the quotas, then the team should return to the centre and walk exactly the opposite direction and continue as before until the desired sample is obtained. If there are not enough households in that community to fulfil the quotas, then another community within the school’s reach will be randomly selected. 30.05.2013 12 DRAFT M&E Framework Plan International UK School Data In each selected village/school selected at the first stage, school data will be obtained by visiting the school itself and finding key informants to provide data and/or directly measuring or crosschecking certain criteria (i.e. physical location, number of students present, size of school, etc.). Therefore we will collect 120 surveys with school data from intervention schools. In addition to these surveys, the team will conduct ‘spot check’ attendance counts on the days that the school is visited. If data collection occurs on a weekend or holiday, then another team member will travel to the school while it is in session in order to complete the spot check and collect any addition school data that was not possible during the household visits. Control In order to best measure the program’s effects, research will be conducted with control groups that will not receive the treatment of Plan’s programme, and will also be outside of the intervention area for similar programs run by DFID and BRAC. Using a quasi-experimental design (implementing the same pre-tests and questionnaires to both control and intervention girls throughout the various stages of data collection including baseline), we will collect data in communities with similar demographic, geographic, socioeconomic and cultural features to those of our intervention communities. These will ideally be within the five intervention districts, but this is dependent upon the project’s actual reach with JSS and PSS (i.e. if any of these schools will be outside the project’s reach). Reflecting the sampling strategy outlined above, there will be two distinct control groups: control villages (providing a counterfactual for the village level household survey) and control schools (for the school/project beneficiary level data collection). Given the constraints identifying appropriate controls in either groups, selection of controls will not be random. Constraints on identifying controls Finding sufficient control schools for the collection of school level data presents a challenge. Plan and partners will be working in 77% of all JSS in the five districts. In some districts, we have selected 100% of eligible JSS schools (where those not selected are either Private schools or Boy’s schools, and so not suitable for a control group). There will be slightly more eligible primary schools not selected in each district, but it is difficult to isolate the communities that feed into different schools. Given the nature of the inter-relationships of schools and communities, and the fact that children from one community may attend a number of different primary schools (in one district for example there is an average of 8 villages that feed into one primary school) and children from one primary school will attend different secondary schools (an average of 3 JSS that children from one primary school may go on to) it is difficult to find a community in the district that will not experience a high degree of spill-over and contamination. We have to date identified a small number of schools and associated communities that we feel are geographically distant enough from intervention areas, but there are not many to choose from. 30.05.2013 13 DRAFT M&E Framework Plan International UK Turning to districts outside the project target districts is one possible solution, but also has a number of issues. One is in the regional differences of the 5 target districts, with the Eastern region (comprised of 3 target districts) being geographically and economically quite different to other parts of the country, and Northern and Southern regions (the two remaining districts being one in each) socially and politically different. More importantly, the GEC project implemented by BRAC is in 13 districts, so all overlapping with Plan’s target districts. The DFID funded Education Consortium, a similar intervention to the GEC, will be targeting 7 districts. Details of exact areas of intervention for this project will not be available for some months but it too will limit possible control sites in other districts. In sum, having a larger number of control clusters is not a viable option financially, nor practically. The easiest way of identifying controls, reducing the number of schools and communities the project intervenes in, is not seen as an option. Nor do we believe that in a three year project it is appropriate to take a phased approach to establishing a control group and do not have time or resources to redesigning the project to achieve this and meet contractually agreed targets. Table 1.7: Location of girls’ education projects in Sierra Leone, 2013 Plan-led GEC Port Loko Moyamba Kono Kenema Kailahun Brac-led GEC Port Loko Moyamba Kono Kenema Kailahun Pujehun Tonkolili Koinadugu Bombali Freetown Kambia Bo IRC Education Consortium Kono Kenema Kailahun Pujehun Tonkili Koinadugu Bombali Freetown The only district not covered is Bonthe, a small costal district in the far south west of the country, relatively distinct from target districts. Approach to selecting controls To overcome the challenges of identifying suitable control clusters within Sierra Leone control sites will be selected at two levels: control schools and control communities. Control schools will be identified in intervention areas that meet the same selection criteria as intervention schools but which have not been selected for project activities. Numbers will vary between districts, but will total 80. These schools will be located in areas close to intervention schools, and at the Primary level will likely be feeder schools for intervention JSS. This makes them unsuitable as sites for the household survey, as at the household level there will likely be spill-over and 30.05.2013 14 DRAFT M&E Framework Plan International UK contamination with other schools located close by. Data collected from control schools will be on enrollment and learning outcomes only. This will allow us to increase the effective sample size for the control group relating to outcome indicators on enrollment, retention and learning, but not for indicators relating to changes in attitudes and behaviors at the household/community level. Control communities (which in some instances may overlap with control schools) will be communities where no effects of the intervention will be felt. Due to the challenges of identifying such areas, they will be smaller in number and will be purposively sampled to ensure, first and foremost, their separation from intervention areas and, to the greatest extent possible to be broadly matched with the socio-economic characteristics of intervention communities. In each control community, the same number of households will be surveyed as with intervention communities. Respondents will be randomly selected using the same method as that for the intervention group and clusters will be the same size and fulfill the same quotas; the same questionnaires and assessments will be used. While the data collected form control groups will not have the same reliability as that collected from intervention groups, because this approach selects representative respondents at random, it means that the results of each data unit gathered can be favourably compared between the experimental group and the control group. B. Quantitative tools: Questionnaires and interview guides Because the purpose of a baseline is to collect data that specifically measures the key indicators that will be tracked over time, the instruments designed will be those that give us information about these particular project outputs. (For any additional data collected because it seems interesting, it would make sense for mechanisms to be put in place in order to collect this data over time, in which case these additional indicators should be added to the log frame in relation to additional outcomes. At the least these indicators need to be more explicitly related to the outcomes as indicated in the log frame). A draft list of all quantitative and qualitative instruments can be found in Attachment 1. Exclusively measuring only those items that relate specifically to the log frame ensures that the actual procedure of baseline data collection is as succinct as possible, which further ensures the validity of the data collected. Any additional data collected simply out of interest for the time being will add to the time that each respondent gives to the research team and if these findings will not be used constructively, it is best to not ask the question in the first place. Validity is ensured with a succinct questionnaire, especially when working with children. Three different questionnaires will be used for collecting quantitative data: For adolescent girls; for household adults; for schools. For households and for girls, DFID questionnaires to which additional questions (not covered by this template) have been added specifically to inform the Sierra Leone GEC. These additional questions will be added throughout the DFID questionnaire. 30.05.2013 15 DRAFT M&E Framework Plan International UK For schools, only a Sierra Leone-specific survey will be used. All survey questions will be aligned with log frame indicators to allow midline and end line measurements of outcomes. The following components must be included in all questionnaires for girls and for adults: Identifiers: geographic area, household ID, date of interview Background characteristics: gender, age, socio-economic status, educational level, educational status, disability status (how the project defines disability is outlined in Attachment 6.) Outcome indicators – (to measure what the program intends to change, i.e. girls’ education) Impact indicators (measures of program’s achievements – for future questionnaires in particular, but at the baseline to determine whether any proposed activities/resources provided by the program already exist to any extent) The Sierra Leone-specific questionnaires will be relatively simple and short (as preliminary field testing and refinement will yield only those items that are useful for measuring our variables). They should be easily understood by those who will be interviewed (i.e. those as young as age 5 and those having no education at all). Consideration should be given to those ‘hard to measure’ indicators such as marginalization and thus measured indirectly using scales or indexes that have been previously validated based upon an agreed upon definition of that indicator. Once developed, the Sierra Leone-specific questionnaires will be pre-tested with appropriate respondents (i.e. individuals with similar traits but who will not be in the program) to ensure that questions are complete and accurate, that response options are relevant, exhaustive and mutually exclusive, and that respondents understand the questions being asked. A first-round test of the questionnaire can be conducted with existing contacts in Freetown, which will also be useful for training the field team; a second-round of testing will be conducted in a rural community during which all fieldwork activities will be performed. Any necessary refinements will be made prior to the beginning of data collection; the Team Leader will determine whether an additional round of testing and/or training is necessary prior to the beginning of data collection. Questionnaires for girls also include a learning assessment component to be completed by all girls who have been interviewed. The ASER test will be piloted in Sierra Leone and adapted accordingly so that it is appropriate in this context. Technology Because of the high volume of questionnaires that will be collected, the research team will use Asus Google Nexus 7 Tablets with EpiSurveyer or Magpi software. This serves to conserve paper, ensure that data is not lost/ruined by weather, that data can be uploaded every day, that manual data entry is minimal, and to allow the research team to remotely analyse data while collection is still on-going. 30.05.2013 16 DRAFT M&E Framework Plan International UK An IT consultant will be contacted in order to assist in programming finalized questionnaires into the device prior to launching, assist in uploading data to computers, and to ensure the survey’s integrity with the software. Data will be uploaded into SPSS for maintenance and analysis. Database Design and Data Processing In developing the database into which the gathered data will be uploaded, the IT consultant will ensure the structure and coding are fully compatible with earlier databases so that cross analyses can be easily achieved. The database developer in the research team must closely collaborate with all aspects of the statistical design and particularly the review and modification to the questionnaires. The database developer will work with Plan in-country staff to help them account for the longitudinal aspects of the project M & E activities so that changes can be clearly measured over time. The mechanisms for data capture and transfer to the databases has to be fully compatible with the database structure and needs of the end-users of the information contained within each. The databases should be in SPSS file format and would be accompanied by a basic guidelines handbook on how best to use them, which will require some basic training on how to use SPSS; this training will be a component of the field training prior to data collection. The team leader will provide further training after data collection has completed, if necessary. C. Qualitative Methodology Qualitative research will complement quantitative data and enable rich longitudinal studies for the duration of the project. There will be three main components of the qualitative research, which will be carried out in one community in each district (5 total communities). To ensure at least one JSS is represented, all JSS schools in the sample will be listed and one will be randomly selected; then each primary school in the sample will be listed according to district (excluding the district of the JSS), and one will randomly be selected. The aim of the qualitative data is explore the lived-experience of marginalized girls in Sierra Leone and to examine this in relation to educational attainment and specific interventions and changes the project intends to bring about. It seeks to expand upon the projects theory of change and begin a process of testing the underlying assumptions of the project and understanding of how and why change does or does not happen that will continue throughout the project. This qualitative research, as part of a mixed methods approach, forms a crucial part of the baseline. It aims to tell us not only if we have met our targets but why change has or has not happened. Similarly, as part of a mixed methods approach, the qualitative data will inform and strengthen quantitative data. Qualitative research will begin before the household survey so as to identify areas of particular concern to marginalized girls in Sierra Leone, which in turn can be used to strengthen the survey to ensure we are measuring – and learn from – what is important to those we are primarily accountable to, as well as what is relevant for our logframe. 30.05.2013 17 DRAFT M&E Framework Plan International UK Qualitative research is of particular importance in relation to Output 4 of the project. The impact of complex empowerment and accountability processes are not easily, or even always usefully, solely measured through quantitative measures. The case studies based on process tracing methodology, alongside focus group discussions, will offer much more appropriate data on the impact of this part of the project (see for example McGee and Shutt, CDI Practice Paper, 2013). Key informant interviews Purposive sampling will be used to identify key informants in five communities who will provide comprehensive and ‘insider’ information surrounding girls’ education. The interviews will be semi-structured (guided by a key informant questionnaire, but conducted with a trained facilitator who will probe when necessary). The interview should last no longer than thirty minutes. Examples of key informants include: o School teachers (male and female) o Head teachers (male and female) o Other school stakeholders (board members, etc.) o Other employees of school (counselor, maintenance) o Members of parent-teacher associations o Local representatives of civil society who deal with children’s/women’s rights o Children with disabilities (age 10-16) Focus group research: The field team will also conduct focus group research in one community per district (5 total communities, the same as those selected for FGDs) with five groups of people as indicated below (30 total focus group discussions). o Girls in school (age 10-16) o Girls out of school (age 10-16) o Boys (age 8-16) o Female adults o Male adults The community that will be selected for a focus group in each district will be selected randomly (numbering each community and then randomly selecting one number out of n communities per district). Attempts will also be made to select a random sample of participants for each focus group; this will be done in a similar manner as that for the household survey: a bottle will be spun in the centre of town and every nth household will be chosen, from which one family member (the first encountered that fits the requirements of any focus group) will become part of the focus group. The process will continue until 6-8 representatives from each group are present. It is also important, in addition to this process, to ensure that various minorities are included in the groups (because they have a lesser chance of having been chosen at random), 30.05.2013 18 DRAFT M&E Framework Plan International UK especially those who are disabled or especially wealthy/poor so purposive sampling will need to be used when necessary. The focus group will involve a combination of close-ended questions (responses collected by eyes closed/hands up method and tallied on a flip chart for participants to see) followed by discussion surrounding the results, and also through a series of open-ended questions. This focus group guide will have been previously developed using with attention to relevant questions that are both within a logical sequence and will also generate conversation, as the goal is to get the group to talk more than the facilitator. The facilitators will be trained to follow the script to an extent, but also to probe further when necessary or to let conversation continue if it is particularly rich. Visual aids will be utilized to help focus group participants visualize certain questions (i.e. color-coded time use charts, tally marks under each response for closeended questions, flow/causal diagrams, pie charts, timelines, ladders of importance). Each focus group will last 1.5 hours long, 2 at a time, in separate venues). A trained facilitator will conduct each; a trained note taker will take exhaustive notes for the duration. Case study In each focus group/ KI community in addition to other communities where the opportunity clearly presents itself, one case study will be conducted. This will involve identifying a person, family, place or event that directly relates to the theme of the study, and will be explored indepth using process-tracing methodology. This involves attempting to highlight the particular event(s) that contributed to some reported outcome. Certain case studies will be useful as longitudinal studies that can be explored over the duration of the project and will help in elucidating other subjects for potential longitudinal data collection. The team leader will make the final decision on what constitutes a worthwhile and appropriate case study (virtually as needed) . All of the above activities will be repeated in both of the control groups. Thus we will have collected qualitative data from the following: Key informants – 5 per community * 5 communities = 20 from intervention groups Focus group participants – (60 per site) * (5 intervention sites) = 300 from intervention groups Case studies – 5 from intervention groups 30.05.2013 19 DRAFT M&E Framework Plan International UK 20 Deliverables and Activities Tentative Work Plan: The timeline given below has been updated but remains provisional. It will be updated once the sampling strategy is agreed with the EM and the baseline can go ahead. Table 1.8: Baseline Data Collection Work Plan Date(s) 23 March 25 - 26 March 26 - 27 March 27 March 27-31 March 04 April 7 April 10-11 April Task virtually meet with project manager and relevant other staff for initial briefing and discussion of tor review project documentation; write evaluation protocol and workplan; discuss with plan review existing data collection tools from plan projects and adapt begin recruitment of local innumerators (virtually) if needed. identify suitable pilot site close to freetown design and review data collection instruments and questionnaires with input from plan staff; test instruments with contacts on the ground in liberia and sierra leone virtual meeting with plan to review and revise sampling strategy; selection of pilot site discuss and coordinate research methods with partners and gec evaluation manager for final agreement international travel days; early arrival in country and firming up logistics; preparing for training events Key Personnel G Heaner and N gardek Deliverable G Heaner and N gardek evaluation protocol and workplan for baseline G Heaner and N gardek inception report and draft instruments M Chea; G Heaner resumes for potential staff G Heaner; N Gardek; C Aldinger draft questionnaires and interview GUIDES; all questionnaires, all qualitative tools for each target group, data entry processes for plan’s review RECCOMENDATOINS ON strategy FROM em G Heaner; N Gardek G Heaner; N Gardek beginning of DISCUSSIONS and revisions to m&e strategy G Heaner; M Van Epp; M Chea 30.05.2013 DRAFT M&E Framework Plan International UK 11 April 12 April 13 April 14-17 April 18 -19 April 20-22 April 22 April 22 April – 14 May 22 April – 14 May 15 May – 22 June team meeting; kick off meeting with local office; device update w/ surveys and testing; meeting with it consultant field team training in methodology and data entry/upload with it consultant first pilot (full) in moyamba for training purposes and final refinement of instruments additional training and pilots as needed; refinement of instruments and sampling strategy qualitative data collectionin moyamba and port loko further recruitment and training for additoinal enumerators and team leaders; finalization of qualitative and quantitative instruments; final upload to tablets refinement of baseline methodology and m&e framework with pwc/coffey/dfid further traiing, recruitment and prepartion for additional tasks for household surveys and learning assessments data collection with support from staff – oversee the start up of PROJECT. field coordinator transition to local team leaders 21 G Heaner; N Gardek (virtual); M van Epp; M Chea G Heaner; M van Epp; M Chea with local field team G Heaner; M van Epp with local field team G Heaner; N Gardek (virtual); M van Epp; M Chea training and capacity building for data collection and entry; guidance document finalized instruments to be used for duration of data collection; uploaded into devices finalized qualitative instruments G Heaner; M van Epp; M Chea with local field team G Heaner; S Kelliher; M Chea; S Tolley G Heaner finalized instruments G Heaner; Plan staff approved m&e framework G Heaner, S Kelliher, Plan staff, local field teams G Heaner; S Kelliher; M Chea; S Tolley; local field team progress reports from field coodinator (by phone) every other day to team leader; uploaded to computer nightly; and internet when possible but minimum 2x per district 30.05.2013 DRAFT M&E Framework Plan International UK 03– 21 June 05– 21 June 15– 25 June 25– 30 June 26– 10 July 05– 15 July 15– 2 August 31 July – 01 August 2 August 12– 16 August data collection for school data and learning assessments in kono and kenema data collection for school data and learning assessmnets in kailahun, port loko and moyamba project tl returns for final baseline data collection / upload with assistance from local field team; transcription of field notes finalized; data entry for cover sheets, learning assessments, paper questionnaires data entry for school survey data from all districts data cleaning; analysis of household baseline data; preparation of summary data for household survey data cleaning; analysis of learning assessment and school data preparation of reports: draft evaluation report; m & E monitoring strategy ethical and child protection protocols; child friendly summary of report submission of draft report to plan uk review of draft reports by plan 22 S Kelliher, local field teams with remote support from G Heaner Plan Sierra Leone staff supported by Plan UK G Heaner; local field team, local data entry assistants all raw household data uploaded for analysis Plan Sierra Leone, local data entry assistants, remote support from g heaner N Gardek; G Heaner; M Van Epp all raw school survey data uploaded for analysis raw data and summary data for key project logframe indicators shared with plan N Gardek; C Aldinger N Gardek; G Heaner; S Peabody; C Aldinger C Aldinger; G protocols and child Heaner; M Van Epp friendly summary G Heaner, Plan UK draft report on 2 august including m&e monitoring strategy, school data analysis, child friendly version of report plan uk comments on draft reports 30.05.2013 DRAFT M&E Framework Plan International UK 16– 20 August 20– 31 August As Needed consideration of plan comments and finalization of report final review of all reports by plan G Heaner, C Aldinger, S Peabody Plan UK verbal presentation of findings in workshop/meeting with plan (virtual or in person possible) G Heaner 23 submission of gk final report by 20 august final report and all supporting materials submitted by plan uk presentation of findings Fieldwork Logistics and Quality Assurance A field team of local and international researchers will conduct the baseline study. Three days of preparation in Freetown will be needed in order to refine methodology, train local field teams, hold meetings with the local Plan office, and prepare devices for data entry. Following that, the team aims to visit a minimum of one site per day to conduct research activities. All efforts will be made in order to ensure validity of data collection while completing all quantitative OR qualitative research in that community during one day. If it appears, following the first days of data collection, that it is impossible to complete all activities in a single day, then more enumerators will be located and hired such that this can be done, given the short timeline for data collection (maximum of 45 days including 1 day off each week as indicated in the work plan). If it appears that data collection can be done with 4-6 enumerators in one day, then the team leader will explore the possibility of completing two communities in one day, either back to back, or splitting the team in half, always ensuring that there is one supervisory team member present to provide on-the-ground. In order to ensure quality in the baseline data collection, the research team will be made up of suitable and well-trained individuals; at every stage in the data collection, at least one senior team member will be present to provide quality assurance and oversee field logistics. One qualified and experienced Sierra Leonean will fulfill this role for at least one week of the data collection upon being trained by the international staff; the Team Leader will return to finalize the data collection and ensure quality in the upload of data. Draft field activities are provided in Appendix 1 (Will be adapted accordingly when sites are selected). Quality assurance of data collection and entry Certain measures will be taken in order to assure quality of the data being collected by enumerators who will, at times, be working unsupervised. We will ensure that not only are enumerators following procedures properly in terms of randomly selecting houses and respondents, strongly following ethical guidelines, and not rushing through questionnaires/making up answers just to get them finished, but also when conducting interviews, are accurately phrasing certain questions that are less straightforward and require 30.05.2013 DRAFT M&E Framework Plan International UK explanation, and correctly coding/recording those answers when responses may not be provided verbatim on the questionnaire answer choices, and thus require a bit more thought. To ensure that data collectors' accuracy in phrasing questions and coding/recording answers, each field team will have a team leader who, while he/she will be collecting data himself/herself, will also periodically visit households in which interviews are being conducted, and observe the enumerator at work, looking over his/her shoulder. This will occur early on in the data collection (and has been ongoing during training activities) to ensure that each enumerator understands exactly how to ask each question, and proper ways in which to probe if that is necessary. Throughout the duration of data collection, 'spot checks' will be conducted in order to observe this as well. If it is found that an enumerator is repeatedly misunderstanding answers or questions, then he/she will be re-trained until corrected and, failing that, replaced. The in-country Field Coordinator, Sharron Kelliher, has good experience training and working with field teams, and has been working with the enumerators from the start. She will be present during the launch and also for the first half of data collection activities with select teams, providing on-the-ground QA. The local Team Leaders (one for each field team) will be present in the villages at the beginning of each field activity in order to ensure that households are being selected randomly, and that certain households are not simply passed-by for a short interview because they do not fulfill some requirement (i.e. girls aged 9-16). This will be assured because prior to interviews, the TL and team will select all houses that will be interviewed, and tape the 'cover sheet' on the door with the name of the first contact in that household, such that the team knows that it is a house to be interviewed during the day. Additional houses will be randomly selected during this time that will be visited in the event that one or more of the selected houses did not have a girl aged 9-16 to interview, or if nobody was present/willing to be interviewed throughout the entire day. At the end of the day, cover sheets will be collected by the TL and matched with the data entered on the devices; where the names do not match, the TL will follow up. Thus it is expected that more than ten cover sheets will be collected at the end of each day because it is likely that at least one of those households will not have a primary caregiver and/or a girl of eligible age present for interview on that day. If teams repeatedly claim that all persons were present in the first ten houses selected, then it will be necessary to investigate further to confirm (i.e. to ensure they've not moved the cover letter to a different house where they knew persons were present). It will also be important to ensure that enumerators are not entering false data and/or rushing through questions such that respondents cannot answer appropriately. In most cases, TLs will not be present to regulate this while in the field; however, it will be their duty every evening when backing up the data to their computers to look at a few questions that will flag instances where data collectors may not have asked questions appropriately. One example of this will be the question to the girl that asks, 'How often are you happy' and the question immediately after 'How often are you unhappy'. An answer that indicates 'always' for both, for example, will suggest that either the data collector did not ask the respondent the question at all, or that 30.05.2013 24 DRAFT M&E Framework Plan International UK he/she was not careful in explaining it such that we obtain valid data. Other questions are strategically repeated throughout the questionnaire, using the exact same wording, so as to confirm that a prior answer matches and again, to ensure data collectors are actually speaking to people to collect data. If the TL notices such answers following a day of data collection, he/she will personally visit that household the following day to find that person who was claimed to be interviewed, and determine whether he/she was interviewed at all and how long the interview took. If he/she was interviewed for the amount of time the interview usually should take, then the interview will be repeated by the TL and answers will be matched. If the answers are different for certain factual questions (i.e. not dealing with perceptions or good/bad, etc), then the previous enumerator will be talked to and possibly released. Also, every week, raw data will be emailed to the Evaluation TL, Gwen Heaner for detailed review. Because each enumerator's name is linked to his/her completed questionnaire, this will enable her to observe certain areas where answers don't make sense and need to be followed up on. This will provide a level of oversight to the TLs as well, who, while GK Consulting has worked with them before with excellent results, will likely benefit from another set of eyes on the numerous questionnaires coming in. Because the teams are using electronic devices to enter data and then upload that data through a hard-wired connection to a computer, we can be certain that the data as it will be exported into SPSS will be correctly entered. Therefore, we will not need to do any double-entry for those questionnaires completed using devices. In the event that paper questionnaires are completed (a contingency plan for each field team), then 5% of those will be double-entered. If we find numerous errors on each questionnaire (+3%) in this method, then all paper questionnaires will be double-entered. Finally, in some cases it will be possible for back checking of data randomly, through having a different enumerator re-visit a previously surveyed household to conduct the interview again, and then match the answers to those of the previous enumerator. While this will be difficult to do often, given the constraints on time that the teams are already facing, whenever a team or individual finishes the targeted number of respondents in a day, then he/she will be tasked by the TL to return to one previous community nearby, to randomly select one of the households from the 10 that were interviewed within each, and to conduct the interview again, in full. Again, because time is limited, and because we have other backstopping mechanisms in place (above), this will not be a frequent occurrence. At this point, we are setting a goal of conducting random back checking in 2% of households per district (29 households in total). As this will occur while data collection is ongoing, any instances where back checking demonstrates inconsistency, then the responsible data collector will be released. 30.05.2013 25 DRAFT M&E Framework Plan International UK Recruiting and Training a Suitable Team Recruitment The local field team will be made of full-time consultants who will be present for the duration of the project in addition to temporary day-hire research assistants (enumerators/interpreters). The full-time team will consist of individuals who are able to work as enumerators and also are able to assist in key informant interviews and focus group discussions. In order to assure respondents will feel comfortable speaking to the research team, both genders must be represented: - Two facilitators, one male and one female, who will also act as enumerators - Two note takers, one male and one female, who will also act as enumerators. Key criteria to consider in recruitment of field team members are: fluency in both English and Krio, in addition to skills in at least one of the regional languages (Temne, Mende and/or Kono) (written and spoken); experience conducting field research including delivering face-to-face questionnaires with people of limited literacy; experience taking part in focus group discussions or at a minimum having leadership roles during group meetings; age (it is critical to include at least one young female to talk to the younger girls who may feel uncomfortable speaking to adults and/or men); basic computer literacy (to assist in uploading data and/or data entry if necessary). CVs are being collected at this stage and there is a strong pool of candidates from which we can choose; final decisions will be made prior to arrival in Freetown. Additional day-hire research assistants (4-6, depending on how long the questionnaires take to administer during the pilot) will be recruited at the district level and will be high school graduates at a minimum, fluent in English, Krio and the regional language, and able to translate questionnaires from English into the appropriate dialect of the respondent, and then record answers back in English. They will be familiar with electronic devices, field research, and dealing with children and adults. Training Training will take place over one half-day in the office and half-day in nearby communities, and will continue in the field during the first rural pilot. In the office, the research team will be briefed on the basics of the project and the goals of the baseline study, and then will be introduced to the research instruments one by one. A ‘run through’ of all survey questions, interview questions and focus group questions will take place, allowing each member of the team to comment on strengths and weaknesses of certain questions in case they need to be removed, rephrased or elaborated upon. Then the local research team will conduct ‘mock’ data collection activities using the international research team as subjects. The international team will give the local team a number of challenging responses in order to enable them to learn how to deal with issues they may face in the field. Special attention will be given to the ethical guidelines that every member 30.05.2013 26 DRAFT M&E Framework Plan International UK of the research team is required to follow and also which they will have to convey and oversee when working with the district-based research assistants. Finally, the local research team will be trained as to the best way to select key informants and identify case studies for further research. While in the field the Team Leader will be responsible for making the final decision via phone or in person when possible, the local team is expected to report to the coordinator their ideas. Pilot Two pilot activities will take place in communities with schools relatively close to the capital (so Plan staff can visit and international staff travel less in their shorter time in-country) during which all field activities will take place (focus group research, household data collection, school data collection, key informant interviews, case studies). The first will be at one targeted rural site on the way to Moyamba or in Moyamba (depending on selected school sites); all collected data will be thrown out. This will serve as both an extended training for the enumerators and also to enable the international research team to observe how questions are received and answered by targeted respondents. After this pilot, further training that will be delivered as needed; the research team will then complete further refinement of questionnaires and interview guides. The second pilot will employ the finalized instruments and will take place in one of the randomly selected sites in Moyamba. If Plan project staff would like to observe the activities, then the site closest to Freetown (Moyamba) can be visited first. While we intend for the second pilot to yield valid data collection, until we are sure that the collected data is reliable and accurate, data will be thrown out and more schools will be randomly selected to obtain the desired sample size as detailed below. Data entry / maintenance while in the field It is expected that throughout fieldwork, quantitative data will be uploaded and qualitative data will be manually entered (if devices are not used) every night after fieldwork activities by the local research team and field coordinator, assuming electricity is available to power computer(s). This includes all quantitative and qualitative (i.e. typed transcripts of interviews / focus group discussions). The field coordinator will be responsible for ensuring this happens and, when possible (though it is not assumed that this will be frequently possible while in the field) uploading the data to an online database that the team leader can access. The data will be reviewed as it comes in for any obvious errors and dealt with accordingly. In the event that it seems the research team requires more guidance, the Team Leader will return early to ensure activities go according to plan and data collected is valid. The field coordinator will also be responsible for keeping track of all paper items through the duration of fieldwork. Whatever data is not entered after fieldwork has been completed will be entered from Freetown so that the team leader can access all data prior to the Team Leader’s. A 30.05.2013 27 DRAFT M&E Framework Plan International UK one-week cushion has been factored into the schedule to allow extra time for data entry or data collection given the inevitable logistical challenges faced during such work. Research ethics plan In any research that deals with vulnerable or marginalized populations, it is imperative to pay close attention to the potential to do harm through asking questions or eliciting conversation. The research team will follow ethical guidelines as prescribed by the British Sociological Association. The team will be acutely aware of the need to obtain vocal informed consent from EVERY participant in the study; each participant will be told that his/her name will not be used in any project documents without their explicit permission; pseudonyms will be used in any narratives, and that we will give a unique ID to each participant, which will be linked to their names but those names will be kept separately in a password protected document. The importance of us keeping track of their individual progress will be explained to them such that they understand its import and can give informed consent. If, at any point in a conversation, it appears that the participant no longer wants to speak, then it is imperative that the researcher a) can identify this easily and b) stop the research immediately. The participant should never be coerced to take part in the first place or to ‘keep answering’ while taking part. Certain questions might deal with sensitive topics; therefore it is important that the researcher, while obtaining informed consent, explain the types of questions that will be asked on the survey, and assure the participant that a) his/her answers will remain totally anonymous (if it will be) and b) that he/she can choose to not answer a question if he/she chooses; c) he/she can stop the interview at any point without question. It will also be important that only females conduct interviews with girls. The team will consist of males and females who will conduct household questionnaires; however, for the interviews with girls, one or two females on the team will be conducting only interviews with girls which are much shorter; therefore, they can visit houses specifically for these interviews. We would also like to stress that we expect DFID and PWC to uphold these ethics as a minimum standard when they are undertaking their own baseline in country. Since they will be considered a ‘project associate’ of Plan’s during this time, it will also be required that Plan’s Child Protection Policy, Attachment 5, is signed by anyone visiting our project sites and communities for the purposes of this project. Data Analysis Data will be analysed using SPSS and disaggregated according to, at a minimum, gender, age, ethnicity, district, degree of isolation from district centre, education level, number of children, disability status, school enrolment, participation in children’s clubs, feelings of support from family and community, exam pass rates, learning assessment results, teachers with sufficient 30.05.2013 28 DRAFT M&E Framework Plan International UK resources, participation in study groups, perception of violence in school, incidence of mistreatment from teacher, hours spent studying, grade girls’ expect to complete, perception of importance of school, and girls’ degree of marginalization, [using marginalization index that indicates parents level of education, disability status, occupation, house type, expendable income]. The marginalization index has been developed in order to gauge girls’ relative level of marginalization, and to track how their opportunities change over time (i.e. not whether they become less marginalized, as this will be recorded at the baseline, but how they are able to have ‘life chances’ despite this marginalization. Likely variables for regression analyses will consist of that are indicated in the logframe and based on tentative intermediate outcomes are: Disability status vs. school enrolment Gender vs. school enrolment Gender vs. school success Gender vs. supportive parents Gender vs. confidence in school to ask questions Gender vs. importance of school Gender vs. grade expected to complete Gender vs. perception of violence in school Marginalisation status vs. school enrolment Marginalisation status vs. supportive parents Marginalisation status vs. school success Mistreatment in school vs. feeling comfortable to report it Disability status vs. enrolment Disability status vs. supportive parents Disability status vs. school success Disability status vs. perception of violence in school Disability status vs. grade expected to complete Disability status vs. importance of school Disability status vs. confidence to ask questions in school Perception of violence in school vs. school success Perception of violence in school vs. enrolment Supportive parents vs. enrolment Supportive parents vs. grade expected to complete Supportive parents vs. school success Participation in study groups vs. school success Degree of isolation vs. school enrolment Degree of isolation vs. supportive parents Degree of isolation vs. good teachers Marginalisation status vs. grade expected to complete Marginalisation status vs. importance of school 30.05.2013 29 DRAFT M&E Framework Plan International UK This list is by no means exhaustive bit is intended to provide an indication of some of the likely areas of significant interest. Qualitative data analysis Further variables for regression analyses will be informed by qualitative data collection (ie when it becomes clear that certain variables may be linked). Qualitative analysis will also become important in understanding any trends that are found in regression analyses, or in case expected trends are not found. In these cases, narratives surrounding these relationships will be explored in order to add rich data to the findings, and provide more nuances to the reported correlations. These findings will also be used to verify reported trends that will ensure that the quantitative data collected is valid. In the event that qualitative data and quantitative data do not agree, this will be presented in the final reports and flagged for follow-up. Other qualitative findings will be coded and logged, and analysed for trends. This is particularly useful when dealing with certain hard to measure indicators (marginalization, goals, confidence), which require a more significant level of explanation and subsequent discussion in order to get valid data. Close-ended questions will also be analysed in order to highlight trends and associate those findings with rich narratives. Finally, qualitative data will be used to highlight significant stories of change for girls, parents, teachers and others; these individuals will be followed over the duration of the project for follow-up interviews in order to provide more data into how their lives are changing, and why. These stories will be compared with quantitative data from those same individuals in order to confirm that findings are valid. Final report outline The final report will be outlined with a dual purpose: a) to provide all the data that is important for the specific log frame indicators that will be measured over the duration of the project; b)to provide background information surrounding the differences in certain project sites, so as to better inform project staff for programme planning. The report will be thus first include a section on the literature surrounding girls’ education globally, and specific information surrounding Sierra Leone. Then the report will describe the methodology for the baseline and the problems encountered and thus any limitations in the data. Next the report will highlight the major correlations that were found that relate specifically to project log frame indicators and enrich the discussion with qualitative findings, including relevant case studies. Then the report will highlight any trends that were not found (but were expected to) and explain reasons based on qualitative data. The report will then discuss any discrepancies within the data. In the second major section of the report, other relevant trends that were found will be highlighted that may be useful information for project planning purposes (i.e. not specifically related to the baseline and/or log frame as it presently exists). Finally, the report will outline broad recommendations for 30.05.2013 30 DRAFT M&E Framework Plan International UK measuring these indicators over the duration of the project, including the mid-term and final evaluations. A comprehensive appendix will provide further tables that may be of interest to project staff. 30.05.2013 31 DRAFT M&E Framework Plan International UK Section 2: Evaluation Plan a. Key Evaluation Questions Project monitoring and evaluation will seek to answer the GEC programme evaluation questions, as well as those posed by Plan’s Programme Accountability and Learning System and Evaluation Standards. To what extent was the GEC successfully designed and implemented? What works to increase the enrolment, retention and learning of marginalised girls? What impact did the GEC funding have on the retention and learning of marginalised girls? To what extent did the GEC represent good value for money? How sustainable were the activities funded by the GEC and was the programme successful in leveraging additional interest and investment? In addition to using GEC questions and OECD-DAC evaluation criteria, the following are crosscutting questions for Plan’s evaluations: Child-centeredness – to what extent were children involved in the project, how were they selected, what was the impact on boys and girls of their participation in the project and how did the project affect girls and boys, directly or indirectly, positively or negatively? Non-discrimination and inclusion – who benefited from the project and who was excluded, and why? How were marginalised/ vulnerable groups included? Gender – to what extent did the project contribute to increased equality between boys and girls, women and men? To what extent was the project gender transformative? During the first quarter of the project Plan and partners will also be looking at a broader learning strategy for the project with specific research questions. b. Logframe Indicators It is recognised that the project logframe needs revising to ensure the appropriate level of outcome and output indicators. This will take place together with further revisions to indicators, targets and assumptions following the baseline (and within 30 days of the baseline report being submitted). The project will measure the additionally of the intervention through comparing treatment groups with control groups. This is discussed in section 1. Outcome level indicators are dealt with in more detail below: Retention: Data on enrolment and repeaters per grade for the 2012/13 academic year and the 2011/12 year will be collected at baseline for project and control schools, allowing calculation of retention. This will be repeated annually at the beginning of each academic year. Data will be collected directly from schools by project staff (as data available from District Education Offices 30.05.2013 32 DRAFT M&E Framework Plan International UK varies in reliability) using the School Selection and Survey Form (attached) and stored on a central school-level EMIS database. Thus baseline figures for enrollment will be set from numbers of girls and boys per grade enrolled in all target schools in the 2012/13 academic year. The same data will also be collected for the 2011/12 academic year, along with numbers of male and female repeaters per grade in 2012/13 allowing calculations of retention. Data on the numbers of children with disabilities in all project schools will also be calculated for the 2012/13 academic year as part of the baseline, disaggregated by sex and disability type though not by grade. More detailed information on children with disabilities enrolled in schools will be collected through HI’s monitoring activities. The same data as above will be collected for all intervention and control schools. The additionally of the project on numbers of girls enrolled and retention, and number of children with disabilities enrolled, will be measured through analyzing changes pre to post-test in intervention schools against pre to post-test changes in control schools. National level EMIS data will provide a further reference point for analyzing changes in enrollment. Attendance: Attendance data was not collected during school selection as requiring substantially more time per school, and historical attendance data is not always available from schools, but will be the focus of various follow-up activities and the baseline survey. Attendance data will be collected from the household questionnaire and questions on estimates of the number of days of school children have missed (in control and target groups). In the five schools randomly sampled for in-depth qualitative research (and in control schools) detailed attendance records will be compiled from school registers, giving an average for all classes in targeted ranges. Spot checks will be carried out on a regular, rotating basis by project field staff. Headcounts of girls and boys in school on a given day will be compared with the number of students officially enrolled to give a snap-shot of attendance. The independent baseline survey team will also carry out spot-checks in all sampled treatment and control schools at baseline. As with all other indicators, the baseline process will be repeated at mid and end-line allowing the additonality of the project’s impact in terms of attendance to be measured in terms of pre to post-test changes in control and intervention schools to be reported on at each milestone. Learning: The project will collect data on exam pass rates (NPSE and BECE) in all intervention and control schools. Such data is only available for the previous academic year and it is recognised that this is not a sufficient measure on its own. In the current absence of EGRA and EGMA tests adapted and validated to Sierra Leone, the project will adapt an ASER test for literacy and numeracy based on the age-appropriate school curricula. This will be piloted as part of the baseline methodology and tools pilot. This test will be administered to all individual children surveyed through the household questionnaire (control and intervention groups) and additional project beneficiaries. The learning assessment for literacy and numeracy is attached. These literacy and numeracy tests were developed by Plan for the GEC in Sierra Leone. They have been informed by 30.05.2013 33 DRAFT M&E Framework Plan International UK international assessments such as the ASER and Uwezo and aim to assess girls’ basic literacy and numeracy skills. They are based on the national curriculum and national literacy and numeracy targets have been taken into consideration when designing the tools. Finally, the design has been informed by an initial pilot with primary school girls in Sierra Leone. The learning assessment will be carried out with girls as part of the household survey and through a random selection of girls in target grades in control and intervention schools, as detailed under the baseline section above. All girls surveyed will form a cohort tracked over time. In addition, all disabled children benefitting from CBR activities will be administered the test, a total of 2,125 girls and boys. Table 1.3 on page 6 summarises the number of girls who will be assessed at each wave in control and intervention groups. Using ASER tests will provide internal validity within the project (and control) over time, but not external validity. The additionallity of learning outcomes will be measured through changes pre to post-test within control and intervention groups. At mid-term and end-line the learning assessment will be re-administered to the same project beneficiaries and the same girls interviewed through the household survey. Within control schools the same cohort of individual girls will also be followed throughout. Re-contact details will be recorded at baseline. If girls in the cohort drop out of school their replacements will be selected randomly, assuming there are sufficient girls still enrolled in the grade (the average number of girls enrolled in selected schools in JSS2 is 34, so at JSS level a high proportion of all girls per grade will be tested). Leverage / sustainability: Specific indicators to be measured regarding structural and systematic changes are outlined in the logframe indicators below. Output 4 in particular (the school score-card) is designed to both bring about and to measure changes in school and district level policies and practices. In summary, the project will assess: Changes in parental / community attitudes to girls education and the value of educating disabled children, evidenced through household surveys pre and post-test in control and intervention communities and qualitative research Changes in the knowledge, attitudes and behavious or teachers and school authorities in relation to gender responsive pedagogy and inclusive education, evidenced through preand post-training questionnaires and triangulated with feedback from girls in schools (through survey, qualitative research and score-cards) Changes in the attitudes of other key duty-bearers (local Chiefs, District Education Officers et.), evidenced through key informant interviews pre and post-test Changes in the existence and implementation of Teacher Code of Conducts and in particular their impact on violence and gender based violence within schools, evidenced through number of schools with a CoC in place triangulated with girls’ feedback and interviews with duty-bearers Changes in, and extent of implementation of, local by-laws at the Chiefdom and District levels, evidenced through local policy mapping to be carried out in first year of project, process tracing, and score-carding activities and follow-up 30.05.2013 34 DRAFT M&E Framework Plan International UK Changing community and individual behaviour and attitudes: This will be measured by tracking the number of local by-laws created and implemented to hold teachers accountable to the teacher Code of Conduct. Additionally, individual and community attitudes and behaviour will be measured through household questionnaires pre and post-test (and in control) and by looking at the number of girls and children with disabilities accessing and being retained in schools. In addition, focus groups at the community level will assess changing attitudes to girls' education, violence and gender equality. Focus group discussions and KIIs with members of parent advocacy groups and board of governors at baseline and end line stages will be specifically designed to assess how attitudes and behaviours have changed through the life cycle of the project. The data collected by the project will enable us to explore changing attitudes of girls towards their own education (and towards gender equity) through questions around their own aspirations and their perceptions of older girls and young women in their communities. Similarly for parents and teachers of girls, by looking at their assessment of girls' futures at the start and end of the future (per se and compared to their assessment of boys' futures) we will be able to comment on the difference our interventions have made towards attitudes. Influencing policy makers: The score-carding and interface meetings process will highlight issues of concern at the school level, district and national level. The Project will document what issues have been acted upon at the district and national levels and who has been involved in the process. The project will also document what interventions are having a positive impact on marginalised girls' learning outcomes. For example, we will track the progress of the learning assistants through the process of attending distance education courses and can thus demonstrate the effectiveness of this as a model for getting more female teachers into the education system. Where the project has demonstrated success, we will advocate for these interventions to be replicated and will track any uptake thus measuring outcome indicator 4. Measuring changes in relation to disabled children Definitions of disability as used by the project are given in Attachment 6. The project will measuring changes in relation to disabled children and the effectiveness of the CBR approach on a number of fronts. Through the household questionnaire, data will be collected on the attitudes of men, women and girls (including some girls with disabilities) to the value of education children with disabilities. This will be strongly supported by the qualitative research. Focus group discussions include questions around attitudes to children with disabilities, their education, and the systems of support available for them. A focus group will also be held specifically with girls with disabilities. While the number of disabled girls interviewed in the household survey will likely be relatively low and will not give a statistically significant sample as a sub-group on its own, relevant questions, and the learning assessment, will be administered to all disabled girls and boys supported by the project (through CBR activities) pre and post-test. There will not be a significantly large control group specifically for children with disabilities, but changes in disabled 30.05.2013 35 DRAFT M&E Framework Plan International UK girl’s learning will be compared to other girls in control and treatment groups allowing a form of ‘internal’ difference-in-difference analysis of changes. The numbers of girls and boys enrolled in primary and JSS (control and intervention schools) will also be measured at baseline through the school survey and annually thereafter. Output Indicators We acknowledge the phrasing of the outputs themselves is perhaps overly ambitious for output level and in relation to the predetermined Outcome. We will consider revising these as below after further discussion with all project partners and prior to the final submission of the logframe. - Output 1: Parental and community attitudes support girls education and financial and disability-related barriers to marginalized girls’ education reduced - Output 2: Girls have positive female role models and are supported to build the skills needed for life - Output 3: School staff and management supported to provide safe and inclusive learning environments - Output 4: Girls voices and needs are listened to and responded to in educational decision-making Plan has tried to achieve a balance of logframe indicators covering direct outputs of the project with indicators measuring changes in attitudes. Plan has also developed a set of draft ‘intermediate outcomes’ for the project, which we see as joining the direct deliverables (measurable annually and not just at base/mid and end line if reliant on the survey data) and the predetermined outcome indicators. These intermediate outcomes do not fit easily within the logframe format as given but are areas we are interested in measuring. Not all of these are suitable for logframe indicators. These changes to the logframe have not been discussed or agreed with all partner organizations and are in no way final. A final logframe will be submitted on 16th August 2013. All relevant indicators will be disaggregated by Primary and Secondary levels. Relevant indicators will, where possible, be disaggregated by disability status. This will not be possible through the household survey alone as to measure this sub-group is beyond the resources of the project. However, the same exact questions will be asked of all children with disabilities engaged in the project through the CBR activity, increasing the effective sample. The control will not be significant for this sub-group, but changes can still be referenced to the target population as a whole. Table 1.9: The Programme Logframe Additional Output and Intermediate Outcomes for Output 1 Indicator Source Baseline Mean # of hours girls spend learning/studying Household Survey - Milestones/ Targets +30% [Estimate based on assumed Notes Added to logframe output 1 to replace current output 30.05.2013 36 DRAFT M&E Framework Plan International UK outside of school during term-time. % of girls receiving bursaries who complete an academic year baseline figure of 5 plus 2 hours of children’s clubs giving a 40% increase for participants] 95% of girls receiving a bursary completing the academic year -30% [bursaries received by approx. 40% of target girls but cannot assume all surveyed HH will benefit] indicator on NPSE and BECE pass rates Will be disaggregated by disability (same questions to be asked of CwD referred through CBR approach) Project Management Data (individual tracking of all bursaryholders annually) Household survey [C98d] 0 % of girls expressing positive attitudes to benefit/right to education Household survey [exact indicator to be refined when survey instrument finalized] - +30% % of parents who take an active interest in their child’s learning / in the performance of their local school % of marginalised girls who feel their parents and community members support their education Household Survey - +20% # of CWD enrolled in target Primary schools Project Monitoring Data (as relating to CBR activities) and School Survey Data School Survey Data of all project and control schools % parents/care-givers with girls of schoolgoing age (appropriate for P5, P6 and JSS1) citing financial barriers as a reason why their girl child is not in school % of CWD as a proportion of school population in PS - Added to logframe. Will be disaggregated by girls in/out of school prior to bursary Added to logframe Current logframe indicator. Will be disaggregated by disability status (see note on disability above) - 2,052 Added to logframe [1.4% from one district] No target Suggest to monitor but not use as indicator as increase in enrolment of nondisabled girls as well as CWD so ambiguous interpretation] Milestones/ Notes Additional Output and Intermediate Outcomes for Output 2 Indicator Source Baseline 30.05.2013 37 DRAFT M&E Framework Plan International UK % of Las passing WACE or TCC entrance exam Project Monitoring Data - % of schools (PS) with at least one female teacher (qualified or unqualified) School Survey data % of girls who believe they have a positive female role-model in their school or community # of girls with a supportive female role model in their school or community Household Survey [needs to be added] [approx.. 45% of schools with one or more female teachers in one GEC district] - % of girls who believe that going to school is important for their future opportunities in life Household survey Targets 80% [LAs in 180 of 540 PS so target change diluted] Added to logframe Project Monitoring Data Relating to mentees and girls benefitting from LAs – note double counting if including both. If just mentees same as current output indicator Likely to be high at baseline, so difficult to see large changes. - Additional Output and Intermediate Outcomes for Output 3 Indicator Source Baseline Milestones/ Targets % of girls agreeing they have been 'mistreated physically, sexually or insulted by a teacher' % of girls (aged 10 or above) who believe that sexual abuse/harassment is never appropriate Household Survey [composite indicator from questions relating to GBV] % of girls (by PS and JSS age) with knowledge of sexual Household Survey [similar questions in PPA2 baseline gave 95% of girls agreeing that sexual abuse not appropriate] [in PPA2 baseline, Added to logframe +20% Notes Currently in logframe – may be removed. Overlap with indicator 3.2 and difficulty in interpreting changes in reported rates of violence. Replaced by indicator below May be added to logframe indicator (output 3) but very low baseline rate makes significant change unlikely +30% 30.05.2013 38 DRAFT M&E Framework Plan International UK and reproductive health and rights (i.e. correct answers to questions about pregnancy or condom use) approx. average of 35% of girls able to give correct answers to questions about SRH knowledge] - % of school board members trained demonstrating improved knowledge of teachers code of conduct Pre and post training questionnaires (inc. questions on appropriate reporting mechanisms for GBC) % of teachers trained Pre and post able to demonstrate training understanding of questionnaires gender responsive (inc. questions on pedagogy and self-reported teachers’ code of classroom conduct practice and GRP techniques) Additional Output and Intermediate Outcomes for Output 4 Indicator Source Baseline % of schools using the score-card with positive change on at least one indicator as a result of students’ voices and actions # of changes in school/local by-laws, education policy or policy implementation in support of girls’ education and/or children with disabilities +60% Replacing indicator 3.3 +60% Replacing indicator 3.4. Notes Not easy to establish targets based on robust data but will be monitored regardless Project Monitoring data and scorecard reporting 0% Milestones/ Targets 50% Qualitative reporting through score-card process, interface meetings, case studies and project reports - - To replace logframe indicator 4.1 c. Evaluation Approach The evaluation will take a quasi-experimental approach, with a control group (as described in Section 1) to allow for difference-in-difference analysis as well as looking at pre- and post-test changes in key outcome indicators. A longitudinal survey will be employed to track changes of individuals surveyed. Future waves will return to the same respondent using the re-contact data from the household questionnaire, GPS coordinates and written descriptions. However, a degree of respondent attrition is inevitable and this has been factored in to the sample size. Plan is committed to the independent evaluation of the project and the transparent and active dissemination of the final results. 30.05.2013 39 DRAFT M&E Framework Plan International UK The final evaluation will be carried out by an external evaluator(s) independent of all partner agencies. The evaluation strategy will be informed by the GEC programme-wide evaluation questions, Plan’s Child Centred Community Development (CCCD) approach and Programme Accountability and Learning System (PALS). In essence, the external consultant will be asked to evaluate how far Plan and partners have gone towards achieving the outcomes and objectives, and to assess the programme in line with the principles of PALS and OECD-DAC criteria. The evaluation will seek to question the assumptions and specific outcome areas of the project in terms of increased and inclusive access and enrolment, improved quality of learning, safety within schools, increased voice of girls and strengthened accountability of school systems. It will evaluate the intervention strategies proposed assessing: (a) whether access was supported through changing attitudes to girls’ education, violence and gender equality and targeted financial support; (b) whether girls retention and learning was supported by mentoring, the presence of learning assistants in classrooms and inclusive, gender-responsive teaching; (c) whether girls safety in school improved and influenced retention and learning outcomes and (d) whether school performance and accountability was influenced by the scorecard process and school, chiefdom, and district level advocacy. In evaluating the influence of project interventions, the evaluations will consider whether the project represents value for money in terms of Economy, Efficiency, Effectiveness and Equity. It will not be possible to consider multiple treatment arms within the impact assessment i.e. to assess the impact of activities relating to disabled children separately from the project as a whole. There are too many possible ‘treatment arms’ (including: bursaries; mentoring; study groups; children clubs; CBR; teacher training; learning assistants; and school score-cards) to quantifiably asses the relative impact of each in relation to each other and the control. On a purely quantitative basis (i.e. data generated through the household survey) the final impact assessment will assess the impact of the project as a whole (where impact is the change pre to post test in relation to the control group, based on a difference-in-difference approach). That is not to say that individual elements of the project will not be assessed, but that the data sources will be more varied. Project monitoring data relating to girls receiving bursaries and children with disabilities enrolled through the CBR approach etc., learning assessments with specific beneficiary groups will provide a strong basis for assessing individual strands of work within the project. In-depth qualitative work (focus groups, key informant interviews, and process-tracing case studies) will further elucidate this data to more fully understand the results of different activities individually as well as the synergies between them. 30.05.2013 40 DRAFT M&E Framework Plan International UK d. Research Methodology Research will be conducted in line with internal guidance on Child Protection and ethical standards. An ethical protocol has been drafted by project partners to ensure a shared commitment and understanding of ethical and child protection considerations specific to the project. The external final evaluator will have a large amount of existing data to draw on, and will be asked to review and assess the quality of data provided, synthesis learning and use it to assess outcome level changes in the target group over time in relation to their level of marginalization, compared with internal differences in intervention exposure, the control group and national level secondary data. Data sources from all MER activities that contribute to the final evaluation research methodology are detailed in the project logframe on ‘Tab 8: MER Data Sources.’ Specific to the final evaluation will be an independent, representative, quantitative communitysurvey and in-depth qualitative research in schools. Primary data collection for the final evaluation will be collected by independent data enumerators, not by project staff. Quantitative household survey from a representative sample of marginalized girls: The quantitative community survey will use the same representative sampling strategy as the baseline and will ensure a 95% confidence level with a ±5% interval. This survey will return to the same communities as sampled in the baseline, though will not explicitly seek to interview the same respondents, instead using a random sampling method to select households within the communities. Certain questions on community attitudes to girls’ education and gender norms, for example, will be the same as in the baseline questionnaire to ensure consistency and provide updated data for logframe indicators. The survey will measure the degree of exposure to the intervention and results analyzed in terms of the vulnerability status of respondents and relative outcome level changes between beneficiaries and non-beneficiaries. Qualitative research data: As at baseline, quantitative survey data will be accompanied by qualitative research, though shifting focus from identifying barriers to education to examine stakeholders’ experiences of the intervention, and intended and unintended, positive and negative outcomes. Qualitative research will include: KIIs with community and religious leaders, district education officials and other duty-bearers regardless of their participation in project activities (and with head teachers of control schools), to examine changes in community attitudes and the education policy context and the extent to which these are attributable to the intervention and/or other political or social changes. KIIs and FGDs will also be held with a larger sample of stakeholders involved in the project. At the community level (in communities already sampled for the qualitative research at baseline), FGDs will be held with parents, girls and boys in separate FGDs. These will be focused on providing contextual understanding in relation to the quantitative data to understand why, or why not, changes in community attitudes and gender norms have come about. The qualitative element of the community survey will not aim to be 30.05.2013 41 DRAFT M&E Framework Plan International UK representative of the total population of the target districts but, as at baseline, will involve a purposive sample for maximum variation against key characteristics. At the school level, research will focus on specific groups of project beneficiaries already involved in baseline and mid-term. The final evaluation will thus follow-up on with the same cohort of groups and so will include girls who have received bursaries (and separate FGDs with their parents); girls with disabilities; members of study groups and children’s clubs; interviews with random samples of Learning Assistants and teachers who have received in-service training, and FGDs with teachers at project schools; sampled Boards of Governors and Parent Advocacy Groups. In addition to the primary data collection, the evaluation will also pull together learning from annual programme reviews (APPRs) and all school and individual beneficiary data, as well as aggregate results from participatory monitoring tools. e. Sampling Framework Details of the sampling strategy to be used at baseline, and repeated at endline, are given in Section 1. Target groups will be tracked over time through the household questionnaire, administered to project beneficiaries, non-beneficiaries, and a control group at baseline and at endline. Project monitoring data specific to girls receiving bursaries, the community based rehabilitation approach, learning assistants and other beneficiary groups will allow for further tracking of individuals over the course of the project. A note on the sampling framework submitted: The sampling frame submitted is complete for all five districts. Population data is, by and large, available only for communities in which schools are physically located, not all feeder communities which are often small villages for which data is not available from the 2004 census (the basis for population data used in the sampling frame). As sample size and stratification is not dependent on community population (stratification is at the first stage of sampling based on numbers of target schools per district) lack of population data does not affect the sampling strategy at this stage. Listings of all feeder communities (for the second stage sampling) are currently within a different data set to the sampling frame and it is proposed to keep these lists separate (so the sampling frame will be used for the first stage sampling, then for selected schools, the second stage sampling will be based on a separate list of schools and their feeder communities). Further, due to small size of some feeder communities that may be randomly selected, more than one community may end up being sampled in some instances to reach the number of households per cluster. In such cases, additional communities will also be randomly selected. Construction of the Sampling Frame The sample frame has been developed from two separate sources of data: a list of all schools selected by Plan for the GEC project and the communities they are in (and some nonintervention schools and communities); and a list of communities with population data from the 30.05.2013 42 DRAFT M&E Framework Plan International UK 2004 Census. Neither data set can be considered wholly complete nor accurate but are the best that is available in Sierra Leone. Selection of schools and control The list of schools was obtained from the school selection and survey process. In one district (Port Loko) more schools were surveyed then were ultimately selected for inclusion in the project as intervention schools. This provides some information on similar schools (that meet the same eligibility criteria) which are not going to be intervention schools, some of which have been identified as suitable control clusters. However, this is only the case for Primary Schools as there are no other eligible JSS in the district (the case in 3 out of the 5 target districts). In Moyamba district, all surveyed schools have been selected. Within Plan’s areas of operation there was little choice as to eligible schools and so field staff did not need to survey schools prior to selection (though all selected schools have been surveyed). This is why there are fewer JSS than planned in Moyamba (43 instead of 45), and thus fewer Primary Schools (as three PS have been selected per JSS), the shortfall being made up from an increase in the number of selected schools in Port Loko. Please see Section 1 above for a discussion of the issue of control clusters. Once control sites have been selected and the sampling strategy agreed, Plan will double-check all control sites to ensure minimal contamination from project areas and investigate more fully to see if other external education initiatives may have a bearing on the controls. Geographic data No data is given for the detailed location or directions to communities in the sample frame. Distance to each community (or rather the school associated with a community) from the District center is given in the school data by way of indication of the ‘remoteness’ of the school/community. Latitude and Longitude are provided but there are a number of errors and no guarantee is given as to the accuracy of GIS data provided at this stage. Geographical data will not be completed for all communities in the sampling frame at this stage as this is not seen to be necessary for sampling nor an efficient use of resources. When sampling points have been randomly selected (for Plan schools and for the EM’s sample), further details for those points only will be added to the sampling frame. Population data Population data given for each community is based on data from the 2004 census. An annual change has been estimated from World Bank data available up to 2011 giving a total increase of 21% in population estimates per community. The estimated number of children aged 5 to 15 has been calculated based on DHS 2008 data giving the percentage of children in the 5 to 14 age range as 33%. 30.05.2013 43 DRAFT M&E Framework Plan International UK 44 Population data and data relating to EAs is not available for all communities. Community names are taken from the school selection survey, which verified the location and name of all communities where there are target schools (and in some cases non-target schools). Not all of these communities can be identified on the list of communities from the 2004 Census. There are a number of possible reasons for this, including: some selected communities are small and may not have been captured on the Census (or have changed since 2004); variations in names and spellings over time and between official records and names used locally. It should also be noted that the Census often contains duplicate listings for the same town/village where a single locale is split across multiple EAs. In constructing the sampling framework, only one EA has been used to give population data (the first listed for a given community, where the first is the lowest numbered EA Geocode). Thus, population estimates drawn from the sampling framework for the total population of all listed communities is likely an under-estimate. As the sampling framework does not include all feeder communities (but only those in which schools are located) it gives a further under-estimate of the total population within all project schools. If such data is needed it would be better sourced from the relevant Section or Chiefdom sub-totals. Data on the size of schools is available for all project schools. f. Value for Money Strategy During our start up workshop in January, Plan with all consortium partners spent time working on the concept of ‘value for money’ and how we could best apply this throughout the project. The presentation for this session is attached in Attachment 3 and a very simple Value for Money Tool like the table below was drafted, and is being used by every partner, so that every expenditure can be analysed against the ‘4 Es’. Ref No. Which E? Decision being made Quantative Unit where possible (money, people, time …) How was the decision made? Result Comparison 1 2 3 4 5 6 7 8 9 10 11 12 30.05.2013 DRAFT M&E Framework Plan International UK g. Evaluation Governance All evaluation processes will be overseen by an Evaluation Steering Group which was established at our M&E Workshop in-country in March. The Steering Group will comprise of the Chief of Party, the CCU M&E Coordinator, the Plan SLE M&E Manager, a representative from each partner organisation, a representative from the Ministry of Education, the Ministry of Social Welfare Gender Children’s Affairs, and district representatives locally. The Plan UK Learning and Impact Officer and GEC Programme Manager will be on the group remotely via telephone or skype. In order to have an independent member on the group, we have invited Gwen Heaner, our baseline and hopefully mid line and final evaluation lead, to be on the group. The Plan UK members of the group joined with the Learning and Impact Assessment Manager and the Education Advisor to form a Baseline Tender Committee in the UK to manage the baseline recruitment process. Going forward the Evaluation Steering Group will have the following ToR (as agreed at the above mentioned M&E Workshop in March during a session dedicated to this issue): ToR for Evaluation Steering Group Frequency of Meetings: Quarterly Membership: Chief of Party, CCU M&E Coordinator, Plan SLE M&E Manager, IRC M&E Lead, FAWE M&E Lead, HI M&E Lead, Pikin to Pikin M&E Lead, GK Consulting Director, Ministry Representatives from Ministries of Education, Social Welfare Gender Children’s Affairs, and local government district representatives. (The Plan UK GEC Programme Manager and Learning and Impact Officer will use phone or skype to take part in each meeting) Tasks: Learning from partners’ experiences, sharing good practice Discussing how to disseminate good practice Ensuring child protection and ethical standards in all M&E activities are understood and adhered to Supporting dissemination of M&E findings with project stakeholders and beneficiaries in appropriate formats Organising learning events Identify capacity needs in M&E and suggesting ways of addressing gaps Drafting ToRs for external research/evaluations etc Reviewing project performance information and making recommendations to resolve under-performance Sharing information on key M&E processes across different activities / districts Facilitating communications and information sharing It was also decided that the GEC project partners should have a Learning Plan. Going forward the Evaluation Steering Group will have the following Learning Plan as a minimum: 30.05.2013 45 DRAFT M&E Framework Plan International UK Quarterly reviews (through evaluation committee) Learning events: bi-annually, lasting 1 to 2 days each. First one to be end of August / start of September* *A bi-annual reflection and learning event will be held with all project stakeholders. The first of these is to be scheduled for late-August, to provide a chance to look back at the work in schools in the first academic year and training carried out over the school holidays, to share experiences and to feed learning into activities scheduled for the next academic year. The reflection and learning meetings will be the responsibility of the CCU M&E Manager supported by the Evaluation Steering Group, who will feed in to the agenda and planning. External stakeholders will be invited such as relevant government officials and staff from the other GEC project in Sierra Leone implemented by BRAC. A small number of teachers, mentors etc. also involved in the project will also be invited (possibly from one district at a time on a rotating basis – so it may be best if the event is held in the district). Representatives of girls and boys involved in the project (perhaps members from one children’s club, again changing on a rotating basis) will also be invited, as long as their meaningful and safe participation can be assured and is adequately planned for. h. Performance Management Framework The Chief of Party and M&E Coordinator will be responsible for collating quarterly performance reports which will be the basis for reports submitted to DFID and progress will be reported through quarterly Project Strategic Group meetings to examine data from all monitoring activities and ensure this is used appropriately to shape project implementation. Annual participatory programme meetings will be held inviting district and national level MEST representatives, donors, UNICEF, NGOs and local stakeholders. These meetings will be used to collectively review and validate annual performance data, as well as share and address project implementation challenges and successes. The project’s design also emphasises shared learning at project level. Community engagement and participation are critical to many project activities and specific budgets have been allocated for sharing events relating to children’s clubs at chiefdom level, chiefdom and district-level interface meetings, and CBR best practice exchange workshops in every district. The project will document the processes and findings of the scorecard tool used, Women into Teaching component and CBR approach to be shared with local and national level stakeholders. Learning from evaluations, and all other data sources, will be actively shared with a view to improving the design and implementation of future girls’ education programmes of Plan and all other partners. Nationally, evidence from the project will be shared with MEST, UNICEF, other NGOs and education networks. Internationally, learning from the project will be used to inform Plan International’s Because I am a Girl campaign, aimed at reaching four million girls globally. All partners will share learning through their own technical networks and will actively seek 30.05.2013 46 DRAFT M&E Framework Plan International UK opportunities for disseminating learning to others in the sector. All relevant project details will be made publicly available in line with Plan UK’s published commitment on IATI. 30.05.2013 47 DRAFT M&E Framework Plan International UK Section 3: Monitoring Strategy a. Data Collection Strategy A detailed Monitoring and Evaluation plan was jointly drafted by all project partners, setting out what data is to be collected, when and by whom. This document is still in draft, while standardised data collection forms are being developed for each and every activity. The basis for the monitoring plan is the detailed implementation plan developed by the project and the project level logframe output indicators. Per output indicator, the project has established key activities to be monitored consistently, and ‘intermediate outcomes,’ for example, employing simple pre and post training questionnaires with teachers to monitor changes in knowledge and attitudes as a direct result of the training component. When finalised, the complete M&E Plan and all data collection tools will be shared with the EM. All data collection tools and approaches are to be jointly agreed by all partners, with standardised formats and pre-agreed definitions. Responsibility for collecting data using these formats is for the agency leading a particular activity in a given district, with oversight by the agency providing technical support for the activity, the relevant M&E Officer in the district and ultimately the project M&E Manager. Output 1.1: Number of bursaries provided to marginalised girls Output 1.2: Number of children's clubs that are established and active Output 2.2: Number of girls participating in girls' study groups Output 2.3: Number of marginalised girls supported with Mentoring Output 2.4: Number of Learning Assistants passing TC entrance exam Output 3.3: Number of board members trained on teacher codes of conduct and the value of girls' education Output 3.4: Number of school staff trained on inclusive education, gender responsive pedagogy and Code of Conduct Output 4.1: Number of JSS schools that have implemented child-led score-carding Output 4.2: Number of events/ interface meetings where members of children's clubs get to discuss issues of concern Key data collection M&E events are presented in the table below. 30.05.2013 48 DRAFT M&E Framework Plan International UK 1 Month J Academic Year Project Milestones 2 F 3 4 M A Inception Period 5 6 M J 7 J O N D J F M A Project review and learning event Beneficiar School y selection Y2 enrolm & assessme ent data nt A M J J A S O N School Year 2013/14 Milestone 1: End of School Year 2013/14 Key Reports Reports 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 S School Year 2012/13 Term 3 Baseline M&E Activities 8 A 49 Y1 exam data B D J F M A M J J A S O N D School Year 2014/15 2015/16 Term 1 Milestone 2: End of School Year 2014/15 Milestone 3: End of Project Mid-term Project review and learning event evaluation Y3 enrolm ent data Y2 exam data 37 J Final impact assessment Y4 enrolm ent data Y3 exam data C D A: Baseline Report B: Revised Logframe C: Mid Line Evaluation Report D: Final Evaluation Report b. Management Information System Data on all project activities will be collected using standardised tools, formats and procedures. Where appropriate for the most data-intensive activities, this will be aided by using digital data collection devices, otherwise data from paper-based tools will be entered into Excel in the first instance at the district level. Excel will be used to collate and store all monitoring data, with databases and coding established at the central level (with the exception of survey data – see Section 1). All project data will share the same coding structures for geographic and other identification information, aiding integration of different activity data-sets. Excel is the most appropriate format given the common level of IT skills at the field level, and allowing data to be exported into other formats or software for further analysis and reporting at the central level as required. Data sharing protocols are shaped by Plan’s Child Protection Policy and the project’s ethical and CP standards. From the latter, the following are of particular relevance: Respect confidentiality and anonymity: Identity of individual participants must be protected, records and reports should be kept confidential, and no responses or quotes should be attributed to an individual without their prior written consent. Data protection: all data (including written notes and reports, digitised data, photographs video and audio recordings) must be handled in such a way as to ensure the confidentiality and anonymity of children is maintained at all times unless permission is given for its use. Sensitive data must be securely stored (secure filing cabinet, password protected electronic files etc.). Photographs and video: We will always ask permission from children (or, in the case of young children, their parent or guardian) before taking images (e.g. photographs, videos) of them. Respect their decision to say no to an image being taken. Ensure that any images taken of children are respectful (For example: children should have adequate clothing that covers up the sexual organs. Images of children in sexually suggestive poses or that in any 30.05.2013 DRAFT M&E Framework Plan International UK way impact negatively on their dignity or privacy are not acceptable). Stories and images of children should be based on the child’s best interest. Monitoring data will be shared between project partners on a regular monthly and quarterly basis based on agreed templates and reporting lines. The Central Coordination Unit will be the primary repository of all monitoring data, responsible for collating information across activities and districts. All partner organisations will be required to pro-actively share data with the CCU along agreed reporting lines. The CCU will be the primary broker on information sharing and knowledge management between partners and external stakeholders. No project specific information or data relating to individual beneficiaries will be released to a third party without the prior agreement of the CCU. Key data sets will also be securely stored on Plan UK’s servers to provide data back-up. Plan UK’s Open Information Policy (in draft) and IATI commitments will further inform the external sharing of information outside of primary project stakeholders. c. Data Verification Process On the basis that prevention is better than cure, project partners have agreed that we will ensure the quality of monitoring data collected by: • Recruit suitable data collectors with the right skills, and provide specific training on all data collection tools • Use, wherever possible, sex and age appropriate data collectors, and take time to build the trust from those we are collecting data from in line with child protection and ethical research standards • Having a clear understanding of the data that needs to be collected, its use and value to those from whom it is being collected, and appropriate means of data collection • Having a clear and common understanding of the definitions of all specific terms and units of measurement used • Having clear lines of communication and data flows • Having user-friendly and clearly explained data collection tools • Wherever possible, and for all large scale data collection processes, pilot data collection methods before full roll-out During the course of the project, we will check and monitor data by: • Carry out peer reviews through joint monitoring visits (at least one per quarter) to support cross-organisation learning and to double check on progress of reported activities • Each organisation, on visiting a school where others are also active, will enquire as to the progress of all GEC activities in that school and will send a report to the CCU for cross-checking with existing records 30.05.2013 50 DRAFT M&E Framework Plan International UK • • • • • • Organisations leading on specific technical activities of the project or training other organisation staff will retain responsibility for ensuring the quality of the activities delivered and the results reported on across all districts Data collected using digital data collection devices will be automatically time and location stamped M&E Manger to carry out random spot checks on school to verify data once a quarter Senior managers (from the CCU and implementing agencies) will regularly carry out field visits to support operational and project quality Compare primary data with secondary data, and with data collected by similar projects in the same areas to look for un-explained differences External financial auditors will also examine project data as per the payment by results framework In verifying data at the organisational and CCU level, we will aim to ensure: • Completeness: making sure there are no gaps in data sets; that the data collected is relevant; and that data is available when needed • Correctness: all staff have the responsibility for ensuring the accuracy of data collected and recorded through, for example, conducting back checks / self-review of own data compared with previous records, ensuring specific numbers and appropriate disaggregation etc. Data clerks will carry out an initial check and clean of data. Data entered will subsequently be reviewed by relevant field staff. M&E Officers at the field level will periodically check paper and electronic copies of data. The CCU M&E Manager will periodically review data submitted and follow up as necessary. • Consistency: data clerks and M&E staff will look for anomalies or unlikely or strange patterns in data sets and unusual outliers over time, between districts or data sets or in comparison to secondary sources Specific details relating to data verification of survey work (at baseline and subsequently) are given in Section 1. d. Stakeholder Engagement and Communication Plan In all MER activities, the project will be informed by an utilisation-focused approach, with the explicit aim of identifying the use of, and primary intended users, of all MER outputs to increase the usability of outputs and increase buy-in from relevant stakeholders. Doing so forms part of the mandate of the evaluation steering group (see below). For example, the baseline report will be actively shared with project stakeholders, and a youth friendly summary of it to facilitate access by different groups and with girls in particular. Results of the baseline will be used to initiate discussions with duty-bearers at the local level and will feed into Plan and partners national level advocacy and campaigning at both national and international levels. This strong emphasis on shared learning will be maintained throughout the project, through participatory monitoring and local level advocacy for example, as well as with midline and evaluation findings, and this is built into the Performance Management Framework above. The project will seek out 30.05.2013 51 DRAFT M&E Framework Plan International UK opportunities for involving and inviting feedback from a range of stakeholders, ranging from Government actors, national and local women’s and child rights organizations and disabled people’s organisations, and other relevant members of civil society. e. Risk Assessment of M&E The table below summarizes the key risks identified for the early stages of the project. The Chief of Party will have responsibility for reviewing this list periodically alongside Plan’s project risk register and updating as necessary. Risk Children put at risk by project MER activities M&E Framework and sampling strategy not agreed in time between Plan, baseline consultant and EM Plan does not have access to the Sierra Leone adapted EGRA and EGMA tests in time to incorporate into the baseline study. M&E needs of the GEC Programme and the Sierra Leone GEC project at odds Inability to identify suitable control groups / or contamination of control groups Impact Potential or real harm to children and/or other project stakeholders; damage to relationships between project partners and individuals or communities; reduced effectiveness of project interventions; damage to organizational reputation of project; legal repercussions against project staff or associates Delay to start of baseline data collection, further impacted by start of rainy season and reduced access to field sites, increased cost of baseline process. Delays to the baseline and project start Necessary to reduce focus / length of additional survey questions to avoid research fatigue and ethical burden, reducing ability to measure context specific barriers to education / impact of project specific activities and intermediate outcomes / unintended consequences of project No counter-factual; reduced ability to demonstrate impact Mitigation Plan International’s Child Protection Policy signed by all project partners and associates Ethical and CP standards agreed by all project partners and followed for all MER activities, with clear reporting and oversight All partners and associates briefed on the CPP and Standards Request that the EM draw sampling points in stages by district Use ASER type tests for the baseline (and endline), and consider subsequent use of EGRA and EGMA tests with specific groups of beneficiaries (i.e. members of girls study groups). Keep project specific questions to minimum focusing on logframe indicators; compliment household questionnaire with qualitative data collection and project monitoring data EM to coordinate between Plan, BRAC and DFID Sierra Leone on identification of possible control 30.05.2013 52 DRAFT M&E Framework Plan International UK Research fatigue and /or attrition of respondents for longitudinal survey / Hawthorne effect in sampled clusters Higher attrition, Reduced ability to track a cohort; Reduced validity of findings due to positive bias within sampled communities Incomplete / inaccurate or noncomparable monitoring data collected Gaps in knowledge of process; inability to report against activities and outputs Monitoring data underutilization / Research fatigue among respondents; inefficient use of resources and time; lack of focus on learning to improve data collection Overlap between Plan target and control areas and Brac and DFID country programme (ISIS project) target and control sites Conducting learning assessments with girls receiving bursaries leads to a biased sample Contamination of control sites reduces counterfactual and ability to prove attribution to a single project Learning assessments at the school level demonstrate changes only within a specific target group of the population not the totality of all project activities f. overload groups; survey and qualitative work to measure exposure to similar activities Ensure focus groups do not overrun; keep survey length to minimum possible, collect respondent ID data to allow recontact, factor in additional time/resources at endline to recontact; Comparison and triangulation of data sources Agree and monitor data collection strategy between all partners; ensure data verification process followed; ensure sufficient resources for data verification Collect data specifically related to measuring quality and effectiveness of the project (and control); Establish project learning strategy and build in time for learning and reflection by project field staff and managers Selection of all schools to be compiled and shared by end of May, coordinated by DFID Sierra Leone Learning assessments at the school level are complimented by learning assessments as part of the household survey, which will provide a (relatively small) sample of target girls not receiving bursaries but involved in. In addition, the exposure of individual girls receiving bursaries to other elements of the project will be measured, allowing some degree of internal differentiation. Budgetary Consideration Please find attached a revised Costed M&E Plan as Attachment 4 30.05.2013 53 DRAFT M&E Framework Plan International UK Appendices Appendix 1: ‘Ideal Case’ Sampling Strategy The assumptions provided reflect how the impact evaluation should be designed ideally with no consideration of operational constraints and imperatives are as follows: 1) The ideal case pursues a two-stage sampling strategy. The following overview presents the different treatments and their main primary unit (i.e. community or school): Community CBR Bursary School Student clubs Parent clubs Girls mentoring school amendment Teacher training Teacher learning assistance Scorecards 2) In the ideal case, we would just want to channel treatment to the beneficiaries either through schools or villages. In the case that is ideal from a purely academic point of view, we would choose schools as the primary unit. 3) In the first stage, we draw a random sample of schools that will participate in the experiment 4) During the second stage, we then list all girls that meet the criteria as specified into a sample frame. Out of this sample frame we then randomly select 3*10 girls that meet the above mentioned criteria. Within each sub group we further stratify by disability status (1/0) to be able to detect impact on disabled vs. non-disabled girls. 30.05.2013 54 DRAFT M&E Framework Plan International UK 5) The expected effect size is based on the main outcome variable ‘learning performance’ that is captured through the learning assessment 6) The main vehicle for data collection is a household level survey. Besides collecting basic socio-economic data, it also administers the learning assessment 7) Ideally, the baseline will be administered before the intervention at the end of the previous school year 8) Upon completion of the baseline, the intervention will be delivered at the beginning of school year 1. For that purpose, we randomly assign primary sampling units to both experimental groups 9) The intervention is followed by the first follow up survey at the end of school year 1 10) A second follow up survey will be administered at the end of school year 2 11) A second follow up survey will be administered at the end of school year 3 12) The final round of data collection is the endline survey that will be administered at the end of the last school year of the programme. Ideally, this should be once those girls that were in PS5 have completed JSS3. 13) This panel data approach allows tracking of performance over time. For this design, the sample size can be calculated using the following formula and values MDE = (t(1-k)+ta) * sqrt(C/(P(1-P)J))*sqrt(1-C/(P(1-P)NJ)) MDE = Minimum effect size, expressed in standard deviation units of the outcome ( we assume 0.2 SD units) K = power of 80% A = significance level of 5% (t(1-k)+ta) = 2.8 C = Intra cluster correlation = 0.2 assuming a design effect of deff = 2.8 P = Portion of girls allocated to treatment (P = 50%) J = Total number of clusters (J = 220) N = Data points within each cluster (N = 10) The values above reflect the assumption that the 3 cohorts do not differ in terms of effect size or intra-cluster correlation. Furthermore, the sub samples are weighed equally. 30.05.2013 55 DRAFT M&E Framework Plan International UK Based on these assumptions, we determine J = 220 with 110 treatment schools and 110 control schools for both PS and JSS. Hence, the total number of schools would be 440. At the household level, we incorporate a response rate of 90%. Hence, we assume around 10% of the sample will refuse the interviews or will be subject to attrition over the course of the programme. This increases the intra-cluster sample size from 10 girls in each group to 12. Hence, the sample size is as follows: - 12 girls in PS5 - 12 girls in PS6 - 12 girls in JSS1, and The total sample size required is thus 5,280. Across these sub samples, we will further stratify girls by disability status. The composition will be then as follows: Nondisabled Disabled Total P5 8 4 12 P6 8 4 12 J2 12 0 12 This will allow us to detect impact of 25% on disabled and non-disabled separately. 30.05.2013 56
© Copyright 2026 Paperzz