130 likes | 232 Views
How Evaluations are Done at NRCan. Strategic Evaluation Division Natural Resources Canada. Purpose. To provide information on how evaluations are planned, costed and conducted once they are on the five-year departmental Evaluation Plan. The Cycle for Evaluation Reports. START
E N D
How Evaluations are Done at NRCan Strategic Evaluation Division Natural Resources Canada
Purpose • To provide information on how evaluations are planned, costed and conducted once they are on the five-year departmental Evaluation Plan.
The Cycle for Evaluation Reports START Evaluation Assessment (2-3 months) Implementing Change/Follow Up Terms of Reference DEC Approval Internet Posting (2-3 months) Large evaluations typically take 12-18 months to complete. Some phases may overlap. Contracting (3-4 months) Management Responses (1 month) Field Work/ Analysis (6-8 months) Report & Recommendations (2-4 months)
Evaluation Assessment • Evaluation Assessments are prepared based on: • program profiles (e.g., objectives; logic models; organization and governance; expenditures; an assessment performance information); • the calibration of the evaluation based on risk criteria, as part of overall 5-year Plan • (i.e., program renewal; materiality; context and need; visibility; management practices and structure; policy and delivery complexity; performance measurement; and past evaluation and audit findings); • the evaluation questions (TBS Policy identifies the generic questions and these may be supplemented by others identified by Program); • an initial plan on how to conduct the evaluation (i.e., methods to be used and levels of effort; contracting strategy; timelines; and estimated costs). • Evaluation Assessments are developed based on experience and professional judgement.
Scope/Costing Considerations • Based on evaluation assessment, the scope and costing are developed, taking into account: • TBS policy’s minimum standards and required "multiple lines of evidence" (e.g., documents; interviews; administrative data; survey results; case studies; focus groups). • SED’s internal risk assessment based on additional methods support quality, rigour & richness of the evaluation. • Sufficient level of evidence to be collected to allow conclusions to be drawn on each program (especially G&C programs). • There are many risk-factors (previous page) and considerations, e.g.: • number of programs, availability of data, past evaluations; • enough interviewees to ensure the full story; • in-person interviews are always better than over the phone • surveys provide less in-depth information from many people; • case studies provide in-depth knowledge of one project and how projects contribute to the achievement of program objectives. • Need for contracting
Contracting • Contractors may be used as part of the evaluation team, based on: • The need for subject matter expertise; • Internal capacity and timing considerations; and • The need for third party (non-NRCan) involvement • Value for money (optimal mix of internal and external resources) is determined through: • The availability of internal staff to meet coverage requirements; and • The competitive procurement process designed to achieve the best value for the Crown • Evaluation has aSupply Arrangement with nine evaluation firms* based on a competitive process to qualify the firm and contractors. • Based on the TOR, an RFP is sent to a minimum of 4 firms. • Contracting takes at least 3 months from SOW to signed contract. • Responses to the RFPs, provide suggestions for altering the SOW and what the company is prepared to do for the available budget. • If a subject matter expert is required, Evaluation will contract directly with the expert to be part of the team (e.g., nuclear expertise). * KPMG; Science Metrix; PMN; PRA; Goss Gilroy;TDV Global; Boulton; Baastel; and CPM.
Terms of Reference • The TOR are derived from the Evaluation Assessment and scoping/costing analysis. • TOR must be approved by the Evaluation Committee. • TOR include: • overview of the entity being evaluated; • evaluation issues and questions, • methods (e.g., interviews; surveys) to be used; • contracting approach (i.e., in house; contracted out; or hybrid); • timelines; • resources required (e.g., estimated contract cost); and • governance (e.g., a working group of program and evaluation officials).
Field Work • Evaluators work on several projects simultaneously. • Field work presents many challenges: • field work requires the input of programs; • contact information for surveys & interviewees not always readily available; • unplanned delays are very difficult for contractors to manage; • work around seasonal cycles: e.g. interviewees not available during the summer; contractors and programs are extremely busy prior to March 31. • Technical reports are usually produced for each method (e.g., interviews). All information is analyzed by evaluation question. • Preliminary findings are presented to programs to confirm findings and seek any additional information.
Report and Recommendations • Preliminary findings serve to highlight and validate key issues from collected evidence, and provide basis for the outline for the report. • Based on the findings, the report is written and the recommendations are developed. • The draft report is vetted with the Program and discussions take place on the recommendations (do they flow from findings; do they make sense; can they be implemented). • Report length is influenced by complexity of subject (e.g., one program or many) and need to present evidence. • Management responses and action plan are drafted by the program and ADM approval is sought.
Approval & Posting of Reports • The DEC advises the Deputy Minister (DM) on the report and management responses action plans. The DM must approve. • DM-approved reports are provided to TBS and may be examined for quality during the annual MAF assessment of evaluation. • The Evaluation Policy requires, complete, approved evaluation reports along with management responses and action plans to be posted in both official languages in a timely manner (i.e., 90 days). • Reports are reviewed by ATIP and Communications. • If required, Communications and the program prepare media lines.
Management Responses & Follow Up • ADMs are responsible for implementing the action plans for each recommendation. • The Evaluation Division follows up with Sectors on the implementation of action plans and updates DEC. • DEC decides if the management responses have been satisfactorily implemented and when the file can be closed. • The activity/program is re-considered as part of Evaluation Plan in the following evaluation cycle.
What Does Evaluation Add? • Accountability (posted reports and input into DPRs) • Strategic input into program decision-making and development (e.g., assistance in developing logic models and performance measurement strategies) • including information for MCs and TB Submissions • Neutral perspectives and additional data collection methods • Evidence for internal reviews (Strategic Reviews/SOR etc.) • Corrective change, starting before the evaluation is over • Follow up on recommendations
Conclusions • We hope this presentation has provided a better understanding of how evaluations are planned, costed and conducted. • Questions? send to dvinette@nrcan.gc.ca