Designing social programs that are evidence-based, result oriented, and culturally relevant.
Ensuring impact through regular monitoring of program performance, enabling mid-course corrections.
Systematic and objective assessment of a program, its design, implementation and results to determine relevance and fulfillment of objectives, development efficiency, effectiveness, impact and sustainability.
Evaluation is a management and learning tool for projects.
Performance monitoring should be based on outcomes.
Effective evaluation is an on-going process.
Impact evaluation should adopt a participatory approach.
Needs assessment studies aligned to the SDGs, put in place the building blocks of an effective monitoring system, improve governance and service delivery, and drive progress towards the shared objective of sustainable development. Specific indicators have been broken down under each SDG, based on the global indicator framework developed by the Inter-Agency and Expert Group on SDG Indicators (IAEG-SDGs).
SDG17 highlights the importance of strong collaborations and partnerships to achieve the other 16 SDGs. We engage in organizing regular workshops and structured input sessions by using participatory methods like cooperative interviewing technique, temperature gauge, five-finger feedback among different stakeholders and program partners. This helps to sustain outcomes and facilitates learning, exchange and action on the ground.
A quasi-experimental design views a social programs as an `intervention` in which a treatment-comprising of the elements of the programming being evaluated- is tested for how well it achieves its objectives, as measured by a pre-specified set of indicators. This design identifies a comparison group as demographically similar as possible to the treatment group (the group receiving intervention). The comparison group captures what would be said to have caused any difference in outcomes between the treatment and comparison groups.