Professional Documents
Culture Documents
& Evaluation
I. Introduction ............................................................................................................................................................... 5
II. The Glossary............................................................................................................................................................... 5
1. Community Development.........................................................................................................................................16
PLANNING A PROJECT....................................................................................................................................................17
THE SPECIFICATION.................................................................................................................................................18
PROVIDING STRUCTURE .........................................................................................................................................19
ESTABLISHING CONTROLS......................................................................................................................................22
THE ARTISTRY IN PLANNING ...................................................................................................................................23
PLANNING FOR THE FUTURE ..................................................................................................................................26
Communication Communication ............................................................................................................35
1.3.4.2 The Rural Works Progaramme (1963-1972)......................................................................................41
Summary ..................................................................................................................................................................41
Tool Number 2: Defining Evaluation.............................................................................................................................47
I. Introduction ...................................................................................................................................................47
II. What is Programme Evaluation?......................................................................................................................47
III. Why evaluate? ....................................................................................................................................................47
IV. What is the Relationship between Monitoring and Evaluation? ............................................................48
V. When do we need Monitoring and Evaluation results during the .........................................................50
VI. What is the relationship between evaluation and audit?.......................................................................50
VII. What is the role of evaluation in RBM? ...............................................................................................................51
Tool Number 3: Purposes of Evaluation.......................................................................................................................51
I. Why define the evaluation purpose? .............................................................................................................52
II. Three common evaluation purposes...........................................................................................................53
1.1 Probing questions .................................................................................................................................................55
1.2 PRA hypothesis ......................................................................................................................................................55
1.3 PRA definition ........................................................................................................................................................55
1.4 Review of development approaches preceding PRA .................................................................................55
1.5 PRA justification.....................................................................................................................................................56
1.5.1 Other justifications ........................................................................................................................................56
1.6 PRA limitations .......................................................................................................................................................56
1.7 PRA Assumptions...................................................................................................................................................57
1.8 PRA focus ...............................................................................................................................................................57
PRA focuses on community: .....................................................................................................................................57
1.9 PRA philosophy .....................................................................................................................................................57
PRA Tools ...........................................................................................................................................................................57
Approach: ...............................................................................................................................................................61
Common mistakes in SSI:.......................................................................................................................................66
During the interview:..............................................................................................................................................67
Tool Number 4: Stakeholder Participation in.............................................................................................................68
Monitoring and Evaluation ............................................................................................................................................68
I. Introduction ...................................................................................................................................................68
II. What is participatory monitoring and evaluation? .........................................................................................68
III. Who are the stakeholders?..............................................................................................................................69
IV. The rationale for stakeholder participation in M&E ....................................................................................69
V. When is it appropriate to use participatory M&E approaches? ..............................................................70
VI. Which stakeholders should participate in evaluation and what role should they play?....................72
VII. What are the steps in a participatory evaluation process?......................................................................74
2
VIII. Elements of successful stakeholder participation in evaluation ..........................................................76
Tool Number 5: Planning and Managing an Evaluation .........................................................................................77
Part II: Defining Evaluation Questions and Measurement Standards ...................................................................77
I. Introduction ...........................................................................................................................................................77
II. Defining Evaluation Questions ......................................................................................................................77
Figure 1. Core Evaluation Objectives ..........................................................................................................................80
Tool Number 5 : Planning and Managing an Evaluation .......................................................................................88
Part III: The Data Collection Process ............................................................................................................................88
I. Introduction .............................................................................................................................................................88
III. Determining methods for collecting additional data ................................................................................89
IV. Analyzing and Interpreting Data ....................................................................................................................92
II. Ensuring the quality of evaluations.................................................................................................................99
Part I: Identifying Output Indicators - The Basic Concepts...................................................................................104
I. Introduction ...........................................................................................................................................................104
II. The Process .............................................................................................................................................................104
Step 1: Define the planned outputs ..........................................................................................................................104
How to identify good indicators?...............................................................................................................................106
3
FOREWORD
In an attempt to implement the Educating the staff & community TRDP – HRD organized & facilitated the
large number of training events focusing on Managerial & technical capacities. HRD team have been
working hand in hand with the other core sections of TRDP. The success of such activities, however,
depends on the devotion and sincerity and above all appropriate skills of the organization implementing
the programmes. This, in turn, depends to a great extent on provision of appropriate training for their
staff.
Considering the needs of the staff & guidelines from Management, Board, & visitors HRD developed a
training kit to build up capability of their staffs engaged in the monitoring & evaluation of the
organizational inputs & outputs.
The contents of the Training Manual include the organization concepts, rationale and characteristics of
competency and skills, lessons of national & international organizations, the successes & the failure of
the organizations.
HRD team deserves appreciation for organizing the whole work for developing the Manuals under the
leadership of Chief Executive Officer & suggestion of Chairperson TRDP in our Training and Materials
Development. I believe would be of great benefit and fulfil the felt needs of training material for
imparting training to the staff for developing capability and enhancing their quality of work. It is our
belief that this manual will play vital role in organizing workshops focusing on M&E staff primarily &
secondarily to managers who are dealing with projects & programme, the developed material is
collection of information from different sources & can be changed as per need of the area.
This Manual is subject to constant improvement. We welcome any comments and suggestions you may
have on its content. We also encourage you to send us information on experiences from other
programmes or partners which can illustrate the issues addressed by this tool.
4
Tool Number 1: Glossary of Planning,
I. Introduction
The glossary responds to the need for a common understanding and usage of results based planning,
monitoring and evaluation terms among development agencies staff and its partners. In this context, the
planning, monitoring and evaluation terminology has been updated to incorporate the definition of terms
adopted by the national & international organization on Simplification and Harmonization.
Activities: Actions taken or work performed through which inputs such as funds, technical assistance and
other types of resources are mobilized to produce specific outputs.
Analysis: The process of systematically applying statistical techniques and logic to interpret, compare,
categorize, and summarize data collected in order to draw conclusions.
Appraisal: An assessment, prior to commitment of support, of the relevance, value, feasibility, and
potential acceptability of a programme in accordance with established criteria.
Applied Research: A type of research conducted on the basis of the assumption that human and societal
problems can be solved with knowledge. Insights gained through the study of gender relations for
example, can be used to develop effective strategies with which to overcome, socio-cultural barriers to
gender equality and equity. Incorporating the findings of applied research into programme design
therefore can strengthen interventions to bring about the desired change.
Assumptions: Hypotheses about conditions that are necessary to ensure that: (1) planned activities will
produce expected results; (2) the cause effect relationship between the different levels of programme
results will occur as expected. Achieving results depends on whether or not the assumptions made prove
to be true. Incorrect assumptions at any stage of the results chain can become an obstacle to achieving the
expected results.
5
Attribution: Causal link of one event with another. The extent to which observed effects can be ascribed
to a specific intervention.
Auditing: An independent, objective, systematic process that assesses the adequacy of the internal
controls of an organization, the effectiveness of its risk management and governance processes, in order to
improve its efficiency and overall performance. It verifies compliance with established rules, regulations,
policies and procedures and validates the accuracy of financial reports.
(B)
Baseline: Facts about the condition or performance of subjects prior to treatment or intervention.
Baseline Study: An analysis describing the situation prior to a development intervention, against which
progress can be assessed or comparisons made.
Benchmark: Reference point or standard against which progress or achievements can be assessed. A
benchmark refers to the performance that has been achieved in the recent past by other comparable
organizations, or what can be reasonably inferred to have been achieved in similar circumstances.
Beneficiaries: Individuals, groups or entities whose situation is supposed to improve (the target group),
and others whose situation may improve as a result of the development intervention.
Bias: Refers to statistical bias. Inaccurate representation that produces systematic error in a research
finding. Bias may result in overestimating or underestimating certain characteristics of the population. It
may result from incomplete information or invalid data collection methods and may be intentional or
unintentional.
(C)
Capacity: The knowledge, organization and resources needed to perform a function.
Capacity Development: A process that encompasses the building of technical abilities, behaviours,
relationships and values that enable individuals, groups, organizations and societies to enhance their
performance and to achieve their development objectives over time. It progresses through several
different stages of development so that the types of interventions required to develop capacity at different
stages vary. It includes strengthening the processes, systems and rules that shape collective and individual
behaviours and performance in all development endeavours as well as people's ability and willingness to
play new developmental roles and to adapt to new demands and situations. Capacity development is also
referred to as capacity building or strengthening.
Causality Analysis: A type of analysis used in programme formulation to identify the root causes of
development challenges. Development problems often derive from the same root causes (s). The analysis
6
organizes the main data, trends and findings into relationships of cause and effect. It identifies root causes
and their linkages as well as the differentiated impact of the selected development challenges. Generally,
for reproductive health and population problems, a range of causes can be identified that are interrelated.
A “causality framework or causality tree analysis” (sometimes referred to as “problem tree”) can be used
as a tool to cluster contributing causes and examine the linkages among them and their various
determinants.
Chain of Results: The causal sequence in the planning of a development intervention that stipulates the
possible pathways for achieving desired results beginning with the activities through which inputs are
mobilized to produce specific outputs, and culminating in outcomes, impacts and feedback. The chain of
results articulates a particular programme theory.
Cost-Benefit Analysis: A type of analysis that compares the costs and benefits of programmes. Benefits
are translated into monetary terms. In the case of an HIV infection averted, for instance, one would add
up all the costs that could be avoided such as medical treatment costs, lost income, funeral costs, etc. The
cost-benefit ratio of a programme is then calculated by dividing those total benefits (in monetary terms)
by the total programme cost (in monetary terms). If the benefits as expressed in monetary terms are
greater than the money spent on the programme, then the programme is considered to be of absolute
benefit. Cost-benefit analysis can be used to compare interventions that have different outcomes (family
planning and malaria control programmes, for example). Comparisons are also possible across sectors. It
is, for instance, possible to compare the cost-benefit ratio of an HIV prevention programme with that of a
programme investing in girls’ education. However, the valuation of health and social benefits in
monetary terms can sometimes be problematic (assigning a value to human life, for example).
Coverage: The extent to which a programme reaches its intended target population, institution or
geographic area.
(D)
Data: Specific quantitative and qualitative information or facts.
7
Database: An accumulation of information that has been systematically organized for easy access and
analysis. Databases are usually computerized.
(E)
Effectiveness: A measure of the extent to which a programme achieves its planned results (outputs,
outcomes and goals).
Effective Practices: Practices that have proven successful in particular circumstances. Knowledge about
effective practices is used to demonstrate what works and what does not and to accumulate and apply
knowledge about how and why they work in different situations and contexts.
Efficiency: A measure of how economically or optimally inputs (financial, human, technical and material
resources) are used to produce outputs.
Evaluability: The extent to which an activity or a programme can be evaluated in a reliable and credible
fashion.
Evaluation: A time-bound exercise that attempts to assess systematically and objectively the relevance,
performance and success, or the lack thereof, of ongoing and completed programmes. Evaluation is
undertaken selectively to answer specific questions to guide decision-makers and/or programme
managers, and to provide information on whether underlying theories and assumptions used in
programme development were valid, what worked and what did not work and why. Evaluation commonly
aims to determine the relevance, validity of design, efficiency, effectiveness, impact and sustainability of
a programme.
Evaluative Activities: Activities such as situational analysis, baseline surveys, applied research and
diagnostic studies. Evaluative activities are quite distinct from evaluation; nevertheless, the findings of
such activities can be used to improve, modify or adapt programme design and implementation.
Evaluation Questions: A set of questions developed by the evaluator, sponsor, and/or other stakeholders,
which define the issues the evaluation will investigate and are stated in such terms that they can be
answered in a way useful to stakeholders.
Evaluation Standards: A set of criteria against which the completeness and quality of evaluation work
can be assessed. The standards measure the utility, feasibility, propriety and accuracy of the evaluation.
Evaluation standards must be established in consultation with stakeholders prior to the evaluation.
Execution: The management of a specific programme which includes accountability for the effective use
of resources.
Ex-post Evaluation: A type of summative evaluation of an intervention usually conducted after it has
been completed. Its purpose is to understand the factors of success or failure, to assess the outcome,
8
impact and sustainability of results, and to draw conclusions that may inform similar interventions in the
future.
(F)
Feasibility: The coherence and quality of a programme strategy that makes successful implementation
likely.
Feedback: The transmission of findings of monitoring and evaluation activities organized and presented
in an appropriate form for dissemination to users in order to improve programme management, decision-
making and organizational learning. Feedback is generated through monitoring, evaluation and
evaluative activities and may include findings, conclusions, recommendations and lessons learned from
experience.
Finding: A factual statement on a programme based on empirical evidence gathered through monitoring
and evaluation activities.
Focus Group: A group of usually 7-10 people selected to engage in discussions designed for the purpose
of sharing insights and observations, obtaining perceptions or opinions, suggesting ideas, or
recommending actions on a topic of concern. A focus group discussion is a method of collecting data for
monitoring and evaluation purposes.
(G)
Goal: The higher order objective to which a development intervention is intended to contribute.
(I)
Impact: Positive and negative long term effects on identifiable population groups produced by a
development intervention, directly or indirectly, intended or unintended. These effects can be economic,
socio-cultural, institutional, environmental, technological or of other types.
Impact Evaluation: A type of outcome evaluation that focuses on the broad, longer-term impact or
results of a programme. For example, an impact evaluation could show that a decrease in a community’s
overall maternal mortality rate was the direct result of a programme designed to improve referral services
9
and provide high quality pre- and post-natal care and deliveries assisted by skilled health care
professionals.
Inputs: The financial, human, material, technological and information resource provided by stakeholders
(i.e. donors, programme implementers and beneficiaries) that are used to implement a development
intervention.
Inspection: A special, on-the-spot investigation of an activity that seeks to resolve particular problems.
Internal Evaluation: Evaluation of a development intervention conducted by a unit and /or individual/s
reporting to the donor, partner, or implementing organization for the intervention.
(J)
Joint Evaluation: An evaluation conducted with other UN partners, bilateral donors or international
development banks.
(L)
Lessons Learned: Learning from experience that is applicable to a generic situation rather than to a
specific circumstance. The identification of lessons learned relies on three key factors: i) the accumulation
of past experiences and insights; ii) good data collection instruments; and iii) a context analysis.
Logical Framework Approach: A specific strategic planning methodology that is used to prepare a
programme or development intervention. The methodology entails a participatory process to clarify
outcomes, outputs, activities and inputs, their causal relationships, the indicators with which to
gauge/measure progress towards results, and the assumptions and risks that may influence success and
failure of the intervention. It offers a structured logical approach to setting priorities and building
consensus around intended results and activities of a programme together with stakeholders.
Logical Framework (log frame): A dynamic planning and management tool that summarizes the results
of the logical framework approach process and communicates the key features of a programme design in a
single matrix. It can provide the basis for monitoring progress achieved and evaluating programme
results. The matrix should be revisited and refined regularly as new information becomes available.
(M)
10
Management Information System: A system, usually consisting of people, procedures, processes and a
data bank (often computerized) that routinely gathers quantitative and qualitative information on pre-
determined indicators to measure programme progress and impact. It also informs decision-making for
effective programme implementation.
Means of Verification (MOV): The specific sources from which the status of each of the results
indicators in the Results and Resources Framework can be ascertained.
Meta-evaluation: A type of evaluation that aggregates findings from a series of evaluations. Also an
evaluation of an evaluation to judge its quality and/or assess the performance of the evaluators.
Methodology: A description of how something will be done. A set of analytical methods, procedures and
techniques used to collect and analyse information appropriate for evaluation of the particular
programme, component or activity.
Monitoring: A continuous management function that aims primarily at providing programme managers
and key stakeholders with regular feedback and early indications of progress or lack thereof in the
achievement of intended results. Monitoring tracks the actual performance against what was planned or
expected according to pre-determined standards. It generally involves collecting and analysing data on
programme processes and results and recommending corrective measures.
Multi-Year Planning, Management and Funding Framework (MYFF): A four-year framework that is
composed of three interlinking components: (1) a results framework, which identifies the major results
that UNFPA aims to achieve, its key programme strategies, and the indicators that will be used to measure
progress; (2) an integrated resources framework that indicates the level of resources required to achieve
the stated results; and (3) a managing for results component that defines the priorities for improving the
Fund's organizational effectiveness.
(O)
Objective: A generic term usually used to express an outcome or goal representing the desired result that
a programme seeks to achieve.
Objectively Verifiable Indicator (OVI) (in Results and Resources Framework): See Indicator.
Outcome: The intended or achieved short and medium-term effects of an intervention’s outputs, usually
requiring the collective effort of partners. Outcomes represent changes in development conditions which
occur between the completion of outputs and the achievement of impact.
11
contributions of a specific organization to the outcome; and identifies key lessons learned and
recommendations to improve performance.
Outputs: The products and services which result from the completion of activities within a development
intervention.
(P)
Participatory Approach: A broad term for the involvement of primary and other stakeholders in an
undertaking (e.g. programme planning, design, implementation, monitoring and evaluation).
Performance Indicator: A quantitative or qualitative variable that allows the verification of changes
produced by a development intervention relative to what was planned.
Performance Monitoring: A continuous process of collecting and analysing data for performance
indicators, to compare how well development interventions, partnerships or policy reforms are being
implemented against expected results.
Process Evaluation: A type of evaluation that examines the extent to which a programme is operating as
intended by assessing ongoing programme operations. A process evaluation helps programme managers
identify what changes are needed in design, strategies and operations to improve performance.
Programme: A time-bound intervention similar to a project but which cuts across sectors, themes or
geographic areas, uses a multi-disciplinary approach, involves multiple institutions, and may be supported
by several different funding sources.
Programme Approach: A process which allows governments, donors and other stakeholders to
articulate priorities for development assistance through a coherent framework within which components
are interlinked and aimed towards achieving the same goals. It permits all donors, under government
leadership, to effectively contribute to the realization of national development objectives.
Programme Theory: An approach for planning and evaluating development interventions. It entails
systematic and cumulative study of the links between activities, outputs, outcomes, impact and contexts of
interventions. It specifies upfront how activities will lead to outputs, outcomes and longer-term impact
and identifies the contextual conditions that may affect the achievement of results.
Project: A time-bound intervention that consists of a set of planned, interrelated activities aimed at
achieving defined programme outputs.
Proxy Measure or Indicator: A variable used to stand in for one that is difficult to measure directly.
12
(Q)
Qualitative Evaluation: A type of evaluation that is primarily descriptive and interpretative, and may or
may not lend itself to quantification.
Quantitative Evaluation: A type of evaluation involving the use of numerical measurement and data
analysis based on statistical methods.
(R)
Reach: the coverage (e.g., the range or number of individuals, groups, institutions, geographic areas; etc.)
that will be affected by a programme.
Recommendation: Proposal for action to be taken in a specific circumstance, including the parties
responsible for that action.
Relevance: The degree to which the outputs, outcomes or goals of a programme remain valid and
pertinent as originally planned or as subsequently modified owing to changing circumstances within the
immediate context and external environment of that programme.
Reliability: Consistency and dependability of data collected through repeated use of a scientific
instrument or data collection procedure under the same conditions. Absolute reliability of evaluation data
is hard to obtain. However, checklists and training of evaluators can improve both data reliability and
validity.
Result: The output, outcome or impact (intended or unintended, positive and /or negative) derived from a
cause and effect relationship set in motion by a development intervention.
Results Based Management (RBM): A management strategy by which an organization ensures that its
processes, products and services contribute to the achievement of desired results (outputs, outcomes &
impacts). RBM rests on stakeholder participation and on clearly defined accountability for results. It also
requires monitoring of progress towards results and reporting on performance/feedback which is carefully
reviewed and used to further improve the design or implementation of the programme.
Results Framework: The logic that explains how results are to be achieved, including causal
relationships and underlying assumptions. The results framework is the application of the logical
framework approach at a strategic level, across an entire organization, for a country programme, a
programme component within a country programme, or even a project.
Risks: Factors that may adversely affect delivery of inputs, completion of activities and achievement of
results. Many risk factors are outside the control of the parties responsible for managing and
implementing a programme.
13
Risk Analysis: An analysis or assessment of factors that affect or are likely to affect the achievement of
results. Risk analysis provides information that can be used to mitigate the impact of identified risks.
Some external factors may be beyond the control of programme managers and implementers, but other
factors can be addressed with some slight adjustments in the programme strategy. It is recommended that
stakeholders take part in the risk analysis as they offer different perspectives and may have pertinent and
useful information about the programme context to mitigate the risks.
(S)
Stakeholders: People, groups or entities that have a role and interest in the aims and implementation of a
programme. They include the community whose situation the programme seeks to change; field staff who
implement activities; and programme managers who oversee implementation; donors and other decision-
makers who influence or decide the course of action related to the programme; and supporters, critics and
other persons who influence the programme environment (see target group and beneficiaries).
Strategies: Approaches and modalities to deploy human, material and financial resources and implement
activities to achieve results.
Success: A favourable programme result that is assessed in terms of effectiveness, impact, sustainability
and contribution to capacity development.
Summative Evaluation: A type of outcome and impact evaluation that assesses the overall effectiveness
of a programme.
Survey: Systematic collection of information from a defined population, usually by means of interviews
or questionnaires administered to a sample of units in the population (e.g. person, youth, adults etc.).
Baseline surveys are carried out at the beginning of the programme to describe the situation prior to a
development intervention in order to assess progress; Mid line surveys are conducted at the mid point of
the cycle to provide management and decision makers with the information necessary to assess and, if
necessary, adjust, implementation, procedures, strategies and institutional arrangements, for the
attainment of results. In addition, the results of midline surveys can also be used to inform and guide the
formulation of a new country programme. End line surveys are conducted towards the end of the cycle to
provide decision makers and planners with information with which to review the achievements of the
programme and generate lessons to guide the formulation and/or implementation of a new programme/
projects.
Sustainability: Durability of programme results after the termination of the technical cooperation
channelled through the programme. Static sustainability – the continuous flow of the same benefits, set in
motion by the completed programme, to the same target groups; dynamic sustainability – the use or
adaptation of programme results to a different context or changing environment by the original target
groups and/or other groups.
(T)
14
Target Group: The main stakeholders of a programme that are expected to gain from the results of that
programme. Sectors of the population that a programme aims to reach in order to address their needs.
Time-Series Analysis: Quasi-experimental designs that rely on relatively long series of repeated
measurements of the outcome/output variable taken before, during and after intervention in order to reach
conclusions about the effect of the intervention.
Transparency: Carefully describing and sharing information, rationale, assumptions, and procedures as
the basis for value judgments and decisions.
(U)
Utility: The value of something to someone or to an institution. The extent to which evaluations are
guided by the information needs of their users.
(V)
Validity: The extent to which methodologies and instruments measure what they are supposed to
measure. A data collection method is reliable and valid to the extent that it produces the same results
repeatedly. Valid evaluations are ones that take into account all relevant factors, given the whole context
of the evaluation, and weigh them appropriately in the process of formulating conclusions and
recommendations.
(W)
Work Plans: Quarterly, annual, or multiyear schedules of expected outputs, tasks, timeframes and
responsibilities.
1. Community Development
1.1 “Community Development is an attempt to influence the course of change at the community level”.
(Warren, 1971: 81)
Explanation: Initiatives, activities and attempts, which directly or indirectly provide opportunity
and facilitate community to act as a catalyst for change at local level. In other words,
all those efforts, which foster the process of positive change in accordance to the
will and need of community. Thematically it can be called effort of community
development.
1.3. Community Development is a process, which involves consultation with and the participation of a
number of groups. The definition of community development will not be static for any community over
time, and will evolve as the community continues to redefine itself.
It clearly indicates that community development is not a static process; therefore, it
Explanation: cannot run by applying a formula approaches. This is a process of evolution, which help
emerging of number of new procedures, methods and course of actions. Community
development programmes must be designed according to the context of community and
nature of interventions. Moreover, Participation may revolve around one single interest
group or number of groups prevailing in the village.
1.4. Community Development promotes local activities and projects designed to improve local
economic conditions. Trenval Community Development and community partners identify and pursue
opportunities in the local economy through the development and implementation of their strategic plans.
1.5 Community Development is an Active participation and if possible on the initiative of the
community, but if this initiative is not forthcoming spontaneously, by the use of techniques for arousing
16
and stimulating it in order to achieve its active and enthusiastic response to the movement. (Colonial office
1958:2)
Community Development is a process through which active participation of community
Explanation: is ensured. Ideally, villagers/community initiates community participation by them selves.
There are number of modes adopted by programmes to stimulate community for taking
collective action without any support of external agent. Stimulation may be due to some
internal and external factors. For example village fire, earth quack, flood, village
development schemes, etc. enforce community to join hand, take collective measures for
their survival, and improve quality of life. Thus all those efforts initiated by external
agencies e.g. Government programmes, NGOs, local CBOs and Donors should emphasis
on sensitization and stimulation of human minds and mobilization of resources with out
influencing over the ideas, perceptions and decisions of community in response for
change process. If any external agency focus on physical incentives without creating
awareness and education among the community could not be considered as community
development initiative.
1.6 What is Community Development strives for is to give ordinary people a voice for expressing and
acting on their extraordinary needs and desires in opposition to the stakes of local economic and political
power, to counter the increasing commodifications of both human beings and human welfare.
PLANNING A PROJECT
17
The success of a project will depend critically upon the effort, care and skill you apply in its initial
planning. This article looks at the creative aspects of this planning.
THE SPECIFICATION
Before describing the role and creation of a specification, we need to introduce and explain a fairly
technical term: a numbty is a person whose brain is totally numb. In this context, numb means "deprived
of feeling or the power of unassisted activity"; in general, a numbty needs the stimulation of an electric
cattle prod to even get to the right office in the morning. Communication with numbties is severely
hampered by the fact that although they think they know what they mean (which they do not), they
seldom actually say it, and they never write it down. And the main employment of numbties world-wide
is in creating project specifications. You must know this - and protect your team accordingly.
A specification is the definition of your project: a statement of the problem, not the solution. Normally,
the specification contains errors, ambiguities, misunderstandings and enough rope to hang you and your
entire team. Thus before you embark upon the the next six months of activity working on the wrong
project, you must assume that a numbty was the chief author of the specification you received and you
must read, worry, revise and ensure that everyone concerned with the project (from originator, through
the workers, to the end-customer) is working with the same understanding. The outcome of this
deliberation should be a written definition of what is required, by when; and this must be agreed by all
involved. There are no short-cuts to this; if you fail to spend the time initially, it will cost you far more
later on.
The work on the specification can seen as the first stage of Quality Assurance since you are looking for
and countering problems in the very foundation of the project - from this perspective the creation of the
specification clearly merits a large investment of time.
From a purely defensive point of view, the agreed specification also affords you protection against the
numbties who have second thoughts, or new ideas, half way through the project. Once the project is
underway, changes cost time (and money). The existence of a demonstrably-agreed specification enables
you to resist or to charge for (possibly in terms of extra time) such changes. Further, people tend to forget
what they originally thought; you may need proof that you have been working as instructed.
• the global context: numbties often focus too narrowly on the work of one team and fail to consider
how it fits into the larger picture. Some of the work given to you may actually be undone or
18
duplicated by others. Some of the proposed work may be incompatible with that of others; it might
be just plain barmy in the larger context.
• the interfaces: between your team and both its customers and suppliers, there are interfaces. At
these points something gets transferred. Exactly what, how and when should be discussed and
agreed from the very beginning. Never assume a common understanding, because you will be
wrong. All it takes for your habitual understandings to evaporate is the arrival of one new
member, in either of the teams. Define and agree your interfaces and maintain a friendly contact
throughout the project.
• time-scales: numbties always underestimate the time involved for work. If there are no time-scales
in the specification, you can assume that one will be imposed upon you (which will be
impossible). You must add realistic dates. The detail should include a precise understanding of the
extent of any intermediate stages of the task, particularly those which have to be delivered.
• external dependencies: your work may depend upon that of others. Make this very clear so that
these people too will receive warning of your needs. Highlight the effect that problems with these
would have upon your project so that everyone is quite clear about their importance. To be sure,
contact these people yourself and ask if they are able to fulfil the assumptions in your
specification.
• resources: the numbty tends to ignore resources. The specification should identify the materials,
equipment and manpower which are needed for the project. The agreement should include a
commitment by your managers to allocate or to fund them. You should check that the actual
numbers are practical and/or correct. If they are omitted, add them - there is bound to be
differences in their assumed values.
This seems to make the specification sound like a long document. It should not be. Each of the above
could be a simple sub-heading followed by either bullet points or a table - you are not writing a brochure,
you are stating the definition of the project in clear, concise and unambiguous glory.
Of course, the specification may change. If circumstances, or simply your knowledge, change then the
specification will be out of date. You should not regard it as cast in stone but rather as a display board
where everyone involved can see the current, common understanding of the project. If you change the
content everyone must know, but do not hesitate to change it as necessary.
PROVIDING STRUCTURE
Having decide what the specification intends, your next problem is to decide what you and your team
actually need to do, and how to do it. As a manager, you have to provide some form of framework both to
plan and to communicate what needs doing. Without a structure, the work is a series of unrelated tasks
which provides little sense of achievement and no feeling of advancement. If the team has no grasp of
how individual tasks fit together towards an understood goal, then the work will seem pointless and they
will feel only frustration.
To take the planning forward, therefore, you need to turn the specification into a complete set of tasks
with a linking structure. Fortunately, these two requirements are met at the same time since the derivation
of such a structure is the simplest method of arriving at a list of tasks.
The reasoning behind this is that the human brain (even yours) can only take in and process so much
information at one time. To get a real grasp of the project, you have to think about it in pieces rather than
trying to process the complexity of its entire details all at once. Thus each level of the project can be
understood as the amalgamation of a few simply described smaller units.
In planning any project, you follow the same simple steps: if an item is too complicated to manage, it
becomes a list of simpler items. People call this producing a work breakdown structure to make it sound
more formal and impressive. Without following this formal approach you are unlikely to remember all the
niggling little details; with this procedure, the details are simply displayed on the final lists.
One common fault is to produce too much detail at the initial planning stage. You should be stop when
you have a sufficient description of the activity to provide a clear instruction for the person who will
actually do the work, and to have a reasonable estimate for the total time/effort involved. You need the
former to allocate (or delegate) the task; you need the latter to finish the planning.
Task Allocation
The next stage is a little complicated. You now have to allocate the tasks to different people in the team
and, at the same time, order these tasks so that they are performed in a sensible sequence.
Task allocation is not simply a case of handing out the various tasks on your final lists to the people you
have available; it is far more subtle (and powerful) than that. As a manager you have to look far beyond
the single project; indeed any individual project can be seen as merely a single step in your team's
development. The allocation of tasks should thus be seen as a means of increasing the skills and
experience of your team - when the project is done, the team should have gained.
In simple terms, consider what each member of your team is capable of and allocate sufficient complexity
of tasks to match that (and to slightly stretch). The tasks you allocate are not the ones on your finals lists,
they are adapted to better suit the needs of your team's development; tasks are moulded to fit people,
which is far more effective than the other way around. For example, if Arthur is to learn something new,
the task may be simplified with responsibility given to another to guide and check the work; if Brenda is
to develop, sufficient tasks are combined so that her responsibility increases beyond what she has held
before; if Colin lacks confidence, the tasks are broken into smaller units which can be completed (and
commended) frequently.
Sometimes tasks can be grouped and allocated together. For instance, some tasks which are seemingly
independent may benefit from being done together since they use common ideas, information, talents.
One person doing them both removes the start-up time for one of them; two people (one on each) will be
able to help each other.
20
The ordering of the tasks is really quite simple, although you may find that sketching a sequence diagram
helps you to think it through (and to communicate the result). Pert charts are the accepted outcome, but
sketches will suffice. Getting the details exactly right, however, can be a long and painful process, and
often it can be futile. The degree to which you can predict the future is limited, so too should be the detail
of your planning. You must have the broad outlines by which to monitor progress, and sufficient detail to
assign each task when it needs to be started, but beyond that - stop and do something useful instead.
Guesstimation
At the initial planning stage the main objective is to get a realistic estimate of the time involved in the
project. You must establish this not only to assist higher management with their planning, but also to
protect your team from being expected to do the impossible. The most important technique for achieving
this is known as: guesstimation.
• make your guesstimates of the simple tasks at the bottom of the work break down structure and
look for the longest path through the sequence diagram
• use the experience from previous projects to improve your guesstimating skills
The corollary to this is that you should keep records in an easily accessible form of all projects as you do
them. Part of your final project review should be to update your personal data base of how long various
activities take. Managing this planning phase is vital to your success as a manager.
Some people find guesstimating a difficult concept in that if you have no experience of an activity, how
can you make a worthwhile estimate? Let us consider such a problem: how long would it take you to walk
all the way to the top of the Eiffel Tower or the Statue of Liberty? Presuming you have never actually
tried this (most people take the elevator part of the way), you really have very little to go on. Indeed if you
have actually seen one (and only one) of these buildings, think about the other. Your job depends upon
this, so think carefully. One idea is to start with the number of steps - guess that if you can. Notice, you do
not have to be right, merely reasonable. Next, consider the sort of pace you could maintain while climbing
a flight of steps for a long time. Now imagine yourself at the base of a flight of steps you do know, and
estimate a) how many steps there are, and b) how long it takes you to climb them (at that steady pace). To
complete, apply a little mathematics.
Now examine how confident you are with this estimate. If you won a free flight to Paris or New York and
tried it, you would probably (need your head examined) be mildly surprised if you climbed to the top in
less than half the estimated time and if it took you more than double you would be mildly annoyed. If it
took you less than a tenth the time, or ten times as long, you would extremely surprised/annoyed. In fact,
you do not currently believe that that would happen (no really, do you?). The point is that from very little
experience of the given problem, you can actually come up with a working estimate - and one which is far
better than no estimate at all when it comes to deriving a schedule. Guesstimating does take a little
practice, but it is a very useful skill to develop.
There are two practical problems in guesstimation. First, you are simply too optimistic. It is human nature
at the beginning of a new project to ignore the difficulties and assume best case scenarii - in producing
21
your estimates (and using those of others) you must inject a little realism. In practice, you should also
build-in a little slack to allow yourself some tolerance against mistakes. This is known as defensive
scheduling. Also, if you eventually deliver ahead of the agreed schedule, you will be loved.
Second, you will be under pressure from senior management to deliver quickly, especially if the project is
being sold competitively. Resist the temptation to rely upon speed as the only selling point. You might,
for instance, suggest the criteria of: fewer errors, history of adherence to initial schedules, previous
customer satisfaction, "this is how long it takes, so how can you trust the other quotes".
ESTABLISHING CONTROLS
When the planning phase is over (and agreed), the "doing" phase begins. Once it is in motion, a project
acquires a direction and momentum which is totally independent of anything you predicted. If you come
to terms with that from the start, you can then enjoy the roller-coaster which follows. To gain some hope,
however, you need to establish at the start (within the plan) the means to monitor and to influence the
project's progress.
For you, the milestones are a mechanism to monitor progress; for your team, they are short-term goals
which are far more tangible than the foggy, distant completion of the entire project. The milestones
maintain the momentum and encourage effort; they allow the team to judge their own progress and to
celebrate achievement throughout the project rather than just at its end.
The simplest way to construct milestones is to take the timing information from the work breakdown
structure and sequence diagram. When you have guesstimated how long each sub-task will take and have
strung them together, you can identify by when each of these tasks will actually be completed. This is
simple and effective; however, it lacks creativity.
A second method is to construct more significant milestones. These can be found by identify stages in the
development of a project which are recognisable as steps towards the final product. Sometimes these are
simply the higher levels of your structure; for instance, the completion of a market-evaluation phase.
Sometimes, they cut across many parallel activities; for instance, a prototype of the eventual product or a
mock-up of the new brochure format.
If you are running parallel activities, this type of milestone is particularly useful since it provides a means
of pulling together the people on disparate activities, and so:
Of course, there are milestones and there are mill-stones. You will have to be sensitive to any belief that
working for some specific milestone is hindering rather than helping the work forward. If this arises then
either you have chosen the wrong milestone, or you have failed to communicate how it fits into the
broader structure.
Communication is your everything. To monitor progress, to receive early warning of danger, to promote
cooperation, to motivate through team involvement, all of these rely upon communication. Regular
reports are invaluable - if you clearly define what information is needed and if teach your team how to
provided it in a rapidly accessible form. Often these reports merely say "progressing according to
schedule". These you send back, for while the message is desired the evidence is missing: you need to
insist that your team monitor their own progress with concrete, tangible, measurements and if this is done,
the figures should be included in the report. However, the real value of this practice comes when progress
is not according to schedule - then your communication system is worth all the effort you invested in its
planning.
At the planning stage, you can deal with far more than the mere project at hand. You can also shape the
overall pattern of your team's working using the division and type of activities you assign.
Ask your team. They too must be involved in the planning of projects, especially in the lower levels of the
work breakdown structure. Not only will they provide information and ideas, but also they will feel
ownership in the final plan.
This does not mean that your projects should be planned by committee - rather that you, as manager, plan
the project based upon all the available experience and creative ideas. As an initial approach, you could
attempt the first level(s) of the work breakdown structure to help you communicate the project to the team
and then ask for comments. Then, using these, the final levels could be refined by the people to whom the
tasks will be allocated. However, since the specification is so vital, all the team should vet the penultimate
draft.
Dangers in review
23
The constant trickle of new information can lead to a vicious cycle of planning and revising which shakes
the team's confidence in any particular version of the plan and which destroys the very stability which the
structure was designed to provide. You must decide the balance. Pick a point on the horizon and walk
confidently towards it. Decide objectively, and explain beforehand, when the review phases will occur
and make this a scheduled milestone in itself.
Even though the situation may have changed since the last review, it is important to recognise the work
which has been accomplished during the interim. Firstly, you do not want to abandon it since the team
will be demotivated feeling that they have achieved nothing. Secondly, this work itself is part of the new
situation: it has been done, it should provide a foundation for the next step or at least the basis of a lesson
well learnt. Always try to build upon the existing achievements of your team.
No plan is complete without explicit provision for testing and quality. As a wise manager, you will know
that this should be part of each individual phase of the project. This means that no activity is completed
until it has passed the (objectively) defined criteria which establishes its quality, and these are best
defined (objectively) at the beginning as part of the planning.
When devising the schedule therefore you must include allocated time for this part of each activity. Thus
your question is not only: "how long will it take", but also: "how long will the testing take". By asking
both questions together you raise the issue of "how do we know we have done it right" at the very
beginning and so the testing is more likely to be done in parallel with the implementation. You establish
this philosophy for your team by include testing as a justified (required) cost.
Another reason for stating the testing criteria at the beginning is that you can avoid futile quests for
perfection. If you have motivated your team well, they will each take pride in their work and want to do
the best job possible. Often this means polishing their work until is shines; often this wastes time. If it
clear at the onset exactly what is needed, then they are more likely to stop when that has been achieved.
You need to avoid generalities and to stipulate boundaries; not easy, but essential.
The same is also true when choosing the tools or building-blocks of your project. While it might be nice
to have use of the most modern versions, or to develop an exact match to your needs; often there is an
old/existing version which will serve almost as well (sufficient for the purpose), and the difference is not
worth the time you would need to invest in obtaining or developing the new one. Use what is available
whenever possible unless the difference in the new version is worth the time, money and the initial,
teething pains.
A related idea is that you should discourage too much effort on aspects of the project which are
idiosyncratic to that one job. In the specification phase, you might try to eliminate these through
negotiation with the customer; in the implementation phase you might leave these parts until last. The
reason for this advice is that a general piece of work can be tailored to many specific instances; thus, if
the work is in a general form, you will be able to rapidly re-use it for other projects. On the other hand, if
you produce something which is cut to fit exactly one specific case, you may have to repeat the work
24
entirely even though the next project is fairly similar. At the planning phase, a manager should bare in
mind the future and the long-term development of the team as well as the requirements of the current
project.
As a manager, you have to regulate the pressure and work load which is imposed upon your team; you
must protect them from the unreasonable demands of the rest of the company. Once you have arrived at
what you consider to be a realistic schedule, fight for it. Never let the outside world deflect you from what
you know to be practical. If they impose a deadline upon you which is impossible, clearly state this and
give your reasons. You will need to give some room for compromise, however, since a flat NO will be
seen as obstructive. Since you want to help the company, you should look for alternative positions.
You could offer a prototype service or product at an earlier date. This might, in some cases, be sufficient
for the customer to start the next stage of his/her own project on the understanding that your project
would be completed at a later date and the final version would then replace the prototype.
The complexity of the product, or the total number of units, might be reduced. This might, in some cases,
be sufficient for the customer's immediate needs. Future enhancements or more units would then be the
subject of a subsequent negotiation which, you feel, would be likely to succeed since you will have
already demonstrate your ability to deliver on time.
You can show on an alternative schedule that the project could be delivered by the deadline if certain
(specified) resources are given to you or if other projects are rescheduled. Thus, you provide a clear
picture of the situation and a possible solution; it is up to your manager then how he/she proceeds.
The most common error in planning is to assume that there will be no errors in the implementation: in
effect, the schedule is derived on the basis of "if nothing goes wrong, this will take ...". Of course,
recognising that errors will occur is the reason for implementing a monitoring strategy on the project.
Thus when the inevitable does happen, you can react and adapt the plan to compensate. However, by
carefully considering errors in advance you can make changes to the original plan to enhance its
tolerance. Quite simply, your planning should include time where you stand back from the design and
ask: "what can go wrong?"; indeed, this is an excellent way of asking your team for their analysis of your
plan.
You can try to predict where the errors will occur. By examining the activities' list you can usually
pinpoint some activities which are risky (for instance, those involving new equipment) and those which
are quite secure (for instance, those your team has done often before). The risky areas might then be given
a less stringent time-scale - actually planning-in time for the mistakes. Another possibility is to apply a
different strategy, or more resources, to such activities to minimise the disruption. For instance, you could
include training or consultancy for new equipment, or you might parallel the work with the foundation of
a fall-back position.
25
Post-mortem
At the end of any project, you should allocate time to reviewing the lessons and information on both the
work itself and the management of that work: an open meeting, with open discussion, with the whole
team and all customers and suppliers. If you think that this might be thought a waste of time by your own
manager, think of the effect it will have on future communications with your customers and suppliers.
With all these considerations in merely the "planning" stage of a project, it is perhaps surprising that
projects get done at all. In fact projects do get done, but seldom in the predicted manner and often as
much by brute force as by careful planning. The point, however, is that this method is non-optimal.
Customers feel let down by late delivery, staff is de motivated by constant pressure for impossible goals,
corners get cut which harm your reputation, and each project has to overcome the same problems as the
last.
With planning, projects can run on time and interact effectively with both customers and suppliers.
Everyone involved understands what is wanted and emerging problems are seen (and dealt with) long
before they cause damage. If you want your projects to run this way - then you must invest time in
planning.
26
2. Community Participation
General
The word participation is defined as “the act of taking part in an activity”
2.1 creations of opportunities to enable all members of a community to actively contribute to and
influence the development process and to share the fruits of development equitably.
2.2 Participation is a process through which stakeholders influence and share control over development
initiatives and the decisions and resources, which affect them. – (Participation Learning Group Final Report)
2.3 Public participation" is the involvement of people in a problem-solving or decision-making process
that may affect or interest them.
2.4 Participation is democracy. Representative democracy as currently practiced is a very limited form of
participation. (UNU Programme)
2.5 Participation means commitment to a more classless society that permits equal access to resources—
not only to land_ but also to food, education, health and other basic human rights. (UNU Programme)
2.6 Participation means building countervailing power, which leads to a healthier democracy. (UNU
Programme)
2.7 What is participation? Involvement, contribution and sharing of all community members at all
stages of development process i.e. Need identification and its prioritization, planning, implementation and
monitoring and evaluation. Participation means:
2.8 Participation is defined as a voluntary contribution by the people in one or another of the public
programme supposed to contribute to national, regional and local development, but the people are not
expected to take part in shaping the programme or in criticizing its contents. (Economic commission for Latin America,
1973)
Community participation is an active process by which beneficiary or client groups influence the
direction and execution of a development project with a view to enhancing their well- being in terms
of income, personal growth, self-reliance, or other values they cherish
(Paul,27
1987)
3. Community
3.1. Whose in the Community?
Community is a word with many meanings. It is a broad set of definitions that allow it to be used in many
ways. Community can be used to describe a large group of people or a smaller distinct group of people
within a larger group. It can even be used to group together people who share similar feelings or beliefs.
The word community is spoken in the news and heard regularly throughout everyday life, uttered by the
public. While having multiple definitions allow the word community to be used frequently, its multiple
meanings may cause confusion for others.
3.1.1 The Oxford English Dictionary defines the word community as the “commonwealth or a multitude
of men”. This definition allows a writer or speaker to use the word community as a way of describing a
group of people.
This denotation of community is most widely known to Americans today. Given this
definition, the word is being used in a very ambiguous sense. When community is being
used in this manual very little is truly understood by the reader or listener, whatever the
case may be. It is only obvious to the reader or listener that a community exists. Any
other traits of the community are unclear. Others remain unaware of the size, location, or
Explanations: any details of the individuals that make up the mentioned community. Therefore, we will
discuss other dimensions of community for clarity.
3.1.2 Another way of defining the word community is that “it is a group of people living together as a
smaller social unit within a larger one.”
3.1.3 The word community is also a way of categorizing people with similar interests as defined by
Webster’s New World Dictionary of American English. Throughout history, the word community has
been used to describe the geographical location of a group of people. Now the word community can
unite people all over the world, from any location, race or gender. Individuals with the same likes and
interests can be placed together into their own categories. An example of this type of community is the
community of faith and belief, which unites people from around the common interest and beliefs.
28
While in today’s language, meaning of the word community is as unclear as it has been
ever before. Community once has a broader meaning describing any group of people
living within the same boundaries. Now the word has evolved to allow the speaker or
Explanations:writer to speak of a smaller/ closer related group of people. Even in this context, an
unclear or incorrect message may be sent. As if the usage of the word community
cannot be difficult enough to understand as it to unite people with similar likes and
dislikes from across the globe. If words and dictionaries could change and language
could evolve, may be one day we will be able to think of a way of saying what we
mean, instead of confusing others with the word community. *
("Community." Webster’s New World Dictionary of American English. 3rd College Ed. NY: Prentice Hall Trade, 1988. )
3.2 A community is a bunch of people who have something in common. Let me try to spell out a broad
conception of community that may be compatible with most or all points of view.
Place. Centrally, a community is a group of people who live in the same place. Depending on context,
that place may be small or large, and the community may include just a handful of people, or millions,
e.g., the rural community. Often but not always, community of place is conceived of in terms of
governmental units: union council, village, mouhalla, hamalat, town, city, and country.
Interest & identity. We also speak of communities based on a shared interest or identity, including
profession, ethnicity, religion, language, hobby, etc., e.g., the local village community, the labor
community, teacher community, Muslim or Hindu community, business community or Sindhi or Punjabi
community etc. It is quite difficult to generalize accurately about communities without specifying what
type of community is being discussed. (M. Jourdain)
3.3. Community is a group of people residing in one geographical proximity and sharing their
resources, and has adhered the same culture, norms and values. Group of people can be categories in
the following:
1. People in area: a group of people who live in the same area, or the area in which they live
e.g. a close-knit fishing, village community
2. People with common background: a group of people with a common background or with shared interests
within society
e.g. the financial, agricultural or labor community
3. People with common history: a group of people with a common history or common economic or political
interests
e.g. the international, national, provincial ,local community
4. Society: the public or society in general
Community:
A specific group of people, often living in a defined geographical area, who share a common
culture, values and norms, are arranged in a social structure according to relationships, which the
community has developed over a period. Members of a community gain their personal and social
identity by sharing common beliefs, values and norms, which have been developed by the
community in the past and may be modified in the future. They exhibit some awareness of their
identity as a group, and share common needs and a commitment to meeting them." (Health
Promotion Glossary of the World Health Organization (1998)
29
Community is defined as;
4. Development
General
Event causing change: an incident that causes a situation to change or progress (often used in the plural)
4.1 Development is a natural as well as a planned process too, by which a change of conditions is
caused to make them better.
4.2 Development is a function of society’s capacity to organize human energies and productive
resources to respond to opportunities and challenges
4.3 Development is the process of positive changes it maybe physical, social or economic.
Development should be perceived as a multi-dimensional process involving the
reorganization and reorientation of entire economic and social systems. In addition to improvement in
Explanations:
incomes and outputs, it typically involves radical changes in institutional, social, and administrative
structure as well as popular attitudes and in many cases, even customs and beliefs.
4.4.1 Human development is development of the people, for the people, and by the people.
Explanations:
Development of the people involves building human capabilities through the
development of human resources. Development for the people implies that the
benefits of the growth must be translated into the lives of people. Development by
the people emphasizes that people must be able to affectivity influence the process
that shapes their lives.
30
1.2 Concept of Community Development
1.2.1 Historical Perception of Community Development
The concept of community development owes a great deal to the efforts of colonial administrators. After
the Second World War, the British Colonial Office became concerned with 'community development'.
Mayo (1975: 130) suggests that administrators 'invented' the term out of their attempts to develop 'basic
education' and social welfare in the UK colonies. For example, a 1944 report, Mass education in the
colonies, placed an emphasis on literacy training and advocated the promotion of agriculture, health and
other social services through local self-help (Midgley et al 1986: 17). This was a set of concerns similar to
those surrounding the interest in rural development and educational 'extension' in North America in the
first two decades of the century. As we discussed earlier about the concept of CD by sharing the different
perceptions, here we select one comprehensive definition to elaborate the concept of CD i.e.
“Active participation, and if possible on the initiative of the community, but if this initiative is not
forthcoming spontaneously, by the use of techniques for arousing and stimulating it in order to
achieve its active and enthusiastic response to the movement.” (Colonial Office 1958: 2)
The concern with community development was, in part, a response to the growth of nationalism, and, in
part an outcome of a desire to increase the rate of industrial and economic development. The notion began
to feature strongly in United Nations documents during the 1950s - and these drew extensively on the
British literature and experiences in Africa and Sub-Continent (Pre-Partition) (Midgley et al 1986: 18).
Within this, there does appear a certain contradiction. Community development emphasizes participation,
initiative and self-help by local communities but is usually sponsored by national governments as part of a
national plan. While from one side it can be seen as the encouragement of local initiative and decision
making, from the other it is a mean of implementing and expediting national policies at the local level and
is a substitute for, or the beginning of, local government (Jones 1977).
1.2.2 Concept of Community Development
Community development can be easily understood if we are able to better familiar with the key essentials
and elements of community development. Therefore, I have tried to elaborate the essentials and elements
of community development for social organizers. Certainly, this can not be an end but will be the mean to
understand the perception of CD.
1.2.3 Essentials of Community Development
Community development has some prerequisites or essentials without which nobody could claim
development as community development. Following are the key essentials of community development:
The participation by the people themselves in efforts to improve their level of living, with as much
reliance as possible on their own initiatives;
31
Community development focuses on the education and awareness of community and
developing their capabilities to ensure active participation of local community. This
process of development increases the community initiatives by relying on their own
Explanations: resources i.e. Human, Natural and Financial. If community initiatives mostly rely on
external resources that will put a big question in front of the development practioniers.
So we need re-designing /re-structuring the social organization approach to rectify the
whole process.
The provision of techniques and other services in ways which encourage initiative, self help and
mutual help and make these more effective
Before discussing the concept of community development, we will briefly share the dimensions of
community, which will help us to comprehend the philosophy of CD.
1.2.2 Community and its dimensions
In our work as social organizer and facilitator, the people of this field have spent many years observing
and interacting with communities of all types. Our experiences have given us significant an insight into
what is a community, how do communities evolve, and what contributes to the life and strength of the
community? Here we discuss the distinctions of community rather than a single, static definition. During
the fieldwork, it has been realized that community is a group of people related by some of each of the
following elements:
a. Common purpose - A fundamental reason for a group of people joining together into a
community. Something that provides the community with some form of shared meaning or
reason for their co-existence. This can range from a common interest or passion to the need to
make the best use of a shared space to the desire to transform the society or the humanity.
b. Common cultural context - Culture is one of those things that is indefinable and takes many
forms depending on who is describing it and for what reason. For the purpose of communities,
I like to think of it as our collective identity - which we are and how we express that. Inwardly,
our identity includes our values, beliefs, attitudes, behavioral norms, and our accumulated
experiences that shape these. Outwardly, culture then becomes the means of expression of our
32
inward identity, including our language, behavior, appearance, rituals, traditions, artifacts, and
events. In a community, a common language that enables clear communication among its
members is particularly important.
c. Co-location - All members of the community share a common physical and/or virtual space.
The relative size of the space and proximity of the members to each other and their shared
resources are major determinants of the degree of cohesiveness of the community. Strong
community relationships usually require some degree of physical co-location, even if just
occasional face-to-face get-togethers. A second-best option is to connect virtually, if they
cannot connect in person. However, we are just beginning to learn that there are certain
inherent advantages to connecting virtually (vs. face-to-face), such as a higher degree of
ambiguity and thus greater willingness to take risks in virtual relationships. The right balance
between virtual and physical interaction depends on the other elements of community, and the
evolving needs of the community and the external environment in which they operate.
d. Common timeframe - Some period of time in which they are all interacting with each other.
Usually, the closer this gets to real-time/simultaneous/immediate, the stronger the sense of
community.
e. Voluntary participation - One of the most powerful aspects of community is that it chooses
who participates in the community and members of the community choose to participate in it -
- a classic win/win scenario. Most work communities are also informal and invisible to the
formal organizational mechanisms, as well as being voluntary. This is the key difference
between a work community and other forms of organizational structure such as departments,
functions, and teams. However, the primary reason for the current focus on work communities
is to change this and make the communities an important stakeholder in the development,
while retaining the voluntary nature and resultant power of the communities.
f. Multiple, shifting, and overlapping membership and participation. Because people
participate in communities by choice and because they usually have any number of purposes,
cultural contexts, locations, timeframes, and combinations of these going on in their lives
simultaneously, they typically participate in more than one community at a time. In addition,
as the circumstances of their work and their lives, change and the environment in which they
work and live changes, people tend to go in and out of different communities and increase or
decrease their participation in the communities to which they belong. E.g. Community
participated in infrastructure project and simultaneously some of the participate in natural
resource management, enterprise/micro credit activity or vise-versa. These are the minimal
requirements. The concept of community development reveals that if we want to do
community development more effectively and efficiently we must need to contribute in the
above-mentioned elements that can be strengthen, enhanced cohesiveness and sustainability of
the community. Those additional elements includes the following:
34
The 6 C's of Relationship Explanations:
Communication Communication
Connection Connection
The relationship now gains some degree of persistent connection as
the people find some topic about which they want to communicate on
an ongoing basis. This is the stage at which communities of interest
(CoI's) exist. They connect to share what they know and care about in
their shared area of interest. At this stage, membership in the
community is usually fairly loose and transitory with members joining
and dropping out of the community quite freely. For the most part,
they have no strong sense of loyalty to the members of the community
or the community as a collective entity, only to their area of interest.
They will often participate in multiple communities revolving around
their specific interest and related topics.
Context Context
The relationship moves into a stage where the scope of shared context
becomes larger than just what the members of the community know
and care about some topic. In the same way we view knowledge as
information with sufficient context to provide real meaning to the user,
this stage of relationship gives the community sufficient context
around the information being exchanged to provide deeper meaning,
that is, knowledge that can be applied to their lives and work. This is
the stage where (CoP's) emerge. It is also the stage of relationship
where all the elements of community listed above down through
"Common Practices" become particularly important to the viability of
the community. For a CoP, the elements of shared knowledge and
common practices are especially critical.
35
Collaboration Collaboration
Commitment Commitment
The parties in the relationship decide that they will commit to a more
permanent relationship or to a temporary relationship with more
significant risks and rewards for all parties such as a joint
entrepreneurial venture. At this point, they become a community of
commitment.
This is where the relationship enters a much more intimate stage and
requires a lot more time, energy, and trust. Just as one of the biggest
problems in many person-to-person relationships is fear of
commitment and the resulting dysfunctional behaviors, fear of
commitment can cause significant problems in a CoC. Anyone that has
ever managed or been involved in a large project knows what a big
part relationship management and maintaining good team spirit and
teamwork are to the success of the project. A community of
commitment is much like a project team, except that they have come
together on a voluntary basis, which makes maintaining the
community collaboration and commitment both easier and harder.
Volunteers can be much more committed to a cause than someone
assigned to it, but they can also be harder to keep aligned around how
to get to the destination they all want to reach.
Communion
This stage transforms the relationship into an intimate spiritual bond that transcends conventional material concerns and attachments. The parties
experience a deep sense of oneness with each other and their environment. Their power grows immeasurably to co-create and to connect on an
intuitive level. This is a spiritual community.
A good example of this level of relationship is what happens to a combat platoon or other small group of people thrown into a life-threatening
situation together. The degree of interdependence and collaboration created by this situation can be a transformative experience for the group. Just
look at the sense of camaraderie and loyalty combat veterans still feel for their mates decades later.
This has also occurred in some non-life threatening situations, but it is more rare. One of the burning questions for those of us deeply
concerned with the long term well being of our organizations, our communities, and our country is -- Is there a way to create and sustain this sense
of communion without having to place the community in a life-or-death situation? Many of us believe there is and are working toward making that
happen on a global scale.
36
1.3 Community Development Practices in Pakistan
Community development initiatives taken in past fifty years in Pakistan are real assets in the field of rural
development. These initiatives provide us a path and direction for future. The major community
development programmes and projects that are quoted repeatedly in the international literature are the
Comilla Project, the Daudzai Pilot Project, the Organgi Pilot Project and the Aga Khan rural Support
Programme, were or still in Pakistan.
One school of thought established an opinion that the fifty years of effort about community development
is not promising and they quoted it as a “vast experience of failures in rural development.” But in my
opinion that all efforts in real terms break the ice and today in 21st centaury Pakistan is moving towards
the real road of development in which every planner and implementer tries to ensure participation of
community up to there desire level. At the national level and Federal Government has taken number of
initiatives to reduce poverty and provide access to disadvantaged community in the process of decision
making by supporting and initiating the following programmes and projects:
Established a District Support Organizations and national micro credit facility, Khushali
Bank (KB)etc
The above salient initiatives clearly indicates that policy makers and civil society organizations convinced
by the efforts of Comilla, Daudzai, OPP and AKRSP. Here we discussed the approaches and experiences
of these programme to understand the concept of participatory development. Before sharing these
projects, very brief insight should be given to the efforts taken in Indo-Pak before partition near about
colonial period.
Community development has a long history; it goes back to colonial period,i.e before partition. The first
initiative was taken by the Christian Mission, which established rural reconstruction centers that trained
multi-purpose village extension workers. The Christian Mission also set up educational, health and social
welfare institutions in rural areas as part of their missionary work.
The famous Indian poet Rabindranath Tagore supported grass-root village organizations through the
services of a rural reconstruction institute. In addition, Mahatama Gandhi advocated and developed new
rural institutions to provide services to the villagers for local development. He advocated self –reliance,
the exclusive use of local resources, and self- help for village development with out external assistance.
Another early experience and generally quoted in the literature of community development resulted from
37
the initiative of the Mr.F.L Brayne,he was the deputy commissioner of Punjab at that time. He tested his
concept of Community development called “Rural Reconstruction Model” in the Gurgaon district during
1920-28.Key activities and there design and assumptions of Brayne Model are discussed below:
Abstract THEME
F.L.Brayne tested his concept of rural development in the Gurgaon district. The brief summary of
experience is highlighted for readers here to understand the origin of community development /rural
development during the period of 1920-28.
38
1.3.2 Key Interventions of Rural Reconstruction Programme
{District Board under the chairmanship of the Deputy Commissioner was the implementing agency}
Coordinated and integrated mechanism developed for all line department activities at district level
Village guides were recruited to serve as multi-purpose extension workers for all line departments
each guide was in charge of a cluster of villages, and a Tehisil level officer and the District
Director supervised his work.
School of Rural Economy established on each cluster level and provided training facilities to
village guides and schoolteachers on rural development activities.
Local participatory organizations established such as the village co-operative societies, better
living societies, cattle breeding societies, women organizations, health centers, etc.
After the Second World War and at the time of creation of Pakistan thereafter, development practitioners
review the experiences and pool the ideas that emerged the base line for future development in Pakistan.
In the early days four main ideas which help in developing the rural development policies in Pakistan:
the concept of rural reconstruction, i.e. the holistic approach of total village development; this
concept went together with the broader development philosophy of developing community, which
in the early 50th was not usual because at that time the term “ Development” was more commonly
associated with the industrial sector;
the concept of Government initiated/guided integrated community development that would bring
together all sectoral line departments at the level to approach the various development problems
simultaneously;
the concept of involving the community and civil society at large in rural development efforts in
the form of localized participatory organizations; and ,
The concept of self-help and community co-operation, i.e. the notion of self-reliance at the local
level.
39
1.3.4 Community/Rural Development Initiatives of Nation Building Institutions {1952-1980}
Before early 80th Government initiated the four major programmes by the support of international donor’s
namely:
Village Agricultural and Industrial Development (V-AID) was the first national attempt to introduce
community/rural development approach on scale. It worked for nine years till 1961. Programme was
operated at two levels i.e.
National Level
Provided policy guidance and advisory assistance to the province responsible for project
implementation.
Provincial Level
Provincial level Director V-AID and at District level the Deputy Commissioners were
responsible for the organization and implementation.
This whole programme was monitored by advisory committees consist of all heads of line
departments to ensure coordination. Chairman of the village councils of elders includes in the
advisory committees.
Formed youth clubs and cooperatives for credit and marketing services for farmers
40
Summary V- AID certainly has created some unforgettable examples of self help initiatives in many areas
of development. It brought rural people together in local organization and at least attempted for
community participation. The implementation of programmes has sown the seeds for the idea of
participation at local level in a situation within the rural areas largely characterized by semi
feudal social structures and authoritarian rule. With the upcoming of basic democracy
programme in 1959, the V-AID gradually lost its role and finally terminated in 1961. Despite that,
V-AID programme is a tremendous initiative of the government that has provided an insight and
direction to the practitioners of development to learn lessons from the success and failures of the
experience.
After the termination of the V-AID, a new programme “The Rural Works Programme (RWP)” 1963-
1972, was designed to utilize the rural unemployment in building and improving rural infrastructure.
Concept of the programme was conceived from Comilla Project. The programme was closely tied up with
basic democracy institutions that were meant to find local remedies to the local problems. Overall, the
Programme operation was same as that of V-AID. The only difference was high level of discretionary
powers vested to district chairman and union councils’ chairman.
RWP was an effort to promote people’s participation to foster the process of development in rural sector.
FOCUS of RWP:
1. Formulation of village / project committees at village level to execute the community based
projects
2. Training of govt. officials and local leaders.
3. Ensuring the optimum utilization of local resources
4. Developing low cost infrastructure projects
5. Improving the institutional capacities of local and district councils for planning and
development functions
6. Promotion of self-help by raising volunteer local contribution in shape of land, labour and
materials
Summary The RWP, as a supporting programme, certainly made some contributions to village
development. It develops the local leadership and through its road programme it opened up
markets to remote villages and created employment opportunities for a part of the idle work force
in the countryside. Numerous good projects were completed through efficient project committees
making fully use of local resources at comparatively low cost in several districts. Due to high level
of political polarization programme could not create a significant impact. The main reasons
included weak community participation, introduction of a contractor system and vested interests
of some power full groups of community. But at the same time the programme proved that if any
community development programme compromises over the principle of “community
participation” it could not attain the real objectives of community development.
41
1.3.4.3 People’s Works Programme (PWP, 1972-80)
In 1972, PWP replaced the RWP. The main domain of the programme ranges as follows:
Road Construction
School buildings
Dispensaries
Tree planting
Brick making
The above-mentioned facilities were expected to arrest the mass exodus of the rural poor to the towns.
Concept of PWP
PWP was designed to mobilize local people for productive projects of local importance at comparatively
low capital cost and providing employment to surplus labor near the villages.
FOCUS OF PWP
1. Emphasized on short period project with active participation of local people
2. Involved a shift from centralized to decentralized planning
3. Established and strengthen the MARKAZ Councils
Summary
The ideology of PWP moved around the popular concept of people’s participation to empower
the vulnerable class of the society. The productive package of the programme covered strong
elements of livelihoods, but at the same time, the operational complexes created bottlenecks to
maximize the outcome. However, the development professionals realized the significance of
common understanding among the all stakes about the philosophy of people’s participation. It
proved the fact that the objective of active participation of people remained an idea instead of a
practice on ground without establishing an effective community infrastructure.
42
1.3.4.1 Integrated Rural Development Programme (IRDP, 1972-80).
IRDP was formally launched across the country in July 1972. The conceptual guidance for IRDP was
conceived from Daudzai project, which was implemented under the leadership of Mr. Shoaib SultanKhan,
who was Director of PARD at that time.
CONCEPT OF IRDP
Concept of IRDP moved around reducing the gap between the urban and rural areas by providing food
self-sufficiency, improving the socio economic condition of low income farmers and enhancing the
quality of life of the rural poor. It was designed as a multi sectoral programme as it worked on integration,
coordination and cooperation among the farmers, line departments, private sector and local bodies.
FOCUS OF IRDP
Involved local leaders and rural groups through the formation of village cooperatives and markaz
cooperatives federations
Empowered local community to voice in local decision making and project management
Emphasized on technical training and up gradation of skills of rural cadres to serve as trainers for
their village cooperatives
Community Development Experiences by Govt. and Non Govt. Organizations 1960 – 1990
The growth of and acceptance towards the participatory programmes in Pakistan dates back to the
experiences of Comila, Daudzai and AKRSP. Mr. Akhtar Hameed Khan (late) and Mr. Shoaib Sultan
Khan are known to be the mentors of the ideology. They developed the model of RSPs (Rural Support
Programmes) with an objective to establish a network of grassroots institutions across the country because
of lessons learnt in OPP, Comila and Daudzai. They provided a complimentary role and strength to the
state institution, civil society organizations by collaboration among the line departments with the
grassroots organization of the rural people. This whole experience in the field of community development
convinced and motivated the policy makers and the govt. to take number of national level initiatives to
promote the concept of participatory approach and institutionalize the state institutions.
Here we briefly discuss the assumptions, problems and approach of Comila, Daudzai and AKRSP. As a
learner in the field of community development, until we understand the approach in the context of its
assumption and problem scenario, we could not be able to understand the essence of community
development.
43
COMILA and DaudZai Model
The nexus of Comila and Daudzai is quoted as “Daudzai pilot project was the brain child of the famous
Comila Project of East Pakistan, now Bangladesh.” These both projects took over a leading function in
disseminating the participatory approach and provided conceptual guidance.
effectiveness of grassroot organizations. This challenge, however lies with the social organizer to identify
and devlop the local leadership for taking over the responsibility of these grassroots institutions.
Generally community co-operatives and group organized in these projects remain ineffective due to
unavailability of activits. This could only form a short period organizations of community gathered
around some physical incentive. Piolt experience of Comilla and Daudzai proved that a clear
commitment and a need based approach could achieve the objective of participatory devlopment either it
is a govt. programme or NGO.
44
Comila and Daudzai Projects
Methodology
Description
Identification of a viable unit for development
As a student of community development, administration
we can easily learn from the methodology Decentralization of Government services
adopted by these projects that the beauty of matched with supplies
participatory development lies in the Formation of cooperatives and community groups
designing of a socio cultural context Fostering a cadre of village specialists
specified effective methodology for Accumulation of capital through savings
creating willingness of community to be the Introduction of innovations
part of development process. Increased productivity packages (NRM)
Two Tier cooperatives
The highlighted methodology covered the
administrative, operational and programme packages effectively and it was further proved that it was
compatible with the needs of the community. This is evident from the lessons learnt from those projects
that an active participation could only be ensured if change agent could rightly identify the common
interest and social capital of the area. Along with, nature of productive package is also very critical to
work as gluing agent for community group. Saving and productive cooperatives worked as strong gluing
agent for these groups in Comila and Daudzai. So let me repeat that formula approach could not replace
the process approach.
AKRSP
ASSUMPTION
Description
Willingness of the community to
AKRSP started in 1982 under the conceptual organize
guidance of Mr. Shoaib Sultan Khan with the Presence of activists
intellectual support of Dr. Akhat Hameed Khan (late). Accumulation of capital through
It is considered as a mother project for the savings
mainstream conceptual developments in the history of Little or no support from government
rural development of Pakistan, particularly
concerning its two basic features local participation
and involvement of NGOs. The programme is based on the assumptions those all move around the local
resources and community i.e. saving, willingness of community and local leadership. It shows the strong
faith of the two legends in the philosophy of active participation and the strengths of community.
45
AKRSP
Problem Scenario
Description
At the time of inception of AKRSP, the major issue was
the low socio economic and suffocated political situation Handicaps of subsistence holders
of the area. In addition, the govt. infrastructure regarding Lack of services
basic services and supplies was inadequate to meet the Lack of facilities
needs of community. The key indexes of development
were below zero and the civil society was living below subsistence level. The socio- cultural situation was
not conducive and apparently, the participation of community was seemed to be a myth. That was the
time when the successes and failures of Comila, OPP and Daudzai worked out and the approach of
support organization emerged.
AKRSP
Methodology
Description
It was the first time when the concept of support Creation of support organization
organization emerged. The development Organizing (VOs, WOs)
practitioners felt the need of support organization Fostering a cadre of village activists
because there was no adequate infrastructure /specialists
available for basic service delivery. However, Development of programme packages
there was a big question for the designers of Physical Productive Physical
AKRSP that what should be the entry point for infrastructure(PPI)
community to mobilize? Therefore, it was Credit; savings and loans
analyzed that for effective mobilization the Increased productivity(NRM)
nueclous that could enforce community to join Introduction of intermediate technology
hand in the process of development would be the Micro Enterprise
common need. If we look over the social Development bank with enterprise support
economic status, the big issue of communities company
was the non-availability of the basic
infrastructure. Therefore, AKRSP adopted PPI approach as an entry point.
Local capital was mobilized in shape of savings and capacitated the social capital by developing a cadre
of village specialists and local activists at the grassroots level. . The approach was dynamic and have
abroad based and holistic in its design. It provide rural community a significant space to enlarge there
choices and avail their opportunity.
46
Planning
Monitoring & Evaluation
Toolkit
I. Introduction
This tool kit is an attempt to provide guidance and options for improvement in planning, monitoring and
evaluation activities in the context of results based programme management. It is also useful for
programme managers at headquarters and for Regional Managers and M&E staff.
This tool defines the concept of evaluation, what it is and why we evaluate, the role of evaluation in
relation to monitoring and audit, and its role in the context of results-based management approaches. The
content is based on a review of a wide range of evaluation literature from academia and international
development agencies such as UNDP, UNICEF, WFP, OECD and bilateral donor agencies such as
USAID & many of others.
1
Definitions of these terms are provided in Tool Number 1: Glossary of Planning Monitoring and Evaluation Terms and are
further discussed in Tool Number 5, Part II: Defining Evaluation Questions and Measurement Standards.
47
To ddemonstrate accountability2 to decision-makers (donors and programme countries).
It is expected that improved decision-making and accountability will lead to better results and more
efficient use of resources.
2
Accountability is the responsibility to justify expenditures, decisions or the results of the discharge of authority and official
duties, including duties delegated to a subordinate unit or individual. Programme Management is responsible for providing
evidence to stakeholders and sponsors that a programme is effective and in conformity with its coverage, service, legal and
fiscal requirements.
48
Monitoring continuously tracks performance against what was planned by collecting and analysing data
on the indicators established for monitoring and evaluation purposes. It provides continuous information
on whether progress is being made toward achieving results (outputs, outcomes, goals) through record
keeping and regular reporting systems. Monitoring looks at both programme processes3 and changes in
conditions of target groups and institutions brought about by programme activities. It also identifies
strengths and weaknesses in a programme. The performance information generated from monitoring
enhances learning from experience and improves decision-making. Management and programme
implementers typically conduct monitoring.
Evaluation is a periodic, in-depth analysis of programme performance. It relies on data generated through
monitoring activities as well as information obtained from other sources (e.g., studies, research, in-depth
interviews, focus group discussions, surveys etc.). Evaluations are often (but not always) conducted with
the assistance of external evaluators.
Monitoring Evaluation
Continuous Periodic: at important milestones such as the mid-term
of programme implementation; at the end or a
substantial period after programme conclusion
Keeps track; oversight; analyses and In-depth analysis; Compares planned with actual
documents progress achievements
Answers what activities were implemented Answers why and how results were achieved.
and results achieved Contributes to building theories and models for change
Alerts managers / team leaders to problems Provides managers with strategy and policy options
and provides options for corrective actions
3
Transformation of inputs into outputs through activities.
49
V. When do we need Monitoring and Evaluation results
during the
Programme Cycle?
During situation analysis and identification of overall programme focus, lessons learned from
past programme implementation are studied and taken into account in the programme strategies;
During programme design, data on indicators produced during the previous programme cycle
serve as baseline data for the new programme cycle. Indicator data also enable programme
designers to establish clear programme targets which can be monitored and evaluated;
Notwithstanding this difference in focus, audit and evaluation are both instruments through which
management can obtain a critical assessment of the operations of the organization as a basis for instituting
improvements.
50
VII. What is the role of evaluation in RBM?
International development organizations currently place strong emphasis on national capacity
development, good governance and public sector transparency. In this context, evaluation, together with
continuous monitoring of programme and project progress, is an important tool for result-based
management. In assessing what works, what does not work and why, evaluation provides information
that strengthens organizational decision-making and promotes a culture of accountability among
programme implementers. The lessons highlighted through evaluation to improve programme and
organizational performance. Demonstration of more and higher quality results through improved
performance can lead to increased funding in different projects and programmes.
Box 3 outlines, in no particular order of priority, some characteristics and expected benefits of introducing
results-based monitoring and evaluation.
IF
Senior management is strongly committed to the use of M&E results in decision-
making – commitment influences the management style;
Staff undertake M&E activities
Tool Number 3: and use M&E data
Purposes of atEvaluation
all stages of the programme
cycle;
Staff apply M&E approaches to all areas of the organizational operations for example
in programme, finance, and human resources management/development;
Staff engaged in monitoring and evaluation activities strive to pursue objectivity.
They make clear the criteria and values on which their judgments are based;
Staff are held accountable for results and take risks to achieve them;
Staff apply lessons learned to programme management;
Staff is recognized by the organization for achieving good results and for their efforts
to counteract risks.
THEN
Organization becomes more efficient and better equipped to adapt to a rapidly
changing external environment;
The quality and effectiveness of organizational assistance increases;
The organization and its partners achieve results;
credibility improves;
Funding for is likely to increase;
Staff has a heightened sense of achievement and professional satisfaction;
productivity improves.
Experience has shown that when the evaluation manager determines the main purpose of the evaluation
together with the intended users of evaluation findings, the chance that the findings will be used for
decision-making is greatly increased.
When planning for an evaluation, the managers should therefore always ask the following questions: Who
wants the evaluation? Why do they want it? How do they intend to use it?
52
II. Three common evaluation purposes
Box 1 highlights the three most common evaluation purposes and a sample of evaluation questions typically asked by the intended users.
What are the programme’s strengths and weaknesses? What kinds of implementation
problems have emerged and how are they being addressed?
What is the progress towards achieving the desired outputs and outcomes? Are the activities
planned sufficient (in quantity and quality) to achieve the outputs?
Are the selected indicators pertinent and specific enough to measure the outputs? Do they
need to be revised? Has it been feasible to collect data on selected indicators? Have the
indicators been used for monitoring?
Why are some implementers not implementing activities as well as others?
What is happening that was not expected?
How are staff and clients interacting? What are implementers’ and target groups’ perception
of the programme? What do they like? Dislike? Want to change?
How are funds being used compared to initial expectations? Where can efficiencies be
realized?
How is the external environment affecting internal operations of the programme? Are the
originally identified assumptions still valid? Does the programme include strategies to
reduce the impact of identified risks?
What new ideas are emerging that can be tried out and tested?
To make an overall judgment about the effectiveness of a completed programme, often to ensure
accountability – A summative evaluation.
Did the programme work? Did it contribute towards the stated goals and outcomes? Were
the desired outputs achieved?
Was implementation in compliance with funding mandates? Were funds used appropriately
for the intended purposes?
Should the programme be continued or terminated? Expanded? Replicated?
What is the assumed logic through which it is expected that inputs and activities will
produce outputs, which will result in outcomes, which will ultimately change the status of
the target population or situation (also called the programme theory)?
What types of interventions are successful under what conditions?
How can outputs/outcomes best be measured?
What lessons were learned?
What policy options are available as a result of programme activities?
53
III- Who uses what kind of evaluation findings?
Certain evaluation findings are particularly suited for decision-making by specific users. For example,
programme managers and staff of implementing partners need evaluation findings related to the delivery
process and progress towards achieving aims. This type of information will help them choose more
effective implementation strategies.
Decision-makers who oversee programmes such as policy makers, senior managers and donors, require
evaluation findings related to effectiveness, impact and sustainability. This type of information will
enable them to decide whether to continue, modify, or cancel the programme or projects.
Data generated through evaluations, which highlight good practices and lessons learned is essential for
those engaged in overall policy and programme design.
It is essential to note that one type of evaluation findings usually constitutes an essential input to
produce other types of findings. For instance, data on programme implementation processes gathered
through a formative evaluation, or through monitoring and review activities, is a necessary input to enable
analysis of programme impact and to generate knowledge of good practices. When no impact of activities
is found, process data can indicate if this occurred because of implementation failure (i.e. services were
not provided hence the expected benefit could not have occurred) or theory failure (i.e. the programme
was implemented as intended but failed to produce the expected results). Data on implementation
processes enable an analysis of which approaches work or do not work and under what conditions. Box 2
highlights an example of theory failure which has affected the impact of UNFPA’s interventions to reduce
maternal mortality rates.
The Evaluation therefore concluded that strategies selected to prevent maternal deaths must be based on
the most up-to-date technical information. Several basic premises are now widely known with regard to
safe motherhood:
Every pregnancy faces risks;
A skilled attendant should be present at every delivery;
Emergency obstetric care should be accessible; and
More emphasis is needed on care during birth and immediately afterwards. Post-partum care should
include the prevention and early detection of complications in both the mother and the new-born.
Source: UNFPA, 1999.
54
Introduction to PRA
• What is PRA?
• When and where were you exposed to PRA?
• What is right and wrong about PRA?
• Does PRA always produce the expected results towards community development?
• Is PRA a miracle approach?
PRA is a community development approach that attempts to empower whole communities to define their
own strategies and actions within the limits of their skills, strengths, and resources in order to realize
sustainable development.
a. End of World War II to the early 70's, the Green Revolution approaches were adopted worldwide but
mainly originated in some present day developed nations. These methods were characterized by
national resources planning and apportionment, and coupled with fortified learning and research
institutions and centralized information banks. These approaches were essentially top-down in
decision-making, planning, and implementation. They were mostly elite-friendly i.e. the elite class
benefited most.
b. Mid-70's to early '80's, the National Tailored Approaches were built in consideration of national
priorities. They were characterized by ministerial, departmental and regional/provincial planning and
apportionment of resources. Regional research institutions and information banks were established but
segregated/restricted from the general public. These approaches were essentially top-
down in decision-making, planning and implementation. They mostly benefited the educated and
politically correct community members.
c.Early 80's to mid-80's. Farming Systems Research (FSR) and Rapid Rural Appraisal (RRA) evolved at
about the same time with the aim of learning and extracting more information from target communities.
FSR has a scientific research orientation where as RRA has a socio-economic survey orientation. FRS
was pursued by agricultural scientists, where as RRA was pursued by sociologists. The main difference
between these two approaches and the preceding approaches is that RRA and FSR bore an emphasis in
learning and extracting information through direct interaction with local communities. However, both
approaches are still top-down with regards to decision-making, planning, resource apportionment and
implementation. Owing the need to elicit community participation in all
55
community development matters, both RRA and FSR lead to the evolution of PRA. Thus PRA has been
advocated as a suitable community development approach, since the mid-80.
Realistic and true development entails the continuous and sustained improvement of the standard of living
of the majority of people. In Kenya and in most developing countries, the majority of people live in rural,
infrastructure-poor locations.
While many financial, material and infrastructural efforts continue to support rural development
endeavors, a critical analysis of demographic, educational, economic and food security trends indicates no
or little tangible improvement in the quality of life for the majority of people, who, in general have, in
fact, fallen below the poverty line.
Many things could be wrong (including the political climate and public administration that mainly lack
transparency, accountability, and responsibility), but a realistic and effective development approach has
also been lacking. For now, PRA has evolved as a promising approach towards true and sustainable
community development.
• PRA emphasis’s true and sustainable development that is community tailored and initiated;
• PRA facilitates community organization towards community development plans and actions;
• PRA upholds community needs and priorities;
• PRA attempted to harmonize community development efforts;
• PRA encourages interactions/partnerships between development agents and community members;
• PRA recreates the wealth of community-based information in order to justify project
interventions;
• PRA encourages continuous follow-up and assessment of project progress and direction at the
community level;
• PRA provides for sustainable exit strategies by development agents;
• PRA provides the basis for transparency and accountability in community development
expenditures.
56
• Community barriers: Language and cultural difference can be a major constraint in the interaction
of outsiders and insiders throughout the PRA process.
• Community, socio-economics/education: The general socio-economic and/or education standing
of the target community greatly influences the outcome of the PRA process. Communities living
in abject poverty or those with little formal education may present a great challenge to the
implementers of the PRA process.
• Community infrastructure: The roads and communication network within a given community can
present a challenge that could be limiting.
• Community harmony: Communities that are rife with clan, ethnic or sectoral differences are areas
generally challenging and limiting to the PRA process.
• General security: Communities laden with insecurities are generally hard to work with.
•
PRA is an ideal and effective approach in community development only if it assumes the following:
• Critical awareness and understanding of the PRA principles, tools, process and ethics by the PRA
team or development agency
• Multidisciplinary yet homogenous PRA team
• Immunity from all possible limitations
• That the PRA process is socially acceptable, economically viable and ecologically sound
• A wealth of untapped, unorganized and under-utilized information and resources lie within the
target community
• Inherent local/community leadership is available to spearhead community development issues
• Definite community needs and priorities form the basis of attracting outside help
In order for realistic and sustainable development to take root, community members must be viewed as
the custodians of all community-based development projects. Therefore, development agents/agencies
must only view themselves as facilitators and partners (not experts) in community development issues.
PRA Tools
57
There are a variety of methods that are used during the PRA exercise in order to elicit community
participation in the generation of information. These methods can be classified into three broad
categories.
1. Visualized analyses
2. Interviewing and sampling
3. Group and team dynamics
These methods can be used singularly one at a time or in a combination of two or three. More often for
the PRA team to get full participation of the community, all the three are combined for best results. The
method(s) to be applied at any particular instance will depend on the type of information required, the
community understands of the method and/or the tool being applied.
1) Visualized analyses
b) Participatory diagramming
People have shown their capacity to produce and understand diagrams of different types, using seeds,
fruits, stones on the ground as well as pens and paper. As for every other tool, it is essential to repeat with
different informants, representing diverse interests of various social groups such as men, women, old,
young, poor and wealthy.
i) Trend lines
Trend lines show quantitative changes over time and can be used for many variables such as:
• yields
• area under cultivation
• livestock population
• prices
• population size
• birth and death rates
• rainfall
• food production
• environmental degradation
• deforestation
• education
• employment rates
• soil loss
58
The trend lines will help the PRA team understand the villagers' perceptions of significant changes in the
community over time. These tools will focus community attention on positive and negative changes over
time in terms of resource use and on its traditional resources management practices.
Data collected on trend lines are important because:
• topics indicate what people consider important;
• the direction of trends is significant even if the change cannot be quantified or is statistically
inaccurate because it shows how villagers view their ever changing situation;
• the process of discussing resources use trends in various groups will bring out important resources
management issues for preparing the Community Action Plan (CAP).
The PRA team should collect information on trends according to what the community deems important,
covering a period of 10 to 40 years. The list will vary from one community to another, but a core set of
trends should include changes that need to be investigated as per the objectives of the PRA. Whereas, the
timeline (historical profile) focuses discussions of elderly people and long-term residents, trend line
exercises can include much younger people/participants. Leaders of community-based groups are
important. As many people, both men and women, should be included in the discussions because their
comments reflect first hand experience. Therefore, full involvement of women is critical. All ecological
zones in the study must be included.
Steps:
1. Explain to the community groups the meaning of trend lines
2. Explain concept of trend lines using appropriate materials
3. When the community has understood the concept, ask one of them to explain to the rest
4. Ask community members the major issues they would like to represent. They can list as many as
they wish but these should be scaled down to 6-10 trends
5. Begin with the most popular resource/topic/issue and ask one of the community members to
represent the status of the resource/topic/issue during the first year of the trend line
6. Repeat for all the following years up to the present
7. Use the discussion of trends to probe the explanation of changes, if any. This will help identify
underlying problems and traditional activities to correct the situation. For example, if the group
agrees that soil erosion is getting worse, ask why. Find out what solutions have been tried in the
past and how well they've worked. Ask what they think might ease the problem.
Seasonal calendars explore seasonal constraints and opportunities month-by-month or season- by-season
throughout the year. They help present large quantities of diverse information in a common time frame.
They identify cycles of activities that occur in a community on a regular basis, like the common periods
of environmental problems or opportunities over the course of a normal year. The yearly cycles are
important in determining for example:
Steps:
i. seasonal calendar data provides information on opinions and attitudes of community towards
certain phenomena/activities, which include attitudes towards gender allocation of labour, gender
ownership and control of certain resources.
ii. shows how much the community depends on existing natural resources to meet their needs using
locally available food resources and what they have to plant on a continuous basis.
iii. shows how the community perceives its problems and assesses existing opportunities.
iii) Venn diagrams
• Venn diagrams are used to gain insight on the community's understanding of linkages and
relationships of different systems e.g. institutions, livelihoods.
• Circles represent people, groups and institutions. They are drawn in such a manner to represent
degree and type of relationship.
• Innovations may include:
o drawing lines between circles and main circle (of village) with the thickness of the line
representing strength of relationship.
o drawing wider circles for stronger relationships
60
o drawing wider and closer circles to the village circle to represent stronger linkages.
Example:
The information collected here is important especially at later stages of the PRA process as it helps:
• to identify possibly community entry points which can be strengthened;
• to inform about the institutional framework within which the CAP is being developed;
• the community to perceive agencies, people, and organizations, reflect their role within the
community and how best they can begin to form closer linkages for further development.
Approach:
61
This task is normally carried out by the facilitators with the community after they have gathered data
(through social mapping for example) on the different stakeholders and their relative interest, or stake in a
given resource.
The different stakeholder groups (e.g. women, pastoralists, poor, saw millers, etc.) are put in a circle in
accordance with the perceived relative importance of their stake. The bigger the circle, the more important
that stakeholder group is with respect to the resource in question.
On the diagram, a campfire is drawn in the centre. Stakeholder groups are analyzed with respect to their
decision-making power and authority. Those who are the most important decision-makers are placed
closest to the campfire (so that they can see, keep warm, prevent others from getting close), and those
who are the least important, further away.
A visual diagram is built up (see example, next page). Lines may be drawn to show the linkages between
the decision-makers and other stakeholders. A circle can be drawn around the main decision making
groups, leaving those stakeholders who do not have decision-making power outside the principal circle.
This tool helps to ensure that we do not ignore important stakeholder groups from either the decision-
making or the actual stake/dependence on the resource.
Pie charts are used to understand different resource contributions to livelihood, constraints, needs and
opportunities for the whole community or individual households.
Steps:
1. Ask the community to identify different resources that contribute to certain constraints or
opportunities
2. Discuss every resource and determine whether it does contribute to the particular constraint or
opportunity.
3. Have the community cut out a big circular paper
4. Have the community divide the contribution of each resource, whether half, quarter, more than
quarter, three quarters, or in percentages. The degree/level of contribution is represented on the
card by drawing a big angle for high and small for low contribution.
v) Mobility map:
62
The above diagram is an example of a mobility map of an imaginary community in Trans Nzoia district.
The map tells you the degree of contact the community has with the outside world. This, to a large extent
will influence the way they deal with changes as a result of their experiences elsewhere. The length of the
line connecting the community and the destination is an indication of the relative distance. The thickness
of the line shows average numbers of people who travel to that place. Thus, the wider it is the more
people go to the particular place.
Indicating the time of the year, week or month along the arrow, it helps the PRA team to recommend
periods when certain activities pertaining to interventions can best be arranged with the community for
the most positive results.
The daily routine diagram helps us to collect and analyze information on the daily pattern of activities of
community members and to compare the daily routine patterns for different groups of people (for
examples, women, men, old, young, employed, unemployed, educated, uneducated) and seasonal changes
in these patterns. Community members should be encouraged to draw their own daily routine diagrams.
It is similar to a seasonal calendar in that it helps identify time constraints (shortages) and opportunities.
For example, it can help in identifying the most appropriate time in the day for a training course for
women.
The daily routine for an individual can be completed through an interview, direct observation, or both. It
is useful to crosscheck results by using more than one method.
The daily activity profile adds a spatial dimension on the activity diagram and shows a person's mobility
during a typical day.
Example of a Daily Activity Profile
vii) Matrices
Matrices are used to study, collect, analyze and compare information on diverse subjects such as
resources use, resources use conflicts, constraints, opportunities, resources availability trends, and many
other topics.
It is a very important tool in PRA because of its flexible and adaptable use covering a wide range of
topics. Matrix construction generates a lot of discussion by the community in coming up with the
variables to be included in the matrices.
Details and explanations of these variables generate important information for the team. Matrices are also
good tools used to rank problems, opportunities, preferences and wealth. See the appendix for several
examples of matrices.
Some examples of ranking matrices include pair wise ranking, direct ranking, preference ranking.
b) Direct observation
Direct observation of important indicators is vital to crosscheck findings. The indicators can also be used
to generate on-the-spot questions to ask community without formal questionnaires.
Methods of direct observation include:
• Measurement with use of tapes, scales to measure field sizes, distances, weight and volumes of
produce in the field
• Indicators such as objects, events, process or relationships which are observable directly e.g. type
of housing related to socio-economy, vegetation showing soil fertility, types of dress;
• Recording in notebooks, record sheets, photographs, sample of objects (weeds, leaves, pests,
infested crops etc..)
• Sites such as markets, places of worship, homes, schools, paces of entertainment, roads, water
points, etc..)
• Use of the senses such as smell, taste, sight, hearing, touch (participate/share in community
activities);
(c ) Transect walks
A transect is a diagram of land use zones. It compares main features, resources use, problems and
opportunities of different zones.
Steps:
• Find community members who are able, willing and knowledgeable to walk the transect through
their village;
• Discuss with them different factors to be drawn;
• Walk the transect;
• Observe, ask and listen;
64
• Identify the main natural and agricultural zones and sketch distinguishing features of each zone;
• For each zone, describe: land slope and soil type, vegetation, land use, water points, forest/agro
forestry, socio-economic indicators, achievement in last 5 years, problems and constraints, and
opportunities;
• Draw the transect;
• Crosscheck the transect.
The transect provides information that verifies details on the resources maps. Responsibilities for
observations and note-taking are assigned to team members and preferably, according to profession. Some
parts of the transect can be walked, others driven. Transects can be sub-divided and assigned to two to
three groups. See appendix for example.
Farm sketches show individual decisions on resources use and practices re: management. They enable the
PRA team to compare resources use, opportunities, problems and strategies in different zones. They
compare, on a relative basis, socio-economic status within the community in terms of standards of
resources management, productivity, and income. Sketches reveal differences in farm size, cultivated
crops, livestock management, and other variables of household and resources use. Whatever families
decide on resources management affects collective decisions, opportunities and environmental impact of
the whole community. See appendix for example.
Steps:
1) Group interviews
These interviews are used to obtain community level information. They have several advantages which
include providing
1) access to a large body of knowledge;
2) immediate cross-checking on information;
3) opportunity to easily single out members with divergent views for further discussion;.
Groups of more than 25 people become difficult to manage for group interviews. At this stage, the logical
step is to further break up the group. Group interviews have limitations, e.g.
65
3) may reveal people's ideas and not what really exists (however, triangulation can reveal the true
picture).
3) Semi-structured interviews
Semi-structured interviewing is guided discussion where only some of the questions are pre-determined
and new questions come up during the interview.
The interviewers prepare a list of topics and questions rather than utilizing a fixed questionnaire.
Semi-structured interviews are held with:
• Individuals: For representative information, interviews are held with a number of different
individuals on the same topic.
• Key informants: For specialized information, these individuals have special knowledge which
others do not possess (e.g. midwives on birth complications).
• Groups: For general community-level information.
• Focus: To discuss a specific topic in detail.
Steps:
• Compose a team of 2 to 4 people of different disciplines
• Begin with the traditional greetings and state that the interview team is here to learn
• Begin the questioning by referring to someone or something visible
• Conduct the interview informally and mix questions with discussion
• Be open-minded
• Let each team member finish their line of questioning
• Carefully lead up to sensitive questions
• Assign one note taker (but rotate)
• Beware of non-verbal signals
• Avoid leading questions and value judgments
• Avoid "closed" questions, those that can be answered with "yes" or "no"
• Keep individual interviews to a maximum of 45 minutes
• Keep group interviews no longer than 2 hours
• Have a list of topics and key questions written down in your notebook
• Use the 5 helpers to ask questions: how, when, where, what and why
67
Tool Number 4: Stakeholder Participation in
Monitoring and Evaluation
I. Introduction
This tool clarifies the significance and different modalities of stakeholder participation in programme
monitoring and evaluation. Its content is based on a review of evaluation literature from academia and
international development agencies and NGOs.
However, as highlighted in Box 1, the principles guiding the participatory approach to M&E clearly
distinguishes it from conventional M&E approaches. Participatory M&E also requires a different
mindset, acceptance of a different way of conducting M&E.
Conventional M&E:
aims at making a judgment on the programme for accountability purposes rather than empowering
programme stakeholders
strives for “scientific”objectivity of M&E findings thereby distancing the external evaluator(s) from
stakeholders
tends to emphasise the needs for information of programme funders and policy makers rather than
programme implementers and people affected by the programme
focuses on measurement of success according to predetermined indicators.
Participatory M&E:
is a process of individual and collective learning and capacity development through which people
become more aware and conscious of their strengths and weaknesses, their wider social realities, and
their visions and perspectives of development outcomes. This learning process creates conditions
conducive to change and action
emphasises varying degrees of participation (from low to high) of different types of stakeholders in
initiating, defining the parameters for, and conducting M&E
is a social process of negotiation between people’s different needs, expectations and worldviews. It is a
highly political process which addresses issues of equity, power and social transformation
is a flexible process, continuously evolving and adapting to the programme specific circumstances and
needs.
Source: Estrella, 1997.
68
III. Who are the stakeholders?
M&E stakeholders are those people
who have a stake in the
programme. They are persons who Box 2. Types of Stakeholders
take decisions using the M&E data
and findings. The community whose situation the programme seeks to
change
Box 2 shows five types of Field Staff who implement activities
stakeholders. They can include Programme Managers who oversee programme
members of the community – men, implementation
women and youth; health clinic Funders and other Decision-Makers(BoD AGM) who
staff, teachers of population decide the course of action related to the programme
education, staff of the Census Supporters, critics and other stakeholders who influence
Bureau who implement the the programme environment.
programme activities; national
counterparts in government and
NGOs at the central and local levels who oversee programme implementation; international and national
programme funders and other decision-makers; community leaders, central and local government
administrators who have a major influence on the “enabling” programme environment.
However, exactly what programme stakeholders are involved in M&E varies according to the purpose of
M&E and the general institutional receptiveness to the use of participatory approaches. In each instance,
programme managers must decide which group of stakeholders should be involved, to what extent and
how.
The extent of stakeholder participation in evaluation, however, depends on the evaluation questions and
circumstances. Participatory evaluations are particularly useful when there are questions about
implementation difficulties or programme effects on different stakeholders or when information is wanted
on stakeholders’ knowledge of programme goals or their view of progress. A conventional approach to
evaluation may be more suitable when there is a need for independent outside judgment and when
specialized information is needed that only technical experts can provide. Such an approach is also more
appropriate when key stakeholders don’t have time to participate, or when such serious lack of agreement
exists among stakeholders that a collaborative approach is likely to fail.
negotiating stakeholder perspectives: participatory M&E may be used as a process that allows
different stakeholders to articulate and present their needs, interests and expectations. The process
of dialogue and negotiation among stakeholders used in participatory M&E facilitates
reconciliation of different stakeholder viewpoints. Difficulties may, however, arise in resolving
competing and conflicting stakeholder perceptions, especially when certain stakeholder groups are
powerless in relation to others.
4
Client-oriented, Provider-efficient. A COPE Handbook can be obtained from AVSC International.
70
ensuring public accountability: participatory M&E can be a way for programme participants and
local citizens themselves to monitor and evaluate the performance of donor and government
institutions. For instance, legal reforms that decentralize decision-making often encourage elected
representatives at district or union council levels to be more proactive in monitoring
implementation of local development plans.
The COPE approach involves a series of activities (self-assessment, client interviews, client-flow
analysis, and action planning) conducted by staff at health service delivery points to help them identify
and solve their own problems and continually improve service quality, using resources already available
at their facility.
In 1994 the NGO SALFA introduced COPE in Madagascar. By 1998 the method was used by 20
provincial level FP centres managed by SALFA and one government health centre. The experience
showed that the method can be applied to many service delivery management processes at clinics and
hospitals: for example in the areas of administration and service delivery such as management of staff and
supplies and provision of preventive and curative services. The opportunity for service providers to
contribute to the assessment and improvement of service delivery increased their sense of responsibility
and the enthusiasm for their work. The self-assessment process increased their awareness of the
importance of better client provider relations. As a result, service utilization improved significantly.
Health care providers tended to think that COPE would resolve all their problems, including the
lack of financial incentives for good service delivery. The introduction of COPE should
therefore emphasise its main purpose of achieving client satisfaction;
In their enthusiasm, health care providers tended to seek the perfect solutions even if sometimes
too ambitious. Therefore, slow achievement of results discouraged them and they lost faith in the
utility of the COPE approach. It is important to ensure that solutions proposed can be
implemented by health care providers themselves, within the health care facility and with
reasonable resources;
Clients interviewed by service providers did not express all their opinions about the services,
leaving out negative aspects. This COPE tool should therefore be applied by data collectors
independent from the service delivery point while ensuring client confidentiality;
The short-term results achieved with the introduction of COPE were not maintained at service
delivery points that did not continuously monitor the use of the COPE approach. Continuous
monitoring of COPE implementation is key to achieving the expected results;
Health care providers at the government facility were demoralized by the fact that their
supervisors rarely visited their health centre, despite official recognition of their excellent work.
Continuous supervision of COPE implementation by higher level supervisors is important to
sustain and improve results;
These health care providers also realized that decisions to solve several of the problems identified
needed to be taken at higher levels of the health administration. The introduction of COPE at the
individual service delivery point should therefore be combined with the necessary related
interventions at higher decision-making levels.
Source: “Le Processus ‘COPE‘ – L’Expérience de Madagascar”, Madagascar, 2000.
71
VI. Which stakeholders should participate in evaluation and
what role should they play?
Participation may be broad to include a wide array of programme staff, communities affected by the
programme, partners and others. It may, alternatively, target one or two of these groups. For example, if
the aim is to uncover what hinders programme implementation, field implementers may need to be
involved. If the issue is the impact of a programme on local communities, they may be the most
appropriate participants. If the aim is to know if all stakeholders understand a programme’s goals and
view progress similarly, broad participation may be best.
Roles may range from serving as a resource or informant to participating fully in some or all phases of the
evaluation.
How can communities be involved in evaluation? Community participation can be constrained by lack of
literacy skills, insufficient time, the intensity of analytical work to be undertaken during the evaluation,
and the fact that many of the issues covered during the evaluation are not directly relevant to community
members. There are different ways to make sure that the community perspective is considered. For
instance, prior to a programme evaluation, complementary evaluation activities could be undertaken with
communities involved in and affected by the programme. Such activities could include interviews with
and collection of data by community members. They could also consist of community members using
PRA and PLA tools5 to analyse programme activities and assess whether they meet their needs.
Alternatively, community members could define their own criteria for evaluating community-based
activities and use these criteria to carry out their own evaluation.
5
Participatory Reflection and Action (PRA) and Participatory Learning and Action (PLA) methods aim at: allowing community
members to express their perceptions, priorities, problems and plans for the future; allowing community development workers
to listen to and dialogue with community members in order to better understand their lives, perceptions, problems, priorities
and plans for the future. PRA/PLA tools include: community mapping; problem ranking; body mapping; role plays and stories
and other tools.
72
Table 1. Responsibilities of participants in “Expert-driven” and Participatory
Evaluations.
73
VII. What are the steps in a participatory evaluation process?
Box 5 illustrates suggested steps and lessons learned based on a participatory evaluation of 15 social
sector, health and agriculture projects implemented by NGOs in Haiti6. The individual project
evaluations were undertaken over a period of three weeks by teams of three to five NGO staff trained in
participatory evaluation approaches.
6
This is a summary of Francoise Coupal, Marie Simoneau. 1997: “Participatory Evaluation: A Case Study of CCIC
Humanitarian Fund Projects in Haiti”. Mosaic.net. 1997.
74
Box 5. (cont’d)
75
VIII. Elements of successful stakeholder participation in
evaluation
It is important to bear in mind that the local environment, the socio-political power structures and socio-
cultural norms and values, influence the evaluation process.
In this context the following are a few important elements of a successful process of stakeholder
participation in evaluation:
the support of programme management (implementing partners) and other direct decision-
makers for the participatory evaluation approach;
the evaluator is committed to the principle of participation; has sufficient group facilitation
and mediation skills to enable effective dialogue and discussion and to ensure that the experience
is both participatory and focused;
representation of the most important stakeholder interests related to the programme being
evaluated;
investigation into the “real” interests and issues of key stakeholders. The “real interests” often
differ from those, which are openly expressed. Box 6 highlights an approach to discover “real
interests” of stakeholders;
76
Tool Number 5: Planning and Managing an Evaluation
Part II: Defining Evaluation Questions and Measurement
Standards
I. Introduction
This part II of tool number 5 discusses the “what” of evaluation: steps to define evaluation questions and
measurement standards. The content is based on a review of evaluation literature from bilateral and other
development agencies such as such as Danida, ILO, Management Sciences for Health etc.
Performance issues relate to the actual programme results (see Box 1). Each of these issues is explained
in greater detail below.
77
Box 1. What do we mean by Result?
A result is a describable or measurable change in state that is derived from a cause and
effect relationship. Results are the effects generated by a programme.
Outputs Products and services that result from the completion of activities within a
development intervention.
Impacts Positive and negative long term effects on identifiable population groups
produced by a development intervention, directly or indirectly, intended or
unintended. These effects can be economic, socio-cultural, institutional,
environmental, technological or of other types.
Source: Tool Number 1: Glossary of Planning, Monitoring and Evaluation Terms, March 2004.
Validity of design
A good programme design guides the implementation process, facilitates monitoring of implementation and
provides a solid base for performance evaluation. In most of the organizations, issues of programme design
are assessed by using the programme logical framework.
Outputs, outcomes and impact (the results): are they clearly stated, describing solutions to
identified problems and needs?
Inputs and strategies: are they identified and are they realistic, appropriate and adequate to
achieve the results?
Indicators: are they direct, objective, practical and adequate (DOPA)8? Is responsibility for
tracking them clearly identified?
7
These questions are illustrative and should not be used as a “blue print”.
8
A Direct Indicator closely tracks the result it is intended to measure; an Objective Indicator is unambiguous about: 1) what is
being measured and data being collected; 2) has a clear operational definition that is independent of the person measuring the
indicator; a Practical Indicator can be gathered at reasonable cost and frequency, and can be available in time for use in decision-
making; an Adequate indicator constitutes the minimum necessary to ensure that progress towards results is sufficiently well
captured. Further details on indicators are provided in Tool Number 6 Part I: Identifying Output Indicators
78
External factors and risks: have factors external to the programme that could affect
implementation been identified and have the assumptions about such risk factors been validated?
Execution, implementation, monitoring and evaluation responsibilities: have they been clearly
identified?
Gender sensitivity: does the programme design address the prevailing gender situation? Are the
expected gender related changes adequately described in the outputs? Are the identified gender
indicators adequate?
Capacity building: does the programme include strategies to promote national capacity building?
Programme approach:
1. In the case of a programme evaluation, does the design clearly establish linkages among
programmes components?
2. In the case of a programme component evaluation, are linkages among its interventions
clearly established to ensure synergy in achievement of programme components results?
Delivery process
An assessment of the delivery process focuses on how the programme is being/was implemented to
determine if the programme has remained on the right track towards the achievement of planned results and
if not, what were the influencing factors.
Outputs: were the planned outputs achieved? Were they achieved within the planned time frame?
Were they of adequate quality? If not, why?
Programme management:
1. Did the programme implementers discharge their roles in a cost-effective and cost-efficient
manner? If not, why not?
2. Were sound financial and equipment management procedures practised? Were the financial,
human and material resources managed responsibly and efficiently?
3. Was the technical assistance provided appropriate and of good quality?
4. Did the monitoring and evaluation systems and processes allow for adequate assessment of
changes in risks and opportunities in the internal and external environments? Did they
contribute to effective decision-making in the course of programme implementation?
79
Performance
When assessing programme performance, evaluations look beyond the delivery process and focus on the
results of inputs delivered and the work done. The outcome of this assessment determines whether or not
the programme has achieved or is likely to achieve its outputs and contribute to achieving programme
outcomes and impact.
The core evaluation concerns to assess programme performance are illustrated in Figure 1 and described
below.
Effectiveness Relevance
Achievement of Programme
results continues to meet
needs
Sustainability
Results sustained after
Efficiency
withdrawal of external
Results vs. costs support
Evaluation is
concerned with
Validity of
Design Causality
Logical and Factors affecting
coherent performance
Relevance
An assessment of programme relevance examines the appropriateness of results in relation to: the needs,
policies, and priorities; the needs and priorities of programme target groups (the local programme context);
organizational policies and priorities and its comparative advantage vis à vis other CSOs & agencies and
development partners. The analysis ascertains whether the programme continues to make sense and
identifies any changes that may have occurred in its context during implementation. The initial problems
and needs may no longer exist and policies and priorities may have changed as a result of political,
80
economic, social and other factors, or even because of programme activities. Ultimately, the analysis
determines whether the results are still valid or should be reformulated.
Needs, mandates, policies and priorities: Do the programme planned results address the national
& local needs? Are they in line with the government’s priorities and policies? Are they in line
with organizations’s mandate? Does the target population consider them useful? Are they
complementary to other donor interventions? Should results be adjusted, eliminated or new ones
added in light of new needs, priorities and policies?
Effectiveness
An assessment of programme effectiveness focuses on the extent to which the outputs have been or will be
achieved and whether the programme is likely to contribute to the stated outcomes and impact. If not, the
evaluation will identify whether the results should be modified (in case of a mid-term evaluation) or the
programme be extended (in case of a final evaluation) in order to enable achievement of stated results.
Outputs: to what extent have planned outputs been or will be achieved? What is the quality of the
outputs?
Data on indicators: have data been collected on the indicators of achievement? Do they provide
adequate evidence regarding achievement of programme outputs and contribution to outcomes and
impact? Is it necessary to collect additional data?
Gender: what were the achievements in terms of promoting gender equity and equality
(planned/unplanned)?
Efficiency
81
Some key questions related to efficiency include:
Costs: did the actual or expected outputs justify the costs incurred? Have the resources been spent
as economically as possible?
Duplication: did programme activities overlap and duplicate other similar interventions (funded
nationally and/or by other donors)?
Alternative options: are there more efficient ways and means of delivering more and better outputs
with the available inputs?
Sustainability
An assessment of programme sustainability ascertains the extent to which the programme results have had or
are likely to have lasting results after programme termination and the withdrawal of external resources. The
factors affecting sustainability are examined on the basis of the priority assigned to the programme by
stakeholders. Their readiness to continue supporting or carrying out specific activities, or even replicate the
activities in other regions or sectors of the country, is particularly relevant. The analysis also assesses the
availability of local management, financial and human resources that would be needed to maintain the
programme results in the long run.
Likely sustainability: is it likely that programme achievements will be sustained after the
withdrawal of external support? Are involved counterparts willing and able to continue programme
activities on their own? Have programme activities been integrated into current practices of
counterpart institutions and/or the target population?
Resources: have they been allocated by the counterparts to continue programme activities?
Causality
An assessment of causality examines the factors or events that have affected the programme results. If the
inputs needed to carry out the planned activities and deliver the expected outputs were available on time, the
implementation and performance would be successful. If, on the other hand, there were significant deviations
from the planned schedules, the analysis would determine the reasons for such changes. The assessment
should also analyse the effect of other factors such as technical, administrative or managerial constraints,
inadequate inputs, failed commitment by programme counterparts, insufficient funds, faulty assumptions or
the effect of unexpected external factors.
What factors: what particular factors or events have affected the programme results?
82
A programme evaluation may find significant unforeseen positive or negative results of programme
activities. Once identified, appropriate action can be taken to enhance or mitigate them for a greater
overall impact.
Were there any unexpected positive and/or negative results of the programme?
How to address them: can they be either enhanced or mitigated to achieve the desired impact?
Alternative strategies
Evaluations examine whether alternative approaches might have had greater impact or might have been
more cost-effective, particularly if the original strategies turn out to be inappropriate. This analysis is
especially valuable when follow- up programmes are planned.
More effective approaches: is there, or would there have been, a more effective way of
addressing the problem(s) and satisfying the needs in order to achieve the outputs and contribute
to higher level aims?
Not all of the above evaluation objectives have to be examined in every evaluation. The final choice will
depend on the purpose of each evaluation9. For instance, a formative evaluation undertaken in the course
of programme implementation with the aim of taking decisions to improve it’s design and/or
implementation would typically emphasise concerns of design, delivery process, efficiency, causality,
unanticipated results, and alternative strategies.
A summative evaluation, undertaken at the end of programme implementation to judge its effectiveness,
would typically concentrate on concerns of relevance, effectiveness, efficiency, alternative strategies and
sustainability.
An evaluation, which aims at extracting lessons learned and best practices or defining policy options
would assess design, delivery processes, causality and efficiency in order to extract those characteristics
which can effectively and efficiently deliver the desired results.
9
For a discussion on the issue of evaluation purpose, consult Tool Number 3: Purposes of Evaluation.
83
III. Methodological Challenges
Evaluator(s) face a number of methodological challenges with respect to the standards10 they use to
measure relevance, effectiveness, efficiency and sustainability. These standards and methodological
challenges are summarized in Table 1 and further discussed below.
Effectiveness Agreed outputs, outcomes and Unclear, multiple, confusing or changing results
impact. statements.
10
A standard is a level of performance according to specified criteria or achievement dimensions.
84
With respect to the relevance of programme strategies, it requires in-depth field analysis by evaluator(s)
to adequately determine the continued relevance of the planned programme results if the context (needs,
priorities and policies) were not clearly defined at the time of programme design or in the course of
implementation. Box 2 provides an example of such a situation from a commissioned evaluation of an
organization.
Box 2. The importance of using contextual information for programme design and
adjustment.
“The second objective stated in the project proposal was that “18,000 new users of family
planning will have been recruited and continuation rates will be improved.”
The problem with this objective is twofold. First of all, the target population was
overestimated and the target recruitment numbers were too ambitious. The number of
target population does not appear to be based on the available scientific evidence
(estimates of the district population published by the Palestinian Central Bureau of
Statistics). Secondly, the manner in which the numbers were determined is not clear from
the proposal document. Furthermore, the project staff did not seem to be aware of this
target and were unable to explain it. Still, the unrealistically high expectations seemed to
put pressure on the staff to generate high numbers of new users and to reflect the numbers
in their reported statistics, to the exclusion of other important activities that were being
carried out.
As for the objective of improving continuation rates, it was not clear how that would be
verified, since no baseline figure for continuation rates in those communities existed.
As a consequence of lack of proper initial situation analysis and adjustment in the course
of project implementation, a large proportion of the budget (80,000 USD) was allotted to
the purchase of contraceptives many of which eventually expired on the shelf due to lack
of demand.”
Source: Halabi, January 2000.
Defining objective efficiency standards is a major challenge for evaluators of UNFPA’s programmes and
projects. In practice the evaluator(s) frequently rely on their expert judgment, which can be subjective.
However, approaches are available to define standards, among others in the health field. The Continuous
Quality Improvement tool to strengthen FP programmes is one such approach, which, if used in the
course of programme implementation, greatly facilitates monitoring and evaluation of programme
efficiency (see Box 3). Another good approach for identifying efficiency standards is “benchmarking”,
analysing the performance of organizations, which excel in areas of work relevant to UNFPA.
Box 3. Using the Continuous Quality Improvement (CQI) approach to define efficiency
standards
CQI is performed by teams of staff at RH service delivery points. The CQI team implements the 7
Step 2: Define a problem within that area, and outline the sequence of activities (the
process) that occurs in that problem area
Step 3: Establish the desired outcomes of the process and the requirements needed to
achieve them
Step 4: Select specific steps in the process to study and for each step, list the factors that
prevent the achievement of the desired outcome
Step 5: Collect and analyze data about the factors that are preventing the achievement of the
desired outcomes of the specific step being studies, and quantify the outcomes of that
step
In step 3 the CQI team defines the standards of efficiency against which services will be
monitored and evaluated. The following is an example of such a standard:
Source: The Family Planning Manager, Volume II, Number 1, January/February 1993.
86
Long-term programme sustainability is hard to foresee as many factors intervene over time. For
instance, governments change and so may policies that are critical to support certain programmes
originally operated by the organization. With a government change, key administrators also change and
with them the institutional memory so necessary to keep particular approaches and programmes running.
A severe economic crisis may appear, jeopardizing funding for the programme. However, programme
designers must ensure that the sustainability factors listed in Box 4 are fully considered at the time of
situation analysis and programme design. Evaluator(s) assess the likelihood of sustaining programme
activities using the same standards.
87
Tool Number 5 : Planning and Managing an Evaluation
Part III: The Data Collection Process
I. Introduction
Part III of tool number 5 discusses the “how” of programme evaluation, namely the data collection
process, including determination of data collection methods, data analysis and interpretation. The content
is based on a review of evaluation and other literature from bilateral and other development agencies such
as such as Danida, Management Sciences for Health, Save the Children, UNFPA, UNICEF, USAID, and
the W.V. Kellogg Foundation.
11
For a detailed discussion of these concepts, see Tool Number 3: Purposes of Evaluation and Tool Number 5, Part II: Defining
Evaluation Questions and Measurement Standards.
88
Box 1. Useful questions to help determine the need for additional data.
How will the additional information be used? It is important to collect only the
information, which will be used and to use all the information collected.
How credible are different types of data to the intended users of evaluation
results? The level of credibility of data sources and data collection methods
determines the acceptance and use of evaluation conclusions and
recommendations by the intended users.
When is the information needed? Time constraints may determine the length
and nature of additional data collection exercises.
What resources are available for the evaluation? The availability of expertise
and financial resources determines the sophistication of additional data
collection.
The next step is to identify how to collect the additional data required. Quantitative and qualitative data
collection methods as well as deductive and inductive analytical approaches can be used for this
purpose.
⇒ questioning people through individual and group interviews such as focus group discussions
and community interviews;
⇒ conducting surveys;
⇒ observing people, processes, objects, conditions, and situations.
Quantitative and qualitative data collection methods each have their strengths and weaknesses and are
suited to answer different types of questions as highlighted in Table 1.
89
Table 1. Characteristics of Quantitative and Qualitative Data Collection
Quantitative Methods Qualitative Methods
Use To numerically measure “who, what, To qualitatively analyse “how and why”
when, where, how much, how many,
how often”
Examples Standardized interviews; surveys using Free and guided interviews (including focus
closed-ended questions; observation. group); surveys using open-ended questions;
observation; interpretation of documents.
Weaknesses May be precise but not measure Are generally not representative; do not allow
what is intended generalizations
Cannot explain the underlying Susceptible to biases of interviewers,
causes of situations. observers and informants.
In answering evaluation questions, to avoid dependence on the validity of any one source, it is highly
recommended to use a combination of different types of qualitative and quantitative data collection
methods such as review of statistics, small-scale surveys, interviews and observation (also called data
triangulation). Observation is an invaluable methodology to collect data that surveys and interviews
cannot accurately capture. For instance, observation is necessary to assess client/provider or
student/teacher interaction on sensitive subjects such as sexual and reproductive practices. Relying only
on surveys and interviews in this situation may not yield accurate information as respondents tend to
report ideal not actual behaviours.
Similarly, carefully study of various materials produced by the programme such as IEC materials (on
90
sexuality, HIV/AIDS prevention etc.), training modules, policies, and guidelines, can provide valuable
information and insights on how the issues are tackled. For example, by reviewing IEC materials, an
evaluation of a UNFPA funded HIV/AIDS prevention project found that brochures designed to increase
awareness on ways to avoid becoming HIV infected did not mention condoms but recommended that
“people not go dancing in places where one can catch HIV/AIDS!”
Finally, quantitative surveys do not enable exploration of underlying causes. Thus, a combination of
methods provides a more complete analysis of the subject matter being evaluated thereby enhancing the
credibility of the evaluation conclusions and recommendations. Box 2 summarizes a few criteria to guide
selection of methods to collect additional evaluation data.
Tie method selection to available resources. This may mean revising the
evaluation design and methods, or determining other options to stay within
budget. It may also mean finding additional resources to fund the most effective
and useful evaluation design.
Evaluator(s) are not only concerned with what data collection methods to use in order to adequately
address evaluation objectives and question. They also need to select a certain analytical approach to
gathering information. When using a deductive approach, evaluator(s) formulate hypothetical answers to
the evaluation questions at an early stage of the evaluation process based on available information and the
evaluator(s) accumulated knowledge of the subject matter being evaluated. Data is then collected to
confirm or refute these hypotheses. When using an inductive approach, the evaluator(s) start with an
open, questioning mind. They gradually uncover issues and themes through iterative field observation,
interviews and data analysis thus leading to a deeper understanding of the subject matter.
While most evaluations rely on a combination of the two approaches, a deductive approach would be
suitable for addressing evaluation objectives of efficiency and effectiveness. A deductive approach would,
for instance, be used to examine whether the best results were achieved with the inputs provided and
activities implemented and whether the planned results were achieved. An inductive approach would be
very useful for addressing evaluation objectives of relevance, impact and sustainability. It is particularly
91
useful for evaluating socio-cultural aspects of a programme because there is limited knowledge about the
cause-effect relationships among programme inputs, outputs and outcomes.
The evaluation information collected must be described, analyzed, interpreted, and a judgment made
about the meaning of the findings in the programme context. Interpretation involves looking beyond the
raw data to ask questions about what they mean, what the most significant findings are, and what
conclusions and recommendations should be drawn from these findings. A few basic techniques for
organizing and analyzing data are described below.
Quantitative Analysis
Quantitative data analysis interprets the numerical findings considering the programme context. As
implementers of programme activities are most knowledgeable about the context, they should work
together with the evaluator(s) to assess whether the figures make sense; whether they adequately reflect
programme results; what possible explanations are for unexpected figures; what conclusions and
recommendations can be drawn from the figures.
While some accounts resulting from in-depth interviews and focus group discussions are stand-alone
illustrations of important themes of the evaluation, it is, in most cases, valuable to analyze qualitative data
more systematically.
Analysis of qualitative data from interview transcripts, observation field notes or open-ended surveys can
identify similarities across several accounts, as well as directions, trends and tendencies. Data can be
categorized into recurrent themes and topics that seem relevant to answer the evaluation questions and to
develop new or test already selected hypotheses.
However, evaluators run the risk of drawing hasty conclusions and making generalizations when breaking
transcripts and field notes up into thematic categories. They can avoid this problem by writing case
studies and narrative summaries, which highlight the context and particular characteristics of key pieces
of the programme being evaluated.
Another problem frequently encountered when analyzing qualitative data is the excessive focus on
“quantifying” qualitative data and interpreting qualitative data as if it were quantitative data. For
example, when analyzing and interpreting focus group discussion data, some evaluators tend to tabulate
the responses and report on them in terms of ratios and percentages rather than exploring further the
information, ideas, opinions and attitudes which can help answer the evaluation questions “why?” and
“how?”
Biases and values inevitably influence both quantitative and qualitative data analysis. Evaluator(s) control
92
for biases and values by triangulating multiple types of data (quantitative and qualitative), data collection
methods, data sources, and perspectives or theories to interpret the data for instance by involving
stakeholders in analyzing and interpreting the data. Evaluator(s) can also ensure that they pay attention to
the influences of biases and values through an ongoing process of writing descriptive memos about the
evaluation process, their data, and their interpretations.
93
Annex 1. Data Collection Methods
The following list and description of data collection methods is not intended to be exhaustive. It is rather
an overview of the key characteristics of the most commonly used data collection methods. These may be
applied not only for evaluations but also at other stages of the programme cycle such as situation analysis,
programme design, monitoring and reviews. Each method may be explored further in the referred
sources.
1. Programme specific information such as reports of programme progress, field monitoring visits,
programme reviews, surveys, research and evaluations.
Such documents enable the evaluator to learn about the history, context, results, and outcomes of a
particular programme. They also provide clues about important shifts in programme development and
implementation. A document review may also be a good way to formulate questions for use in a
survey or interview.
2. Other information not directly related to the programme such as research studies; government data
such as clinic based statistics; and evaluations of similar programmes and projects. Evaluation
databases such as the UNDP CEDAB and IFAD EKSYST12 are good sources for increasing
knowledge of lessons learned on issues which are present in all development programmes such as
gender, capacity-building, and collaboration with NGOs.
It should be noted that written documents do not necessarily provide comprehensive or correct
answers to specific problems, as they may contain errors, omissions, or exaggerations. They are
simply one form of evidence, and should be used carefully and together with other types of data.
B. Questioning People.
1. Interviews such as Key Informant, Focus Group Discussion and Community Interviews, and Nominal Group
Technique.
General Characteristics
94
Interviews, together with document reviews, are the most frequently used data collection method in
UNFPA evaluations. Unstructured and guided interviews yield qualitative data. In unstructured
interviews, the interviewer’s only guide are the evaluation objectives. Unstructured interviews are a
good tool for exploring the opinions of respondents and uncovering unexpected factors. In a guided
interview, the respondent is asked to provide information about items on a prepared checklist.
Standardized interviews yield quantitative data. They use a questionnaire with a fixed number of
questions and sometimes a pre-selected range of possible answers.
In general, the more open-ended the interview the more deeply the respondents’ feelings and
perspectives can be understood; the more structured the interview, the more comparable the data.
Many reports based on questionnaires provide an array of facts (percentages, breakdowns) but shed
little light on people’s motivations that could be built on to improve practices.
One of the first steps in interviewing is to identify knowledgeable informants, people who can provide
pertinent and reliable information. Informants can be clients at service delivery points, programme
implementing partners, community members, local leaders, politicians, or health professionals.
Depending on the type of information needed, informants can be interviewed individually or in
groups.
In-depth Interview
If the evaluator(s) are concerned about maintaining the informants’ anonymity or simply want to
make sure that they feel free to express controversial ideas, it is best to interview informants
individually. This also allows the evaluator(s) to compare various perspectives of an event, which is
particularly useful when exploring sensitive topics.
A key informant interview is a form of in-depth interview often used. Key informants are selected for
their first-hand knowledge about the topic of interest. For example, the head of an HIV epidemiology
unit may act as a key informant on information relating to the incidence of HIV/AIDS. Traditional
birth attendants would be key informants for information on traditional management of pregnancy and
delivery13.
Group Discussion
When confidentiality is not a concern, and the evaluator(s) are interested in quickly
sampling a range of opinions on a topic, a group discussion is preferable. There are several types of
group discussions. Focus group discussions, community and other types of group interviews are
among those frequently used.
A Focus group discussion is an inexpensive, rapid appraisal technique through which a facilitator
guides 7-10 people in a discussion of their experiences, feelings and preferences about a topic. The
facilitator raises issues identified in a discussion guide and uses probing techniques to animate the
discussion and promote in-depth reflection among focus group participants. Sessions typically last
one to two hours. The facilitator’s discussion guide should contain few items thereby allowing some
time and flexibility to pursue unanticipated but relevant issues. In order to maximize exchanges
95
among focus group participants they should share certain common characteristics, i.e. be of same sex,
age group, and social background and have similar concerns. Many participants in focus group
discussions find the interaction stimulating and mention things they would not have thought of
individually14.
In community interviews, which usually take the form of community meetings open to all, interaction
is between the interviewer and the participants. Such meetings are susceptible to manipulation from
the more powerful members of the community and are less suitable to discuss sensitive issues.
In this technique, from five to seven people are asked by a leader to generate ideas on a single topic.
Through discussion, a consensus is reached on a list of most important ideas. A single session, which
deals with a single question, usually takes about 60-90 minutes. The nominal group technique was
developed to facilitate efficient group decision-making by busy private sector executives. It may also
be useful in evaluation, particularly when groups composed of experts, community members, or
programme staff are making recommendations for ongoing programmes.
Box 3. To ensure reliability, validity and avoid bias when questioning people:
96
2. Surveys
Formal large-scale sample surveys (for instance DHS surveys): a randomly drawn representative
sub-group from which researchers generalize about the whole population.
Informal small-scale sample surveys (for instance KAP surveys): a small non-random (such as
purposeful selection of people in different categories on the basis of easy accessibility) sample of 30-
50 individuals who are asked a few questions (10-20).
Large-scale technically complex surveys should be avoided in programme evaluation as they are
expensive and time-consuming. Informal, small-scale sample surveys can, however, provide useful
quantitative data, for instance, on use of and access to RH services to complement other qualitative
evaluation data.
Survey questions can be open-ended or closed-ended. Open-ended questions might ask: How do you
feel about the program? What do you want to see happen in your community? Open-ended questions
provide relatively rich information about a topic and allow participants to report thoughts, opinions
and feelings. However, there are disadvantages. Sometimes people are reluctant to express opinions,
or the survey may be time-consuming to complete and analyze.
Closed-ended questions have the advantage of uniformity and easy translation for statistical analyses.
Surveys with closed-ended questions can easily be administered to large groups of people and are
usually easy to complete. However, they tend to impose a set of fixed ideas or values on the
respondent by forcing choices from a limited array of options. As a result, they are less likely to
uncover new and unexpected information, and they limit the emergence of in-depth understandings
and nuances of meanings. In general, written survey questions are inappropriate if the respondents
have low literacy or are unfamiliar with the conventions of survey completion. A survey administered
in person might be more appropriate for this population.
A survey is only as good as the people administering it, so care should be given to selecting, training
and supervising surveyors.
97
C. Observation
Evaluator(s) record what they see and hear at the project site using an observation checklist.
Observation may be of physical surroundings, ongoing activities, processes and discussions.
When performance monitoring data indicate that results are not being accomplished as
planned, and when implementation problems are suspected, but not understood. Direct
observation can help identify whether the process is poorly implemented or required inputs are
absent;
When details of an activity’s process need to be assessed, such as whether tasks are being
implemented according to standards required;
When an inventory of physical facilities and inputs is needed and not available from existing
sources;
When interview methods are unlikely to elicit needed information accurately or reliably, either
because the respondents don’t know or may be reluctant to say;
D. Rapid Appraisal
Rapid appraisal is essentially the use of a mix of the above-described methods in order for decision-
makers to obtain timely, relevant, accurate and usable information on development programmes and
projects. Key informant, focus group, and community interviews, observation and informal surveys
are the methods most commonly used by rapid appraisal16.
98
Tool Number 5: Planning and Managing an Evaluation
Part VI: Evaluation Standards
I. Introduction
This part VI of the tool number 5 lists evaluation standards to be applied throughout the evaluation
process to ensure the quality of the evaluation product. In addition to a review of the literature from
bilateral and development agencies such as OECD, DANIDA and U.S. Centers for Disease Control and
Prevention, the content is based on Program Evaluation Standards developed by the Joint Committee on
Standards for Educational Development (1994, 1999) and the adapted evaluation standards recommended
by the Swiss Evaluation Society (SEVAL).
For evaluations to be useful they should meet certain standards of quality. The international community of
evaluators has established standards for sound and fair evaluation, which can be applied while planning
an evaluation and throughout its implementation. Some of these standards are considered as universal
while others are perhaps more unique to certain cultural settings. As such, their application should be
adapted taking into account the specific situation. These standards are organized around four important
attributes of evaluation:
Utility Standards
The Utility standards should ensure that an evaluation is guided by the information needs of its
users. These standards are as follows:
Those who decide upon the future of the programme (often the
funding agency)
Those who are responsible for the planning and design of the
programme
99
Other groups with an interest in the evaluation findings (e.g.,
decision makers who plan similar programmes, evaluators,
and the general public).
Evaluator Credibility - Those conducting an evaluation should be both trustworthy and competent, so that
the evaluation findings achieve maximum credibility and acceptance. To be found credible by different
stakeholder groups, the following characteristics are crucial: professional competence, integrity,
independence, as well as social and communication skills.
Information Selection - The information collected should be comprehensive enough to address pertinent
questions about the programme and be responsive to the interests and needs of stakeholders. When
planning an evaluation, it is also important to distinguish information that is essential versus information
that is desirable.
Transparency of Assessment - The perspectives, rationale, and procedures used to interpret the findings
should be carefully described, so that the bases for value judgments are clear.
Report Clarity - Evaluation reports should clearly describe the programme being evaluated including its
context, and the purposes, questions, procedures, and findings of the evaluation. The language should be
precise (e.g., clear definitions of the most important terms and consistent use of terminology) and easily
understood by the intended audience.
Report Timeliness - Significant interim findings and final reports should be brought to the attention of
intended users, so that they can be used in a timely fashion. Evaluations are most useful when planned to
fit into the stakeholders’ decision-making processes. For many evaluations it is sensible to share interim
findings with the stakeholders, especially when these results might have an impact on their future actions.
Evaluation Impact - Evaluations should be planned, conducted, and reported in ways that encourage
stakeholder participation to varying degrees, so that the likelihood that evaluation results will be used is
increased. The more involved stakeholders are at the different stages of the evaluation process, the greater
the likelihood they will act on the evaluation recommendations.
Feasibility Standards
The Feasibility standards should ensure that an evaluation is carried out in a realistic,
thoughtful, tactful, and cost-effective manner. These standards are as follows:
Practical Procedures - Evaluation methods and instruments should be practical to keep disruption to a
minimum while the needed information is collected. It is crucial to discuss with stakeholders the
advantages and disadvantages of chosen methods.
Political Viability - The evaluation should be planned and conducted taking into account the different
positions of the various interest groups, in order to obtain a balanced presentation of different points of
view. It should enlist their cooperation and avert or counteract possible attempts to curtail evaluation
activities or to bias the results.
100
Cost Effectiveness - Evaluations should produce information of sufficient value for informed decision-
making, learning and accountability so that the resources expended can be justified.
Propriety Standards
The Propriety standards should ensure that an evaluation is conducted legally, ethically, and
with due regard for the welfare of those involved in the evaluation as well as those affected by its
results. These standards are as follows:
Formal Agreement - Obligations of the formal parties to an evaluation (what is to be done, how, by whom,
when) should be agreed to in writing, so that they are obligated to adhere to all conditions of the agreement
or to renegotiate it. Such a formal written agreement should at least regulate budget, time, personnel,
design, methodology and report contents.
Protection of Individual Rights - Evaluations should be designed and conducted in a way that respects
and protects the rights and welfare of human beings. If an evaluation leads to well-founded conclusions
that pose a threat to the welfare of individuals, the extent to which these findings are disseminated needs
to be carefully considered and justified.
Human Interactions - Evaluators should respect human dignity and worth in their interactions with other
persons associated with an evaluation so that participants are not threatened or harmed. This is not only a
matter of human dignity but also relates to practical considerations. It is therefore necessary to be familiar
with the cultural practices (i.e. beliefs, manners and customs) of those involved.
Complete and Fair Assessment - Evaluations should be complete and fair in their examination and
recording of strengths and weaknesses of the programme being evaluated, so that strengths can be built
upon and problem areas addressed. If, for whatever reason (e.g., because of time or budgetary
constraints), there are issues that cause difficulties for the evaluation (e.g., it was impossible to collect
certain data) these should be brought to light.
Disclosure of Findings - The formal parties to an evaluation should ensure that the full set of evaluation
findings is made accessible to the persons affected by and/or interested in the evaluation.
Conflict of Interest - Conflict of interest should be dealt with openly and honestly so that it does not
compromise the evaluation process and results. It is therefore crucial that evaluators be able to clarify
their roles and make a distinction between facts and opinions. The integrity of the evaluation cannot be
compromised just to accommodate conflicts of interest.
101
Accuracy Standards
The Accuracy standards should ensure that an evaluation would reveal and convey technically
adequate information about the features that determine the value of the programme being
evaluated. These standards are as follows:
Programme Documentation - The Programme being evaluated should be described and documented
clearly and accurately. The description should be sufficiently detailed to ensure an understanding of
programme aims and strategies. It is especially crucial to note differences between the planned and the
actual performance of the programme.
Context Analysis - The context in which the programme exists should be examined in enough detail so
that its likely influences on the programme can be identified. Understanding the setting in which a
programme functions will help in the accurate interpretation of evaluation findings and in assessing the
extent to which they can be generalized.
Described Purposes and Procedures - The purposes and procedures of an evaluation should be monitored
and described in enough detail so that they can be identified and assessed. It is important that the
evaluation process focus on the issues of greatest concern to stakeholders while using time and resources
as efficiently as possible.
Defensible Information Sources - The sources of information used in a programme evaluation should be
described in enough detail so that their adequacy can be assessed. The criteria used for selecting sources
should be stated clearly so that users and other stakeholders can interpret the information accurately and
assess if it might be biased.
Valid and Reliable Information - The information gathering procedures implemented should provide
assurance that the interprtetation arrived at is valid and reliable. Validity is defined by the extent to which
methodologies and instruments measure what they are intended to measure. A data collection method is
reliable to the extent that it produces the same results repeatedly.
Systematic Review of Information - The information collected, analyzed, and reported in an evaluation
should be systematically reviewed and any errors found should be corrected.
Analysis of Qualitative and Quantitative Data - Qualitative and quantitative data should be analyzed in an
appropriate, systematic way so that the evaluation questions can be effectively answered. Data analysis
should follow rules of methodological soundness.
Justified Conclusions - The conclusions reached in an evaluation should be explicitly justified so that
stakeholders can assess them. Evaluation conclusions are justified when they are based on a synthesis of
empirical findings derived from the information collected. Evaluation information must be interpreted to
appreciate the practical significance of what has been learned. Conclusions can be both positive and
negative. Controversial conclusions should be substantiated.
Impartial reporting - Reporting procedures should guard against distortion caused by personal feelings
and biases of any stakeholder group. All relevant perspectives need to be fairly represented.
102
Meta evaluation - The evaluation itself should be subject to an assessment of the evaluation’s process and
quality upon its completion using these and other pertinent standards to determine its strengths and
weaknesses.
103
Tool Number 6: Programme Indicators
Part I: Identifying Output Indicators - The Basic Concepts
I. Introduction
With adoption of results-based management as a guiding principle for programming, indicators have
become important instruments for organizations to measure the results of its development assistance at the
output, outcomes and goal levels. Indicators, particularly those at the output level of results, for which
organizations programmes are responsible, provide essential information for successful programme
management. In-house reviews of programme logframes have found that the Logical Framework/ Results
and Resources Framework indicators at the output level and their associated MOVs are often not well
identified17.
Output indicators should tell us how the programme is performing. They are the detailed expressions of
the programme results for which organization is responsible. Thus, before programme stakeholders
identify output indicators, they must reach a consensus on the content of the output statements. Box 1
shows how to.
Box 1. How to clarify the outputs?
Review the wording and intention of the output. What exactly does it say?
Avoid broad output statements. They make indicator identification difficult.
Be clear about what type of change is implied in the output. What is expected to change – a
condition, level of knowledge, attitude?
Be clear about where change should appear. Among individuals, counterpart organizations,
communities, regions?
Specify in more detail the targets for change. Who or what are the specific targets for change?
What kind of individuals, organizations etc.?
Source: Adapted from USAID TIPS 1996, number 6.
“ improved awareness and knowledge on reproductive rights and reproductive health of women, men and
adolescents”
These statements lack clarity and specificity; they leave considerable room for interpretation by
programme implementers and managers:
What is meant by “access”? “quality” RH services? What elements of RH services are being targeted?
What is “strengthened planning and management capacity”? What is the difference between “awareness”
and “knowledge”? What is “reproductive rights? Are these outputs to be achieved for the whole country
or special target areas? The public or the private sector? Any target population?
Some output elements, such as target groups or geographical location, can be specified very briefly in a
footnote to the logframe/results and resources framework or in the indicators. Other output elements,
such as quality of care or specific knowledge and attitudes expected from the target groups, are more
complex and may require more work after elaboration of the logframe/results and resources framework
18 Women, men and adolescents and sub-groups of these sharing common characteristics, socio-economic groups.
19 Essential obstetric care; family planning; prevention and management of RTI/STD and prevention of HIV infection;
management of the consequences and complication of unsafe abortion; information, education and counselling.
21 Provider performance/service delivery according to protocols; performance of systems such as training, supervision,
management, health information system, logistics; client satisfaction.
22 Such as the presence of a strategic plan; of a system for preparing yearly operational plans for the organization; of a regular
system to assess the needs and preferences of clients and to adjust services in response to identified changes; of a manager
whose job description includes assessing clients’ needs, developing the strategic and operational plan, revising and assessing
the operationalisation of the plan
105
for example to identify specific standards and to incorporate them in checklists that can be used for
monitoring indicator progress in the course of programme implementation.
Step 2: Identify the best indicator or cluster of indicators and the performance targets to track each
output
Various organizations use different criteria to select relevant indicators23. DOPA criteria (explained in
Box 2) encapsulate the most important requirements of useful indicators. They are a simple tool to guide
us in the indicator selection process.
Direct
closely measure the intended change.
Objective
unambiguous about what is being measured and which data to be collected.
clear operational definition that is independent of the person conducting the measurement.
Practical
reasonable in terms of data collection cost, frequency, and timeliness for decision-making
purposes.
Adequate
the minimum number of indicators necessary to ensure that progress towards the output is
sufficiently captured.
23 Many organizations use SMART which stands for: Specific (measures exactly the result); Measurable (so that the result can
be tracked); Attainable (realistic); Relevant (to the intended result) and Timebound (indicates a specific time period ).
106
Box 3 illustrates common problems with respect to donor funded programme logframe/results and
resources framework output indicators. Problems 1 and 6 highlight that output indicators do not
correspond to the output level: they are not direct.
If it is not possible to use a direct measure, one or more proxy indicators might be appropriate. A proxy
indicator is an indirect measure that is linked to the result by one or more assumptions. For example, the
contraceptive prevalence rate can be considered a proxy indicator for access to and utilization of RH
services. Problem 2 highlights that output indicators are often ambiguous or lacking detail: they are not
objective. Therefore there may not be consensus among programme and project implementers and
managers about what is being measured and what are the results of an intervention.
Problems 4 and 5 highlight that organizations logframe matrices/results and resources framework often
include too many and/or unrealistic output indicators, which it would be impossible to adequately manage
and which are not all necessary to capture progress in achieving the output: they are neither practical nor
adequate. Instead, programme stakeholders should select one or a cluster of two or three indicators
which are practical because they allow managers to track the output in the most direct and timely fashion
and require the least effort in terms of time, human and financial resources, and adequate because
together they represent the most important dimensions of the planned output.
A desk review of UNFPA Country Programme documents highlighted some weaknesses in specifying
output indicators:
1. Indicators do not correspond to the output level. For example, for an output of “improved
availability of RH services” use of an activity indicator, such as quantity of equipment
procured, or of an outcome indicator, such as the contraceptive prevalence rate. Quantity of
equipment procured does not alone ensure that RH services are available. Additionally,
availability of RH services does not necessarily lead to increased use of such services and
therefore increased CPR.
2. Indicators do not include an objective standard against which achievement can be assessed. For
example, an indicator of “system developed” is used for an output such as “system for
coordination, monitoring and evaluation of population programmes”. The standard needs to
be defined explicitly.
4. Too many indicators with little consideration of the time, human resources and cost required to
collect the indicator data.
5. Indicators that seem unrealistic due to lack of data to construct the specified indicator and/or
because the indicator is very difficult to measure.
6. Inconsistency between the universe of the output and the indicators. For example, the output
relating to a few sample areas but the specified indicators relating to the entire country.
Indicators tell us what we are measuring; targets are the results expected in the context of the specific
programme and within a certain time frame (see example in Box 5). While organization’s guidelines
require that the programme logframe/results and resources framework output indicators include targets,
this has often not been possible for lack of sufficient knowledge on the baseline situation at the time of
preparing the programme logframe matrix/results and resources framework. Thus, the output indicators
currently stated in most of Organizations’s logframe matrices/results and resources frameworks do not
include targets. It is expected that it will be easier to identify targets in the course of programme
implementation and in the development of future country programmes as the required surveys and/or
research will then have been completed.
108
What are different types of
indicators and targets?
Indicators and targets may express Box 6. Common ways of expressing Quantitative,
quantity (how much), quality (how Qualitative and Efficiency Indicators/Targets
good), or efficiency (best output at
lowest cost). Box 6 illustrates Quantitative indicators/targets are statistical measures
common ways of expressing these Number
different types of indicators and Percent
targets. Each type of indicator and Rate (ex. birth rate - Births per 1,000 population)
target conveys a different dimension Ratio (ex. sex ratio – Number of males per number
of the planned output. For example, of females).
quantitative indicators and targets
Qualitative indicators/targets imply qualitative
provide “hard data” to demonstrate
assessments
results achieved. They also facilitate
Compliance with
comparisons and analysis of trends
Quality of
over time. Qualitative indicators and
Extent of
targets provide insights into changes
Level of.
in organizational processes,
attitudes, beliefs, motives and
behaviours of individuals24. Efficiency indicators/targets are unit cost
Qualitative indicators and targets measures
must be expressed quantitatively in Cost per unit of (clients treated, student, school
order to illustrate change. This can etc.).
for instance be done by using a
scoring system. A scoring system to
track improvement in the quality of RH services over time could include an indicator such as “the percent
of existing SDPs with a score of four out of a total of five points on a quality of care checklist increased
from X to Y”.
Efficiency indicators should tell us if we are getting the best value for our investment. In order to
establish such an indicator, we need to know the “market”, i.e. the current price of desired outputs
considering both quantity and quality aspects.
109
How can we identify targets?
In setting targets it is important to be realistic about the outputs that are feasible to achieve given
contextual constraints and past experiences in a particular sector. Box 7 provides a few suggestions of
useful information for target setting.
Step 3: Identify the Means of Verification (MOVs), timing and reporting responsibility
The indicator MOV is the data that is needed to determine the value of the indicator. MOV data can be
collected through review of documentation, facility observation, in-depth interviews, focus group
discussions, small surveys such as facility based quality of care surveys. In order to save time and effort,
the feasibility of using existing data systems and sources, such as Demographic and Health Surveys,
Health Information Systems, government or NGO administrative records, to provide some of the output
indicator data should be explored. Essential output data for which there are no existing sources should be
collected as an activity, for instance rapid appraisal surveys, in the context of the funded programme. It is
important, before planning a data collection activity in the donor programme, to check the data collection
plans of other actors or partners in the concerned sector. For instance, Donor may negotiate the collection
of data useful for tracking programme progress as part of other CSO’s/ agencies’ data collection efforts
such as the UN sponsored Multi-cluster indicators surveys, thereby saving time and money.
110
Time, money and responsibilities must be allocated in all Organizational projects for these important
indicator and data management activities. The process of detailing responsibilities and a budget for data
collection and management activities will often lead to a re-examination of data priorities.
III. Conclusion
Box 8 summarizes good practices in identifying and managing indicators that have been discussed in this
tool. Some of these practices also apply to efficient management of outcome and impact indicators.
Start with programme design. Implications for data collection need to be fully integrated in the design
of the programme, including a budget to cover data collection costs;
Baseline information. Where change is being assessed obtain baseline data at the start of the
programmes, and, if possible, data on past trends;
Use existing data sources and reporting systems where possible. However, if data is not available,
cost-effective and rapid assessment methodologies should be considered for supplementary data
collection;
Partnerships. Establish partnerships with government, NGOs, bilateral donors and other key
stakeholders to collect the data so as to reduce costs;
Information management. Plan how the flow of information relating to the indicators will be managed,
stored and retrieved in a user-friendly data base within all the Offices or in the main office.
111
Sources
By Gerard M Blair
OECD/DAC. “Glossary of Key Terms in Evaluation and Results Based Management”, The DAC
Working Party on Aid Evaluation, 2002.
Scriven, Michael. “Evaluation Thesaurus – Fourth Edition”, Sage Publications, 1991.
The United Nations Development Group (UNDG). “Results Based Management Terminology”, June
2003.
The World Bank. “Measuring Efficiency and Equity Terms”.
OECD. “Improving Evaluation Practices: Best Practice Guidelines for Evaluation and Background
Paper”, 1999.
Patton, Michael Quinn. “Utilization- Focused Evaluation – The New Century Text”, 3rd Edition, Sage
Publications, 1997.
USAID. “Managing for Results at USAID”, presentation prepared by Annette Binnendijk for the
Workshop on Performance Management and Evaluation, New York, 5-7 October, 1998.
WFP. “WFP Principles and Methods of Monitoring and Evaluation”, Executive Board Annual
Session, Rome, 22-26 May, 2000.
Aubel, Judy. “Participatory Program Evaluation Manual – Involving Program Stakeholders in the
Evaluation Process”, Catholic Relief Services, Child Survival and Technical Support Project, Second
Edition, December 1999. Available in English, Spanish and French
Coupal Francoise, Simoneau Marie. “Participatory Evaluation: A Case Study of CCIC Humanitarian
Fund Projects in Haiti”, Mosaic.net, 1997.
Davies, C.T. “An introduction to Advocacy”, Addis Ababa, January 1998.
Estrella, Marisol and Gaventa, John. “Who Counts Reality? Participatory Monitoring and
Evaluation: a Literature Review”, IDS Working Paper 70, 1997.
Ryan, Katherine E., DeStefano, Lizanne Eds. “Evaluation as a Democratic Process: Promoting
Inclusion, Dialogue, and Deliberation”, New Directions for Evaluation, A Publication of the American
Evaluation Association, Number 85, Spring 2000.
UNDP. “Who are the Question-makers – A Participatory Evaluation Handbook”, OESP, 1997.
112
UNICEF. “EVALUATION – A UNICEF Guide for Monitoring and Evaluation – Making a
Difference?”, Evaluation Office, 1991.
Halabi, Hanan; Salem, Ruwaida; Wick, Laura. “Jenin Community Based RH Education Project”,
Project-end Evaluation for UNFPA, Birzeit University, Institute of Community and Public Health,
January 2000.
ILO. “Guidelines for the Preparation of Independent Evaluations of ILO Programmes and
Projects”, Evaluation Unit, Bureau of Programming and Management, November 1997. Available in
English, French and Spanish
Management Sciences for Health. “The Family Planning Manager”, Volume II, Number 1,
January/February 1993.
Ministry of Foreign Affairs, Danida. “Evaluation Guidelines”, February 1999.
Gosling, Luisa; Edwards, Mike. “Toolkits – A Practical Guide to Assessment, Monitoring, Review
and Evaluation”, Development Manual 5, Save the Children, 1995.
UNFPA. “Readings in Population Research Methodology - Volume 6, Advanced Basic Tools”, Social
Development Center, Chicago, Illinois. 1993.
USAID. Center for Development Information and Evaluation. Performance Monitoring and Evaluation
TIPS.
“Conducting Key informant Interview”, 1996, Number2.
“Using Direct Observation Techniques”, 1996, Number 4.
“Using Rapid Appraisal Methods”, 1996, Number 5.
“Conducting Focus Group Interviews”, 1996, Number 10.
113
CIDA. “CIDA Evaluation Guide”, Performance Review Branch, 2000.
DANIDA. “Evaluation Guidelines”, Ministry of Foreign Affairs, 1999.
OECD. “Improving Evaluation Practices: Best Practice Guidelines for Evaluation and Background
Paper”, 1999.
Rossi, Michael H., Freeman, Howard E., and Lipsey, Mark W. “Evaluation - A Systematic Approach”,
6th edition, Sage Publications, California, 1999.
Swiss Evaluation Society (SEVAL). “Evaluation Standards”, December 2000. Available at:
Western Michigan University. “The Program Evaluation Standards in International Settings”, The
Evaluation Center, Occasional Papers Series, May, 2000.
Bertrand, Jane and Amy Tsui. “Indicators for RH Program Evaluation”, The Evaluation Project.
University of North Carolina at Chapel Hill, 1995.
USAID. “Health and Family Planning Indicators: Measuring Sustainability”, Volume II, Africa
Bureau’s Office of Sustainable Development (AFR/SD).
UNFPA. “Measuring Outputs in Population and Development with Illustrative Indicators”, Technical Support Division,
November 2001.
USAID. Performance Monitoring and Evaluation TIPS, Centre for Development Information and Evaluation.
TIPS Number 6, 1996: Selecting Performance Indicators.
TIPS Number 8, 1996: Establishing Performance Targets.
TIPS Number 12, 1998: Guidelines for Indicator and Data Quality.
This tool is subject to constant improvement. We welcome any comments and suggestions
you may have on its contents. We also encourage you to send us information on
experiences from other programmes or partners which can illustrate the issues addressed
by this manual/tool. Please send your inputs to:
This tool is based on experiences & different published & un published material from
other national & international organizations
114