Professional Documents
Culture Documents
May2010
TABLE OF CONTENTS
TABLE OF CONTENTS
ii
ABBREVIATIONS
iv
1.0
BACKGROUND
2.0
OBJECTIVES
3.0
3.1 Content
3.2 Geographical Focus
3.3 Subject of Pilot Test
4.0
5.0
1
2
2
3
3
3
3
4
4
4
5
FINDINGS
5.1 System Efficiency
5.1.1
Appropriation of the Selection in Criteria
5.1.2
Feasibility of the Process
5.1.3
Utilization of System Implementation Tools
5.2 Effectiveness
5.2.1
Appropriateness of Personnel Selected
5.2.2
Appropriateness of Tools Used for M&E Processes
5.2.3
Utilization of the M&E Reports to Education Planning
5.2.4
Utilization of M&E Reports for Provision of Technical Assistance
5.3 Relevance
5.3.1
Relevance of QAA Reports in the Provision of
Information on Quality Schools
5.3.2
SBM Level of Practice
5.4 Facilitating & Hindering Factors
5
5
5
7
14
15
15
16
16
16
16
17
18
19
P a g e | ii
6.0
RECOMMENDATIONS
6.1 Recommended Changes in the System
6.2 Recommendation for Support Requirements for Systems Institutionalization
6.2.1
Policy recommendations
6.2.2
Capability Building programs
6.2.3
Support Requirements
6.3 Support for the Roll-Out of the System to other Regions
21
21
22
22
22
23
23
ANNEXES
24
25
25
27
29
31
31
32
34
37
37
37
39
39
43
43
44
45
49
E. Reports Templates
E.1 School QMT M&E Report Template
E,2 Sub-Division QMT M&E Report Template
E.3 Division QMT M&E Report Template
62
62
63
63
40
40
40
41
41
43
P a g e | iii
ABBREVIATIONS
BEIS
BESMEF
BESRA
B/T
CLC
DepED
DO
Division Office
DQMT
E-BEIS
EFA
EMIS
FD
Functional Divisions
ICT
INSET
LR
Learning Resource(s)
LRMDS
M&E
NAT
NO
Negros Occidental
NS
Northern Samar
PPRD
QA
Quality Assurance
QAAS
QAAD
QMT
REXECOM
RO
Regional Office
SBM
School-based Management
SDQMT
SQMT
SY
School Year
TA
UIS
P a g e | iv
1.0
BACKGROUND
The Department of Education (DepEd) is purposively and progressively taking leaps towards the implementation of RA
9155 to address the forefront issue of insuring quality performance in education. In its efforts to create a basic
education sector that is capable of achieving the goals and objectives of Education for All (EFA) 2015, DepEd is carrying
out policy reforms with renewed emphasis on bringing about quality learning outcomes through improved strategies of
its educational systems. This is referred to as the Basic Education Sector Reform Agenda (BESRA), which has five key
reform thrusts (KRT) that revolve around the school, the teachers, the learning environment, complementary
interventions and the organizational culture.
STRIVE 2 with School based Management Support Systems as a project component, intends, by and large, to directly
support KRT 1 popularly referred to as Continuous School Improvement . By design, Quality Assurance and
Accountability (QAA) System is one of the major outputs of this component. By BESRAs framework however, it
considers QAA as part of KRT 3, being a social support to attainment of desired learning outcomes. This present
mindset goes to illustrate the significant role of QAAS as a priority systemic and integrating strategy to bring about
quality in all education systems. It is a cross-cutting theme for all the components of Strive 2 as well as the BESRA
reform programs.
Implementation of a quality assurance and accountability mechanism interfaces with an equally significant machinery as
monitoring and evaluation to insure its functionality. Thus, these two sub-systems (i.e. QAA-M&E System) have been
developed by STRIVE DepED implementers in a synergistic operation towards nurturance of a culture of quality and
accountability in the organization.
2.0 OBJECTIVES
Piloting the implementation of the QAA-M&E System primarily aims at obtaining constructive information useful for
improving/revising the system via testing the application of its processes in the implementation of the school
improvement plan (SIP) as a core process in school.
The pilot specifically examines the functionality of the systems basic elements: the structures installed in terms of the
Quality Management Teams with their Terms of Reference (TOR), the QAA-M&E processes applied and the tools
used for operation. The information derived is valuable for spawning recommendations on improving system
implementation and on identifying the organizational support requirements for its institutionalization. Specifically, the
pilot run intends to:
2.1 Obtain information on the systems efficiency, effectiveness, and relevance serve as basis for its revision and
enhancement
2.2 Identify the critical factors that have facilitated or hindered the implementation of the system to serve as input
for improving system implementation
2.3 Generate suggestions from the experiences and perception of the system implementers in order to identify the
necessary support requirements for its institutionalization
3.0
3.1. Content
The pilot focuses on the applicability of the M&E system consisting of its structures, processes and tools in the
implementation of SIP in schools and in how the districts, divisions and regions utilize the information obtained from
them.
For the structures, the pilot looks at the Terms of Reference followed in the establishment of the implementing units
who are the Quality Management Teams (QMTs) at various levels. Specifically, it reviews the criteria for selection, the
functions as indicated and the specific tasks related to M&E of SIP implementation.
Table 1 summarizes the intended coverage in the pilot run of the system within seven (7) calendar months (i.e.
September to March) of SY 2009-2010. The scope matrix provides a snapshot of the selected QAA-M&E mechanisms
to quality assure SIP implementation and SBM assessment of schools.
Page |1
DISTRICT/
DIVISION
REGION
Content
Performance Criterion
1. Progress M&E
2. Results M&E
3. Self-Assessment
1. Process Review of
Progress M&E
2. Process Review of
Results M&E
38
32
18
14
12
Elementary
Negros Occidental
32
Bohol/Tagbilaran
39
Northern Samar
38
Secondary
18
14
12
Page |2
4.0
4.1 Processes
4.1.1 The field testing of the QAA-M&E System actually commenced with the installation of the needed structures of
the system with the application of the Terms of Reference developed for the selection of the process
implementers at the different functional levels and the definition of their functions and responsibilities relevant
to the QAA-M&E system. See original Terms of Reference in Annex A
4.1.2 Table 4 shows the major processes followed to pilot the system. Implementation of all these processes was
supported by capability building activities for the QMTs involved and which were carried out by the core team
of the QAA-M&E component of STRIVE 2 coming from the pilot regions and divisions. While these processes
were being carried out, simultaneous development activities for the other elements of the system were
undertaken like development of implementation processes and tools as well as monitoring tools.
Table 2
Processes Field Tested by Level
Pilot Processes
Level
Target Schedule
School, District
and Division
June 2009
3. Establishment and orientation of the School Quality Management Teams School, District, July 2009
(SQMT) on their Terms of Reference by the DQMT and SDQMT.
Division
4. Capability building of the SQMT in undertaking progress M&E of their SIP School, District, July August 2009
Division
5. Progress M&E of SIP implementation by the SQMT. Monthly processing School and
and reporting of Schools physical outputs, SIP implementation issues
District
and challenges, lessons learned to SDQMT
September 2009
March 2010
6. Progress M&E / Initial Gains through Self-assessment by Schools of their School, District, October 2009
SBM level of practice
and Division
7. Quarterly consolidation of M&E findings on SIP implementation of
schools by district and submission of SDQMT report to the Division
9. End-of-cycle results evaluation of SIP by schools (i.e. from NS and B/T) School, District, March- April 2010
Division
10. Processing of M&E reports from DQMT by QAAD for planning,
technical assistance and quality assurance
Piloting the system necessitated two layers of M&E activities: the first involved the actual implementation of the
M&E activities as required by the QAA-M&E system and as implemented by the QMTs and the second entailed
M&E of the system itself by the SMT who played a dual role of also being their mentors in carrying out the former.
4.2 Source of Information
4.2.1 Sampling of Quality Management Teams
With the limited time allowed for piloting the system as well as for data gathering, only a sample of SQMTs (i.e. 20)
were purposively selected to provide comprehensive information on the functionality of the QAA-M&E processes
targeted for try-out. The QMTs however, of the district, division and region from the pilot sites are all represented.
Table 3 gives the actual number of teams involved in the data gathering across the 3 divisions.
Page |3
Table 3
Quality Management Teams by Pilot Site
Quality
Management Team
A.
B.
C.
D.
SQMT
SDQMT
DQMT
RQMT(QAAD)
Number
of Teams
20
9
3
3
Region 6/ Negros
Occidental
6
3
1
1
Pilot Site
Region7/
Bohol/Tagbilaran
8
3
1
1
Region 8/
Northern Samar
6
3
1
1
Contents of the forms are basically focused on the teams TOR, their functions and tasks, facilitating and hindering
factors affecting the performance of the M&E processes, the process tools, support received, and support
requirements for system institutionalization. Similarity of content of the forms is intentionally considered to allow
comparison across levels.
In addition to the survey questionnaires, another source of data came from FGDs carried out by the SMTs of their
respective division. The information obtained either validated or supplemented the data gathered from the
questionnaires.
4.2.5 Data Processing and Analysis
Data gathered through the set of M&E questions contained in the survey questionnaires were processed and
analyzed by the SMT with the end view of arriving at a regional/division consensus regarding the systema
efficiency, effectiveness and relevance. Two levels of quantitative analysis occurred in the process: first, the
counting of team responses to represent a division or regional response and second, the processing of
division/regional responses to synthesize the overall perception or response to an M&E item. A regional/division
Page |4
teams response to an item is counted if it gets at least 75 % level of agreement among the QMT respondents, e.g.
if there are 5 SQMT respondents from a division, four of them should show the same response in order to register
a count ; and an overall consensus is established, if at least 2 out of 3 divisions/regions register agreement to the
same response. The processed information obtained from the various QMTs are separately consolidated using
Excel files to facilitate obtaining the needed frequency and percentage for each item thus, the tables presented in
the Findings portion were extracted from these consolidation files.
For the qualitative data gathered to identify issues or the facilitating and hindering factors influencing various
aspects of the system, sorting and categorizing the entries into themes or categories was carried out to arrive at a
synthesis.
5.0
FINDINGS
Division
Total
1
1
1
1
3
3
3
3
100%
100%
100%
100%
100%
100%
100%
NO
B/T
NS
1
1
1
1
1
1
1
1
It is worthwhile to note that the Sub-Division Quality Management Teams (SDQMT), who has been tasked to
facilitate the organization of the SQMTs in the schools have appropriately articulated the selection criteria to
the schools. All seven items could contribute to the effective performance of the SQMT functions
5.1.1.2. Sub-Division Quality Management Team (SDQMT)
Page |5
The set of criteria used in the selection for Sub-Division Quality Management Team (SDQMT) members is
presented in Table 1b. Similarly, the frequency of responses recorded for the SDQMTs of each site have all
reached the criterion of agreement for each item. This clearly indicates that all the items of the criteria set in
the selection of the SDQMT have been highly considered.
Table 1b
Selection Criteria in the Selection of the Sub-Division Quality Management Team
SDQMT Selection Criteria
Division
Total
NO
B/T
NS
100%
100%
100%
100%
.
5.1.1.3 Division Quality Management Team (DQMT)
Table 1c shows the criteria in the selection of the DQMTs as reflected in the TOR. The items are very similar
to what are required for SDQMTs except for the additional requirement of being involved in the DEDP
development process in the division. This item has been proposed since the system intends to support
division planning through the M&E reports collected from the schools /district. It is interesting to note that
among the five items included, only this item failed to get 100 %. This could have occurred due to the actual
composition of DQMT at the pilot division of Region 8 which was largely a mix of division and non-division
personnel. The non-division personnel had no experience in the planning process. The rest of the items have
been highly considered which is indicative of their appropriateness in the selection of DQMT.
Table 1c
Selection Criteria in the Selection of the Division Quality Management Team
DQMT Selection Criteria
1. Involved in the crafting of and in the enhancement process
of the Division Educational Development Plan (DEDP).
2. Has knowledge of monitoring and evaluation
3. Has knowledge on SIP development and appraisal
4. Has knowledge on providing technical assistance to
schools
5. Can implement and monitor national and regional
standards and guidelines in the division and schools
Division
NO
B/T
NS
Tot
al
1
1
1
1
1
1
3
3
66.67
%
100%
100%
100%
100%
Page |6
Table 1d
Selection Criteria for the Selection of the Regional Quality Management Team
RQMT Selection Criteria
1. Well informed and has updated knowledge on the basic
educational policies and standards
2. Involved in the DEDP review and REDP development
3. Has knowledge of Monitoring and Evaluation
4. Familiar with national and regional standards and guidelines
5. Capability in coordinating technical assistance to divisions
6. Familiar with the regulatory and accreditation policies and
processes
7. Has working knowledge of ICT.
Region
Total
100%
1
1
1
1
1
1
1
1
3
3
3
3
100%
100%
100%
100%
100%
100%
VI
VII
VIII
1
1
1
1
The over-all criteria for the selection of the QMT have clearly identified the tasks to be performed efficiently
and effectively in the attainment of the objectives. The selection criteria are appropriate to identify the
composition of QMTs to perform well their roles. Following the established criteria facilitate performance of
roles and functions.
In summary, the findings indicated in the above tables clearly say that the criteria set for the selection of the
QMTs at the school, sub-division, division, and regional levels are appropriate and have been highly
considered. Having met all the criteria, the QMTs across levels were able to perform their respective tasks with
utmost efficiency.
5.1.2
The efficiency of the processes is measured in terms of how the QMTs have regarded the feasibility or doability of
their assigned functions and tasks. Results of the directed group interview conducted with the selected Quality
Management Teams of the pilot sites revealed that, in general, the functions and tasks expected of them were
efficiently performed during the pilot implementation.
5.1.2.1 Assigned Functions Performed by the SQMTs
Tables 2a, 2b and 2c, respectively, show the functions of the Quality Management Teams at the school,
sub-division, division and the regional levels. It can be gleaned from Table 3a that all of the required
functions of the SQMTs of the three (3) divisions as specified in the Terms of Reference (TOR) were properly
executed for the whole duration of the pilot implementation. However, while they were able to perform all the
assigned function, they found it difficult to accomplish all the functions within the specified period of time. This
was revealed in follow-up discussions with them.
Table 2a
Assigned Functions Performed by the SQMT
SQMT Functions
Manage a mechanism for quality assurance and accountability
in the school
Process and review data gathered to make inferences to
enable school management to arrive at sound decision-making
Communicate results of monitoring and evaluation to improve
the implementation of SIP and SBM practice
Ensure that inputs, processes, outputs and outcomes of the
school are at par with quality standards and targets set
NO
B/T
NS
Total
100%
100%
100%
100%
Table 2b likewise shows that the SDQMTs of the three pilot divisions have efficiently performed the required
functions. Out of nine (9) teams of SDQMT coming from the three pilot divisions, only one team disclosed its
inability to manage the Quality Assurance & Accountability System in the assigned district due to overlapping
Page |7
activities. It also revealed that considering proximity and accessibility, it was practical for the team to submit
the required monthly reports directly to the Division Office instead of to the SDQMT as instructed.
Table 2b
Assigned Functions Performed by SDQMT
SDQMT Functions
NO
B/T
NS
Total
100%
100%
100%
Table 2c reveals that the DQMTs of the pilot divisions have executed almost all of the assigned functions as
indicated in the primary Terms of Reference (TOR). However, DQMT from Bohol divulged during the FGD that
on the function of ensuring that inputs, processes, outputs and outcomes of the schools to be at par with the
quality standard sets, the members commented that they could not indicate a positive response here since the
pilot schools do not meet the quality standards set as manifested in the result of the SBM Practice
Assessment. The item could have been interpreted as an assessment question of whether or not the schools
have been meeting the standards rather than whether or not the team has been doing something to assist
schools meet the standards. At any rate, all functions are reported to have been performed by the SDQMTs.
Table 2c
Assigned Functions Performed by DQMT
DQMT Functions
NO
B/T
NS
Total
100%
67%
100%
100%
Table 2d similarly illustrates the efficiency of the Regional Quality Management Teams (RQMT) to execute the
different functions they need to achieve within the extent of the pilot implementation. However, in the course of
discussion, RQMT from Region 7 disclosed that managing quality assurance and accountability in their region
to ensure compliance to standards of quality basic education is still in progress and they still need time to
reach out some RO personnel who are not yet oriented of the system. Furthermore, this team also unveiled the
necessity to strengthen the mechanisms of communicating status and progress of the region to CO for policy
adjustment and to stakeholders for resource mobilization.
Generally speaking, Regional Quality Management Teams (RQMT) of the pilot regions are proficiently fulfilling
their functions.
Table 2d
Assigned Functions Performed by RQMT
RQMT Functions
Manage quality assurance and accountability in the region to ensure
compliance to standards of quality basic education
Formulation and review of education standards in the region.
Evaluation of regional performance to inform technical assistance to
divisions
Communicating status and progress of region to CO for policy
adjustment and to stakeholders for resource mobilization
R6
R7
R8
Total
100%
100%
100%
100%
Page |8
NO
B/T
NS
Total
1
1
1
1
1
1
1
1
1
1
1
1
3
3
3
3
100%
100%
100%
100%
100%
100%
100%
100%
100%
100%
The School Quality Management Team (SQMT) of the three regions were likewise asked to enumerate
additional roles/responsibilities/tasks relevant to Quality Assurance & Accountability System they performed,
aside from those listed and the reasons that triggered the teams to perform them. Overwhelming confirmations
were obtained as indicated in their responses such as:
Engaging themselves in the development of project proposals for CLASS & SOBE Schools to make
certain of the needs-based interventions suited to the necessities of the clients
Being involved in the decision making and modifications of system's process, M & E Tools and timing
of SQMT reports
Intensifying e-tracking system to come up with appropriate interventions to prevent SARDO from
dropping out and to retrieve OSC/OSY/OSA
Reviewing performance rating of teachers after their self-rating
Sitting with the Appraisal Review Committee(PARC) to ensure proper performance appraisal
Validating school statistical reports (Forms 20, BEIS, etc.) to ensure quality of reporting
Engaging in the preparation of the action research on identified least-learned skills to identified nonreaders to improve reading performance of the school
Serving as member of the ranking committee for teacher-applicants and promotion of teachers to
ensure quality selection/promotion process
Participating in the preparation of parallel test items for least learned competencies/skills to ensure
validity & reliability of test items and the test construction processes
Facilitating the conduct of information dissemination to grade leaders to communicate results, issues
and concerns, hindering and facilitating factors encountered during the implementation
phase in
order to communicate the results properly for immediate action.
It is worthwhile to note that the additional tasks and responsibilities identified above are closely related to the
QAAS. These are seen to be facilitating factors to the efficient performance of the SQMTs.
Considering the conditions affecting the performance of the different tasks, the SQMTs were asked to
ascertain the tasks that were relatively easy for them to perform during the pilot testing. From among these
tasks, the SQMTs of the three pilot divisions identified the following as relatively easy to perform:
Page |9
On SIP Implementation
o Analyzing the identified issues and concerns to resolve problems
o Identifying issues and concerns during the SIP implementation within a given period
o Identifying the facilitating & hindering factors of SIP within a given period of time
In addition, the SQMTs of the three divisions indicated in their responses some critical factors that
facilitated/supported the efficient performance of the tasks, as follows:
On SIP Implementation
o Active participation of stakeholders (SGC, PTA, LGU, etc.) in coming up with a sound and
o realistic interventions to solve issues within the control of the school and for those which are to be
elevated to the higher managerial levels
o Readily made available QAA/M & E Tools
o Presence of Technical Assistance Providers (i.e. SDQMT & DQMT)
The SQMTs also determined the tasks that were difficult to perform. These are the periodic tracking of
physical outputs of SIP implementation and the task of submitting monthly M&E report to the SDQMT/DQMT.
What hindered them to perform the task are the multi-tasking of School Heads and other SQMT members,
overlapping of activities (STRIVE vs. organic & STRIVE against STRIVE). Another reason which could have
triggered difficulty in the periodic tracking of physical outputs was the vagueness of targets. This means that if
the targets are not so clear to them, tracking could really be difficult.
5.1.2.3 M&E Tasks Performed by the SDQMT
The positive responses reflected in Table 3a clearly reveal that the Sub Division Quality Management Teams
from the pilot divisions have efficiently performed their part. All teams reported having monitored and assisted
the SQMTs in their M&E tasks, consolidated their M&E reports on a quarterly basis for submission to the
division. Not a single task was reported as not having been performed.
Table 3b
Tasks Performed by the SDQMT for the Duration of the Pilot Implementation
SDQMT Tasks
NO
B/T
NS
Total
1
1
1
1
1
1
3
3
100%
100%
100%
100%
1
1
1
1
1
1
1
1
1
3
3
3
100%
100%
100%
P a g e | 10
Additional roles/responsibilities/tasks relevant to QAA-M&E system performed by the SDQMT were provision
of appropriate Technical Assistance to SQMTs to ensure the proper operationalization of the system and
acting as sub-division review committee for ranking of teachers and school heads.
Results of the FGD conducted with the SDQMTs revealed that the relatively easy tasks common to the three
divisions are consolidation of the physical outputs of SIP implementation in schools in the district and analyzing
the identified issues reported by the SQMT. These tasks became easy to them because of the training they
have received on collection and analysis of both qualitative and quantitative data and because of their
familiarity with SIP processes and the updated data from the
SQMT of the schools.
While some of the tasks have been reported to be easy for the teams to perform, the SDQMTs also identified
some tasks which they found difficult to perform due to certain critical hindering factors. These tasks included
difficulty in validating the SQMT reports for accuracy due to the delayed submission of needed reports. This
consequently resulted to just assuming the validity of the report for timely submission to the DQMT. Other
factors are non-availability of travel allowance to visit the different schools to validate the SQMT reports and
other documents which are required by the Division Office. Monitoring and process-checking the SBM
assessment of schools were also seen to be difficult to undertake . In spite of the training orientation given the
team, they admittedly commented that they still have difficulty in the standard-based execution of the SBM
Assessment processes. Conflict of schedule is another reason given by most of the respondents. With these
issues, it is high time that the Schools Division management consider some resolutions to address these
perennial problems.
At any rate, when asked of the Division Office support on the organization of Quality Management Teams at
the District and School levels, the SDQMT mentioned the conduct of orientation-workshop on Quality
Assurance and Accountability System as having facilitated their work. Organization of the SDQMTs with clear
TOR, the identification of focal persons for QAA to coordinate with the QAAD in the Region, and the
establishment of the technical assistance mechanism, have also been identified as substantial support of the
Division which contributed to the raised performance of the SDQMT.
Additionally, capability building and training-workshops on QAA-M & E System (Streamlined processes,
structures, & instruments), on data harvesting and analysis, orientation workshop on consolidating uantitative
& qualitative reports of the SQMTs, assessment of SBM Practices, and Provision of Technical
Assistance,
were
mentioned to have contributed much in the performance of the SDQMT.
In connection with Technical Assistance, the SDQMT emphasized that having a Functional Technical
Assistance Plan, timely sharing of information ( DepED Memo, Orders, Guidelines, Policies, etc.), Identifying
Promising Practices /Experiences to be elevated to the higher managerial levels for Technical
Assistance
Preparations and Policy Recommendations, coaching and mentoring on the
Development
of
Technical Assistance Plan, on-site coaching on the Implementation, monitoring & evaluation of technical
assistance, and intensive monitoring to
the schools are seen to facilitate the performance of their roles
and functions.
SDQMTs also articulated some enabling mechanisms to support Quality Assurance and Accountability System
implementation. These are the conduct of timely advocacy on the QAAS, on-site coaching and mentoring,
monitoring and evaluation of QAA Pilot Schools, gathering, interpreting and analyzing data, prioritization of
the improvement areas which are needs-based, process check of the SIP and AIP adjustment, and finally, the
timely provision of technical assistance.
5.1.2.4 M&E Tasks Performed by the SDQMT
The tasks performed by the DQMTs are reflected in Table 3c across divisions. It can be seen that the division
QMTs had varied opportunities to perform the tasks and only close to 50% could be performed. Evidently only
Items 1 ( utilizing appropriate M&E instruments and tools)and 6 ( Coordinating development of TA plans using
M&E reports from the SDQMT as input) are done by all the teams in the three divisions. Item 7 was not done
at all by any team. The DQMTs do not interact with the SDS/ASDS nor communicate findings in matters
related to heir M&E tasks. This task tho prescribed in their TOR, apparently, has not been established in the
division.
The DQMTs cited that the reasons for the non-performance of Items 3 (consolidation of M&E reports from the
districts), and Item 4 (validating information received from the SDQMTs) are the delayed submission of reports
from the SDQMTs, overlapping of activities, and no organized SDQMT particularly in Tagbilaran Division. This
scenario somehow suggests lack of coordination between the SDQMTs and the DQMTs.
P a g e | 11
Table 3c
Tasks Performed by the DQMT for the Duration of the Pilot Implementation
DQMT Tasks
1. Utilizing appropriate instruments and tools in conducting
M&E
2. Generating quantitative information from the School
Education Management Information System for preparing
M&E reports
3. Consolidating and analyzing the M and E reports
submitted by the Division Sub M and E team (quantitative
and qualitative)
4. Validating information received from the Sub-division
QMTs
5. Communicating findings and recommendations to
MANCOM
6. Coordinating development of TA plans using M&E
reports rom the division sub-M&E teams as input
7. Submit division sub level M and E report to the Division
M and E Team
8. Interacting with ASDS/SDS and the Sub-division M&E
teams
NO
B/T
NS
Total
100%
67%
33%
33%
33%
100%
100%
0%
Ibn addition, the DQMTs itemized additional roles/responsibilities/tasks they have performed that they
considered relevant to Quality Assurance & Accountability System along with the reasons that pushed the
teams to perform them. Along this line, the DQMTs mentioned the following:
Following-up the proper implementation of the Technical Assistance Plan
Providing information to school on where and how to generate, mobilize and manage funds and other
resources e.g., MOOE. SEF, donations, Project Proposal Development, Involvement in the
development of Barangay/Municipal Resolutions, etc.
Facilitating referrals to the appropriate resources that could address their existing problems e.g.
Training and Development Team for NCBTS and T & D related issues, and to LRMD teams for
access on Teaching and Learning Resources
Preparing tests materials and monitoring the administration of DAT & RAT
Appraising SIPs and coaching and evaluating project proposals of SOBE & CLASS
Providing appropriate technical assistance to schools.
Conversely, the DQMTs of the three pilot regions were also asked to identify the tasks that were difficult for
them to perform and the factors that contributed to the difficulty. Significant to note was that not even one of
the tasks is common to the three teams. The DQMT from Northern Samar expressed that, primarily, the most
difficult task for them to do is consolidating and analyzing the M & E reports submitted by the sub- M & E team
both quantitatively and qualitatively. Most of the team members are so engrossed with the overwhelming
organic functions to seriously do this task. Moreover, the lack of ICT skills among the members of the SDQMT
resulted to too much delay in the submission of reports to the DQMT.
On the other hand, DQMT from Bohol/Tagbilaran conveyed that because most of the members have no
background on the development and localization of the M & E tools, they found it hard to develop the needed
division M & E instruments. While the DQMT from Negros Occidental signified that it is on interacting with the
SDS/ASDS and the sub-division M & E Teams that is most difficult for them to tackle because of the conflict in
schedules.
With the above-stated negative issues, it may be suggested that the SDS/ASDS consider seriously including
these in the agenda during the Mancom meetings with the concerned personnel in order to arrive at
agreements and resolutions to have efficient and feasible QAA-M&E System.
P a g e | 12
Given the experience of the DQMT in managing the QAA System, they were also made to identify other
significantly critical tasks they have done in addition to what have been delivered, and the reasons that were
seen to have contributed in some ways to the facility in doing their work.. The team enumerated the following:
Division M & E team taking the lead in the preparation of the technical assistance plan based on the
consolidated reports of the SDQMT to fine tune the present set up of Technical Assistance
Mechanism pilot run.
Division M & E Team facilitating the annual improvement plan review & adjustments in school
because the team is in better position to validate issues from the data in the BEIS and in prioritizing
TA plan based on the availability of the resources.
Division M & E Team helping the SDQMT in consolidating the SQMT reports as Technical Assistance to
SDQMT for immediate actions on the issues found from the schools that need instantaneous solution from the
division. This serves as an avenue also in coming up with a common understanding, interpretation, analysis,
TA Plan & recommendations on the findings from the SDQMT reports.
In the course of the pilot, DQMTs also articulated some enabling mechanisms to support Quality Assurance &
Accountability System implementation to wit:
Intensive monitoring and supervision of the QAAS through TEAM Monitoring & Supervision scheme,
an existing division initiative wherein ES is paired with a PSDS, primarily to track down the progress
of the implementations of the different DepED programs and projects, SIP at the school and to
provide Technical Assistance based on the findings.
Development, implementation and monitoring of the Technical Assistance Plan in partnership with
the SDQMT as part of the organic function of the Team Monitoring and Supervision scheme.
Table 3d
Tasks Performed by the RQMT for the Duration of the Pilot Implementation
RQMT Tasks
NO
B/T
NS
Total
100%
100%
100%
100%
100%
100%
0%
100%
The above data indicates that although most of the tasks required of them were efficiently delivered, the teams
of the three pilot regions have come up with a common response of not implementing yet the reward,
recognition and incentive system in the region thus, resulted to only 87.5% of the total percentage of the tasks
performed by the RQMT. Quality Management Teams from Regions 6 & 7 have the same reason of not
implementing the said task because of the absence of the operational guidelines to run the system. While QMT
from Region 8 has its own ground of not implementing it because the system, according to them, is already
embedded in the function of the Human Resource/Administrative Division Unit of the region.
The group of Quality Management Team (SQMT) of the three regions were likewise asked to specify additional
roles/responsibilities/tasks relevant to Quality Assurance & Accountability System they performed,
notwithstanding from those registered in the tool and the reasons that triggered the teams to perform them.
The group has come up with a confirmation as indicated in their response and that is attendance to different
meetings, conferences and other related activities called by relevant agencies to further strengthened linkages
to other external stakeholders.
P a g e | 13
Taking into consideration on the diverse conditions affecting the performance of the different tasks, the RQMTs
were asked to determine the tasks that were relatively easy for them to perform during the pilot testing. It was
extensively noted that teams of the pilot regions were all agreeable when they averred that all the tasks
obliged to them are not difficult to achieve.
In addition, the RQMTs pointed out in their responses some critical factors that facilitated/supported
efficient performance of the tasks, as follows:
Easy to follow policy and guidelines
Updated REXECOM report submitted
The support of TA Division and the availability of Technical Assistance Plan
Availability of M & E Tools and M & E Results
Clear Guidelines & Set of Standards
Involvement in the Crafting of RMEAF
Technical Know-How of the Team
the
The RQMTs were also made to determine the task/s that was/were difficult to perform. But RQMTs of the three
pilot regions consistently responded that there was even no single instance in the course of execution of their
task/s that made them hard to comply.
5.1.3
For the Pilot Implementation of the QAA-M&E System, two categories of process tools were utilized.
One
was for the SIP implementation which used seven tools; namely, Physical Outputs Accomplishment Form/Tracking
Form, Issues/Problems and Concerns Log, Log Sheet on Resolution of Issues/problems, Log Sheet on
Sustainability
Measures, Log sheet on Lessons Learned, SIP end-Cycle Terminal Report Template, and
Monthly M&E Report Template. See Annex C for the tools.
For the Assessment of School-Based Management, , the SBM Assessment Tool (printed) and SBM Assessment
Module (EBEIS) which can be accessed on the world-wide web (http://beis.deped.gov.ph) were used.
Table 4 reflects the results per division on the utilization of the Process Tools in the performance of the SQMTs
roles/ responsibilities/tasks.
Table 4
Process Tools Used by the SQMT in Performing their Task
Process Tools
A. For SIP Implementation
a. Physical Output Accomplishment Form/Tracking Form
b. Issues/Problems & Concerns Log
c. Log Sheet on Resolution of Issues/Problems
d. Log Sheet on Sustainability Measures
e. Log Sheet on Lessons Learned
f. SIP End-Cycle Terminal Report Template
g. Monthly M&E Report Template
B. SBM Assessment
a. SBM Assessment Tool
b. SBM Assessment Module (EBEIS)
Pilot Division
Negros
Occidental
Bohol/
Tagbilaran
Northern
Samar
The table clearly indicates that almost all of the nineteen (19) SQMTs have signified that they used the tools in
undertaking the M&E processes for SIP implementation and SBM assessment. However, the Division of
Northern Samar was not able to use the SIP End-of-Cycle Terminal Report Template due to low turn-out of
SQMT reports submitted to the SDQMTs (PSDS). Only 42 percent or 21 schools out of the 50 QAA pilot
schools had submitted needed reports within the prescribed period. The non-submission was due to the bulk
of end-of-the-year reports required of the School Heads who were also members of the SQMT, and the
disparity between the monthly report format requirement by the division and the QAAS pilot implementation.
P a g e | 14
The two other divisions might have used the same supplementary tools, however, during the conduct of FGD
they failed to disclose the same. The SQMT may opt not to use the above-named tools unless these are
requisites or connected to their targets during the 3-year SIP/Annual Implementation Plan.
5.2 SYSTEMS EFFECTIVENESS
System effectiveness is the extent to which a system achieves its objectives. It requires the functionality, dependability,
and capability of the system
In the present report, system effectiveness refers to the suitability or aptness of the personnel selected to compose the
QMTs, the appropriateness of the tools used to capture the needed information, and the ability of the M&E processes to
yield information for educational planning and decision making.
5.2.1 Appropriateness of Personnel Selected
An indicator of the effectiveness of the selection criteria for the different QMTs is the extent to which the teams can
meet the skills required to perform the TOR. This was gauged by obtaining the percentage of M&E tasks done by
the teams representing their level of performance. The higher the percentage, the more able are the teams in
performing what is expected of them thus showing the effectiveness of the criteria used in selecting them. The
acceptable level of performance for a task is 2 out 3 thus obtaining the frequency of tasks performed means
counting those that registered at least 66%.
Table 6 shows the frequency and percentage of tasks performed by the teams by level. It is heartening to note that
the teams involved have shown remarkable evidence in its ability to meet the skills required for QAA-M%E. This is
especially demonstrated by the SQMT who are in fact, playing the most significant role in the chosen areas for
piloting QAA-M&E, i.e. SIP Implementation and SBM Assessment. They were able to perform 100% of the tasks.
Likewise are the SDQMTs who have substantially demonstrated 100% level of performance. A closer look at their
performance by item however, reveals three (3) of the items not being carried out by all teams in their respective
division. Nonetheless, they have met the acceptable level.
Among the teams, the DQMT relatively obtained the lowest level of performance as they were able to do 5 out of 8
tasks (62%) only. The tasks not performed by two-thirds of the teams were those contingent on the performance of
the SDQMTs in the division, e.g. delayed submission of M&E reports by SDQMT. This particular finding illustrates
the interdependence of the QMTS to successfully implement the system.
Of the eight tasks of the RQMT-QAAD, seven or 88% were performed by the three regional QMTs, again
suggesting the appropriateness of the selection criteria in choosing the members of the team who could do the
tasks. Data likewise show are fully aware they are of their tasks as stipulated in R.A. 9155 where the one of the
regions primary functions is to provide Technical Assistance to the divisions. These TA to the divisions cannot be
effectively carried out if the M&E system is not in place since submission of reports will not be acted upon.
Effectiveness of Criteria
Region QMT
Division QMT
No of Tasks Performed
Sub-division QMT
No of Tasks Assigned
School QMT
10
Figure 2
Number of M&E Tasks Assigned and Performed by QMT
P a g e | 15
It is worth mentioning that the selection criteria for the various QMT levels were effective because majority of the
expected tasks to be performed by the QMTs were accomplished..
5.2.2. Appropriateness of Tools Used for M&E Processes
The development of the system implementation tools is in compliance with the requirements of the BESRA
Implementation Plan on KRT 1. These tools are to be used in monitoring and tracking the progress of their SIP
implementation and in evaluating their SBM practices.
Positive responses were given by the SQMT, indicating that the tools developed and used during the pilot
implementation captured the needed information expected from the M&E Reports. These tools had given the
school planning teams vital information in adjusting the SIP and the district/division in developing their TA plans to
address the needs of the schools in the course of their SIP Implementation.
5.2.3 Utilization of the M&E Reports to Education Planning
Table 8 shows the extent of the utilization of the M&E reports for SIP/AIP adjustments by the target schools. It
indicates the number and percentage of target schools of the pilot divisions that have utilized M&E information in
adjusting their AIP2.
Table 8
Number of Schools Using M&E Information in SIP/AIP Adjustment
Level of
Utilization
Utilized
Not Utilized
Neg. Occ.
(N = 50)
F
%
50
100%
0
0%
Bohol/Tag
(N= 53)
F
%
49
92%
4
8%
N. Samar
(N = 50)
F
%
48
96%
2
4%
TOTAL
(N = 153)
F
%
147 96%
6
4%
The foregoing table shows that one hundred forty seven (147) out of one hundred fifty three (153) or 96% of the
QAA target schools across the three (3) regions utilized the M&E reports in adjusting their SIP/AIP2. As reported
by the SBM coordinators who are members of the SMT, the adjustment has been based on the information on the
undelivered outputs of the school.
In general, the QAA System provides the schools an opportunity to utilize information from M&E reports as basis for
educational planning.
5.2.4. Utilization of M&E Reports for Provision of Technical Assistance
Monitoring and Evaluation (M&E) is expected to generate information relative to the situation of target schools.
M&E reports submitted by the SQMT include issues/problems that need to be resolved. This provides an avenue
for the TA Teams to consider these issues/problems in the development of TA Plan.
The results of the survey conducted are reflected on Table 9 that shows the number and percentage of QAA target
schools in the STRIVE Pilot Regions utilizing/not utilizing M&E information in the provision of TA.
Table 8a
Utilization of M&E Information in the Provision of Technical Assistance
Utilization of M&E
Information in the
Provision of TA
R6 Neg. Occ.
(N = 50)
R7 Bohol/Tag
(N = 53)
R8 N. Samar
(N = 50)
TOTAL
(N = 153)
Utilized
49
98%
52
98%
48
96%
149
97%
Not Utilized
2%
2%
4%
3%
5.3 Relevance
Relevance is a term used to describe how pertinent, connected, or applicable something is to a given matter. A thing is
relevant if it serves as a means to a given purpose. The Quality Assurance and Accountability System being piloted
shows its social applicability in becoming to be a means to lead schools in improving their quality indicators, in
P a g e | 16
progressively advancing their SBM level of practice and in influencing and informing management of the technical
assistance they need.
5.3.1 Relevance of QAA Reports in the Provision of Information on Quality of Schools
5.3.1a. Performance Indicators
After three years of SIP implementation, the SQMT able to provide information on the performance
indicators attained by the pilot schools. The quality implementation and monitoring of the SIP implementation
in every school is manifested in every SIP End-of- Cycle Report.
Table 9
Number of Schools with Improved Performance Indicators
as shown in the SIP End-of- Cycle Reports
Division
Bohol
Negros
Occidental
Northern
Samar
Achieve
ment
Rate
Net
Enrol
ment
Rate
25
12
NDA
NDA
29
NDA
NDA
25
27
16
19
22
NDA
NDA
21
21
Prom
otion
rate
Com
pletio
n
rate
18
11
NDA
27
NDA
22
23
NDA
Table 9 shows that in Bohol and Tagbilaran City Divisions, 25 schools have shown improvement in
Achievement Rate, 12 schools showed increase in Net Enrolment Rate, 19 schools with improved Graduation
Rate, 8 schools with improved Participation and Survival Rates, 20 schools showed improvement in Failure
and Repetition Rates, 18 schools have improvement in Promotion Rate while 11 showed improvement in
Completion Rate.
In Negros Occidental Division, 29 schools showed improvement in Graduation Rate, 25 schools have
improved in Failure Rate, while 27 schools have improved in Drop-out and Promotion Rates.
While in Northern Samar Division, 16 schools with improvement in Achievement Rate, 19 schools showed
improvement in Net Enrolment Rate, 22 schools are reported to have improved in Graduation and Repetition
Rates; 21 schools have showed improvement in Failure and Drop-out Rates and 23 schools have improved
their Promotion Rate.
Table 9a
Consolidated Number of Schools with Improved Performance Indicators during the
Current Year as shown in the SIP End-of- Cycle Reports
Division
Bohol & Tagbilaran City
Negros Occidental
Northern Samar
Number of
Pilot Schools
53
50
50
30
43
26
29
40
26
97%
93%
100%
Table 9a presents the number of schools with improved performance indicators during the current year as
shown in the SIP End-of- Cycle Report. Of the 53 QAA pilot schools in Bohol and Tagbilaran City Divisions ,
only 30 schools with SIP End-of-Cycle Reports. Of the 30 only 29 or 97% schools showed improvement in the
performance indicators.
P a g e | 17
In Negros Occidental Division out of the 50 schools, only 43 schools with SIP End-of-Cycle Reports. Of the
43 schools, only 40 or 93% showed improvement in the performance indicators as shown in the QAA report.
While in Northern Samar, there are 50 QAA pilot schools. Of these, only 26 schools with SIP End-of-Cycle
Reports and the 26 or 100% showed improvement in the performance indicators as reflected in the SIP Endof- Cycle Report s.
5.3.2. SBM Level of Practice
There were two different versions of assessment tools used in the conduct of SBM Assessment on the Level of
Practice among the schools. In the first assessment, the tool used was STRIVE version, while the second
version was SBM TWG version of assessment tool. The SQMT was able to provide information on the level of
SBM practice of the QAA pilot schools.
Table 9b
Number of schools with Results of their Assessment of SBM Practice
and with Improved Level of SBM practice as shown in the SIP End-of- Cycle Report
First Assessment
Second Assessment
Level 1
Level 1
Level 2
Schools
with
SBM
Division
Moving Standa
Moving Standa Gearin
Progre
Advanci
Assessme Starting Toward rd Starting Toward rd
g Up
ssing
Stage
Stage
ng Stage
nt
Stage Stage
Stage Stage Stage
Stage
Bohol & Tagbilaran City
27
13
4
1
2
11
1
0
2
0
Negros Occidental
42
13
23
0
10
16
0
1
4
5
Table 9b presents that in Bohol and Tagbilaran City Divisions, 27 schools have reported to have assessed on
SBM Level of Practice as reflected in the SIP End-of- Cycle Report. On the First Assessment, 13 schools
considered themselves in Level 1- Starting Stage, 4 schools are in Level 1-Moving Toward Standard while only 1
school is in Level 1- Standard Stage. On the Second Assessment, 2 schools considered themselves Starting Stage
yet, 11 schools are in Moving Toward Standard and 1 is in Standard Stage and 2 schools considered themselves
to be in Level 2 Advancing Stage.
In Negros Occidental Division, 42 schools have reported to have assessed on SBM Level of Practice as shown in
the SIP End-of- Cycle Report. On the First Assessment, 13 schools considered themselves in Level 1- Starting
Stage while 23 schools are in Level 1-Moving Toward Standard. On the Second Assessment, 10 schools
considered themselves Starting Stage yet, 16 schools are in Moving Toward Standard and 1 is in Level 2
Gearing Up Stage and 4 schools considered themselves to be in Level 2 Advancing Stage and 5 are reported to
be in Level 2- Progressing Stage.
Table 9c
Number of schools with Results of their Assessment of SBM Practice and with
Improved Level of SBM practice as shown in the SIP End-of- Cycle
Report in Northern Samar Division
First Assessment (STRIVE Version)
Schools
with SBM
Division
Level 1
Level 2 Level 3
Assessmen
Almost
Progressiv
t
Starting
Ready
Mature
Ready
e
N. Samar
26
12
21
Table 9c shows that in the Division of Northern Samar, there are 26 schools reported to have assessed on SBM
Level of Practice as shown in the SIP End-of- Cycle Report. On the First Assessment, 12 schools are Level 1Starting Stage, 5 Level 1- Almost Ready, 1 school is considered on the Ready Stage, 4 schools are in the Level 2Progressive Stage and 4 schools are in the Level 3-Mature Stage. While on the Second Assessment, all schools
are classified in the Level 1, 3 schools are in Starting Stage and 21 schools are in Moving Towards Standard Stage.
P a g e | 18
Northern Samar Division utilized different tools in assessing Schools SBM Implementation Level of Practice.
During the first assessment, the SBM Assessment Tool being utilized was the STRIVE version, while the SBM
Tool developed by the SBM TWG was used in the second assessment as mandated through the national roll
out of SBM Assessment Policy.
Table 9d
Consolidated Number of Schools with Results of their Assessment of SBM Practice
and with Improved Level of SBM Practice as shown in the SIP End-of- Cycle Report
Division
Number
of Pilot
Schools
53
50
50
Table 9d shows the number of schools with results of their assessment of SBM practice and with improved level
of SBM practice. In the divisions of Bohol and Tagbilaran City, of the 53 QAA pilot schools, only 27 or 51% of
the schools have indicated the results of their assessment of SBM practice. Of the 27 schools, only 16 or 59%
of the schools showed improvement in their level of SBM practice in Bohol and Tagbilaran City Divisions.
In Negros Occidental Division, of the 50 schools, 42 or 84% have showed the results of their assessment of
SBM practice as shown the QAA report. Only 18 or 43% of schools with improved level of SBM practice in Negros
Occidental Division.
While in Northern Samar Division, there are 50 QAA pilot schools. Of these, 26 have reported the results of their
assessment of SBM practice and the 26 or 100% have showed improvement in the level of SBM practice
as
shown the QAA report.
5.4 Critical Helping and Hindering Factors Influencing Implementation of the QAA-M&E System
A summary of the facilitating and hindering factors that have influenced implementation of the system by the
respondents is shown in Table _ . The list was derived from the those given by the QMTs across pilot sites by functional
level. ( See Annex _ for the listing by QMT level). For processing of these data, qualitative analysis was undertaken to
sort the entries and identify the themes that govern them.
Among the enabling factors, seven (7) categories have emerged ranging from the characteristics and features of the
pilot system in the form of its simplified structures, processes, tools and guidelines, capability building provided to the
QMTs, the management support mechanisms provided by the DO to SDQMTs/SQMTs in performing their tasks, the
observed commitment and dedication of the QMTs to the stakeholders support given to SIP implementation. The list is
quite useful in identifying potential recommendations on what should be sustained in implementing the system
particularly in determining what needs to be given emphasis in the systems implementation manual.
The hindering factors, quite interestingly, focused more on implementation concerns that have caused delay in
performing their tasks. They had issues on the delayed submission of reports, absence of consolidation tools on the
part of the SDQMT to do their work, lack of communication between DQMT and SDQMT, some management issues
brought about by change in assignment at the DO or transfer of school heads to other stations adversely affecting SIP
implementation. A number of problems were also cited but were more STRIVE project-related rather than the system
itself. These were on conflicting schedules of organic and STRIVE activities, work overload of those involved, and
absence of financial assistance from STRIVE, which may not resurface anymore after project completion.
P a g e | 19
Table 10
Summary of Critical Helping and Hindering Factors Influencing
Efficiency and Effectiveness of QAA-M&E System
Hindering Factors
Enabling Factors
system
a. Clear Terms of references of the SQMT
b. Selection criteria for the QMTs helpful in organizing
QMTs
c. Mobilized SQMT, SDQMT and DQMT
d. Simplified/localized M & E instruments
e. Sufficient QAA-M&E tools and templates
1.
2. Technical Assistance
2. Clear guidelines on the implementation of QAA-M & E a. No provision of periodic consultation/interaction with the
DQMT on QAA matters
system
b.
Inadequate Mentoring/Coaching processes to assist
a. Common understanding of the system delivery at all
QMTs
levels (SQMT, SDQMT & DQMT)
b. Presence of framework in every mechanism
developed
3. Administrative / Management Issues
a. Discontinuity of SIP activities as result of movement of
3. Provision of technical know-how on M & E & ICT
SHs due to promotion, transfer and retirement.
a. Capability building conducted by STRIVE
b.
Conflicting schedule of trainings/workshops conducted by
b. Availability of ICT support (facilities) for M&E
DepED and STRIVE.
c. Training received by the PSDS who are members of
c.
Using the same work force in the implementation of the
the SDQMT
different STRIVE activities.
d. Updated division BEIS & SIS for data gathering
d.
Absence of policy/Memorandum/DepEd Order for the
purposes & for prompt preparation of M & E reports
implementation of the system during the pilot period
e. Technical know-how of the members on the national &
e.
No financial assistance given from the division or
regional standards, indicators and performance
STRIVE
measures
f.
New set of functions because of the restructuring of the
f. Advocacy and education
division where some of the Division M & E Team
g. Continuing orientation of other members of QMT
members are assigned in other unit requiring other
responsibilities
g. Absence of Awards/Incentives to performing SQMTs
h. Lack of upgraded ICT equipment
i. Unfinished REDP
4. Commitment and dedication of implementing units
4.
a. Positive attitude of the SDQMT to perform their tasks
and functions
b. Harmonious relationship between SDQMT and
SQMTs
c. Maximized participation/involvement of the members
belonging to the QAA division unit
d. Teamwork
e. Support of SDQMTs to SQMT in complying with report
requirements
f. Coaching by DQMT in the report consolidation
processes of SDQMT
g. Utilization of personal resources for travel purposes
5. Management support
5.
a. REXECOM actions in resolving issues
b. Involvement of SDS or ASDS in QAA related activities
e.g. analysis of M & E reports and in the formulation
of technical assistance plan
c. Immediate actions on the recommendations coming
from M & E reports
P a g e | 20
6.0 RECOMMENDATIONS
The pilot implementation of the Quality Assurance and Accountability System (QAAS) in the three pilot regions and
divisions started School Year 2009-2010. Based on pilot results, the Quality Management Teams (QMTs) at the
Region, Division, District and School levels have identified two sets of recommendations: a) recommendations to
improve system implementation and b) recommendations to improve sustainability through identified support
requirements for its institutionalization.
6.1 RECOMMENDED CHANGES IN THE SYSTEM
The result of the pilot test revealed that there are needed changes for systems improvement. The people involved in
this pilot test felt there was need to institutionalize at the three STRIVE target regions the technology of the Regional
Monitoring, Evaluation and Adjustment before its roll-out nationwide. There is also a need to engage the Public Schools
District Supervisor (PSDS)
1. Terms of Reference of QMTs. Selection of personnel that will compose the Quality Management Teams
(QMTs) at the region, division, district and school levels was based on criterion identified by the STRIVE
QAA/M&E System Monitoring Team which were deemed to contribute to the attainment of the objectives of
the system.
Reformulation of the TOR of District QMTs and acceptance by the Division Office. The pilot
implementation revealed the need for the Public School District Supervisors (PSDS) to be the Subdivision Quality Management Team (SDQMT) at the District Level. Hence a review of functions and
responsibilities of PSDS in relation to QAA/M&E is needed. It is suggested that the Regional Office
make an issuance mandating all Public School District Supervisors (PSDS) to be the SDQMT in the
District level.
Review of TOR of SQMT. Part of the study was the identification of hindering factors or those that
make their functions difficult to perform. The SQMTs of the pilot schools agreed that lack of timely
advocacy to the stakeholders made it difficult for them to coordinate needed tasks with them. Thus,
aside from the seven (7) identified roles and functions of the SQMT it is recommended to include:
a. Timely advocacy of the QAA System to internal and external stakeholders. - This was seen
as necessary to sustain the over-all support of these stakeholders in the schools programs
and projects. This is also seen as one mechanism to heighten the interest of the stakeholders
to participate in coming up with appropriate interventions to address the issues and concerns
identified.
b. Submit monthly report to the Sub-division QMT.
2.
Development of consolidation tools for district and division. Although the report revealed that the QMTs
were able to perform their roles and functions, it was also noted that some of their functions were not
accomplished within the specified time due to overlapping of activities. Some SQMTs and SDQMTs disclosed
that they find it difficult to submit monthly reports to the SDQMT and DQMT, respectively. Reason being that is
their schedule and proximity of their respective offices. To ensure that there is interdependency of M&E
processes at both the division and the district levels, the QMTs found that it will be necessary to develop a
consolidated tools for both the District and Division levels.
3.
Harmonization of required reports from the division. As mentioned above, timely submission of required
reports is as necessary as doing the tasks assigned to QMTs. Ideally, the School QMTs submit their monthly
report to the District or the Sub-division QMT, the SDQMTs submit their monthly report to the Division QMT.
The DQMT then consolidates the report submitted by the SDQMTs and submit their report to the Region QMTs
P a g e | 21
on a quarterly basis. Reports are main source of information for educational planning and decision making.
These reports are also used for identification of priority areas for technical assistance.
As mentioned in the preceding report, some SQMT find it easier to directly submit their report to the DQMT or
the Division SBM Coordinator/Chair rather than to the SDQMT. This process took more time because the
reports were then turned-over to the proper SDQMT for consolidation before submission to the DQMT. A
systematic way of gathering data for reporting purposes may be initiated as this affects effectiveness and
efficiency of the system. If reports are not harmonized, DQMTs could not get information that can be used for
TA planning.
4.
Improvement of flow of Division M&E processes to integrate TA of Div./District QMT to School QMT.
Some of the M&E tasks identified by the DQMTs to be difficult for them to perform are as follows:
Neg. Occidental Interacting with SDS/ASDS and the Sub-division M&E teams because of conflict in
their schedule.
Bohol/Tagbilaran Most of the DQMT members have no background in the development,
indigenization and localization of M&E tools. They found it hard to develop M&E tools and
instruments to suit the need of the division.
N. Samar Consolidating and analyzing M&E reports, both quantitative and qualitative. This is
because most of the team members are so focused on their organic tasks that they find it hard to
accomplish their M&E assignments.
Thus, it is suggested that QAA/M&E System processes and tools be part of the agenda during
Division Mancom. Use the team monitoring and supervision scheme wherein an ES is paired with a
PSDS in monitoring the progress of implementation of various education programs & projects. The
pilot report also suggest to include in the organic function of the PSDS (SDQMT) the development,
implementation and monitoring of the Technical Assistance Plan.
5.
QAA System as part of the SIP. Inclusion of a plan on how to quality assure the School Improvement Plan
(SIP) was seen as an indicator of a good plan. During the crafting of the SIP each planning team shall
prepare a plan on how to monitor and evaluate the implementation of the SIP.
b.
c.
6.2.2 Capability building programs QMTs should be provided with CB programs that will help them
perform their respective duties and responsibilities efficiently. They must be equipped with capabilities and
technical know-how to provide technical assistance to education managers at their jurisdiction. Some of the
CB programs identified during the pilot implementation are the following:
In-depth training on the utilization of M&E Reports for Technical Assistance planning for QMTs at the
Region, Division and District Levels (for effectiveness and relevance).
P a g e | 22
Fund Support
MOOE for the M&E activities; training of QMT members, procurement, repair and maintenance
of ICT equipment, travelling expenses
Travel allowance must be given to PSDS from the Division fund
Strict implementation of guidelines and intensified monitoring of downloading of MOOE of the
Division to schools
6.3 SUPPORT FOR THE ROLL OUT OF THE SYSTEM TO OTHER REGIONS
1.
2.
3.
Finalization of QAA-M&E Operations Handbook (will contain structures needed, processes & tools)
Development of Installation Guide for QAA-M&E System (for regions who would like to start
implementation of the system)
Refinement of Session guides for capability building
P a g e | 23
P a g e | 24
ANNEX A
POSITION
PSDS
PSDS
PSDS
NAME
Mariolina Jalwin
Ma. Lourdes Tison
Rosario Baes
POSITION
PSDS
PSDS
PSDS
EDUCATION SERVICE 3
SOUTH NEGROS OCC.
POSITIO
NAME
N
Caridad Ablao
Dolores Ramirez
Marilyn Cogonon
PSDS
PSDS
PSDS
P a g e | 25
School Head
M&E Coordinator
SPT Representative
Teacher
Parent
LGU
Department. Chairs / Subject Chair
ICT Coordinator/Teacher
Others : (Please specify)
E. List of Pilot Schools
Area (Sub-Division)
ESA 1
(North Negros Occ.)
Total = 17
ESA 2
(Central Negros Occ.)
Total = 17
ESA 3
(South Negros Occ.)
Total = 16
/
/
/
Elementary School
SECONDARY
SCHOOLS
/
/
/
/
/
/
/
/
Secondary School
Balogo ES
Calapnusan ES
Canubgan ES
Dolis ES
Escalante Central ES
Jonob-Jonob ES
Pandanon ES
Talisay South ES
Victorias North ES
Calatrava II CES
Bug-ang NHS
Calatrava NHS
Don Simplicio Lizares NHS
EB Magalona NHS
Escalante NHS
Manapla NHS
Tigbao NHS
Pontevedra North ES
Isabela East Elem School
La Castellana ES
Lopez Jaena ES
Mambalio ES
Mansablay ES
Moises Padilla ES
Sibucao ES
Soriano ES
Valladolid ES
Pontevedra South ES
Delicioso ES
Aguisan ES
Bacuyangan ES
Binalbagan ES
Binulig ES
Bulwangan NHS
Nabulao NHS
Hinigaran NHS
Himamaylan NHS
Negros Occidental National Agro-Industrial
School & Home Industries (NONAISHI)
Candoni CES
Cauayan Central School
Himamaylan CES
Hinigaran ES
Ilog Central School
Pilar ES
P a g e | 26
POSITION
PSDS
PSDS
PSDS
NAME
EDITO YANA
NOEL DUAVIS
LORNA TORREGOSA
POSITION
PSDS
PSDS
PSDS
EDUCATION SERVICE 3
NAME
POSITION
ESTRELLITA BARAJAN
URBANO BERNASOR
DELFINA OLAA
PSDS
PSDS
PSDS
P a g e | 27
SECONDARY
SCHOOLS
/
/
/
/
/
/
/
/
/
/
/
/
/
/
Elementary School
Antequera Central School
Balilihan Central School
Bingag Elementary School
Catigbian Central School
Cogon Elementary School
Hanopol Elementary School
Tagbilaran City Central School
Toril Elementary School
Tubigon Central School
Secondary School
Busao High School
Cahayag High School
Dr. Cecelio Putong National High School
Manga High School
Mayor A. Tuazon School of Fisheries
Tagbilaran City Science High School
ESA 2
Total = 23
Aguining Elementary
Banlasan Elementary School
Baud Elementary School
Bugang Elementary School
Buenavista Central School
Cataban Elementary School
Cawayan Elementary School
Clarin Central School
Dait Norte Elementary School
Hinlayagan Ilaud Elementary School
Lawis Elementary School
Pangpang Elementary School
Pres. C.P.G. Central School
Sto. Nio Elementary School
Tipolo Elementary School
ESA 3
Total = 15
P a g e | 28
NAME
DR. THELMA C. QUITALIG
DR. GORGONIO G. DIAZ, JR.
FERMINA VILLANOBOS
EMILY DELACRUZ
MARITES GUILLERMO
ANTONIO BARANDINO
JUN GEPOLLO
POSITION
AREA / DISTRICT 2
NAME
AREA / DISTRICT 3
NAME
POSITION
POSITION
AMADA MARQUITA
PSDS
CARLOS BALANQUIT
PSDS
NENA GARDUCE
PSDS
GLORIA MATIC
PSDS
FE TIU
PSDS
DODITO FLORALDE
PSDS
BERNARDINO BALANQUIT
PSDS
CLARITA CELAYA
PSDS
/
/
SECONDARY
SCHOOLS
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
P a g e | 29
Balicuatro Area
Total = 13
Secondary School
Laoang National High School
Laoang National Technical H S
Las Navas National High School
Mapanas Agro Industrial School
San Roque Pambujan Vocl. School
P a g e | 30
ANNEX B
REVISED TERMS OF REFERENCE
Quality Management Teams
A
Composition of SQMT:
For Small Elementary Schools:
School Head
School M&E Coordinator
SPT Representative
P a g e | 31
P a g e | 32
Composition of DQMT
Chair
Co-Chair
:
Members (Regular)
(On-call)
:
SDS
ASDS
: Division M and E Team (5)
Chair: Education Supervisor
Members:
2 Education Supervisors
Planning Officer
Division Statistician
District M and E Team*
Members:3 PSDSs per sub-division
:
(* The number of District M&E Teams will depend on the clustering of districts in the division. Each
Cluster will have its own-M&E team)
Terms of Reference
Chair: SDS
Function: Manages the division QAA M and E system
Ensures compliance of quality standards for basic education programs)
Responsibilities:
1. Establishes the DQMT following guidelines set
2. Manages the operation of the Division QAA M and E system
3. Acts on the recommendations from M and E reports
4. Submits quarterly Division QAA M & E report to the region
5. Submits division annual accomplishment report to the region and division stakeholders
6. Interacts with Region and Division Stakeholders, ASDS, Division M and E team
Co-chair: ASDS
Function:
Co-manage the division QAA M and E system
Responsibilities:
1. Takes the lead in the preparation of appropriate instruments and tools for M and E
2. Reviews findings and recommendations submitted by the Division M and E team
3. Communicates the reviewed M and E report to the SDS for appropriate action
4. Takes the lead in the preparation of TA Plans based on findings and recommendations of division M&E
reports
5. Submit reviewed findings and recommendations to the SDS for appropriate action
6. Interacts with SDS and D M&E team
Division M&E Team
Function:
Assures the conduct of quality planning and implementation of Technical Assistance to schools and
strengthening of SBM implementation such as:
- SIP Appraisal
- Curriculum Implementation
- Instructional Supervision conducted by SHs to teachers
- Technical Assistance provided by SHs to schools
- HRM processes
- School based training programs
- Learning Resource Materials utilization
- Resource generation and mobilization
P a g e | 33
P a g e | 34
Communicating status and progress of region to CO for policy adjustment and to stakeholders for resource
mobilization.
Composition of RQMT
Chair
Co-Chair
Members
(On-call)
:
ARD
QAA Division Chief
:
QAA Specialist (4)
Regional UIS Coordinator
:
Terms of Reference
Chair: ARD
Function:
Manages the regional QAA System following the set goals and directions of the region
Responsibilities:
1. Coordinates the development of strategic QAA-M&E Plan of the various units based on
QAA priorities of the region
2. Leads in the installation of units for the continuous operations of the QAA system to
ensure compliance to standards and processes in the region
3. Reviews and finalizes recommendations from QAAS unit for the improvement of regional
projects and programs for submission to RD.
4. Interacts with RD, Regional QAAS unit, and SDS
Co-chair: Regional QAA Division Chief
Function:
Manages the functions, operations and performance of Quality Management teams to
determine the efficiency & effectiveness of delivery and for accountability.
Responsibilities:
1. Monitors the operation of QAA-M&E in the regional office and divisions to assure the
quality of processes to be employed.
2. Provides efficient technical support to organized QAA-M&E team of regional units and
division for improved performance of their roles and responsibilities
3. Provides feedback on the monitoring and evaluation results to concerned entities to
recommend appropriate interventions.
4. Submits recommendations from M&E reports to ARD
5. Interacts with REXECOM
Regional QAA Specialists
Function:
Ensure the implementation of national and regional standards in all processes, outputs
and outcomes of the region and divisions
Responsibilities:
1. Provide technical assistance to divisions on the implementation of national & regional
standards for the improvement of division performance
2. Review QAA M & E Tools to ensure validity and reliability.
3. Validate and integrate findings of all M & E Results in the region
4. Ensure that the operation/ establishment of public/private schools is based on standards
set.
5. Conduct accreditation of public, vocational/technical, private schools and CLC ,e.g.
Project Sterling, PAASCU, APES.
P a g e | 35
To formulate
national education
policies and
standards
Region
Division
School
To manage the
QAAS in the entire
region
To ensure
compliance with the
standards in the
planning,
implementation and
monitoring and
evaluation of
education plans
To monitor and
evaluate progress of
regional initiatives
and performance
To communicate
M&E results to
inform policy and
technical assistance
P a g e | 36
ANNEX C
SYSTEM IMPLEMENTATION TOOLS
SIP IMPLEMENTATION TOOLS
1
SIP Three-year Report on Physical Output and Financial Status of Projects & Programs
10
11
12
P a g e | 37
C.1
Item
SIP Outputs
Indicator/s
Data Source
Identify the source where the measure of the attainment of objectives can be
Method of
Collection
Timing of Collection
Data Gatherer
Report Schedule
5
6
7
Instructions
project objective. This is a qualified/quantified parameter which details the
extent to which an objective has been achieved
C.2
RESULTS M&E
PROGRESS
Outputs
P a g e | 38
C.3
District: ______________________
School Year: _______________________________
Physical Output
Target
Actual
Financial Target
%
Budget
Expenditure
Note: Write only those outputs with physical and financial targets.
Legend: % - Percentage of Accomplishment (TARGET/ACTUAL x 100)
C.4
Plan Actual
Cumulative As of
Quarter
Accomplishment to date
%
Plan
%
%
Slippa
%
Actual
% Plan
% Plan
Slippage
Accomp (Total)
Accomp
Actual
ge
Actual
C.5
SIP Three-year Report on Physical Output and Financial Status of Projects & Programs
PHYSICAL OUTPUT
Program/Project
Outputs
Year 1
Target Actual
Year 2
FINANCIAL
Year 3
Year 1
Year 2
Year 3
% Budg
%
%
%
% Budg Actua % Budg
Target Actual
Target Actual
Actual
Actual
Acmp
Acmp
Acmp et
l Exp et
et
Exp
Exp
P a g e | 39
C.6
Report Period:
Accountable
Unit/Person
LESSONS
LEARNED:
SUSTAINABILIT
Y MEASURES
TAKEN:
C. 7
School :____________________________________
Implementation Period Covered: ______________
Division: ___________________________________
Region: ____________
P a g e | 40
C. 8
Take and analyse each of the issues listed in Part A as to whether they have been
resolved/addressed by the school.
Write in Column 1 these resolved issues and in Column 2, the solution applied.
In Column 3, list the issues which are beyond the schools control and which will need
the assistance of the district/division for its resolution.
2
Solution
applied
3
Issues Needing
Assistance
4
Remarks
C.9
2
Action taken to
apply strategy
3
Result of
strategy
P a g e | 41
P a g e | 42
C.10
Write in Column 1 those experiences/factors which have been productive, useful or influencing the
successful implementation of the project.
In Column 2, indicate those which have caused delay or unsuccessful delivery of the planned target.
In Column 3, write the Lessons Learned derived from both facilitating and hindering factors. Lessons
learned from facilitating factors are stated as values/ good practices/ principles. Lessons learned from
hindering factors are stated as what could have been.
1
Facilitating Factors
2
Hindering Factors
3
Lessons Learned
Objective
Source of Information
SCHOOL LEVEL
1. To evaluate the outcomes of a 3year SIP implementation.
DIVISION LEVEL
1. To evaluate school
performance based
accomplishments
2.
on
To determine appropriate Analysis of the SIP Terminal report of schools to SIP M & E Report
information for DEDP adjustment identify information for DEDP adjustment
Report on SBM Assessment
and directions for the succeeding
NCBTS/ TNASH Report
P a g e | 43
SIP preparation
3.
To determine areas for technical Analysis of the list of problems/ hindering factors DEDP Accomplishment Report
assistance to schools
encountered by schools in SIP implementation; SIP 3 year Accomplishment
List of identified best practices and Schools Report
Performance Assessment
SIP M & E Report
Achievement Results (NAT)
4. To obtain aggregated information Analysis of List of Recommendations and Adjusted SIP/ AIP
in aid of local policy formulation Sustainability Measures/ Lessons Learned/
SDQMT Report
Hindering Factors
Progress M & E Report
REGIONAL LEVEL
1. To check on the process of the Process check of Divisions utilization of SIP Process Audit checklist
division in appraising terminal
terminal reports
reports of schools
2. To obtain aggregated information Consolidation of:
DQMT reports
for policy review/
Facilitating Factors
formulation/localization/ standard
Hindering Factors
setting
Sustainability Measures
Recommendations
D.2
SIP End-of-Cycle Report
Introduction
A. SIP Accomplishments
a. Level of accomplishment of SIP targets
b. Trend Analysis of School Performance Indicators
c. Trend Analysis of Learners Achievement Performance
B. SBM Implementation
a. Current level of SBM practice
b. SBM indicators needing technical assistance
C. Input for Next SIP
a. List of delivered and undelivered outputs and reasons for backlogs
b. Lists of outputs to be carried to the next SIP
D. Lessons learned
a. Summary of Facilitating Factors
b. Summary of Hindering Factors
c. Summary of Sustainability Strategies Applied
E. Recommendations for Continuous Improvement of the School
P a g e | 44
D.3
Scaling Up
Progressing
Moving towards Progressing
Meeting Standards
Starting
Evaluation Area
(1)
I.
If response in Column 3 is
YES, check the extent of
increase in this column.
Response
(2)
Within the
Planning
Standard
(4a)
(3)
Exceeds
Planning
Standard
(4b)
1. Access
YES
NO
P a g e | 45
YES
2. Retention
NO
YES
NO
NO
YES
NO
YES
NO
YES
NO
YES
NO
4. Achievement
YES
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
P a g e | 46
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
2. Improved
teachers
performance in
managing the
teaching and
learning
process
3. Improved school
heads
performance in
school based
management and
instructional
supervision
4.
Improved
learning
environment
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
P a g e | 47
No. of Items
(1)
(2)
Number of YESResponses
with Accomplishment that
Exceeds Planning
Standard/Criteria
(3b)
2. Retention
3. Completion
4. Achievement
15
Sub-total:
% Score:
Sub-total:
% Score:
4
2
13
Sub-total:
Sub-total:
% Score:
% Score:
Overall Score:
Overall Score:
% Score:
% Score:
Descriptive Rating :
_________________
Descriptive
Rating
_________________
28
P a g e | 48
D.4
The assessment results can be converted into quick quantitative ratings by getting sub-total and over-all percentage
scores based on number of items for each part or for overall number of items. These numerical indices more or less
provide the region ideas on where the divisions stand along the indicators.
Percentage scores could be interpreted using the following descriptive ratings that suggest how a division has
performed after a full cycle of DEDP implementation.
95-100 %
85- 95 %
65 84 %
35- 64 %
34 % and below
Starting
Scaling Up
Progressing
Moving towards Progressing
Meeting Standards
Evaluation
Area
(1)
II.
M&E Questions
Response
(2)
(3)
If response in Column
3 is YES, check the
extent of increase in
this column.
Within the
Planning
Standard
(4a)
Exceeds
Planning
Standard
(4b)
1. Access
YES
NO
YES
NO
YES
NO
P a g e | 49
2. Retention
3.
Completion
c) Is there an increase in number of 3-5 yearold children availing pre-school and early
childhood care development programs?
(d) Are all learners who entered the elementary
schooling system ready?
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
4.
(a) Is there an increase in number of Grade 3
Achievement
pupils demonstrating the basic competencies in
reading and numeracy?
(e) Is there an increase in number of Grade 6
pupils demonstrating the desired competencies
as evidenced by the national achievement
assessment results?
(f) Is the school able to improve the achievement
of 2nd year high school learners in math,
science, English?
(d) Is there an increase in number of 4th year high
school learners/ ALS secondary learners
passing the functional literacy test?
5. Universal
(a)Is there an increase in number of Out-of-School
coverage of
Youths and adults participating (enrolled) in the
Out of School
basic literacy program, accreditation and
Youth and
equivalency program and ALS programs?
adults in the
(d)
Is there an increase in number of IPs
provision of
and Muslims participating in the basic literacy
basic learning
program, accreditation and equivalency
needs
program and ALS programs?
YES
NO
YES
NO
YES
NO
P a g e | 50
3. Management
and technical
assistance
systems are in
place and
operational
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
YES
NO
M&E Area
(1)
No. of
Items
(2)
2. Retention
3. Completion
4. Achievement
20
Sub-total:
% Score:
Sub-total:
% Score:
P a g e | 51
1
6
Sub-total:
Sub-total:
% Score:
% Score:
Overall Score:
Overall Score:
26
% Score:
Descriptive
Rating : _____________
% Score:
Descriptive
Rating : _____________
REPORT TEMPLATES
E.1
SCHOOL QMT M&E Report Template
Introduction
(Monthly)
A. On SIP/AIP Implementation
1. Physical Outputs
2. Financial Report
3. Balance and gaps
B. Other school projects (if applicable)
1. Training & Development
2. Learning Resources
3. Others ( SOBE, etc)
C. Qualitative Information
a. Issues and Concerns
b. Sustainability Measures Applied
c. Lessons Learned
d. Recommendations
(Quarterly )
A. Cumulative SIP Accomplishments for the Quarter
B. Learners Performance Indicators
1. Enrolment
2. Drop-out during the quarter
3. Failures during the quarter
(Annual)
1. Performance indicators
2. SBM Level of Practice
P a g e | 52
E.2
SUB-DIVISION QMT M&E Report Template
Introduction
Part I: (Quarterly)
A. On SIP Implementation-Consolidation of Level of Accomplishment of Schools in the District/ Zonal Area by
priority area
1. Physical Outputs
2. Financial Status
B. Status of Accomplishments on Special Projects of Schools in the area (if not included in SIP)
1. Training & Development
2. Learning Resources
3. Others
C. Qualitative Information
1. Consolidation of Issues & Concerns of Schools and their Resolutions
2. Consolidation of Sustainability Measures Applied by Schools
3. Consolidation of Lessons Learned by Schools
Part II:
Consolidated Learners Performance for the Quarter
Part III:
A. Recommended areas needing attention and possible solutions applied by the district.
B. Recommended areas needing technical assistance from the division.
E.3
DIVISION QMT M&E Report Template
Introduction
Part I: (Quarterly)
A. On SIP Implementation
Consolidation of SIP/AIP implementation of districts in the division
B. Status of Accomplishments on Special Projects in the Division
1. Training & Development
2. Learning Resources
3. Others
C. Qualitative Information
1. Issues & Concerns of Districts
2. Consolidation of Sustainability Measures Applied by the Districts
3. Consolidation of Lessons Learned in the Division
Part II:
A.
B.
C.
(Annual)
SBM Level of Practice Profile in the Division ( End of Year)
Consolidation of Learners Outcome Indicators
Consolidation of Summary of School Outcome Indicators
P a g e | 53