Friday, March 24, 2017

GRADING AND REPORTING

NEW K -12 GRADING SYSTEM

 
Components of
Summative Assessment
 

           ·           Written Work (WW)
           ·           Performance Tasks (PT)
           ·           Quarterly Assessment (QA)

Components
Purpose
Written
Work
(WW)
1.      Assess learners’ understanding of  concepts and application of skills
2.     Prepare learners for quarterly assessments

Components
Purpose
Performance Tasks (PT)
1.      Involve students in the learning process
2.     Give students opportunities to demonstrate and integrate their knowledge understanding and skills
3.     Give students the freedom to express their learning appropriate and diverse ways.
4.      

Components
Purpose
When Given
Quarterly Assessment (QA)
1.      Synthesize all the learning skills, concepts and values learned entire quarter.

Once at the end of the quarter

Grading System
The K-12 Basic Education Program (BEP) uses a  standards-based and competency-based grading system.
All grades will be  based on the weighted score of the learners summative assessments.
The minimum grade to pass a specific learning area is 60, which transmuted to 75 in the Report Card.
Grading System
The lowest mark that can appear on the Report Card is 60 for Quarterly Grades and Final Grades
Learners from Grades 1 to 12 are graded on Written Work (WW), Performance Tasks(PT), and Quarterly Assessment (QA) every quarter.
These three are given specific percentage weights that vary according to the nature of the learning area.  
Weight of the components for SHS(Senior HS)


METHODS OF INTERPRETING THE RESULTS last:
1.      Norm-referenced Interpretation
Ø  Describes student’s performance or progress in relation to others of the same peer group, age or ability
Ø  May involve ranking or scaling a pupil to help with streaming classes
Ø  May look at cross-school achievements to compare achievement in particular groups, subjects and years wit local and national levels of attainment

Norm-referenced Interpretation
e.g.
Ø  Sam ranked 5th in the unit test in Physics
Ø    Scion’s percentile rank in the Math achievement test is 88.

METHODS OF INTERPRETING THE RESULTS
2.     Criterion-referenced Interpretation-describes student performance according to a specified domain or clearly defined learning tasks e.g. divide three-digit whole numbers correctly and accurately, multiply binomial terms correctly
Ø  Concerned with national examination and other assessment bodies
Ø  Used in the assessment of vocational and academic qualifications
Ø  Results are given on a pass/fail,  competent/not competent basis
Ø  Results are conclusive and usually open to review

Criterion-referenced Interpretation
e.g.

Ø  Hyacinth scored 7 out of 10 in the spelling test
Ø    Ophie can encode an article with no more than 5 errors in spelling.

EDUCATIONAL EVALUATION

Quality
  • The European committee for the standardization agreed on the following definition in the non-governmental and business fields, published as a European standard EN ISO 9000. “Quality is a degree to which a set of inherent characteristics fulfils requirements. The term quality can be used with adjectives such as poor, good or excellent.
  • Despite the large number of reference and the growing body of literature about quality in education (quality control, auditing, valorization, policies and public funding provisions..) the concept of quality is still not very clearly and articulated.
           
           With this background, Harvey and Green (1993), have identified and analysed four different dimensions of quality, which are, in our opinion, a substantial contribution for the understanding of quality in non formal education.

The quality management cycle and youth work
Dr. William Edwards Derning (1900-1993), known as the father quality management developed the so called P-D-C-A Cycle.
The quality management cycle
Every project is like a circle with four parts or phase:
           ·           Planning
           ·           Doing
           ·           Checking/Studying
           ·        Acting

 

  The Cycle Theory In Youth Work
           ·           Derning’s theory was a starting point for developing different models of quality management system for different fields (industry, public administration, the primary sector..)
           ·           In later adaptations, the circle developed into a spiral to underline the ongoing characteristics of a project. As we know every project has its own dynamic and there is never a real beginning and never complete end, because the concrete end is often already the first step into a “new”.
           ·           

              Quality Criteria: characteristics chosen to define quality in coherence with the objectives of a project. The assumption is that if a criterion is fulfilled, the quality is good and if it is not fulfilled not so good or bad. Quality criteria are relevant and valid.

An example of quality criteria in youth work could be:
Involvement of Participants
-          “Involvement of participants” is a characteristic  of an educational project.
-          It is coherent with the objectives of the project. Youth projects in general have as objectives the promotion of active participation, engagement, co-responsibility and the use of active methodologies.
-          The assumption is that if there is involvement of participants in good and if it there is no involvement it is bad.
-          “involvement of participants” is relevant and valid criterion.

¨  Quality Standard: these are conditions set to criteria. Standards are often expressed  as the highest or lowest value or allowed range. They define the expected amount, the intensity and the manner of the criteria.

 An example of quality standard in youth work for the quality criteria “involvement of participants” could be:

¨  Participants self-management and organise their projects and the free time
-          “the self management and organisation of their projects and free time” is a condition, an amount, a manner of looking at the criteria “involvement of participants”
-           there are other criteria to evaluate the involvement of participants but this standard (about their project and the free time) what is expected.
       Quality indicators. Indicators are measures that indicate, if expected quality described in the standard is achieved.

¨  Participants self-management and organise their projects and the free time could be:
o    Number and kind of projects
o    Project team composition
o    Number and kind of self-organized activities during the free time
o    Social committee’s way of working
o    Sharing responsibilities

¨  Example of quality criteria, standards and indicator
In addition to the example given defining quality criteria, standards and indicators, let us imagine youth exchange project with the objective of promoting mutual understanding and co-operation between participants…
  •     A quality criterion could be:
  •             a communication between the participants
  •     A quality standard could be:
  •             communication in the project happens:
  • -          Between  all participants of the project
  • -          All through the different phases of the project
  • -          In the different activities and moments of the day.
  •        Some quality indicators could be:
  • -          Interaction between participants during the free time and informal time
  • -          Direct communication/interaction of participants in the “working spaces”
  • -          Initiatives to overcome communication

¨  Quality  in youth: steps and debates in the Europe institutions:

Quality in youth work at European level
        What do we consider “good” or “bad” in our work in educational youth projects? Establishing evaluation criteria is an essential and sensitive part of the evaluation process of an educational activity.
The quality of non formal education and youth work is ingoing discussion among researchers, trainers, and youth workers in Europe, which has been stimulated by the council of Europe and by the European commission.
Quality is also relevance-with regard to life skills, cognitive learning and understanding, living in groups and communities.

¨  European level initiatives on quality in the youth field
Both, the European Commission and the Council of Europe put a strong emphasis on quality assurance and quality development in their policies and activities.
q  The  long term training course Advanced Trainers in Europe (ATTE) has been developed to meet the increasing need for qualified trainers in this field and to enlarge and further develop the European networks of trainers, who have the competence and the motivation to develop and implement European level training activities in the youth field.

¨  Develop your own evaluation
Experienced evaluators agree that there are no “golden rules” for the development of a perfect educational evaluation. We believe that the definitions, models and questions explored in this “theoretical” part of the T-Kit (“Ingredients of Educational Evaluation”) can help a lot but they cannot guarantee an evaluation process free of resistance and mishaps.

Who and for whom? The actors of educational evaluation
       Having reviewed the possible definitions aims, operational objective and fields, it is not difficult to identify the actors of educational evaluation. The questions “who” and “for whom” are deliberately formulated together because everybody involved in the educational process should participate in its evaluation and should be informed about its results as follows:
  • The participants: as learners and the target group of the activity.
  • The Facilitator: leaders or team members, responsible persons for animating the activity.
  • The organizers and the partners: as promoters of the activity.
  • The funders: as supporters of the activity.
  • The decision-makers: as those “responsible” for considering the results of the evaluation in further decision making process.

When to evaluate? The timing of educational evaluation
       Evaluation is an ongoing and continuous process! However, “when” we evaluate is also an important decision to make in the planning of the evaluation process.

Main types of evaluation:
1.                Initial evaluation – is the evaluation done at very beginning.
2.                Mid-term evaluation – is our chance to check how things are going.
3.             Final evaluation – using a variety of methods, the participants and the team evaluate the whole project in the light of the outcomes.

Typologies of evaluation
       We have seen in the previous section different types of evaluation according to their timing.

Personal/Interpersonal/Group
       Personal evaluation – is the kind of evaluation in which each individual or actor involved in the educational process.

Interpersonal evaluation – happens when more than one individual actor involved in the educational process.

Quantitative/Qualitative Evaluation:
      Quantitative evaluation – focuses on the quantity of experience.

  Qualitative evaluation – relates to the quality of the program and of the experiences.




VALIDITY AND RELIABILITY


VALIDITY OF A TEST
Validity – (Airasian, 2000) is concerned whether the information obtained from an assessment permits the teacher to make a correct decision about a student’s learning.


Types of Validity:
   1.   Content Validity – refers to the relationship between a test and the instructional objectives, establishes content so that the test measures what it is supposed to measure.
     2.   Criterion-related Validity – refers to the extent to which scores from a test relate to theoretically similar measures.
a.   Concurrent Validity – the criterion and the predictor data are collected at the same time.
b.   Predictive Validity – refers to a measure of the extent to which a student’s current test result can be used to estimate accurately the outcome of the student’s performance at a later time.
        3.    Construct Validity – refers to the measure of the extent to which a 
                                test measures theoretical and unobservable variable qualities such as intelligence, math achievement, performance anxiety, and the like, over a period of time on the basis of gathering evidence.
a.   Convergent Validity – is a type of construct validation wherein a test has a high correlation with another test that measures the same construct.
b.   Divergent Validity – is a typeof construct validationwherein a test has low correlation with a test that measures a different construct.
c.    Factor Analysis – is another method of assessing the construct validity of a test using complex statistical procedures conducted with different procedures.
Important things to remember about Validity:
              1.      Validity refers to the decisions we make, and not to the test itself or to the measurement.
           2.     Like reliability, validity is not an all-or-nothing concept; it is never totally absent or absolutely perfect.
         3.   A validity estimate, called a validity coefficient, refers to specific type of validity. It ranges between 0 to 1.
       4.   Validity can never be finally determined; it is specific to each administration of the test.
Factors affecting the Validity of a Test Item:
1.   The test itself.
2.   The administration and scoring of a test.
3.   Personal factors influencing how students response to the test.
4.   Validity is always specific to a particular group.

Reasons that reduce the Validity of the Test Item:
1.   Poorly constructed test items.
2.   Unclear directions.
3.   Ambiguous test items.
4.   Too difficult vocabulary.
5.   Complicated syntax.
6.   Inadequate time limit.
7.   Inappropriate level of difficult.
8.   Unintended clues.
9.   Improper arrangement of test items.

Guide Questions to Improve Validity:
1.   What is the purpose of the test?
2.   How well the instructional objectives selected for the test do represents the instructional goals?
3.   Which test item format will be the best measure achievement of each objective?
4.   How many test items will be required to measure the performance adequately to each objective?
5.   When and how will the test be administered?

Validity Coefficient – is the computed value for the rxy.
Coefficient of determination – indicates how much variation in the criterion can be counted for by the predictor (teacher test).

Example:


Interpretation:
          The correlation coefficient is 0.94, which means that the validity of the test is high, or 88.36% of the variance in the student’s performance can be attributed to the test.

RELIABILITY of a TEST

Reliability – refers to the consistency with which it yields the same rank for individual who takes the test more than once (Kubiszyn and Borich, 2007).


Factors Affecting Reliability of a Test:
1.   Length of the test.
2.   Moderate item difficulty.
3.   Objective scoring.
4.   Heterogeneity of the student group.
5.   Limited time.

Four Methods of Establishing Reliability of a Test:
1.   Test-retest Method – a type of reliability determined by administering the same test twice to the same group of students with any time interval between the tests.

FORMULA:

Rxy=     (n) (Σxy) –(Σx)(Σy)
              √ [(n)( Σx²) - (Σx)²] [(n)(Σy²) – (Σy)²]
2.   Equivalent Form – a type of reliability determined by administering two different but equivalent forms of the test to the same group of students in close succession.
3.   Split-half Method – administer test once and score two equivalent halves of the test.
4.   Kuder-Richardson Formula –administer the test once. Score the total test and apply the Kuder-Richardson formula.

Reliability Coefficient – is a measure f the amount of error associated with the test scores.

Description of reliability Coefficient:
a.   The range of the reliability coefficient is from 0-10.
b.   The acceptable range value is 0.60 or higher.
c.    The higher the value of the reliability coefficient, the more reliable the overall test scores.
d.   Higher reliability indicates that the test items measure the same thing.



Interpreting reliability coefficient



1.    The group variability will affect the size of the reliability coefficient.
2.   Scoring reliability limits the test score reliability.
3.   Test length affects test score reliability. As the length increases, the test’s reliability tends to go up.
4.   Item difficulty affects test score reliability. As test items become very easy or very hard, the test’s reliability goes down.


Level of Reliability Coefficient:



Example solving:

students
FT (x)
ST (y)
xy
X2
Y2
1
36
38
1368
1296
1444
2
26
34
884
676
1156
3
38
38
1444
1444
1444
4
15
27
405
225
729
5
17
25
425
289
625
6
28
26
728
784
676
7
32
35
1120
1024
1225
8
35
36
1260
1225
1296
9
12
19
228
144
361
10
35
38
1330
1225
1444
N=10
∑x=274
∑y=316
∑xy=9192
∑ X2=8332
∑ Y2=10400

  



   
Analysis:
          The reliability coefficient using the Pearson r = 0.76 means that it has  a high reliability. The scores of the 10 students conducted twice with one-week interval are consistent. Hence, the test has a high reliability.

RERFLECTION:
          Testing the reliability of a test and its importance is very important to us. Questions, exams, procedures must be measured to the public to avoid using unreliable items. Learning of students mostly takes place when they are assessed so it is important that a question in the assessment is very reliable and valid.