ASSESSMENT OF LEARNING [CBRC]

DEFINITION OF TERMS Process of gathering, describing or quantifying information about student performance. May include paper and pencil test, extended responses and performance assessment are usually referred to as “authentic assessment” tasks

 

MEASUREMENT A process of obtaining a numerical description of the degree to which an individual possesses a particular characteristic. It answers the question “How much?”

 

EVALUATION Process of examining the performance of student. It also determines whether or not the students has met the lesson instructional objectives.

 

TEST Instrument or systematic procedure designed to measure the quality, ability, skill or knowledge of students by giving a set of questions in a uniform manner. It answers the question “How does an individual student perform?”

TESTING Method used to measure the level of achievement or performance of the learners. It also refers to the administration, scoring and interpretation of an instrument designed to elicit information about performance in a sample of a particular area of behavior. 


 

NORM REFERENCED

CRITERION REFERENCED

·         Designed to measure the performance of a student compared with other students.

·         Each individual is compared with other examinees and assigned a score, usually expressed as a percentile, a grade equivalent score or a stanine. The achievement of student is reported for broad skill areas, although some norm-referenced tests do report student achievement for individual.

·         The purpose is to rank each student with respect to the achievement of others in broad areas of knowledge and to discriminate high and low achievers.

·         Designed to measure the performance of students with respect to some particular criterion or standard.

·         Each individual is compared with a pre-determined set of standards for acceptable achievement. The performance of the other examinees is irrelevant. A student’s score is usually expressed as a percentage and student achievement is reported for individual skills.

·         The purpose is to determine whether each student has achieved specific skills or concepts and to find out how much students know before instruction begins and after it has finished.

·         AKA Objective-referenced, domain-referenced, content referenced and universe referenced

DIFFERENCES

Typically covers a large domain of learning tasks, with just a few items measuring each specific task.

Typically focus on a delimited domain of learning tasks, with a relative large number of items measuring each specific task

Emphasize discrimination among individuals in terms of relative of level of learning

Emphasize what individuals can and cannot perform

Favor items of large difficulty and typically omits very easy and very hard items

Match item difficulty to learning tasks, without altering item difficulty or omitting easy or hard items

Interpretation requires a clearly defined group

Interpretation requires a clearly defined and delimited achievement domain

TYPES OF ASSESSMENT


Placement Assessment

·        Concerned with the entry performance of student.

·        Purpose is to determine the prerequisite skills, degree of mastery of the course objectives and the best mode of learning.

Diagnostic Assessment

·        Type of assessment given before instruction.

·        Aims to identify the strengths and weaknesses of the students regarding the topics to be discussed.

·        Purposes:

1.     Determine level of competence of the students

2.     Identify students who already have knowledge about the lesson

3.     Determine the causes of learning problems to formulate a plan for remedial action.

 

Formative Assessment

·        Used to monitor the learning progress of the students during or after instruction.

·        Purposes:

1.     Provide feedback immediately to both student and teacher regarding the success and failures of learning.

2.     Identify the learning errors that is in need of correction

3.     Provide information to the teacher for modifying instruction to improve learning.

Summative Assessment

·        Usually given at the end of a course or unit.

·        Purposes:

1.     Determine the extent to which the instructional objectives have been met

2.     Certify student mastery of the intended outcome and used for assigning grades

3.     Provide information for judging appropriateness of the instructional objectives

4.     Determine the effectiveness of instruction.


MODES OF ASSESSMENT

 

Traditional Assessment

·        Assessment in which students typically select answer or recall information to complete the assessment. Tests may be standardized for teacher-made, and these tests may be multiple choice, fill in the blanks, matching type.

·        Indirect measures of assessment since the test items are designed to represent competence by extracting knowledge and skills from their real-life context.

·        Items on standardized instrument tend to test only the domain of knowledge and skill to avoid ambiguity to the test takers.

·        One-time measures are used to rely on a single correct answer to each item. There is a limited potential for traditional test to measure higher order thinking skills.

 

Performance Assessment

·        Assessment in which students are asked to perform real-world tasks that demonstrate meaningful application of essential knowledge and skills

·        Direct measures of student performance because tasks are designed to incorporate contexts, problems, and solution strategies that students will use in real life.

·        Designed ill-structured challenges since the goal is to help students prepare for the complex ambiguities in life.

·        Focus on processes and rationales. There is no single correct answer; instead, students are led to craft polished, thorough and justifiable responses, performances and products.

·        Involve long-range projects, exhibits and performances are linked to the curriculum

·        The teacher is an important collaborator in creating tasks, as well as in developing guidelines for scoring and interpretation.

 

Portfolio Assessment

·        A collection of student’s works, specifically selected to tell a particular story about the student.

·        A portfolio is not a pile of student work that accumulates over a semester or year.

·        A portfolio contains a purposefully selected subset of student work.

·        It measures the growth and development of students

FACTORS TO CONSIDER: GOOD TEST ITEM

VALIDITY

RELIABILITY

ADMINISTRABILITY

SCORABILITY

Degree to which the test measures what it intends to measure. It is the usefulness of the test for a given purpose. A valid test is always reliable

Consistency of scores obtained by the same person when retested using the same instrument or one that is parallel to it.

Test should be administered
uniformly to all students so that the scores obtained will not vary due to factors other than differences of the student’s knowledge and skills.

There should be a clear provision for instruction for the students, proctors and the scorer.

Test should be easy to score.

Directions for scoring should be clear. The test developer should provide the answer sheet and the answer key.

APPROPRIATENESS

ADEQUACY

FAIRNESS

OBJECTIVITY

Mandates that the test items that the teacher construct must assess the exact performances called in for in the learning objectives. The

test items should require the same performance of the student as specified in the learning objectives

The test should contain a wide sampling of items to determine the educational outcomes or abilities so that the resulting scores are

representative of the total performance in the areas measured.

Mandates that the test should not be biased to the examinees. It should not be offensive to any examinee subgroups. A test can only be good if it is also fair to all test takers.

Represents the agreement of two or more raters or a test
administrators concerning the score of a student. If the two raters who assess the same student on the same test cannot agree on the score, the test lacks objectivity, and the score of neither the judge is valid, thus, lack of objectivity reduces test validity in the same way that the lack of reliability influences validity.


BACK