0% found this document useful (0 votes)
4 views

Four-Key-Components-of-Writing-Tests

The document outlines the principles and steps involved in test construction, emphasizing the importance of validity, reliability, and appropriate item types. It discusses the planning, preparation, trial, and evaluation phases of creating effective tests that accurately measure student knowledge and abilities. Key elements include determining objectives, preparing test specifications, and conducting item analysis to ensure the quality of the test.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
0% found this document useful (0 votes)
4 views

Four-Key-Components-of-Writing-Tests

The document outlines the principles and steps involved in test construction, emphasizing the importance of validity, reliability, and appropriate item types. It discusses the planning, preparation, trial, and evaluation phases of creating effective tests that accurately measure student knowledge and abilities. Key elements include determining objectives, preparing test specifications, and conducting item analysis to ensure the quality of the test.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF or read online on Scribd
You are on page 1/ 38
Jemima Sultana Department of Education Aligarh Muslim University Introduction A test refers toa tool, technique or a method that is intended to measure students knowledge or their ability to complete a particular task, In this sense, testing can be considered as a form of assessment. Tests should meet some basic requirements, suchas validity and reliability. testing is generally concerned with tuming performance into numbers. [Baxten,1998] class are caused by faulty test questions [World watch. The philadelphia trumpet, 2005) + [1 isestimated that 90% of the testing items are out of quality {[Witen Ww,1992) % The evaluation of peoples progress is a major aspect of teachers job. [Omaldo & Antario, 1995] 13% of students who fi Characteristics of a good test at) oseeUTol Orie asad omwenwe Principles of test construction Measure all instructional objectives Objectives that are communicated and imparted to the students. Designed as an operational control to guide the learning sequences and experiences. Harmonious to the teachers instructional objectives. Cover all learning tasks Measure the representative part of the learning task. Appropriate testing strategies or items Items which appraise the specific learning outcome. Measurements or tests based on the domains of learning. Principles of test construction 4, Make test valid and reliable * reliable when it produce dependant, consistent, and accurate scores. * Valid when it measures what it purports to measure. *Test which are written clearly and unambiguous are reliable. *Tests with fairly more items are reliable than tests with less items. *Tests which are well planned, covers wide objectives, & are well executed are more valid. Principles of test construction 5. Use test to improve learning * tests are not only an assessment but also it is a learning experience. «Going over the test items may help teachers to reattach missed items. * Discussion and clarification over the right choice gives further learning. 6. Norm referenced and Criterion referenced tests «Norm Referenced: higher and abstract level of cognitive domain. *Criterion Referenced: lower and concrete levels of learning. Types of Test Help educators place a student into a particular level or, section of a lang curriculum or s\ Lgeoite (ita) Patty They measure’ learner’s level o! language. They help teachers and learners to Btu iis ioied after some kind of instruction. 4 main steps of Test Construction = Step # 1. Planning the Test Planning of the test is the first important step in the test construction. The main goal of evaluation process is to collect valid, reliable and useful data about the student. Therefore before going to prepare any test we must keep in mind that: (1) What is to be measured? (2) What content areas should be included and (3) What types of test items are to be included. Therefore the first step includes three major considerations. + Preparing test specifications. 1. Determining the Objectives of Testing: +A test can be used for different purposes in a teaching learning process. It can be used to measure the entry performance, the progress during the teaching learning process and to decide the mastery level achieved by the students, *Tests serve as a good instrument to measure the entry performance of the students. *It answers to the questions, whether the students have requisite skill to enter into the course or not, what previous knowledge does the pupil possess. * Therefore it must be decided whether the test will be used to measure the eniry performance or the previous knowledge acquired by the student on the subject. 2. Preparing Test Specifications: In order to be sure that the test will measure a representative sample of the instructional objectives and content areas we must prepare test specifications. One of the most commonly used devices for this purpose is ‘Table of Specification’ or ‘Blue Print.’ Preparation of Table of Specification/Blue Print: +Preparation of table of specification is the most important task in the planning stage. «It acts, as a guide for the test construction. *Table of specification or ‘Blue Print’ is a three dimensional chart showing list of instructional objectives, content areas and types of items in its dimensions. It includes four major steps: eal a een (i) Determining the weightage to different instructional objectives: For example if we have to prepare a test in General Science for Class X we may give the weightage to different instructional objectives as following: Table Showing weightage given to different instructional objectives in a test of 100 marks: Objectives Weightage in % No. of Questions (marks) Knowledge 15% 15 Understanding 45%, 45 Application 30% 30 Skill 10% 10 Total 100% 100 (ii) Determining the weightage to different content areas: For example if a test of 100 marks is to be prepared then, the weightage to different topics will be given as following. Weightage of a topic = No. of pages in the topic x total number of items or Total number of pages in the book marks If a book contains 150 pages and 100 test/items (marks) are to be constructed then the weightage will be given as, following: 1 1 to 25 17 16.7% 2 26 to 75 33 33.3% 3 76 to 110 23 23.3% 4 111 to 150 27 26.7% Total 100 100% (iii) Preparation of item types to be included Items used in the test construction can broadly be divided into two types like objective type items and essay type items. For some instructional purposes, the objective type items are most efficient where as. for others the essay questions prove satisfactory. Appropriate item types should be selected according to the learning outcomes to be measured, For example: when the outcome is writing, naming then supply type items are useful. If the outcome is identifying a correct answer selection type or recognition type items are useful. So that the teacher must decide and select appropriate item types as per the learning outcomes, (iv) Preparation of the table of Specification: Class: VI ‘Time: 80-min ibe Retin BLUEPRINT M40 fobjectives Knowledge Understanding | Application (20%) (52.5%) (37.5%) Total Contents vss [SA] 1 | vsa | SA | tA [vsA] SA] tA Structure of plants 342) | siz) 20/8) (a) The Leaf 442) | 343) 8a) 3) ‘The Flower i) 54) 312) 12/4) TOTAL a2) 20) 1519) 40(16) ‘VSA: very short Answers * Number in bracket shows number of questions SA: Short Answers. * Number outside the bracket shows marks given LA: Long Answers to each questions Step # 2. Preparing the Test After planning, preparation is the next important step in the test construction. In this step the test items are constructed in accordance with the table of specification. Each type of test item need special care for construction. The preparation stage includes the following three functions: god -la ly 4 ag Lala i 8 a eae Ld clean aaa) ea é scoring key Preparation of test items is the most important task in the preparation step. Therefore care must be taken in preparing a test item. The following principles help in preparing relevant test items, 1. Test items must be appropriate forthe learning oulgome to be mensured. 2. Test items shoul measure all types of instructional objectives and the whole content area, 3. The test items should be free from ambiguity. Example: Poor item —Where did Gandhi born? Better —In whieh city did Gandhi born? 4, The test items should be of appropriate difficulty level: The items should not be so easy that everyone answers it correctly and also it should not be so difficult that everyone fails to answer it. The items should be of average difficulty level. 5, The test item must be five from technical errors and imelevamt clues: For example: grammatical inconsistencies, verbal associations, extreme words (ever, seldom, always), and mechanical features (correct statement is longer than the incorrect). Therefore while constructing a test item careful step must be taken to avoi most of these clues. 6, Test items should be free from ra While portraying 2 of all the facies of the society shold be given equal importance. The terms used in the test item Should have an universal meaning to all members of group. ethnic and sexual biasnesss The validity and reliability of the test items to a great extent depends upon the instructions for the test. N.E. Gronlund has suggested that the test maker should provide clearcut direction about; ait g i) Ra Lita ODirections must be given whether the scoring will be made by a scoring key(when the answer is recorded on the test paper) or by a scoring stencil and how marks will be awarded to the test items. QA scoring key helps to obtain a consistent data about the peoples performance. So the test maker should prepare a comprehensive scoring procedure along with the test items, Point Method Rating Method Step #3. Try Out of the Test Once the test is prepared now it is time to be confirming the validity, reliability and usability of the test. Try out helps us to identify defective and ambiguous items, to determine the difficulty level of the test and to determine the discriminating power of the items. Try out involves two important functions: (a) Administration of the test: ¥ Administration means administering the prepared test on a sample of pupils. So the effectiveness of the final form test depends upon a fair administration. ¥Gronlund and Linn have stated that- ‘the guiding principle in administering any class room test is that all pupils must be given a fair chance to demonstrate their achievement of learning outcomes being measured.’ Vit implies that the pupils must be provided congenial physical and psychological environment during the time of testing. Any other factor that may affect the testing procedure should be controlled. One should follow the following principles during the test administration: Talk as less as possible No interruption at the time of test Not give any hints to any student Proper invigilation to prevent cheating (b) Scoring the test: Once the test is administered and the answer scripts are obtained the next step is to score the answer scripts. A scoring key may be provided for scoring when the answer is on the test paper itself Scoring key is a sample answer script on which the correct answers are recorded. Scoring stencil Correction for guessing: When the pupils do not have sufficient time to answer the test or the students are not ready to take the test at that time they guess the correct answer, in recognition type items. In that case to eliminate the effect of guessing the following formula is used: Score=R - __W n-l where R= No. of right responses W= No. of wrong responses n= No. of alternatives But there is a lack of agreement among psychometricians about the value of the correction formula so far as validity and reliability are concerned. In the words of Ebel - “neither the instruction nor penalties will remedy the problem of guessing.” Step #4. Evaluating the Test Evaluation is necessary to determine the quality of the test and the quality of the responses. Quality of the test implies that how good and dependable the test is? (Validity and reliability). Quality of the responses means which items are misfit in the test. It also enables us to evaluate the usability of the test in general classroom situation, Evaluating the test involves following functions: (i) Item Analysis Procedure Item analysis procedure gives special emphasis on item difficulty level and item discriminating power. The item analysis procedure follows the following steps: 1. The test papers should be ranked from highest to lowest. 2, Select 27% test papers from highest and 27% from lowest end, For example if the test is administered on 60 students then select 16 test papers from highest end and 16 test papers from lowest end. 3, Keep aside the other test papers as they are not required in the item analysis. 4, Tabulate the number of pupils in the upper and lower group who selected each alternative for each test item, 5. Calculate item difficulty for each item by using formula: Item difficulty= Rx 100 7 Where R= Total number of students got the item correct. T =Total number of students tried the item. Example: out of 32 students from both the groups 20 students have answered the item correctly and 30 students have tried the item. The item difficulty is as following: Item difficulty = x 100 Rk T 20. x 100 = 66.67 3 rc) It implies that the item has a proper difficulty level. Because it is customary to follow 25% to 75% rule to consider the item difficulty. It means if an item has a item difficulty more than 75% then is a too easy item if it is less than 25% then item is a too difficult item. 6. Calculate item discriminating power by using the following formula: Item discriminating power = Ru - RE T/2 Where Ry= Students from upper group who got the answer correct. R,= Students from lower group who got the answer correct. T/2 = half of the total number of pupils included in the item analysis. Example: Out of 32 students 15 students from upper group responded the item correctly and 5 from lower group responded the item correctly. Item discriminating power = Ru - RE TR =15-5 =0.63 32/2 ‘A high positive ratio indicates the high discriminating power. Here 0.63 indicates an average discriminating power, Qf all the 16 students from lower group and 16 students from upper group answers the item correctly then the discriminating power will be 0.00. It indicates that the item has no discriminating power. UiIf all the 16 students from upper group answer the item correctly and alll the students from lower group answer the item in correctly then the item discriminating power will be 1.00 it indicates an item with maximum positive discriminating power. 7. Find out the effectiveness of the distracters A distractor is considered to be a good distractor when it attracts more pupil from the lower group than the upper group. The distractors which are not selected at all or very rarely selected should be revised. Total No. of | Item Pais Pret ca Ae i upper 16 (OO) 2255) fw) 0) 16 lower 16 5 45 00 14 neree we In our example (Table below) distractor ‘D* attracts more pupils from Upper group than the lower group. It indicates that distracter ‘D" is not an effective distracter. “E’ is a distracter which is not responded by any one. Therefore it also needs revision. Distracter ‘A’ and ‘B* prove to be effective as it attracts more pupils from lower group. (a) Preparing a test item file: It can be done with item analysis cards. The items should be arranged according to the order of difficulty. While filing the items the objectives and the content area that it measures must be kept in mind. This helps in the future use of the item. (b) Determining Validity of the Test: At the time of evaluation it is estimated that to what extent the test measures what the test maker intends to measure. (c) Determining Reliability of the Test: Evaluation process also estimates to what extent a test is consistent from one measurement to other. Otherwise the results of the test can not be dependable. (d) Determining the Usability of the Test: Try out and the evaluation process indicates to what extent a test is usable in general class-room condition, It implies that how far a test is usable from administration, scoring, time and economy point of view.

You might also like