Automated Item Banking and Test Development Model used at the SSAC.

Similar documents
University of Moratuwa

MASTER OF ENGINEERING PROGRAM IN INFORMATION

Preparing for the CHT Examination

Hand Therapy Certification and the CHT Examination

MASTER OF SCIENCE IN COMPUTER SCIENCE

Announcement date: December 1, 2009 New program launch date: May 1, 2010

Title Core TIs Optional TIs Core Labs Optional Labs. All None 1.1.6, 1.1.7, and Network Math All None None 1.2.5, 1.2.6, and 1.2.

Engineering Graphics Concept Inventory

PECT Test Development Process and Test Preparation. May 2013

Scope and Sequence for the Maryland Voluntary State Curriculum for Mathematics

STUDY GUIDE. Illinois Licensure Testing System. Computer Science (038) This test is now delivered as a computer-based test.

Instructors Experiences in the Cisco Networking Academy: Impact of the Curriculum on Teaching

Eight units must be completed and passed to be awarded the Diploma.

PMP Certification Preparatory Course

Bachelor of Engineering Technology (Electronics & Controls) Curriculum Document. Australian College of Kuwait. (September 2015) BEEF15 - Version 5.

PE for Software Engineers

Cisco Networking Academy Program

ASOPE PROCTOR TRAINING AND GUIDELINES

POSTGRADUATE CERTIFICATE IN LEARNING & TEACHING - REGULATIONS

Connecticut Alternate Assessment: Individual Student Report Performance Literals Mathematics

Title Core TIs Optional TIs Core Labs Optional Labs. 1.1 WANs All None None None. All None None None. All None 2.2.1, 2.2.4, 2.2.

ALABAMA FIRE COLLEGE AND PERSONNEL STANDARDS COMMISSION ADMINISTRATIVE CODE CHAPTER 360-X-2 REQUIREMENTS FOR CERTIFIED FIRE FIGHTER TABLE OF CONTENTS

BSc (Honours) Computer Science Curriculum Outline

ANNUAL NATIONAL ASSESSMENT 2015 ASSESSMENT GUIDELINES MATHEMATICS GRADE 7

PMI-ACP Certification Preparatory Course

PMI-PMP CERTIFICATION. (You Initiate, We Close)

Program Report for the Preparation of Social Studies Teachers National Council for Social Studies (NCSS) 2004 Option A

Student Orientation to Skills Testing

1 st Grade Math Curriculum Crosswalk

PHYSICAL EDUCATION (ADVANCED)

ANNUAL NATIONAL ASSESSMENT 2014 ASSESSMENT GUIDELINES MATHEMATICS GRADE 7

1. BACHELOR OF SCIENCE IN COMPUTER SCIENCE EDUCATION

PMP Certification Preparatory Course

CORE BODY OF KNOWLEDGE MATH GRADE 6

MIAMI-DADE COUNTY PUBLIC SCHOOLS District Pacing Guide GEOMETRY HONORS Course Code:

3 Continuation. 5 Continuation of Previous Week. 6 2, 7, 10 Analyze linear functions from their equations, slopes, and intercepts.

Programme Regulations

5 th International Symposium 2015 IntSym 2015, SEUSL

STEM Scholars Radius: STEM Readiness

b. Remediation will only be offered for courses offered at the home school.

The course makes up the first semester of the BSc programme in Design of Information Systems or can be taken as a freestanding course.

Virtual Training for the Flexographic Professional

NLN PAX TESTING INFORMATION, MATC REQUIREMENTS AND HOW TO REGISTER

Loyola Marymount University School of Education

COURSE TITLE. Introduction to Java LENGTH. One Semester Grades DEPARTMENT. Computer Department Barbara O Donnell, Supervisor SCHOOL

Field Testing. *NEW Benefits of Field Testing

Bachelor of Science Information Studies School of Information Program Summary

IT 204 Final Project Guidelines and Rubric Database Proposal and Implementation Plan Report

GLS UNIVERSITY. Faculty of Computer Technology Master of Computer Applications (MCA) Programme

customer service

CRITERIA FOR ACCREDITING COMPUTING PROGRAMS

MOKASA JOINT EXAMINATION Kenya Certificate of Secondary Education (K.C.S.E)

LIS 2680: Database Design and Applications

Diploma Of Computing

INFC20, Informatics: Advanced Database Systems, 7.5 credits Informatik: Avancerade databassystem, 7,5 högskolepoäng First Cycle / Grundnivå

INFORMATION AND COMMUNICATION TECHNOLOGY (ICT)

ASSESSING CRITICAL AND CREATIVE THINKING

Work-ready skills in Business, Administration and IT

School of Engineering and Built Environment. MSc. Applied Instrumentation and Control (Oil & Gas)

North Carolina Testing Program

Teacher Education: Reading Certificate, Reading Language Arts Certificate

The University of Jordan. Accreditation & Quality Assurance Center. Curriculum for Doctorate Degree

Topics covered 10/12/2015. Pengantar Teknologi Informasi dan Teknologi Hijau. Suryo Widiantoro, ST, MMSI, M.Com(IS)

This table connects the content provided by Education Perfect to the NSW Syllabus..

Objective 1 : The student will demonstrate an understanding of numbers, operations, and quantitative reasoning.

SYLLABUS. Departmental Syllabus. Structured Query Language (SQL)

How does Algebra and Geometry help us make sense of our world?

ANNUAL NATIONAL ASSESSMENT 2014 ASSESSMENT GUIDELINES MATHEMATICS GRADE 8

Document Control Information

Title Core TIs Optional TIs Core Labs Optional Labs. All None 1.1.4a, 1.1.4b, 1.1.4c, 1.1.5, WAN Technologies All None None None

A New Baccalaureate Program in Wireless Engineering. Victor P. Nelson 1, Richard O. Chapman 2, Richard C. Jaeger 1 Auburn University, Alabama

This course supports the assessment for Scripting and Programming Applications. The course covers 4 competencies and represents 4 competency units.

PNC.com, Weather.com & SouthWest.com. Usability Analysis. Tyler A. Steinke May 8, 2014 IMS 413

KOMAR UNIVERSITY OF SCIENCE AND TECHNOLOGY (KUST)

Honours in Information Systems

New Edexcel International GCSEs

COURSE PLAN Regulation: R11 FACULTY DETAILS: Department::

ASSIUT UNIVERSITY. Faculty of Computers and Information Department of Information Technology. on Technology. IT PH.D. Program.

Edexcel Maths Mock Paper December 2013

Regulatory Affairs Certification

Industrial Engineering Education in the Philippines Eugene Rex L. Jalao, Ph.D.

Academic Program Plan for Assessment of Student Learning Outcomes The University of New Mexico

PRODUCT SAFETY PROFESSIONAL CERTIFICATION PROGRAM DETAILS. Overview

Liberal Arts and Science Test (LAST)

PeoplogicaSkills Test Centre User Guide. How to Use Your Account on the PeoplogicaSkills Application

Bachelor of Science in Software Engineering (BSSE) Scheme of Studies ( )

Autodesk Academic Certification Program Overview. Worldwide Education Programs 1

Form 18 Heriot-Watt University Graduate/Postgraduate Course Structure and Course Notes Template (RAY) Version 3.0 (2007/2008)

WHO SHOULD ATTEND COURSE OUTLINE. Course Outline :: PROJECT MANAGEMENT PROFESSIONAL (PMP) EXAMINATION PREPARATORY COURSE::

Buros Center for Testing. Standards for Accreditation of Testing Programs

Course Outline. [ORACLE PRESS] OCA Java SE 8 Programmer Course for Exam 1Z

NATIONAL BOARD OF MEDICAL EXAMINERS Subject Examination Program. Introduction to Clinical Diagnosis Examination Paper Version

Introduction to CFCM Exam

European Master s in Translation (EMT) Frequently asked questions (FAQ)

PMP (Project Management Professional)

Effective and Efficient Use of the Fundamentals of Engineering Exam for Outcomes Assessment

ALABAMA FIRE COLLEGE SKILLS EVALUATOR TRAINING

Instructions for the Conduct of the Examination and Coursework (ICE) Physical Education. Entry Level Certificate (ELC) 8930

North Carolina Fire and Rescue Commission Certification Programs POLICY FOR TESTING

Fundamentals of Database Systems (INSY2061)

Transcription:

Automated Item Banking and Test Development Model used at the SSAC. Tural Mustafayev The State Student Admission Commission of the Azerbaijan Republic Item Bank Department Item Banking For many years tests have been administered for various reasons, such as educational testing or assessing particular skills. Results of these tests are used, for instance, as a factor in determining whether an examinee should be admitted to study at an educational institution. Such tests are typically administered to a large number of examinees. Due to some concerns, including administration costs and security, automation of the entire testing process becomes an issue of great importance. Automated item banking and test development make a significant contribution to this process. Test constructed from item bank is a very useful tool in educational and psychological measurement. An item bank is a large collection of test items all measuring the same domain of knowledge, stored in a computer storage together with their parameters. Known item parameters allow the test constructor to have strict control of them and make it possible to search for items according to a number of criteria. The goal of the testing is to construct a test efficiently for the purpose of measuring a skill, ability, etc. Therefore, each test is constructed to conform to a test specification, which defines the rules and/or constraints for selecting the items. In constructing a test, items are selected from item bank, so that the combination of selected items satisfies the test specification. A test is typically divided into sections of questions. The test specification generally defines the number of items to be presented in the test, the number of test sections, the number of items in each section, and the time set for taking the test. The test specification also defines criteria for item selection. These are based on item characteristics, which include, for example, item content, e.g., mathematical questions relating to arithmetic, algebra, or geometry and parameters of items, difficulty of the selected items as an example. The approach to design of an automated item banking and test development system can be envisioned as an integration of three phases. Phase I involves the development of an item bank that contains each item s content and parameters, and has full editing capabilities. The parameters include, but are not limited to, item statistics, such as item difficulty, and statistics of the test where the item appeared. Phase II provides the capability to retrieve items from the item bank in order to construct tests with prespecified characteristics. Item parameters are used for this purpose. Finally, in Phase III, the actual test booklets are published with specific format. 1

In order to provide these three phases SSAC has developed the in-house Automated Item Banking and Test Development system. This system consists of two parts. The main part of the system is the item bank consisting of items stored in computer storage as MS Word files. This item bank is subject-based, i.e. items are grouped by subjects. This bank can also be considered as a set of independent subject-based item banks. The organizing structures of the bank have been designed and developed by discipline committees of teachers. Paying particular attention to what is now being taught throughout the Republic, they have developed comprehensive course outlines. Test items have been collected from a large number of teachers. The second part of the system is the software suite designed to sort and register items, to get statistical report about the state of item bank, to retrieve items from the bank in order to construct tests with prespecified characteristics, to provide item editing capabilities, to compile and format test booklets, to get a list of answers to a test (key) in different formats, and to add usage information and statistics of the test where the item appeared. Since 1994 SSAC (State Students Admission Commission of the Azerbaijan Republic) has been developing item bank for administering entrance examinations to Bachelor s level of higher schools. Today SSAC is running the item bank consisting of over 100 thousand items in 17 subjects and this bank is in the process of continuing development. Besides, starting from 2005, SSAC has been developing item bank intended for use at entrance examinations to Master s level of higher schools. Classification of item parameters and item format Each item in SSAC s item bank has 5 parameters, namely a subject and the topic within that subject, degree of difficulty, cognitive level called relevancy, indicator of previous usage called usage marker, and indicator of similarity called analogousness. As stated before, items are grouped by subjects and each subject domain is divided into a number of topics in accordance with school curriculum. For each item subject experts determine the most corresponding topic. They also assign the degree of difficulty and cognitive level to each item. Subject experts use 3 difficulty levels (easy, moderate and difficult) to characterize item difficulty. Further, their estimations are checked and revised when an item is used on a test administered either at preparatory or admission examination. Subject experts use 7 cognitive levels to distinguish the different thought processes required to respond. The cognitive level reflects the manner in which knowledge is being assessed, rather than the technical difficulty of the content. Items coded to cognitive level I (knowledge of terminology) elicit knowledge of terminology. Items coded to cognitive level II (knowledge of facts) elicit knowledge of specific facts, methodology, principles, theories and structures, and ability to memorize, repeat and identify. Items coded to cognitive level III (reasoning based on generalization) elicit the ability to generalize and differentiate specific facts, principles, theories and structures, and the skills such as comparison and classification. Items coded to cognitive level IV (interpretative reasoning) 2

elicit the ability to interpret events or facts rather than naming them, often involving cause/effect relationships. Items coded to cognitive level V (predictive reasoning) elicit ability to predict consequences. Such items describe a specific or hypothetical situation, and require forecasting consequences when certain period of time passes or when some factors determining a situation are changed. Items coded to cognitive level VI (modeling or application) elicit ability to use information, methods, concepts and theories in new situations, to solve problems that have single or best answers by applying acquired knowledge, and techniques and rules in a different way. Items coded to cognitive level VII (calculation) elicit ability to solve problems using mathematical principles and formulas. In this sense calculation can be regarded as one of the forms of application. Each item has also a usage marker assigned on the basis of its latest usage and indicator of similarity assigned to indicate whether an item has an analog or parallel, i.e. an item or items with content, difficulty and cognitive level similar to its. Generally, a used item is not included in the tests administered in the next two years. Analogous or parallel items are used to build parallel (equivalent) test forms. At present SSAC administers paper and pencil tests using traditional multiple-choice items. Items are written by teachers and reviewed comprehensively by the experts in the field. Each item consists of a unique number called item code that identifies the item and all of its parameters, a stem containing the introductory question or partial statement that the examinee must answer or complete, five alternatives, marked by the letters A, B, C, D, and E with a comment. One of the alternatives, the key, is the correct, or best, response to that stem. A comment contains encoded information about an item writer and reviewers, as well as usage information and statistics of the tests where the item has appeared. It may also include solution of a problem presented in the item. Automated test development The SSAC staff has developed a software titled Test for Windows that consists of several integrated procedures designed to handle an entire spectrum of test development process. This software enables to sort and register items, to get statistical report about the state of item bank, to retrieve items from the bank in order to construct tests with prespecified characteristics, to edit drawn items, to build parallel test forms, to compile and format test booklets, to get a list of answers to a test (key) in different formats, and to add usage information and statistics of the test where the item has appeared. Prior to using the item bank, it is processed by special software to verify that each item has a correct code and all the required elements (i.e. stem, alternatives and comment). After that each item in the bank is registered. Registration consists of performing code-based sort on the items (optional) and marking all elements of all items for an easy access. Registration procedure also builds a database containing full information about each item necessary for running the 3

following procedures. This database is first used to gain statistical report called map of the bank a three dimensional table reflecting the distribution of items by difficulty and cognitive levels within each subject topic. This information together with the information received from experts in the field is used for automated test construction. First, a group of subject experts prepares a blueprint of the objectives across a given content area and defines the number of items representing each difficulty and cognitive level at the test, as well as subject topics or topic groups that should be represented in the test. Generally, for each subject the number of topics is greater than the number of items that represent given subject in the test. For this reason experts assemble close topics into groups so that the test covers the whole curriculum. On the basis of information provided by subject experts a special software randomly generates a request called technical task using map of the bank. This request contains parameters of each item to be included in the test. Based on this request, procedure called Generator randomly retrieves items from the bank. Generator is capable of simultaneously developing parallel test forms, i.e. test forms containing items that are similar in terms of content and parameters by making use of analogous items. Test forms generated by Generator are called initial or "original forms. These forms are reviewed by the experts once more. After this final expertise these forms are processed by the procedure called Generator of Forms. Generator of Forms generates a specified number of test forms called derivative forms from each initial form by randomly changing both the position of items and the position of alternatives within each item in the initial form. This contributes to preventing cribbing during an examination while providing identical items on the test. Generator of Forms also provides for control over the frequency of occurrence of each of the alternatives as a key. Generator of Forms generates derivative forms for each subject administered in the test. After that a procedure called Booklet Compiler compiles test booklets from derivative forms. This procedure also formats test booklets and makes them ready for use at the examination. Test for Windows also contains three procedures that are used after an examination. Key Generator, first of them, enables to get a list of answers to a test (key) in different formats. First format is designed for experts who are invited immediately after the examination to confirm correctness of items and the key. Second format is designed for publishing and, thus, is more compact than the first one. Third format is designed to provide the source information for software which handles scoring of each examinee's responses. Usage Markers, the second procedure, enables to add usage information to items used on the test. As stated before, a used item is not included in the tests administered in the next two years. Third procedure, Post-examination statistics enables to add statistical data to the item reflecting the performance of that item on the test. This data consists of the percentages of examinees responding correctly and incorrectly to that item and is later used to revise the difficulty level of that item. 4

Future plans As stated, SSAC s Automated Item Banking and Test Development system consists of two parts: item bank stored in computer storage as MS Word files, and item bank management and test development software. Storing test items in MS Word files was preferred due to the ease of use as MS Word provides tools for text editing, creating mathematical formulas (Equation Editor) and graphics (Word Picture), and insertion of bitmap images (OLE standard). But on the other side, this model has several limitations. It is difficult to maintain the rapidly growing item bank using this model. Another difficulty is with adding new fields of data associated with the item. These limitations can be overcome by building central item database using one of the DBMS software (Oracle, Sybase, etc.). This solution will contribute to designing a more flexible structure for item bank, and provide more advanced data integrity and data security mechanisms. In the nearest future SSAC is going to build a central item database on the Oracle platform. Associated with each item stored in the central item database, will be the data stored in fields related to the revisions the item has undergone, a list of the test forms in which the item has previously appeared, and statistical data indicating how the item performed at each previous administration. Other descriptive data fields will include information related to the item type (i.e., multiple choice or open-ended), the item's author and editors, solution of a problem presented in the item, subject and topic information, and both difficulty and cognitive levels. At present, SSAC administers only paper and pencil tests. But in the nearest future it is going to provide computer-based tests for admission to master s level of higher schools. Computer-based testing has several advantages over traditional paper and pencil testing. The biggest advantage of using computers to deliver tests is the level of control over the conditions of the testing session it gives to instructors. For example, the amount of time used for the test and the type of feedback (either on individual questions or on the entire test) can be varied. Such timing information can assist test designers in eliminating bad or poorly worded questions and in comparing the knowledge level of test-taking candidates. A second set of advantages includes the variety of multimedia learning objects that can be incorporated into test questions and the fact that most question types can be scored instantly. Using computer based testing also enables the use of adaptive tests. Adaptive tests are tailored to the individual ability of the test-taker and are able to provide somewhat better estimates of ability with shorter tests. SSAC staff is working on implementation of a computer-based testing in admission examinations in a few years to come. 5