Translations of this page:

User Tools

Site Tools


knowledge_assessment:e-assessment_system_functionality

e-Assessment system functionality

General

The basic functionality of an assessment system includes the following1):

  • support for different question/item generation,
  • support for different assessment types and assessment delivery models,
  • support for automatic answer assessment,
  • support for assessment feedback.

Question/item generation

Today a large variety of assessment items is used for the e-assessment process. A recently proposed taxonomy or categorization suggests 28 item types that can be used in e-assessment, classified by the level of constraint they impose on the examinee's options for answering. Those item types, sorted from most to least constrained are the following2):

  • multiple choice items (including: true/false, alternate choice, standard multiple choice, multiple choice with media distractors),
  • selection/identification (including: multiple true/false, yes/no with explanation, multiple answer, complex multiple choice),
  • reordering/rearrangement (including: matching, categorizing, ranking and sequencing, assembling proof),
  • substitution/correction (including: interlinear, sore-finger, limited figural drawing, bug/fault correction),
  • completion (including: single numerical constructed, short answer and sentence completion, cloze-procedure, matrix completion),
  • construction (including: open-ended multiple choice, figural constructed response, concept map, essay),
  • presentation/portfolio (including: project, demonstration, experiment, performance, discussion, interview, diagnosis, teaching)

Assessment types and assessment delivery models

A good assessment system should support various types of assessment including:3)4):

  • continuous or end-of-the-course credit bearing assessments;
  • authenticated or anonymous self-assessments;
  • diagnostic assessments;
  • support for other or emerging types of assessment (competence assessment, performance assessment, portfolio assessment and peer assessment);

Application of computers in assessment also enabled a comfortable use of adaptive tests supported by the item response theory. Adaptivity here means that the next test item (question, task or request) or group of items will depend on the correctness of the answer on the current test item or group of items. Some of the assessment or test delivery models based on the level of their adaptability are5):

  • linear tests, which represent a computer version of common pen and paper tests and not adaptively administered,
  • linear-on-the-fly tests, which are also not adaptively administered, yet every examine gets a unique test composed out of a pool of test items,
  • testlets, which are “a group of items related to a single content area that is developed as a unit and contains a fixed number of predetermined paths that an examinee may follow6), or simplified, small tests of single units, small enough to manipulate, but big enough to carry their own context and be used to balance contents of a test7),
  • mastery models, which usually refer to tests designed of testlets presented using an adaptive mechanism which finally distinguishes mastery from non-mastery knowledge or skill level,8) and
  • adaptive tests (also computer-adaptive tests - CATs), in which not just a testlet, but the very next item to be presented is dependent.
knowledge_assessment/e-assessment_system_functionality.txt · Last modified: 2012/05/13 22:20 by jpetrovic