22/9/22

 PROCEDURE FOR MEASURING ATTAINMENT OF COURSE LEARNING OUTCOMES – DIRECT MEASURES The assessment process used to measure attainment of CLO’s is described as under: The assessment process uses both direct and indirect measures to measure the attainment of each outcome. The examples of such measures are given below: Direct Measures  Student Assignments  Projects  Examinations In-direct measures: CLO Surveys To assess each course, we use CLO’s defined for that course. For example in Course A, we defined four CLO’s (LO1 to LO4) that need to be met to successfully achieve that outcome at a minimum target performance level for a course. In each course, we assess the level of achievement of each course outcome. The data are then combined to analyze and evaluate the program level achievement of each program outcome. If any student outcomes are not met, action is taken for improvement. In the section below, the assessment of course Inorganic Chemistry (PCY102) using CLO 1 is explained as an example. For example, at the course level, CLO1 reads CLO 1: Chemistry of main group elements, and synthesis and properties of few main group compounds. Attainment Assessment Process The step by step process for assessing CLO’s is tabulated below. Step-by-step process for assessing Student Outcomes Step 1: The Program coordinator analyses each course by breaking down into course learning outcomes and weightage and rating scale has been defined for each course. In addition, well designed surveys were used to assess each outcome. Step 2: For each outcome define performance indicators (Assessment criteria) and their targets. Step 3: The module coordinators collects the qualitative and quantitative data and is used for outcome assessment in a continual process. Step 4: The Program Assessment Committee analyzes the collected data. If the assessed data meets the targeted performance value specified in step 2, the outcome is attained. Step 5: The Department of Academic Affairs Committee recommends content delivery methods/course outcomes/ curriculum improvements as needed. In case the targeted performance for some outcome is not met, a corrective action plan is put in place which serves as a feedback to the process for continuous improvement. The procedure followed at the course level is depicted in the following flowchart diagram

24/11/21


 NHỮNG VỊ TRÍ ĐẶT CAMERA THỨ 2 

KHÔNG ĐÚNG QUY ĐỊNH TRONG THI TRỰC TUYẾN














25/9/21

https://www.e-assessment.com/glossary-of-terms/


Adaptive Feedback

Feedback given to students on an assessment item, where the feedback is modified according to the student’s response , particularly where the response was marked as incorrect, and typically aims to provide guidance on why the answer was wrong.

 

Advanced Question Types

Advanced Question types are a level above Standard Question types. They typically require complex setup and/or marking, or they contain complex marking algorithms. Additionally, the level of interaction within the item may also be over and above that of a Standard Question Type, and therefore they could be considered as separate applications within a test.

 

Application Programming Interface

The software mechanism by which one computer program makes its functionality and data available to another computer program. For example, a test-marking program might use an API to request the correct marking key from the item bank program. APIs are widely used in internet applications where programs are distributed across different servers (for example in multi-vendor systems).

 

Assessment

An instrument (e.g. an on-screen examination) used to make a judgement about learning, skills acquisition or educational readiness or need against a set of pre-determined criteria. Assessment instruments include examinations and tests, but also portfolio and observation-based approaches to judgement.

 

Assessment Delivery System

A combination of software, hardware and communications components, as well as human processes that manage the end-to-end process of administering e-assessments. Note: this includes delivering the test package to centres, supporting the examination administration functions, delivering the assessment to the candidate, recovering their response, managing the marking and verification processes, setting standards and providing assessment outcomes. It generally does not include the writing and production of the assessment.

 

Assessment Engine

The software application that delivers an electronic test on a computer. The engine usually consists of generic components (such as a web server) and proprietary software (such as a specific e-assessment program). An e-assessment system typically consists of an assessment engine, an item bank and an item and test creation system.

 

 

 

Audio Capture Question Type

The audio question type enables the capture of sound created by the candidate. This could be spoken sentences such as in a language test, or music from a performance. Audio question types can sometimes start automatically as part of a question workflow, or be controlled by the candidate as part of evidence capture. If being used as part of language testing, it would be appropriate to combine the audio capture question with either on-screen text or sound files. Visual feedback on-screen is important so that the candidate is clear that the capture has started and is recording audio. Capture can be open ended, however this would generate larger file sizes. Consideration of file size is important, as language testing could be being conducted in a location with poor internet or power stability. Use of an on-screen avatar can sometimes help to improve the quality of the conversation, if the candidate feels like they are speaking to a real person, then the quality of the audio is less likely to suffer. Giving the candidate the opportunity to playback the capture can also be useful. Audio quality can be impacted by the confidence of the candidate, the quality of the microphone or the background noise from other candidates. Simple techniques of placing candidates completing an audio capture section of a test next to candidate undertaking a written component can be good ways of reducing background interference. The more practice a candidate has with this type of question, the better they will perform. Classed as an Advanced Question Type, audio capture questions are not typically auto-marked.

 

Authoring Tool

Software used to create items and tests for e-assessment. Note: the term is used more widely to cover tools used for creating any content for delivery on-screen (e.g. e-learning, general web pages, etc.).

 

Classical Test Theory (CTT)

CTT is a set of statistical measures relating to the performance of items within a test, and the test itself, and used to provide evidence about the quality of a test and the items within it. CTT’s most commonly used measures include facility (the difficulty of a question), discrimination (the extent to which candidates’ performance on a question mirrors their performance on the test as a whole) and Internal Reliability (the extent to which the test assesses a single construct). CTT is relatively easy to implement and understand, but it cannot distinguish between facets of the candidates and the questions (eg whether a question is hard or a candidate cohort is weak), and for this reason Modern Test Theory, including Latent Trait Theory, Rasch and Item Response Theory are increasingly widely used.

 

Competency-Based Assessment

An assessment process based on the collection of evidence on which judgements are made concerning progress towards satisfaction of fixed performance criteria which describe the competency. Note: the competency-based assessment of an individual takes no account of the performance of others in the wider group being assessed (as is the case in norm-referenced assessment), and is typically limited to a pass/fail grading (also called mastery/non-mastery). Competency testing is typically used for licence to practice assessments (for example for doctors, pilots, etc.).


 

Computer Adaptive Test

A CBA test in which successive items in the test are selected for presentation by a computer algorithm drawing primarily on the properties and content of the items, and the test taker’s response to previous items. Adaptive tests are typically used in summative tests where a more accurate test result (compared to a linear test) can be determined for a given test duration (or a shorter test can be provided). For diagnostic purposes, adaptive tests allow a more detailed exploration of strong and weak areas in a test taker’s knowledge within a given time for a test. Two key features of adaptive tests are that (a) test takers cannot return to questions once they have moved on and (b) the test taker’s responses to items must be computer-marked.

 

Computer Based Training

A term from the 1980s describing learning which is mediated by a computer (typically a stand-alone computer). Note: the modern term ‘e-learning’ differs in that there is an expectation with e-learning that more of modern computing’s capabilities (rich media, interactivity, etc.) will be used, as well as the computer’s connection to internal networks and the internet.

 

Computer-Assisted Assessment

Computer-aided assessment (or “computer-assisted assessment”) describes assessments delivered with the help of computers. This includes assessments delivered to the candidate on-screen, developed on computer but delivered on paper, marked on-screen or electronically (eg using OMR).

 

Computer-Based Assessment

A subset of CAA where the candidate is presented with the question on-screen and responds to the question using the computer.

 

Content Management System

A general web term for a software tool which manages and maintains the technical and editorial content of a website. Note: in an e-assessment setting, it describes a software tool for managing assessment content (items and tests) and will generally comprise an Authoring Tool/System and an Item Bank.

 

Delivery Method

Delivery method refers to the way in which the test is taken by the candidate, this can be online, offline, secure locked down or open book.

 

Delivery Platform

The platform being used on the client device, for example PC running windows, Linux running Ubuntu, Smartphone running Android, Apple Mac running iOS or windows etc.


 

Diagnostic Testing

Non-accredited, assessment used as part of a learning programme to identify a learner’s strengths and weaknesses with a view to providing an appropriate learning programme. Generally undertaken at the start of a programme, diagnostic assessment therefore needs to evaluate learners’ existing levels of attainment across the range of relevant knowledge, skills and understanding, so as to inform personalised learning.

 

Digital Logbooks

An online record of practice experience and skills that are assessed. The logbook often complements formal examinations.

 

Drag-And-Drop  Question Type

The Drag and Drop question type requires candidates to drag answer options into the relevant drop zones; this can be useful for a number of question scenarios.Drag and drop items can be text, video or image-based. Classed as a Standard Question type, auto-marking is possible with this question.

 

e-Assessment

E-assessment describes a range of activities where technology is used to enhance some or all elements of an educational assessment process.  Common activities are on-screen tests (the candidate reads and answers the question on-screen), electronic marking (the marker marks on screen – either a scanned candidate script or an on-screen response), remote proctoring (candidates taking on-screen tests are invigilated remotely), item banking (test questions are stored in a database and assembled into tests electronically), and e-portfolio (candidates assemble digital evidence of their learning for skills assessments).

 

E-portfolio

An e-portfolio contains digital items – ideas, evidence, reflections, feedback etc., captured on the web which someone can present to a selected audience as evidence of their learning and/or ability.

 

E-portfolio Management system

An e-portfolio assessment system enables the learning and skills of a number of individuals to be captured and assessed online.

 

Either/Or Question  Type

The Either/Or question type requires candidates to select the correct answer from two different answer options and is most often used for questions where the options are true/false or yes/no. Classed as a Standard Question type, auto-marking is possible with this question.

 

End Point Assessment

The final assessment in the UK’s reformed apprenticeship, which takes place after the student’s employer has signed off the student as ready for assessment. Typically the assessment is graded and designed to show that the student is competent across the range of skills and knowledge that have been covered during their apprenticeship.

 

Equation Entry  Question Type

The Equation Entry question type provides candidates with an equation creation tool, which allows them to enter complex equations into the answer box. Classed as a Standard Question type, auto-marking depends on the complexity of the equation and the capabilities of the technology.

 

Essay Style Question  Type

An Essay question requires the candidate to write an extended response in answer to a question. Typically, this question type will provide the candidate with text formatting features, and sometimes the ability to insert symbolic information such as equations or symbols. Due the length of time a candidate may spend on this question, it is important that their responses are securely stored at regular intervals to prevent loss of data from due to computer error. Options around the permitted response length can often be set to control the amount of information the candidate provides. It is important that the candidate is able to view all information provided either through an expanding window or scrolling. Classed as a Standard Question type, auto- marking is not possible with this question and human-marking is required.

 

Extended Matching  Question Type

The Extended Matching question type allows item authors to create questions where candidates are required to respond by making correct links between two lists of options. There are a number of linking options available within this question type: these include a one to one relationship, one to many, or many to many. Classed as a Standard Question type, auto-marking is possible with this question.

 

Familiarisation Test

A test used by either the candidate or the centre. The candidate would use the test to simulate the real assessment so that they can become familiar with the delivery interface, question types and test structure. A centre may use the test to highlight any technical or process problems ahead of the exam day.

 

File Attach Question  Type

The File Attach question type allows test providers to present candidates with a file which they can change or edit before uploading back into the test for marking. This question type is perfect for any assessment that requires candidates to use external software, such as Microsoft Word or Microsoft Excel. The files can be pre-populated with information or can be left blank. The files can be launched securely to prevent the candidate accessing additional files or other applications. Typically both the candidate’s device and the marker’s device will require a licence to the file type’s software, otherwise access to view or edit the file might not be possible. Classed as a Standard Question type, auto-marking is not normally possible with this question type and therefore requires human-marking.

 

Fill in the Blank  Question Type

The Fill in the Blank question type requires candidates to fill in blank spaces in a passage of text. Multiple answer options can be set for multiple blank spaces within a passage of text. Fill in the Blank can be used as an alternative to using the Select from a List question type. The space provided for the response typically expands to the answer given so as not to provide the candidate with a clue to the response required. Classed as a Standard Question type, auto-marking is possible with this question.

 

Hotspot Question Type

A Hotspot question requires the candidate to select one or more areas within an image. Variations of this include providing a single point answer with options for marking tolerance around the point. Multiple points can also be used along with area mapping. In most cases it would be unreasonable to expect the candidate to pick an exact pixel, therefore marking tolerance is normally recommended. Classed as a Standard Question type, automarking is possible with this question.

 

Interoperability

A feature of computer systems components which allow the components to interact to exchange information according to technical standards which define functionality useful to the user. The IMS QTI specification is an example of an interoperability specification within the e-assessment domain. IMS QTI allows the transfer of test questions (and their associated metadata) from one system to another.

 

Item Response Theory

IRT is a commonly used Modern Test Theory statistical approach to measuring the performance of candidates and test questions. Part of a wider group of Latent Trait Theory approaches (including Rasch analysis) it is based on the idea that the probability of a correct response to a question is a mathematical function of both the candidate’s ability and the item’s difficulty (in contrast to CTT where the difficulty of a question is fixed). IRT analysis is more complex to implement and interpret but provides analysts with information to distinguish between the ability of a candidate and the difficulty of an item, thereby allowing comparisons of different candidate groups and/or different tests.

 

Likert Scale Question  Type

Typically associated with surveys, the Likert Scale Question Type provides the candidate with the means to answer according to a scale (for example “Strongly Agree, Agree, Neutral, Disagree, Strongly Disagree”). These question types are generally used in surveys. The scale may have fixed points and is therefore a horizontal MCQ, or have variable points which would therefore need a degree of marking tolerance. Classed as a Standard Question type, auto-marking is possible with this question.

 

 

Locked Down

Locked Down refers to the delivery method used during the assessment. An assessment in Locked Down mode should prevent the candidate gaining access to applications or internet sites not permitted during the assessment.

 

Manual Test  Generation

The manual (human) process by which a test instance (a test form) is generated from a bank of items (according to a formal or informal set of rules which may involve a selection algorithm and randomisation). Note: this is widely used in e-assessment settings where the candidates do not all take the test at the same time, hence a variety of tests are required (to reduce the likelihood of test and question exposure) but where known comparability is required between the tests so that fair results can be given. Also see Automated Test Generation, which is increasingly used where large numbers of test instances are required.

 

Metadata

Reference data about a piece of information (e.g. an assessment item) that enables it to be systematically stored in and retrieved from a database (e.g. an item bank) according to a variety of selection criteria. In the context of assessment, metadata might typically refer to aspects such as qualification or test specifications, curriculum content and performance statistics. Note: metadata is most useful when it conforms to an open standard, e.g. IMS LOM or QTI.

 

Mobile Learning

A type of e-learning where the learning is undertaken using a mobile ICT device (e.g. a PDA, mobile phone or smartphone, handheld computer, etc.). The availability and popularity of handheld portable devices has led to some research into the use of m-learning techniques for e-assessment.

 

Multiple Choice  Question Type

A Multiple Choice question (MCQ) gives candidates a number of answer options (“distractors”) to choose from with only one correct answer (the “key”). These are the most common on-screen question types in use within Computer-Based Testing. Text is typically used in the answer options, but images or equations may also be used. In some versions of MCQ tests, candidates may be asked to provide their confidence level that the selected answer is correct, which can then be used as part of the scoring calculation. Typically answer options are randomised to reduce assessment malpractice, however care has to be taken not to use punctuation on the last item only, or to use terms like ‘All of the above’ if the intention is to randomise the answer order. Due to the correct answer being presented on-screen, consideration is required when authoring to prevent the correct answer from standing out, e.g. candidates can sometimes guess at the ‘longest answer option’ using the logic that the author has had to be explicit in the answer. This question type may also use weighted marking, where the award for one option may be greater than another. Classed as a Standard Question type, auto-marking is possible with this question.

 

 

Multiple Response  Question Type

 

A Multiple Response question is similar to a Multiple Choice question, except more than one answer option is correct and candidates may be asked to respond by selecting all of the correct options. This type of question can sometimes use images or equations. Variations may also include combination responses and controls on how many responses can be given. Classed as a Standard Question type, auto-marking is possible with this question.

 

Navigation

In an e-assessment context, the on-screen buttons and other controls that move candidates from screen to screen in an on-screen assessment (typically, from question to question), and provide access to other non-question specific features such as on-screen help, print functions, exit, etc). They are generally visually separate from controls that relate to the specific question.

 

Non-question Items

Non-question items include: introduction pages, seen at the start of a test; information pages, which may be seen by candidates during a test; and finish pages, seen by the candidate after they have submitted completed their tests and submitted their responses. Non-question items will not have a mark assigned to them. Typically non-question items may be visible outside of a timed section of a test.

 

Non-scored Questions

A non-scored item is a question (of any type) which has been included in a test to gather performance data, but does not impact the score or result of the test for the candidate. For example, it is a common practice to pilot new questions embedded within a live test, but where the scores on the pilot items do not count toward the candidate’s score.

 

Numeric Entry  Question Type

A Numerical Entry question allows candidates to only enter numbers as the response to a question. This can either be an exact value, or within a range set by the item author. Consideration should be taken regarding any symbols which are expected with the answer, such as $, £ or €. When considering international delivery the use of commas or decimals should be considered. Classed as a Standard Question type, auto-marking is possible with this question.

 

Objective Question

A test item in which the response is evaluated against objective criteria. This might be a single simple response, such as in a multiple choice item where the criterion is whether the student has made a correct selection. It might be that the student’s response has certain properties which can be established objectively. In CAA this is usually done automatically.

 

Offline Assessment

An on-screen assessment which is conducted without using an internet connection during the test (although an internet connection may well be used to deliver the test to the client computer prior to the test starting, and to upload the candidate responses once the test has completed).

 

On programme assessment

Assessment in the UK’s reformed apprenticeship, which takes place during the apprenticeship learning programme, and prior to the student being signed off by the employer for the End Point Assessment

 

On-Demand  Assessment

Used in public examinations. Assessments where there is a high degree of flexibility in the date and time that tests can be offered to suit the student or their learning programme (although it may not necessarily include all days, times and dates). In contrast to many traditional assessments which are provided on a fixed date and time (or a limited range of dates and times).

 

On-Screen Assessment

An assessment delivered to the candidate on a computer screen, and where the candidate provides their response on-screen (for example by typing, or clicking on the correct response).

 

Online Assessment

An on-screen assessment which relies on an internet connection during the test to download subsequent questions and upload candidate responses. Sometimes termed “conducting a test live over the internet”.

 

Online/Offline  assessment application

The ability to capture and assess evidence of learning and skills to an e-portfolio when àn internet connection is unavailable and for this evidence and assessments to be added to the e-portfolio once an internet connection is established.

 

Open source e-portfolio

This is a personal learning environment mixed with social networking, that allows an individual to collect, reflect on and share their achievements and development online in a space they control.

 

Open Standards

Shared, freely available and internationally agreed standards for computer-based systems, designed to enable communication and interoperability.


 

Optical Character  Recognition

Optical Character Recognition (OCR) is a means by which a computer can recognise text and other marks in handwritten responses on paper that have been scanned, and convert these to digital format. OCR is often used in assessment to electronically mark paper responses to multiple choice tests (for example, on bubble sheets)

 

Optical Mark Reader

A device that scans paper-based tests and converts marks made by the student using pen or pencil into digital data.

 

Parameterised Item Type

An item where parts of the question are generated according to a formula embedded within the question. This randomisation is usually undertaken dynamically, using the formulae, as the test is generated or delivered to the student, in contrast to cloning where item variants are generated during authoring by use of the randomisation parameters. The term parameter refers to the variables that are used by the formula to create the item instances. Many different types of questions (e.g. MCQ, gap fill, etc.) may use parameterisation.

 

Personalisation

The configuring of an IT system by students to suit their personal requirements (e.g. selecting preferred font sizes and colours, volume levels for audio, et al). Also refers to more complex customisations of the user experience to meet personal learning needs.

 

Polytomous Item

An item having more than two response categories. For example, a 5-point Lykert type scale where items can be scored 0, 1, 2, 3, 4, or a partial credit question where candidates can score between 0 and 4 marks. Polytomous is a term typically associated with IRT and other forms of latent trait analysis.

 

Portfolio Assessment

An assessment where a student’s portfolio of assembled work is assessed. This type of assessment is distinct from a test (which is administered on a single occasion)

 

Practicability

The feasibility of an assessment in terms of operational efficiency and viability. A valid and/or reliable assessment may not be practical due to the cost or time required to carry it out. High quality assessments are valid, reliable and practicable, although there are typically trade-offs to be made between these three key elements of test quality.

 

Proprietary Software

Software requiring a licence (for which a charge is usually made) from a particular company.

 

Psychometric Test

Provides one or more measures of a candidate’s personality traits. It is most commonly used as part of the selection process for employment, or to support careers guidance. Typically it seeks to place candidates on a number of scales according to their preferred behaviours or aptitudes for aspects such as working with others, managing pressure, preferred working environment and thinking style.

 

QTI Lite

A simpler-to-implement version of the QTI technical interoperability specification for tests and items which allows tests developed in one system to be delivered to candidates on other systems (currently at version 1.2). See www.imsglobal.org/question/index.cfm#version1.2lite

 

 Question and Test Interoperability

A technical specification for tests and items which allows tests and test items to be authored and delivered on multiple systems interchangeably. It specifically relates to content providers (that is, question and test authors and publishers), developers of authoring and content management tools, assessment delivery systems and learning systems. It is designed to facilitate interoperability of assessment content between systems.

 

 Rich Feedback

Feedback which goes beyond providing the correct or model answer to an item, and a simple explanation of why the student’s selected response was wrong. Rich feedback is usually personalised to the candidate’s response and designed to deal with the underlying misconception.

 

 Select from a List  Question Type

The Select from a List question type requires candidates to select missing word(s) from a passage of text by choosing from a drop-down list of answer options. Select from a List items can often be used multiple times within a passage of text. Random ordering of the answer options may be enabled to help prevent malpractice opportunities during the assessment. Classed as a Standard Question type, auto-marking is possible with this question.

 

 Short Answer Question Type

A Short Answer question will allow candidates to respond to a question with a word or short phrase, as specified by the test author. Multiple answer variations can be provided and settings such as case sensitive can be included. Classed as a Standard Question type, auto-marking is possible with this question, although some responses may need to be human-marked.

 

 Source Materials

Source materials are files such as images, PDFs or editable material that are made available to a candidate during an assessment (for example containing reading material, formulae, etc.). It is useful for the candidate to be able to interact with this material by making notes or highlighting sections.

 

Spreadsheet Question

The Spreadsheet question type is common in assessments of accountancy. By embedding a spreadsheet within a question, the candidate does not need to access a third party application. The question can be pre-populated with data to be statically visible to the candidate or enabled for editing by the candidate. The question can also be provided to the candidate as blank so that they may present their data in a form deemed suitable. The spreadsheet question type allows the formatting of text and data, and also supports the inclusion of formulae and other common spreadsheet features. The spreadsheet question may also provide candidates with the ability to justify their responses with annotation. Classed as an Advanced Question type, auto-marking can be achieved by logic-based marking rules which can mark a fully correct answer or an answer where the calculation is correct however the source data is incorrect.

 

Standard Question  Types

Standard question types are typically QTI supported items; however, definitions vary. In some locations, standard question types like Drag and Drop can be considered Advanced. A preferred way to define standard question types is where the authoring and marking do not require complex coding, and that the item can typically be auto-marked by the computer, or marked by a human without the need for complex marking tools.

 

Table Question

Tables are commonly used for the presentation of data or to provide a means by which a candidate can record their responses in table form. The table question type can either be created by the test author and then presented as static information for the candidate, or provided as a partially completed table for editing by the candidate. The question type could also be provided as a response option alongside other options such as graphs and charts. The table question may also feature formatting of text and alignment of cells. Classed as a Standard Question type, automarking is possible with this question type.

 

Test Package

The electronic package of files in an eassessment system which includes the test content as well as any embedded applications, resources, and sometimes the test player, which is delivered to the Client PC for the student to undertake. The test package may have travelled from a central server at the Awarding Body/Certification Agency, to a local server in the test centre and then to the Client PC, and will be subject to considerable file and channel security in the case of a high stakes assessment.

 

 

Test Player or Test Driver

A piece of software which resides on the client PC and “runs” the test content package, i.e. displays the questions on-screen, collects the student responses, controls the exam clock, and access to other resources, etc. Some test players are considerably more complex than others. For example, a test player for MCQ items may be little more than some scripts embedded in standard HTML pages, whereas the test player for innovative items or a sophisticated test made up of simulations and embedded applications would be a substantial computer application in itself. Sometimes test players are pre-installed on client PC’s and sometimes they are delivered to the client PC along with the test content itself.

 

Video Capture Question Type

The video capture question type enables the capture of audio and video. Associated more with evidence-based assessment, files can either be recorded directly into the assessment or attached as evidence. Classed as an Advanced Question Type, marking would typically be undertaken on-screen by a human marker.

 

Web-Based Assessment

An assessment delivered from a server via the Internet or an Intranet (such as a centre or education authority Intranet) and where candidates access the assessments using a standard browser. If Internet-based, the assessments could be online with candidate responses delivered to the server in real time for automatic marking and immediate feedback.