Test Development Quick Start

Introduction

Often, teachers and examiners (in higher education, these two functions are usually combined) need information to help them to design tests. They  especially need practice advice. This article seeks to meet this need.

It is a checklist of key points which anyone who needs to design tests needs to have at hand when doing so. The theory behind testing is omitted insofar as possible. You can find the theory in various other resources.

It should be underlined that the contents of this article are principally concerned with summative written exams (tentamens), nothing more and nothing less. It is certainly possible that the article may also be relevant for other kinds of assessments, but this was not the aim of the authors. 

The article is arranged chronologically. The sections correspond to the different steps in the process of constructing a test. 


This article is based on the following article: Van Berkel, H. J. M., & Draaijer, S. (2011, maart). Gids voor toetsontwikkeling [Practical guide for Testdevelopment]. EXAMENS, Tijdschrift Voor de Toetspraktijk, 8(1), 8.

1.  Create a Test blueprint, choose a question type and the duration of the test 

A Test blueprint is made up of the main topics of the course which you want to ask questions about and, secondly, the cognitive level at which you wish to target with the questions. 

Test Development

1.1 Subject and number of questions

The number of questions that can be asked depends on the form and duration of the test. For a discussion on this topic we refer to Combining closed and open questions in a test and Rule of thumb: 40 questions in a 4-choice multiple-choice test – a short follow-up…


1.2 Open Questions

Test Development

1.3 Closed Questions

Test Development

Extra points to consider - Some misconceptions and tips

2. Construct questions

2.1 Finding inspiration

Sources of inspiration for questions: 

Verbs

If you want to follow the usual sequence of knowledge/reproduction - application - insight, you can use a particular set of verbs to ask your question.

Knowledge - Reproduction

Name, Describe, Quote, Define, , Identify, Distinguish, Sum up, Paraphrase, Summarize, Estimate, Select, Explain, Translate, Explain in your own words

Application

Calculate, Demonstrate, Use, Make, Develop, Solve, Organize, Produce, Relate, Transfer, Change, Prepare, Change, Extrapolate, Interpret

Insight

Criticize, Categorize, Compose, Conclude, Contrast, Deduce, Formulate, Rewrite, Illustrate, Interpret, Create, Differentiate, Support, Design, Justify, Relate, Summarize, Outline, Explain, Validate, Defend, Compare, Value

Question Shells

The following question shells can also give you a good starting point when designing questions, especially the ones that aim to assess a high cognitive level.

Knowledge: Determine the student's knowledge of subjects

Evaluating critical thinking: Determine whether the student can use the characteristics of facts, procedures, principles or theories.

Critical thinking (predicting): Determine whether the student can deduce implications, consequences etc. on the basis of facts, procedures, principles or theories.

Solve a problem (given scenario): Determine whether the student can provide solutions or evaluate solutions on the basis of a given problem.

Additional tips

2.1 After the construction phase

2.2 Model answer sheets

A model answer sheet is a tool for the person marking the exam or open questions, which serves to increase reliability of the test. It contains the following:

2.3 Checklists

Use the following checklist after you have devised the questions 

Contents

Form

Process

Specific for Open Questions

Specific for Closed Questions


3. Assembling and administering the test 

3.1 Assembling the test

A test is more than just a collection of questions.

To prevent cheating, make at least two versions of the test, changing the order of questions. In e-Assessment sytems such as VU's TestVision, consider using randomisation.

3.2 Administering the test

The test instructions for students for pen-and-paper tests must include the following information: 

4. Analysing and reviewing the test

Open questions

Closed questions

5. Assigning grades

There are many methods of assigning grades which can be put into three categories: absolute, relative and compromise.

Absolute

Under the absolute method, you determine the pass mark yourself in advance. The point of the pass mark is to determine whether the students have met the requirements of the course which, in turn, are derived from the learning objectives of the course.

Relative

The relative method is based on the idea that the test should be geared to the level of the students who are entitled to attend classes, the majority of whom should in principle be capable of passing the test. Because it is not known in advance what these students are capable of on the basis of the classes that have been given and the test that has been set, the standard cannot be determined in advance. The results of the test must be known before the pass mark can be set.

Compromise

The compromise method tries to bridge the fundamental differences between absolute and relative standards and is preferred in educational practice. The compromise method is generally based on an absolute standard, but it also identifies circumstances under which exceptions can be made to the absolute standard. One compromise method is described below.

Open questions

1. Begin by defining the pass mark. It is not unusual for this to be set at 55% or 60% of the maximum achievable score.

2. If you find that too many students fail your test, it is possible to make the following adjustments to the procedure (after consultation with the Examination Board).

     a. Set the limit at 55% of the maximum score instead of 60%, and/or

     b. You do not take the maximum score as a starting point, but an average of the five highest scores.

3. You now have the raw score that corresponds to a 0 and the raw score that corresponds to a 10: you can determine the remaining grades by dividing the raw scores into 10 equal units, each of which will correspond to a particular grade. 

Closed questions

1. Begin by defining the pass mark. This is usually set at 60% of the highest mark, taking into account the random answer score.

     - The random blind guess score is the number of valid questions in the test divided by the number of alternatives given per question. For example, the random answer score for 105 true / false questions is 105 / 2 = 52.5.

     - In this example, if the highest score is 96 points, the pass mark will be 60% of the score between 52.5 and 96 = 78.6.

2. Students with the highest score (96) will be awarded a 10; those who score the random answer score will be awarded a 0. The other grades can be calculated using linear regression.

3. If you find that too many students fail your test, the following adjustments to the procedure are possible (after consultation with the Examination Board).

     a. You set the limit at 55% of the maximum score instead of 60%, and/or

     b. You do not take the maximum score as the starting point, but an average of the five highest scores. 

6. Evaluating your test

Additional analysis can be carried out at your request. You can consult the test and item analysis service about this. The report includes quality indicators that relate to the test as a whole and to individual questions. Explanations are provided. 

Studying the report carefully will yield information that will enable you to make improvements to next year's test: 

Additional Resources

Van Berkel, H. J. M., Bax, A., & Joosten-ten Brinke, D. (2014). Toetsen in het Hoger Onderwijs. (3rd ed.). Bohn Stafleu Van Loghum.

Shrock, S. A., & Coscarelli, W. C. (2008). Criterion-referenced test development: Technical and legal guidelines for corporate training. John Wiley & Sons.

Downing, S. M., & Haladyna, T. M. (Eds.). (2006). Handbook of test development. Lawrence Erlbaum Associates.

Download this article

A PDF document of this article can viewed and downloade below

Quick Start Test Development v26 jan 2020.pdf