By Izah, SC; Odubo, TC; Ajumobi, VE; Osinowo, O (2022).

Greener Journal of Biological Sciences

Vol. 12(1), pp. 11-22, 2022

ISSN: 2276-7762

Copyright ©2022, the copyright of this article is retained by the author(s)

http://gjournals.org/GJBS

 

Description: Description: Description: C:\Users\user\Pictures\Journal Logos\GJBS Logo.jpg

 

 

 

Item Analysis of Objective Structured Practical Examination (OSPE) Used as an Assessment Tool for First-Year Microbiology Students.

 

 

Sylvester Chibueze Izah1*, Tamaraukepreye Catherine Odubo1, Victor Emeka Ajumobi1 and Olugbenro Osinowo2

 

 

1Department of Microbiology, Faculty of Science, Bayelsa Medical University, Yenagoa, Bayelsa State, Nigeria.

2Department of Surgery, Faculty of Clinical Sciences, Bayelsa Medical University, Yenagoa, Bayelsa State, Nigeria.

 

 

 

ARTICLE INFO

ABSTRACT

 

Article No.: 112621137

Type: Research

Full Text: PDF, HTML, PHP, EPUB

 

Item analysis is used to examine students' responses to items to determine the quality of an assessment tool. This study aimed at assessing the quality of the objective structured practical examination (OSPE) on Introductory Microbiology. 41 first-year Microbiology major students of Bayelsa Medical University, Nigeria took part in the OSPE containing 40 items. Marks were not deducted for wrong answers as decided by the department and each item carried one mark, and 40% was the pass mark. The items were analyzed for the difficulty index, discrimination index, distractor efficiency, and reliability index (Cronbach’s alpha). Also, key distributions, numbers, and percentage passed were determined. Results showed that the discrimination index had 15 (37.5 %), 2 (5.0 %), and 3 (7.5 %) translated as excellent, good, and acceptable. The difficulty index revealed that 4 (10.0 %) of the items were ideal, while the remaining 36 (90.0 %) were difficult. The 40 items had 160 distractors, of which 71 (44.4 %) and 89 (55.6 %) were functional and non-functional distractors, respectively. The difficulty index indicated positive significant relationship with the discrimination index (r = 1.000) and distractor efficiency (r = 0.408) at p = 0.01. The overall reliability analysis of the item was 0.754, an indication that it is good for classroom assessment, but some items need improvement. Too easy items and poor distractors may have caused the poor difficulty index. Therefore, there is the need for item flaws and technical pitfalls to be carried out to correct the errors in subsequent assessments. From the findings in this study it is recommended that OSPE be more widely adopted among more science-based departments and item analysis be a standard practice in departments of every university.

 

Accepted:  29/11/2021

Published: 31/03/2022

 

*Corresponding Author

Sylvester Izah

E-mail: chivestizah@ gmail.com

Phone: +2347030192466

 

Keywords: Assessment tool, Internal Consistency, Item analysis, Microbiology items, OSPE.

 

 

 

 

 


1.0   BACKGROUND

 

Assessment is a very vital key in the education process. It is crucial in determining the nature of the learning person, their way of knowing, what the learner knows, need to know, and what learning and mediating processes are associated with effective teaching and learning for each learner (CERI, 2008, Armour-Thomas and Gordon, 2013). In recent years, educational research and policies have increasingly focused on standardized forms of learning-centered assessments to improve knowledge, skills, and disposition for living in a competitive global society. An assessment provides feedback about the strengths and weaknesses of the students’ performances about a given task which can inform subsequent decisions about curriculum and instruction (Armour-Thomas and Gordon, 2013). In addition, the assessment tools also need to be evaluated because the assessment method can influence students learning (Vishwakarma, 2016).

A significant part of a microbiology curriculum is an appropriate assessment of the student's practical capabilities in the laboratory. For general student assessments, different types of examinations are used, such as Multiple choice examinations (MCQ), short answer examination (SAQ), and theory (Essay) examinations, all of which are vital for only assessing the knowledge, i.e., cognitive domain of the students (Relwani et al., 2016). On the other hand, practical examinations are essential for determining the cognitive, psychomotor, and affective domains, and thus an effective system of evaluation should be applied (Relwani et al., 2016). Assessment drives learning; therefore, adopting a suitable assessment tool would ensure constructive alignment between goals and learning outcomes (Bhat et al., 2020).

In the clinical sciences, the examination of practical skills and competence is critical for proper medical education. Still, conventional clinical and practical examinations had several limitations, especially in terms of their outcome. The traditional practical assessments lacked objectivity (Frantz et al., 2013). Although grading should depend on the student’s competence variability in the experimental process, standardization and examiners' decision were also concerns (Jaswal et al., 2015). These defects in traditional practical examinations, especially in clinical sciences, gave rise to the development of new examination systems which can test all the objectives (Munjal et al., 2011). 

The objective structured practical examination (OSPE) was modified from Objective Structured Clinical Examination (OSCE) in the 1970s (Harden and Cairncross, 1980; Frantz et al., 2013; Mard and Ghafouri, 2020), and has gained worldwide acceptance as a method for clinical skills assessment because of its uniformity, reliability, validity, and practicability (Nigam and Mahawar, 2011), standardization, and absence of variability in an experiment, scores, etc. In addition, it provides a more objective method of assessment, covers a broad scope, maximizes reliability, allows individual students to display a full range of their knowledge, skills, and abilities, which are evaluated in a comprehensive, structured and consistent manner and provides a measurement of skills (Mokkapati et al., 2016, Vijaya and Alan, 2014, Frantz et al., 2013).

The OSPE is a suitable examination system comprising a series of stations where students work through different tasks to test various skills applying to all learning domains (Munjal et al., 2011). All students are expected at all stations and spend equal time at each station, eliminating group work by students. All the students take the same examination, which is standardized.

The use of OSPE has been reported in many Universities with great benefits (Vishwakarma et al., 2016). OSPE maximizes reliability in assessment and allows students to display the full range of their knowledge, skills, and abilities. OSPE is very popular in the medical sciences. Information about OSPE in non-health sciences is very scanty, especially in developing countries like Nigeria. The Department of Microbiology Bayelsa Medical University, Bayelsa State, Nigeria, adopted OSPE as a tool for assessing students in their practical examination because of its merits over the traditional practical assessments. The practical assessment of introductory microbiology, a first-year microbiology course for microbiology major students was carried out using OSPE. Since this is the first time an OSPE is being conducted in the institution, item analysis should be carried out to ascertain the reliability of the examination. Hence, the focus of this paper is to carry out an item analysis of OSPE as an assessment tool for first-year microbiology major students. The findings of this study will be helpful to the scientific community, especially the Department of Microbiology, in decision-making about effective assessment methods for first-year Microbiology students.

 

 

2.0   METHODS

 

Design techniques and organizational structure of OSPE in Bayelsa Medical University, Nigeria

 

Before the OSPE, the course lecturers informed the students of the processes that would be carried out for the practical examination. The exercise was carried out by the course lecturers (examiners) with the assistance of eight Laboratory Technologists and a Professor in the University who have carried out OSPE/OSCE before now.

 

A total of forty-one first-year microbiology major students participated in the study. The students were divided into two batches. Three laboratories were used for the exercise; two of the three were arranged with twelve stations each. The third laboratory was used as a quarantine room for the students who belonged to the second batch. The questions were pasted on each of the stations before students entered the examination hall or laboratories. Twenty-four students belonging to the first batch were moved into the two designated laboratories, twelve students in each examination room.  Hence there were twelve stations (including ten workstations and two rest stations). The rest stations were arranged at stations 6 and 12.  The students were placed in such a way that each student in a station spends only five minutes.  The duration spent for each workstation was regulated by a laboratory technologist who served as the timekeeper during the OSPE. At the end of the exercise for the first batch, the answer scripts were retrieved, and they were moved to the third laboratory, which served as the quarantine room. At the same time, the students for the second batch were moved to the two laboratories used for the OSPE itself. The movement was carried out in a way that both groups of students did not meet. Twelve students were placed in a laboratory from the second batch, and the remaining five students were placed in the other laboratory. The laboratory with only five students was invigilated by four laboratory technologists and supervised by one examiner.

 

Structure of OSPE Questions

 

The OSPE consisted of forty single response stems, four alternatives (distractors), and one key (correct answer). Each correct response was awarded one mark, and no marks were awarded for blank or incorrect answers. Negative marks were not awarded for wrong answers. The maximum possible score in the examination was forty. The scores of all students were arranged in the order of merit and divided into three groups. The upper one-third students (H) (27%) will be considered to have high ability, and the lower one-third (L) (27%) will be deemed to be lower ability (Hingorjo and Jaleel, 2012). Out of the 41 students, 11 will be in the high group and eleven in the lower group, while nineteen will be in the middle group and will not be used in this study. Based on the data, the discrimination index (DIS), difficulty index (DI) or (p-value), and distractor efficiency (DE) were calculated following the method previously described by Hingorjo and Jaleel (2012), Gajjar et al. (2014).

 

p-value or difficulty index (DI) = [(H+L)/N] x 100 -------[1]

 

Discrimination index (DIS) = 2x [(H-L)/N]----------------[2]

 

Where N is the total number of students in both high and low groups, and H and L are the correct responses in high and low groups, respectively.

 

Items with P< 30% are considered difficult, p=30-70% are ideal, and P>70% are considered easy. Discrimination index (DIS) is the ability of an item to differentiate between students of higher and lower abilities, and it ranges between 0 and 1.

D=Negative (Defective item/wrong key), D=0-0.19 (Poor discrimination), D=0.2-0.29 (Acceptable discrimination), D= 0.3-0.39 (Good discrimination) and D>0.4 (Excellent discrimination). An item contains a stem, five options, including one correct option (Key), and four incorrect alternatives (Distractors). Non-functional distractors (NFD) were options other than the key selected by <5% of students, and functional distractors were the option (distractors) selected by 5% or more students. DE ranged from 0% to 100% and was determined based on the number of NFDs in an item. If an item contains four, three, two, one, and zero NFDs, the DE would be 0%, 25%, 50%, 75%, and 100%, respectively.

 

Statistical analysis

 

SPSS version 20 and Microsoft Excel were used to carry out the statistical analysis. Descriptive statistics (mean, mode, median, and standard deviation), Pearson's correlation, Chi-square, analysis of variance, and reliability (Cronbach’s alpha) were carried out at varying levels. Charts were used to show the total score using histograms with binomial distribution curves, bar charts for key distribution and percentages of students that passed, and interpolation lines for criteria for the different indices (DIS, DI, and DE).

 

 

RESULTS

 

Table 1 shows the discrimination and difficulty indices of OSPE among first-year Microbiology major students. The DI and DIS ranged from 0.00 – 50.00 with a mean ± standard deviation of 14.66±12.79 and 0.00 – 100.00 with mean ± standard deviation of 0.29±0.26, respectively. However, the DIS and DI mean value was within acceptable and very difficult ranges, respectively. The percentage distribution of the DIS and DI criteria are shown in Figures 1 and 2, respectively. Out of the 40 items, 15 representing 37.50%, 2 representing 5.00%, 3 representing 7.50%, and 20 representing 50.00% showed excellent, good, acceptable, and poor discrimination, respectively (Figure 1). These frequencies showed significant variations (X2 =58.72, P=0.000). For the DI, 4 representing 10.00% and 36 representing 90.00% showed that the items were ideal and difficult, respectively (Figure 2). Statistically, there were variations (X2 =64.00, P=0.000) between the two criteria that the values of this study fall within.


 

 

Table 1: Discrimination and difficulty indices of OSPE among first-year Microbiology major students

Items

Difficulty index

Criteria

Discrimination index

Criteria

1.       

50.00

Ideal

1.00

Excellent

2.       

13.64

Difficult

0.27

Acceptable

3.       

0.00

Difficult

0.00

Poor

4.       

4.55

Difficult

0.09

Poor

5.       

22.73

Difficult

0.45

Excellent

6.       

22.73

Difficult

0.45

Excellent

7.       

0.00

Difficult

0.00

Poor

8.       

0.00

Difficult

0.00

Poor

9.       

0.00

Difficult

0.00

Poor

10.    

0.00

Difficult

0.00

Poor

11.    

22.73

Difficult

0.45

Excellent

12.    

13.64

Difficult

0.27

Acceptable

13.    

9.09

Difficult

0.18

Poor

14.    

9.09

Difficult

0.18

Poor

15.    

4.55

Difficult

0.09

Poor

16.    

13.64

Difficult

0.27

Acceptable

17.    

22.73

Difficult

0.45

Excellent

18.    

22.73

Difficult

0.45

Excellent

19.    

4.55

Difficult

0.09

Poor

20.    

22.73

Difficult

0.45

Excellent

21.    

0.00

Difficult

0.00

Poor

22.    

9.09

Difficult

0.18

Poor

23.    

4.55

Difficult

0.09

Poor

24.    

4.55

Difficult

0.09

Poor

25.    

27.27

Difficult

0.55

Excellent

26.    

36.36

Ideal

0.73

Excellent

27.    

9.09

Difficult

0.18

Poor

28.    

4.55

Difficult

0.09

Poor

29.    

4.55

Difficult

0.09

Poor

30.    

4.55

Difficult

0.09

Poor

31.    

27.27

Difficult

0.55

Excellent

32.    

18.18

Difficult

0.36

Good

33.    

27.27

Difficult

0.55

Excellent

34.    

9.09

Difficult

0.18

Poor

35.    

40.91

Ideal

0.82

Excellent

36.    

22.73

Difficult

0.45

Excellent

37.    

22.73

Difficult

0.45

Excellent

38.    

36.36

Ideal

0.73

Excellent

39.    

18.18

Difficult

0.36

Good

40.    

0.00

Difficult

0.00

Poor

Mean

14.66

Difficult

0.29

Acceptable

Standard deviation(±)

      12.79

-

0.26

-

Minimum

0.00

-

0.00

-

Maximum

50.00

-

1.00

-

Source: Authors

 

 

Figure 1: Percentage distribution of the discrimination index criteria.

 

 

Figure 2: Percentage distribution of the difficulty index criteria

 


 

 

Table 2 shows the distractor efficiency and implication of the outcome of OSPE among first-year students of the Microbiology programme. The DE ranged from 0.00 – 100.00%, with a mean of 44.74±30.74%. Out of 160 distractors, 71 (44.38%) and 89 (55.62%) were functional and non-functional distractors respectively (Table 3). Thus, the majority of the items need one or more distractors revised. The percentage distribution of the distractors is shown in Figure 3. Among the 40 items, 3 (7.50%), 8 (20.00%), 15 (37.50%), 5 (12.50%) and 9 (22.50%) had 0, 1, 2, 3 and 4 non-functional distractors, respectively. These frequencies showed significant variations (X2 =25.75, P=0.000).

Table 4 shows Pearson’s correlation between the DI, DIS, and DE. The indices (DI, DIS, and DE) showed very strong positive significant relationships at p=0.01. The key distribution of OSPE among first-year microbiology major students is shown in Figure 4. The keys had equal (8 each 20.00%) distribution.

The distribution of the OSPE score according to the grading criteria is shown in Figure 5. Out of the 41 students that participated in the OSPE 18, 11, 9, 2 and 1 representing 43.90%, 16.80%, 22.00%, 4.90% and 2.40% scored A (70 – 85%), B (60 – 69%), C (50 – 59%), D (45 – 49%) and E (40 – 44%), respectively. Statistically, there was variation (X2 =58.90, P=0.000).  The distribution (histogram showing the normal curve) of the students' scores is shown in Figure 6. The histogram showed skewness ± standard error of skewness of -0.248 ± 0.369. The mean and standard deviation was 26.46 ±4.70. The least score was 16, which is the pass mark.

 

 


Table 2: Distractor efficiency and options distribution (4 distractors + 1 key) of OSPE among first-year students of the Microbiology programme

Items

Options  (4 distractors + 1 key)

 

Distractor efficiency, %

A

B

C

D

E

1

2(4.90)

2(4.90)

29(70.70)

4(9.80)

4 (9.80)

50

2

6(14.60)

16(39.00)

3(7.30)

15(36.60)

1(2.40)

75

3

8(19.50)

10(24.40)

13(31.70)

7(17.10)

3(7.30)

100

4

0(0.00)

1(2.40)

0(0.00)

0(0.00)

40(97.60)

0

5

0(0.00)

2(4.90)

33(80.50)

6(14.60)

0(0.00)

25

6

5(12.20)

30(73.20)

2(4.90)

1(2.40)

3(7.30)

50

7

100(100.00)

0(0.00)

0(0.00)

0(0.00)

0(0.00)

0

8

1(2.40)

8(19.50)

15(36.60)

1(2.40)

16(39.00)

50

9

100(100.00)

0(0.00)

0(0.00)

0(0.00)

0(0.00)

0

10

0(0.00)

0(0.00)

1(2.40)

39(95.10)

1(2.40)

0

11

4(9.80)

1(2.40)

2(4.90)

28(68.30)

6(14.60)

50

12

0(0.00)

9(22.00)

3(7.30)

8(19.5)

21(51.20)

75

13

5(12.20)

1(2.40)

29(70.70)

5(12.20)

1(2.40)

50

14

25(61.00)

3(7.30)

8(19.50)

3(7.30)

2(4.90)

75

15

0(0.00)

38(92.70)

1(2.40)

2(4.90)

0(0.00)

0

16

18(43.90)

15(36.60)

2(4.90)

1(2.40)

5(12.20)

50

17

1(2.40)

9(22.00)

9(22.00)

5(12.20)

17(41.50)

75

18

0(0.00)

2(4.90)

8(19.50)

26(63.40)

5(12.20)

50

19

8(19.50)

12(29.30)

0(0.00)

15(36.60)

6(14.60)

75

20

1(2.40)

22(53.70)

5(12.20)

6(14.60)

7(17.10)

75

21

0(0.00)

0(0.00)

41(100.00)

0(0.00)

0(0.00)

0

22

35(85.40)

2(4.90)

3(7.30)

1(2.40)

0(0.00)

25

23

1(2.40)

0(0.00)

40(97.60)

0(0.00)

0(0.00)

0

24

0(0.00)

3(7.30)

0(0.00)

5(12.20)

33(80.50)

50

25

6(14.60)

26(63.40)

0(0.00)

4(9.80)

5(12.20)

75

26

25(61.00)

4(9.80)

3(7.30)

4(9.80)

5(12.20)

100

27

0(0.00)

2(4.90)

0(0.00)

38(92.70)

1(2.40)

0

28

0(0.00)

0(0.00)

40(97.60)

1(2.40)

0(0.00)

0

29

2(4.90)

3(7.30)

1(2.40)

0(0.00)

35(85.40)

25

30

6(14.60)

2(4.90)

1(2.40)

0(0.00)

32(78.00)

25

31

3(7.30)

1(2.40)

11(26.80)

26(63.40)

0(0.00)

50

32

19(46.30)

10(24.40)

4(9.80)

3(7.30)

5(12.20)

100

33

2(4.90)

0(0.00)

27(65.90)

10(24.40)

2(4.90)

25

34

2(4.90)

13(31.70)

23(56.10)

3(7.30)

0(0.00)

50

35

26(63.40)

1(2.40)

6(14.60)

8(19.50)

0(0.00)

50

36

15(36.60)

1(2.40)

19(46.30)

4(9.80)

2(4.90)

50

37

28(68.30)

1(2.40)

7(17.10)

1(2.40)

4(9.80)

50

38

5(12.20)

24(58.50)

1(2.40)

2(4.90)

9(22.00)

50

39

4(9.80)

26(63.40)

4(9.80)

0(0.00)

7(17.10)

75

40

6(14.60)

0(0.00)

5(12.20)

30(73.2)

0(0.00)

50

Mean

-

-

-

-

-

44.38

Standard error

-

-

-

-

-

30.74

Minimum

-

-

-

-

-

0.00

Maximum

-

-

-

-

-

100.00

Source: Authors

 

 

Table 3: Summary of Distractor efficiency of OSPE among first-year students of Microbiology programme

Parameters

N

%

Total OSPE

40

-

Distracters (Total)

160

-

Functional Distractors

71

44.38

Non-functional distractors

89

55.62

Source: Authors

 

Figure 3: Percentage distribution of the distractor efficiency criteria

 

Table 4: Pearson’s correlation between difficulty index, discrimination index, and distractor efficiency

Parameters

Difficulty index

Discrimination index

Distractor efficiency

Difficulty index

1

 

 

Discrimination index

1.000**

1

 

Distractor efficiency

0.408**

0.408**

1

**. Correlation is significant at the 0.01 level (2-tailed).

Source: Authors

N=40

 

Figure 4: Percentage distribution of the keys

 


 

The Cronbach's Alpha of all the OSPE items is 0.790. When some items are deleted, the internal consistency of the items (Cronbach's Alpha) decreases but is not less than 0.766 (Table 5). The analysis of variance between the items showed that p=0.000 (Table 6).


 

Figure 5: Distribution of OSPE scores according to the grading criteria

 

 

Figure 6: Distribution of OSPE total scores

 

 

Table 5: Cronbach's Alpha if each item is deleted

Item-Total Statistics

Items

Scale Mean if Item is Deleted

Scale Variance if Item is Deleted

Corrected Item-Total Correlation

Cronbach's Alpha if Item is Deleted

Q1

27.2667

26.638

.604

.778

Q2

27.2000

28.314

.000

.791

Q3

27.8000

28.600

-.100

.802

Q4

27.2000

28.314

.000

.791

Q5

27.4667

25.838

.487

.776

Q6

27.3333

26.667

.419

.781

Q7

27.2000

28.314

.000

.791

Q8

27.9333

29.495

-.280

.807

Q9

27.2000

28.314

.000

.791

Q10

27.3333

26.952

.339

.783

Q11

27.4667

25.981

.455

.778

Q12

27.6000

25.543

.491

.775

Q13

27.6000

27.543

.097

.794

Q14

27.8000

27.457

.113

.793

Q15

27.2667

28.210

.014

.792

Q16

27.8000

28.886

-.152

.804

Q17

27.7333

25.781

.432

.778

Q18

27.4000

25.257

.693

.769

Q19

27.9333

28.924

-.166

.803

Q20

27.4667

26.695

.298

.784

Q21

27.2000

28.314

.000

.791

Q22

27.4000

27.686

.105

.792

Q23

27.2000

28.314

.000

.791

Q24

27.4667

26.410

.360

.782

Q25

27.6000

24.686

.669

.767

Q26

27.5333

26.838

.245

.787

Q27

27.2667

26.638

.604

.778

Q28

27.2000

28.314

.000

.791

Q29

27.3333

27.095

.299

.785

Q30

27.4000

27.400

.171

.789

Q31

27.6667

25.524

.484

.775

Q32

27.7333

25.781

.432

.778

Q33

27.5333

25.695

.481

.776

Q34

27.6000

29.114

-.193

.806

Q35

27.6000

24.400

.730

.764

Q36

27.4667

25.981

.455

.778

Q37

27.5333

25.838

.451

.777

Q38

27.4000

24.971

.766

.766

Q39

27.4667

27.410

.145

.791

Q40

28.2000

28.314

.000

.791

 

 

Table 6: Overall analysis of variance between the OSPE items

 

Sum of Squares

df

Mean Square

F

Sig

Between People

9.910

14

.708

 

 

Within People

Between Items

33.718

39

.865

5.817

.000

Residual

81.157

546

.149

 

 

Total

114.875

585

.196

 

 

Total

124.785

599

.208

 

 

Grand Mean = .7050

Source: Authors

 

 

 

 


DISCUSSION

 

Item analysis is usually carried out to examine students’ responses to each item (OSPE). Its major focus is to determine the quality of the OSPE items and the overall test (examination). The discrimination index, which is usually used to show the high and low ability students, showed that 50.00% of the items had poor discrimination, while the DI is very high. Also, the non-functional distractors account for 55.62% of the total distractors. The DI, DIS, and DE suggest a need to improve the items because a weak design (especially for the distractors) may have contributed to the overall DI. Thus, these criteria are not providing adequate assessment (Izah et al., 2021).

In this study, item difficulty is seen in terms of frequencies with which those taking the OSPE chose the single best response instead of the distractors are associated with intrinsic characteristics of the items. It could also be seen that the items were too easy and answered correctly by nearly all the students. In contrast, the very difficult questions were responded to wrongly by both students (students with high and low abilities). This could be seen in 7 (17.50%) of the items. According to Chhaya et al. (2018), items answered correctly or wrongly by both groups of students should be removed. The mean difficulty index recorded in this study is lower than the values of 39.40% (Gajjar et al., 2014), 50.16% (Rao et al., 2017), 55.90% (Patel, 2017), 58.74% (Mahjabeen et al., 2017) and 57.62% (Chhaya et al., 2018) from MCQ examinations. The variation may be associated with intrinsic factors in the OSPE design.

 The discrimination index ranges from 0 to 1.0. The findings of this study showed that 50.00% of the items had a poor discrimination index. According to Charanja et al. (2015), poor discrimination index is caused by ambiguous items, wrong keys, many correct answers, too easy or difficult items, and failure of teaching and learning sessions. From these criteria, the poor discrimination observed in 50.00% of items may be associated with the fact that the items were too easy, as shown in Figure 5, which showed that 38 students (92.70%) of the students scored ≥50.00% of the total score and the remaining three students (7.30%) scored ≥40.00% to <50 of the total score.

Even though the mean DIS is within the acceptable range, the values were within previously reported ranges of 0.29 (Patel, 2017), 0.22 (Charania et al., 2015), but lower than the values of 0.34 (Rao et al., 2017) and 0.35 (Mahjabeen et al., 2017) recorded in some MCQ examinations.

The DE showed that approximately 44.00% of the distractors were effective, while the non-functional distractors accounted for over 55.00% of the total distractors. However, the findings of the non-functional distractors are related to the results of previous studies that have values of 15.00% (Patel, 2017), 5.00% (Rao et al., 2017), 11.4 (Gajjar et al., 2014), and lower than the value of 28.00% reported by Mahjabeen et al. (2017) in some MCQ examinations. However, the non-functional distractors also influenced the DI and DIS. As seen in Table 4, there was a strong significant correlation between the three indicators (DE, DI, and DIS). The findings also agree with the work of Rao et al. (2017). According to Rao et al. (2017), reducing the number of distractors increases DIS and reliability level.

The study further revealed that no students failed the examination, but the distribution did not follow the standard binomial distribution curve.  The skewness, which was less than -1, indicates the distribution is highly asymmetrical. However, the mean and standard deviation values provide estimates of the actual parameters of the curve. Thus, it gives information about the symmetrical distribution of the number of students that passed. But in this study, the skewness (degree of asymmetry or not symmetrical) is negative, having a longer tail in the left of the central maximum with mean values less than the mode and median values, an indication that most of the students are high scorers.

The reliability (internal consistency) is within 0.70 – 0.80 and 0.80 – 0.90, which are classified as suitable for a classroom assessment with few items requiring improvement and very good for classroom assessment (Patel, 2017). However, a value of 0.790 recorded in this study indicates that the overall test is still good and reliable. However, deletion of some of the items reduces the Cronbach’s alpha value, which suggests that such items strongly influenced the reliability of the test. The Cronbach's Alpha value recorded in this study is slightly higher than the value of 0.702 previously reported by Patel (2017) in an MCQ examination. Again, the analysis of variance showed that p=0.000, an indication that the reliability of the items is quite different statistically.

 

 

CONCLUSIONS

 

Item analysis as an evaluation tool is beneficial to both students and teachers. This study assessed the quality of OSPE served to first-year microbiology major students, and the results showed that the mean DIS was within the acceptable range. However, 17.50% of the items need to be removed because of their poor discrimination power. The DI showed that 90% of the items are difficult, and 10% are ideal. None of the students scored below the 40% pass mark, indicating that most of the items were too easy. The DE showed that over 55.00% of the distractors were not effective. The Cronbach's Alpha showed that the items are classified as good for an assessment though improvement is required. Based on the DI and DE, there is a need to carry out item flaws and technical pitfalls analysis to improve the assessment scores, identify the difficult items, discriminate among the students, and remove or revise the non-functioning distractors. From the findings in this study it is recommended that this exercise be repeated for the next four years and the results compared to ensure assessments items are standard. It is also recommended that OSPE be more widely adopted among more science-based departments and item analysis be a standard practice in departments of every university.

 

Acknowledgments

 

The authors would like to express their thanks to the following Laboratory Technologists of Bayelsa Medical University that that participated in the invigilation of the OSPE; Ms. Timipre Grace Tuaboboh, Ms. Biembele Virtuous Temple, Ms. Sebhaziba Benjamin Ezem, Ms. Blessing Muji Olagoke, Ms. Ann Tugwell Ototo, Mrs Christy Koroye, Mr Henry Ebiowei Alpha and Mr Samuel Philemon Bokene.

 

Ethical approval

 

Ethical approved was obtained from the Research and Ethics Committee of Bayelsa Medical University, Yenagoa, Bayelsa State, Nigeria with Ethical approval number REC/2021/0009.

 

Competing interests

 

The authors declare that they have no competing interests.

 

 

REFERENCES

 

1.      Armour-Thomas, E., and E.W. Gordon, 2013. “Toward an understanding of assessment as a dynamic component of pedagogy”.  Gordon Commission on the future of assessment in education. Retrieved June 30 2020. https://www.ets.org/Media/Research/pdf/armour_thomas_gordon_understanding_assessment.pdf.

2.      Bhat, D., P. Murugesh, and NB. Pushpa, 2020. “Objective structured practical examination: As an assessment tool in newly introduced competency based anatomy curriculum”. Indian Journal of Clinical Anatomy and Physiology7(1): 81-86.

3.      Centre for Educational Research and Innovation (CERI), 2008. “Assessment for learning formative assessment”. Retrieved from oecd.org/site/educeri21st/40600533.pdf. Accessed December 16, 2020.

4.      Charania, JS 2015. "Item analysis of multiple choice questions given to first year medical students-concept building and MCQs". International Journal of Research in Humanities and Social Sciences, 3(9): 18 – 29.

5.      Chhaya, J., H. Bhabhor, J. Devalia, U. Machhar, and A. Kavishvar, 2018. “A study of quality check on multiple choice questions (MCQs) using item analysis for differentiating good and poor performing students”. Healthline Journal, 9(1): 24 – 29.

6.      Frantz, J.M., M. Rowe, D.A. Hess, A.J. Rhoda, B.L. Sauls, and L. Wegner, 2013. “Student and staff perceptions and experiences of the introduction of objective structured practical examinations:  A pilot study”. African Journal of Health Professions Education 5(2):72-74.

7.      Gajjar, S., R. Sharma, P. Kumar, and M. Rana, 2014. “Item and test analysis to identify quality multiple choice questions (MCQs) from an assessment of medical students of Ahmedabad, Gujarat”. Indian journal of community medicine: official publication of Indian Association of Preventive & Social Medicine39(1): 17-20.

8.      Harden, R.M., and RG Cairncross, 1980. “Assessment of practical skills: The objective structured practical examination (OSPE)”. Studies in Higher Education 5(2): 187-196.

9.      Hingorjo, M.R., and F. Jaleel 2012. “Analysis of one-best MCQs: the difficulty index, discrimination index and distractor efficiency”.  Journal of the Pakistan Medical Association62(2): 142 - 147.

10.   Izah, S.C., T.C. Odubo, V.E. Ajumobi, and K.E. Torru, 2021. “Item analysis of Multiple Choice Questions (MCQs) from a formative assessment of first year microbiology major students”. Research and Review Insights, 5:1-6. doi: 10.15761/RRI.1000166.

11.   Jaswal, S., J. Chattwal, J. Kaur, S. Gupta, and T. Singh, 2015. “Assessment for learning with Objectively Structured Practical Examination in Biochemistry”. International Journal of Applied and Basic Medical Research 5:S71-S75.

12.   Mahjabeen, W., S. Alam, U. Hassan, T. Zafar, R. Butt, S. Konain, and M. Rizvi, 2017. “Difficulty index, discrimination index and distractor efficiency in multiple choice questions”. Annals of Pakistan Institute of Medical Sciences, 3(4):310 – 315.

13.   Mard, S.A., and S. Ghafouri, 2020. “Objective Structured Practical Examination in Experimental Physiology Increased Satisfaction of Medical Students”. Advances in Medical Education and Practice 11: 651—659

14.   Mokkapati, A., G. Pavani, S.M. Dass, and M.S. Rao, 2016. “Objective structured practical examination as a formative assessment tool for IInd MBBS microbiology students”. International Journal of Medical Science and Public Health, 4(10): 4535-4540.

15.   Munjal, K., P.K. Bandi, A. Varma, and S. Nandedkar, 2012. “Assessment of medical students by OSPE method in pathology”. Internet Journal of Medical Update, 7(1):2-7.

16.   Nigam, R., and P. Mahawar, 2011. “Critical analysis of performance of MBBS students using OSPE & TDPE–A comparative study”. National Journal of Community Medicine2(3), 322-24.

17.   Patel, R.M. 2017. “Use of item analysis to improve quality of multiple choice questions in II MBBS”. Journal of Education Technology in Health Sciences, 4(1): 22 – 29.

18.   Rao, C., H.K. Prasad, K. Sajitha, H. Permi, and J. Shetty, 2016. “Item Analysis Of Multiple Choice Questions: Assessing An Assessment Tool In Medical Students”. International Journal of Educational And Psychological Researches2(4): 201.

19.   Relwani, N.R., R.A. Wadke, S. Anjenaya, and P.N. Sawardekar, 2016. Effectiveness of objective structured practical examination as a formative assessment tool as compared to traditional method for MBBS students. International Journal of Community Medicine and Public Health, 3(12):3526-3532.

20.   Vijaya, D.S., Alan, 2014. “Comparative Study to Evaluate Practical Skills in Physiology Among 1st Phase Medical Under Graduates At Jnmc Belgaum: Traditional Practical Examinations Versus Objective Structure Practical Examinations (TPE V/S OSPE)”. International Journal of Educational Research and Technology, 5(1):126-134.

21.   Vishwakarma, K., M. Sharma, P.S. Matreja, and V.P. Giri, 2016. “Introducing objective structured practical examination as a method of learning and evaluation for undergraduate pharmacology”. Indian Journal of Pharmacology 48(Suppl 1): S47–S51.

 


 

 

Cite this Article: Izah, SC; Odubo, TC; Ajumobi, VE; Osinowo, O (2022). Item Analysis of Objective Structured Practical Examination (OSPE) Used as an Assessment Tool for First-Year Microbiology Students. Greener Journal of Biological Sciences, 12(1): 11-22.