Hostname: page-component-586b7cd67f-t8hqh Total loading time: 0 Render date: 2024-11-27T20:21:00.508Z Has data issue: false hasContentIssue false

Development and Validation of a New Tool to Measure Performance Knowledge and Communication Skill of Multidisciplinary Health Science Learners on Radiation Emergency Preparedness and Response Management

Published online by Cambridge University Press:  29 June 2023

Shivanand Bhushan*
Affiliation:
Manipal College of Health Profession (MCHP) MAHE Manipal-Karnataka, India
Sumeet Suresh Malapure
Affiliation:
KMC, MAHE Manipal-Karnataka, India
Nisha Rachel D’Souza
Affiliation:
Manipal College of Health Profession (MCHP) MAHE Manipal-Karnataka, India
Pankaj Tandon
Affiliation:
Head TSS, Radiological Safety Division AERB Mumbai, India
Sibi Oommen
Affiliation:
Manipal College of Health Profession (MCHP) MAHE Manipal-Karnataka, India
Srinidhi G.
Affiliation:
KMC, MAHE Manipal-Karnataka, India
Ashwani Kumar Verma
Affiliation:
School of Pharmaceutical and Populations Health Informatics, DIT University, Uttarakhand, India
Akhilesh K. Pandey
Affiliation:
KMC, MAHE Manipal-Karnataka, India
William Wilson
Affiliation:
KMC, MAHE Manipal-Karnataka, India
Muralidhar M. Kulkarni
Affiliation:
KMC, MAHE Manipal-Karnataka, India
*
Corresponding author: Shivanand Bhushan, Emails: [email protected]; [email protected].
Rights & Permissions [Opens in a new window]

Abstract

Objective:

The purpose of the study was to design, develop, and validate a newer tool on radiation emergency preparedness responses (RadEM-PREM IPE tool) to measure communication, knowledge, performance skills in multidisciplinary health science learners.

Methods:

The study design is a prospective, single centric, pilot study. Five subject experts designed, analyzed, and selected items of the instrument for relevant content and domain. Psychometrics that the tool assessed were content validity, internal consistency, test-retest reliability, and intraclass correlation coefficient. Twenty-eight participants completed test-retest reliability for validation of 21 sorted out items calculated percentage of agreement >70% I-CVI/UA (item content validity index with universal acceptability) and S-CVI/UA (scale content validity index with universal agreement method).

Results:

Items with percentage agreement >70% and I-CVI over 0.80 were kept, ranged from 0.70 to 0.78 were revised, and below 0.70 were rejected. Items with kappa values ranging from 0.04 to 0.59 were revised and ≥0.74 were retained. Internal consistency assessed using Cronbach’s alpha was 0.449. Positive correlation between attitude and communication (r = 0.448), between performance and communication (r = 0.443) were statistically significant at 0.01 level. Overall, intraclass correlation coefficient for all the measures is 0.646, which is statistically significant at 0.05 level (P < 0.05).

Conclusions:

Study concludes that the RadEM-PREM IPE tool would be a new measuring tool to assess knowledge, performance, and communication skills of interprofessional radiation emergency response team learner’s evaluation.

Type
Original Research
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of Society for Disaster Medicine and Public Health, Inc.

Emergency preparedness and response (EPR) to handle radioactive materials is concerned with occupational health and safety at the work place. Enriching the high standard, quality work culture by adopting good radiation safety practices in a nuclear medicine setup is a vital step in minimizing hazards and protecting the health of people. Lack of knowledge, behavior (attitude), and practice process for handling the various awkward radiation emergency situations is alarming among all the stakeholders and creates panic situations and consequences that are hazardous to professional, public, patients, and environments. To assess and evaluate learners’ pre-existing level of learning, the learning acquired in course of time, continuous improvement in the delivery of content, and modes of training administration as per the Kirkpatrick level 2 evaluation, Reference Kirkpatrick and Kirkpatrick1,Reference Kirkpatrick, Brown and Seidner2 there is a need for an appropriate assessment tool, which is lacking at present for radiation emergency response team evaluation. Assessment of learners for psychometric parameters analysis, the close-ended questionnaires design, formatting, validating, and reliability are the systematic process and vital to analyze educational deliverable programs. In the same line, developing an evidence-based radiation emergency response assessment in the hospital setup is vital to design and develop a reliable and valid assessment tool.

Aims and Objectives

The aim of this present study is to develop a valid and reliable tool on radiation emergency response preparedness and response management (RadEM-PREM) also to explore the psychometric measures of the instrument. Once this instrument is ready for intervention, it will help to evaluate the net learning and the learning gap in future-ready multidisciplinary health science graduate students. Objective of the study is to design, develop, validate, and check the reliability of the items for “Radiation emergency preparedness response management (RadEM-PREM) Inter professional evaluation (IPE) assessment tool,” which is lacking at present.

This RadEM-PREM IPE evaluation tool would be significantly helpful to assess the readiness of health science interprofessional learners in team building for radiation emergency preparedness and response.

The process adopted in this work is designing the items’ content, content validation, face validation, analyzing test-retest of the items to check for their reliability for ready to use as assessment tool recommended by Considine et al. in 2005, Tsang et al. in 2017, Calonge-Pascual et al. in 2020. Reference Considine, Botti and Design3Reference Calonge-Pascual, Fuentes-Jiménez and Casajús Mallén5

This work in the long-run will be very effective and productive to assess and evaluate the pre-existing learning level, net learning attributes and learning lag to bridge the continuous improvements in future training to check the interprofessional learner’s readiness for building multidisciplinary radiation emergency response team.

Methods

Ethical Committee Approval

This work is duly approved from KMC and KH Institutional Ethics Committee vide IEC :1017 /2019 in December 2019.

Study Design

Study design is prospective, cross-sectional, single centric, pilot study conducted in India.

Participants

Items were generated by 5 subject experts for instrument design for relevant content, and domain specifications followed the stages of questionnaire design and development systematic good practice process. The tool is hypothesized for initiating team work capacity building in radiation emergency preparedness and response area in a hospital setup. Psychometrics of the tool was assessed in terms of content validity, internal consistency, and test-retest reliability.

Of the 3 domains of knowledge, performance skill, and communication skill, items were sorted out based on the high percentage relevance of the content-domain compatibility. In the next step of validation, the 5 subject experts validated the contents of items, substituting to remove ambiguousness in the item contents, modification–remodification, screened for the levels of learning in the radiation emergency response area. Content validity was assessed to review the scale by a panel of 5 content experts with professional expertise in radiation protection. These 5 content experts were experienced either teaching faculty, a professional member of Nuclear Medicine Physicists Association of India (NMPAI)/Association of Medical Physicist India (AMPI), or having at least more than 10 y of professional practicing experience in radiation safety and the protection relevant field of medical radioisotope applications handling. All the item content experts and validators have at least a postgraduate degree in science holding with postgraduate degree/diploma in medical physics or nuclear medicine relevant professional qualifications and training in radiation protection.

Twenty-eight radiation safety expert professionals completed test re-test reliability of items for validation of RadEM preparedness management 21-item questionnaire Tool (RadEM-PREM tool) designed and developed for multidisciplinary interprofessional education (IPE) health science learners.

Study Size

In the present study, 92 items were screened out to 42 items in the first stage. In next stage, a further 21 items were sorted out for 3 domains such as knowledge learning attributes (KS), performance (PS), and communication (CS) skills based on the content validity index and high relevance in their respective domains. A total of 21 items were sorted out having percentage of agreement more than 70%, and each item’s content validity index with universal acceptability (I-CVI/UA) and scale content validity index with universal acceptability (S-CVI/ UA) is calculated. Of 21 items, 7 belong to the knowledge domain, 7 were from the performance domain, and another 7 were from the communication domain.

The inclusion criterion for the present study was working radiation professional experts in India. Entry-level future ready radiation professional were excluded.

Measurement Tool Used

Measurement tool includes close-ended questionnaires having 5 options which was curated and sent to the subject experts. The experts were then asked to rate each item based on relevance, clarity, grammatical /spelling, ambiguity, and structure of the sentences. Substituting content body, editing, correcting is adopted for domain-wise relevance of the items. Item-level content Validity index (I-CVI), Scale-Content Validity index (S-CVI)/Universal agreement (UA), Cronbach alpha coefficient, Pearson’s correlation coefficient (PC), intraclass correlation coefficient, percentage relevance of the tools are either discussed or tabulated.

Results

The high percentage relevance of the content-domain compatibility on the 4-rater scale initiated for instrument design process by 5 validators is shown in Table 1. In this table, validators’ percentage agreement for relevancy of the 21 items is tabulated in percentage relevance. The I-CVI, S-CVI, and modified kappa were analyzed for all the questions using Microsoft Excel (Microsoft Corp., Redmond, WA) (Table 2).

Table 1. Initial validators percentage agreement domain-wise rater scale

Table 2. Item-wise 21 item I-CVI, SCVI-average, and SCVI /UA

Lynn et al. Reference Lynn6 recommended that I-CVIs should be no lower than 0.78 to accept that item content valid. Researcher used I-CVI information in revising, deleting, or substituting items. The items that had a CVI over 0.80 were kept, and those ranged from 0.70 to 0.78 were revised. Items with CVI score of less than 0.70 were rejected and not considered for further inclusion or revision. The S-CVI calculated using universal agreement (UA) was calculated by adding all items with an I-CVI equal to 1 divided by the total number of items. The S-CVI/UA of 21 items ranged between ≥0.8 and ≥1.0, considering that the items had excellent content validity. The modified kappa was calculated using the formula: kappa = (I-CVI-Pc)/ (1-Pc), where Pc = [N!/A!(N-A)!]*0.5N. In this formula Pc = the probability of chance agreement; N = number of experts; and A = number of experts that agree the item is relevant. Reliability testing was done, and items with a kappa value ≥0.74 were retained and those with values ranging from 0.04 to 0.59 were revised.

Reliability of the developed tool done by doing test-retest of these questionnaire items measured Cronbach’s alpha coefficient. The test-retest reliability of all 21 items were administered on-line through MS form to 28 radiation safety professionals working in India within a 4-week interval. The internal consistency of the subscale was assessed using IBM SPSS-16 software in which Cronbach’s alpha was reported to be 0.449.

Pearson’s product-moment correlation was used to assess the correlation between the subscales of the questionnaire for item discrimination analysis as advocated by Haladyna 1999. Reference Haladyna7 A highly positive correlation was observed between communication and knowledge (r = 0.728). Moderate correlation was observed between performance and knowledge (r = 0.544) and also between performance and communication (r = 0.443). All 3 correlations among knowledge, communication, and performance were statistically significant at the 0.01 level (Table 3).

Table 3. Pearson correlation coefficient for the questionnaire

** Correlation is significant at 0.01 level (2-tailed).

Overall, intraclass correlation coefficient for all the measures was 0.646, which was statistically significant at 0.05 level (P < 0.05).

Limitations

This tool has been developed and validated in consultation with content experts and peers of radiation protection. Furthermore, the tool is checked for reliability and internal consistency on another group of radiation safety professionals working in India. However, this tool is not tested among the interprofessional learner sample group from multidisciplinary health sciences including medical students.

Discussion

The present study describes a newer tool to assess the knowledge along with performance and communication skills of interprofessional radiation emergency team learners, with 5 subject experts as recommended by Lynn 1986 Reference Lynn6 followed. In this tool design and development, all the steps have been taken into account (Polit and Beck 2006, Reference Polit and Beck8 Polit et al. 2007 Reference Polit, Beck and Owen9 ) to obtain content validity evidence. The example provided was taken from a construction process instead of an adaptation process. Thorndike 1994 Reference Thorndike10 advocates that both reliability coefficients and alternative form correlation be reported, so it is followed in this study too. Low reliability coefficients may indicate that the test done on 28 professionals are small. Another possible reason of less coefficiency, as Rattray and Jones in 2007 Reference Rattray and Jones11 reported, may be close-ended questions, which mainly restrict the depth of participants response and, hence, diminish or render incomplete the quality of the data. Another possible reason of low reliability coefficient may be that the professional who participated in the re-test either did not pay required attention on the items, were disinterested, or got distracted. A solution to this could be item discrimination analysis to conduct on a larger group of panel experts. To overcome, in this study, an alternative form of correlation has also been calculated following the guidelines and it is greater than the reliability coefficient (r = 0.66). Peer review practices using guidelines led to improved psychometric characteristics of items. Reference Abozaid, Park and Tekian12 This instrument has been constructed it has undergone the testing by limited experts as pilot data collection, and their results are very promising; need to check test-retest reliability in either a larger panel expert group or multidisciplinary health science learners group after the sufficient span of time.

The new “RadEM-PREM IPE assessment tool” would be very useful for assessment of multidisciplinary team-building in the area of radiation emergency preparedness and response in entry-level and intermediate-level emergency response team members.

Conclusions

This study concludes that the 21-item “RadEM-PREM IPE assessment tool” is a valid and reliable new measuring tool. “RadEM-PREM IPE assessment tool” is to assess knowledge, performance, and communication skills of interprofessional radiation emergency response team learners in hospital settings.

Acknowledgments

This project was done as a part of the Fellowship Program at MAHE-FAIMER International Institute for Leadership in Interprofessional Education (M-FIILIPE), Manipal Academy of Higher Education, (MAHE) Manipal, India. We thank Dr. Ciraj Ali Mohammed, Course Director M-FIILIPE and CCEID MAHE Manipal, all MAHE FAIMER faculty and MCHP Manipal for support and guidance.

Author contribution

B.S., M.K., P.T., S.S.M., W.W. Data curation: B.S., P.T., A.V., S.G.C., N.R.D. Formal analysis: A.V., A.K., B.S. Methodology: Project administration: B.S., S.G.C., N.R.D. Visualization: B.S., S.G. N.R.D. Writing–original draft: B.S., M.K., S.O., A.K.P., A.V. Writing–review and editing: B.S., M.K. S.O., A.V.

Funding

None.

Competing interests

None.

Appendix

References

Kirkpatrick, DL, Kirkpatrick, JD. Evaluating Training Programs: The Four Levels. Berrett-Koehler Publishers; 2006. Accessed May 24, 2023. https://bkconnection.com/static/Evaluating_Training_Programs_EXCERPT.pdf Google Scholar
Kirkpatrick, DL. The four levels of evaluation. In: Brown, SM, Seidner, CJ, eds. Evaluating Corporate Training: Models and Issues. Springer Netherlands; 1998:95-112. Accessed May 24, 2023. https://doi.org/10.1007/978-94-011-4850-4_5 CrossRefGoogle Scholar
Considine, J, Botti, M, Design, Thomas S., format, validity and reliability of multiple choice questions for use in nursing research and education. Collegian. 2005;12(1):19-24.CrossRefGoogle ScholarPubMed
Tsang, S, Royse, CF, Terkawi, AS. Guidelines for developing, translating, and validating a questionnaire in perioperative and pain medicine. Saudi J Anaesth. 2017;11(Suppl 1):S80-S89.CrossRefGoogle ScholarPubMed
Calonge-Pascual, S, Fuentes-Jiménez, F, Casajús Mallén, JA, et al. Design and validity of a choice-modeling questionnaire to analyze the feasibility of implementing physical activity on prescription at primary health-care settings. Int J Environ Res Public Health. 2020;17(18):6627.CrossRefGoogle ScholarPubMed
Lynn, MR. Determination and quantification of content validity. Nurs Res. 1986;35(6):382-385.CrossRefGoogle ScholarPubMed
Haladyna, TM. Developing and Validating Multiple-choice Test Items. Routledge; 2004. Accessed May 24, 2023. https://www.taylorfrancis.com/books/mono/10.4324/9780203825945/developing-validating-multiple-choice-test-items-thomas-haladyna CrossRefGoogle Scholar
Polit, DF, Beck, CT. The content validity index: are you sure you know what’s being reported? Critique and recommendations. Res Nurs Health. 2006;29(5):489-497.CrossRefGoogle ScholarPubMed
Polit, DF, Beck, CT, Owen, SV. Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Res Nurs Health. 2007;30(4):459-467.CrossRefGoogle ScholarPubMed
Thorndike, RM. Book Review: Psychometric Theory (3rd ed.) by Jum Nunnally and Ira Bernstein New York: McGraw-Hill, 1994, xxiv + 752 pp. Appl Psychol Meas [Internet]. 1995 Sep 1;19(3):303-5. https://doi.org/10.1177/014662169501900308 CrossRefGoogle Scholar
Rattray, J, Jones, MC. Essential elements of questionnaire design and development. J Clin Nurs. 2007;16(2):234-243.CrossRefGoogle ScholarPubMed
Abozaid, H, Park, YS, Tekian, A. Peer review improves psychometric characteristics of multiple choice questions. Med Teach. 2017;39(supp 1):S50-S54.CrossRefGoogle ScholarPubMed
Figure 0

Table 1. Initial validators percentage agreement domain-wise rater scale

Figure 1

Table 2. Item-wise 21 item I-CVI, SCVI-average, and SCVI /UA

Figure 2

Table 3. Pearson correlation coefficient for the questionnaire