Hostname: page-component-cd9895bd7-dk4vv Total loading time: 0 Render date: 2024-12-18T21:15:42.978Z Has data issue: false hasContentIssue false

Revalidation of the Mentoring Competency Assessment to evaluate skills of research mentors: The MCA-21

Published online by Cambridge University Press:  01 April 2022

So Hee Hyun*
Affiliation:
Institute for Clinical and Translational Research, University of Wisconsin–Madison, Madison, Wisconsin, USA
Jenna G. Rogers
Affiliation:
Wisconsin Center for Education Research, University of Wisconsin–Madison, Madison, Wisconsin, USA
Stephanie C. House
Affiliation:
Institute for Clinical and Translational Research, University of Wisconsin–Madison, Madison, Wisconsin, USA
Christine A. Sorkness
Affiliation:
Institute for Clinical and Translational Research, University of Wisconsin–Madison, Madison, Wisconsin, USA
Christine Pfund
Affiliation:
Institute for Clinical and Translational Research, University of Wisconsin–Madison, Madison, Wisconsin, USA Wisconsin Center for Education Research, University of Wisconsin–Madison, Madison, Wisconsin, USA
*
Address for correspondence: S. H. Hyun, PhD, Institute for Clinical and Translational Research, University of Wisconsin–Madison, 750 Highland Ave, Suite 2112, Madison, WI 53705, USA. Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Introduction:

The Mentoring Competency Assessment (MCA) is an example of a validated instrument for measuring mentor skills for postsecondary Science, Technology, Engineering, Mathematics, and Medicine research. The purpose of this study was to revalidate the MCA scale using a larger, more diverse population since the original MCA was validated on a small sample of predominantly senior white male faculty.

Methods:

The MCA was completed by 1626 mentors from a survey data set of 1759 respondents who participated in eight or more hours of face-to-face Entering Mentoring-based training between 2010 and 2019. We conducted principal component analysis (PCA) with varimax rotation to investigate the internal structure of the MCA and Hatcher’s criteria were applied. After a team of mentoring experts independently interpreted the PCA results and reached a consensus on the interpretations of the components, factor analysis and internal consistency reliability analysis were applied to assess the construct validity and the reliability.

Results:

While the 26-item MCA instrument was originally validated with six subscales, through the factor and reliability analyses, all the parameter estimates for each item of seven components of 24-item MCA were significant and had relatively high internal consistency; the alpha coefficient for the components ranged from 0.77 to 0.86.

Conclusions:

Five items from the MCA have been dropped, leaving a condensed 21 item scale (MCA-21) which loads onto six competencies, and should now be used to effectively measure mentoring skills. We provide recommendations for furthering the scale development and validation of common measures.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
© The Author(s), 2022. Published by Cambridge University Press on behalf of The Association for Clinical and Translational Science

Introduction

Evaluating mentoring effectiveness requires psychometrically sound and reliable measures. Such measures can be used to elucidate the ways in which specific factors, such as mentor knowledge and skills, contribute to the quality and impact of mentoring relationships. These measures can also provide information to mentors to optimize their mentoring practices in ways that contribute most significantly to positive outcomes. Unfortunately, review of the mentorship assessment literature indicates a shortage of measures [1]. A meta-review of mentoring assessment tools in internal medicine between 1990 and 2019 revealed that many “prevailing assessments of mentoring … fail to contend with mentoring’s longitudinal, competency-based, evolving, adapting, entwined, goal-sensitive, context-specific, mentor-, mentee-, mentoring relationship and host organization-dependent nature” [Reference Ng, Koh and Yap2].

The Mentoring Competency Assessment (MCA) is an example of a validated measure of mentor skills, developed specifically for postsecondary Science, Technology, Engineering, Mathematics, and Medicine research contexts [Reference Fleming, House and Hanson3]. The MCA was originally developed by a cross-institutional working group to serve as the primary outcome of a multicenter randomized controlled trial (RCT) to assess the effectiveness of the Clinical and Translational Research version of the Entering Mentoring curricula [Reference Pfund, House and Asquith4]. The group proposed 26 items to align with the learning objectives of the six curriculum mentoring competencies : (a) Maintaining Effective Communication (six items), (b) Aligning Expectations (five items), (c) Assessing Understanding (three items), (d) Fostering Independence (five items), (e) Addressing Diversity (two items), and (f) Promoting Professional Development (five items).

The initial validation of the MCA instrument was based on the baseline trial data collected from 283 mentor–mentee pairs across 16 academic health centers, 15 of which were connected to their Clinical and Translational Science Awards (CTSA). The mentors were directed to assess their own skills with all of their mentees; their paired mentees were directed to assess the skills of their mentor in the study. The MCA data were collected via interviews conducted by trained research assistants at each site as part of an expanded protocol. This trial enrolled a relatively homogeneous group of participants in terms of age, career stage, and race. The majority of the mentors were tenured faculty (88%), male (60%), and white (91%); the “average” mentor was a 51-year-old white male full professor in the health sciences with 15 years of mentoring experience. While their mentees were primarily female (58%) and more racially/ethnically diverse, the majority were white (74%) [Reference Fleming, House and Hanson3,Reference Pfund, House and Asquith4]. The analysis identified six subscales that were aligned with the six mentoring competencies in the curriculum [Reference Fleming, House and Hanson3].

In the Entering Mentoring-based mentor training RCT, mentors randomized to the training intervention group showed significant gain compared to the control group in their MCA composite score, as well as in all six subscores. Mentees of the trained mentors reported higher MCA scores for these mentors compared to untrained [Reference Pfund, House and Asquith4]. Since the original publication, the MCA has been widely used as an assessment of mentor training interventions, largely in clinical fields, as well as to assess mentoring programs more broadly [Reference Lau, Ford, Van Lieshout, Saperson, McConnell and McCabe5-Reference Padek, Mir and Jacob7]. The MCA has also been adapted for other contexts and uses, such as a means of assessing the importance of each competency to mentees [Reference Padek, Mir and Jacob7-Reference Harker, O’Toole, Keshmiripour, McIntosh and Sassen10]. Researchers have used a version of the MCA as a needs assessment tool and as a mentor self-assessment tool [Reference Orsini, Daroub, McAuslane and Stedman11]. For example, a study of 135 faculty at an academic health sciences center reported that higher MCA self-assessment ratings were more frequently associated among faculty members with greater academic rank when compared to mentoring experience and clinical/non clinical experience [Reference Mickel, Wiskur, James, VanWagoner and Williams12].

The purpose of this analysis is to revalidate the MCA scale using a larger, more diverse population beyond senior faculty mentors in academic medical settings. This study revalidates the MCA using a much larger (N = 1626) and more diverse (in terms of participants, institutions, and facilitators) sample. In addition to a different sample, the mentor training curricula to which the MCA was aligned has been refined and expanded beyond the original version utilized in the RCT. This analysis was intended to assess whether the MCA works on a more generalizable population as a measure of mentorship skills.

Method

Respondents and Procedure

Research mentor training survey data were collected nationally from 2010 to 2019. This data set included 1759 respondents who participated in an Entering Mentoring-based in-person training for 8 h or more to improve their skills as research mentors. The MCA was completed by 1626 mentors after 166 research mentor training events hosted by 54 institutions/organizations. Table 1 shows sample characteristics of the respondents on which the psychometric properties of the measure were tested in this study. The mean age of the respondents was 41.2 years. In the gender distribution, 664 men participated as mentors (41.7%) while 882 women participated (55.4%). The majority of the ethnicity group was non-Hispanic/Latino (n = 1152; 84.2%) and 1056 respondents (67.2%) self-identified as white. Most respondents (n = 889; 50.5%) were faculty. The mean years of experience as a formal research mentor was 5.7 years, and only 297 respondents (20.1%) had participated in a prior mentor training workshop.

Table 1. Characteristics of respondents

Instrument

The MCA instrument consists of 26 items on a Likert-type scale (1 = Not at all skilled, 4 = moderately skilled, 7 = Extremely skilled). The 26-item MCA served as the primary outcome for the trainings and asked mentors to rate a retrospective pre score and post score for their own skills in mentoring. The post scores were analyzed for the revalidation of the MCA. The MCA was developed and validated by Fleming et al. [Reference Fleming, House and Hanson3] using the baseline data collected in 2010. Six subscales were originally identified: (a) Maintaining Effective Communication (six items), (b) Aligning Expectations (five items), (c) Assessing Understanding (three items), (d) Fostering Independence (five items), (e) Addressing Diversity (two items), and (f) Promoting Professional Development (five items) [Reference Fleming, House and Hanson3]. The draft instrument was reviewed by the UW Survey Research Center and tested using both cognitive and standard pilot interviews before being finalized for use in RCT [Reference Fleming, House and Hanson3].

Analysis

To investigate the internal structure of the MCA, principal components analysis (PCA) with varimax rotation was applied. PCA is a multivariate technique typically used to reduce a large set of variables to a small set, while containing as much of the information (variation) as possible. Several criteria were applied to determine how many components should be retained to explain most items of the MCA using Hatcher’s criteria [Reference Hatcher and O’Rourke13]:

  1. 1. The point of inflexion displayed by the scree plot.

  2. 2. The eigenvalues criterion. The eigenvalue-one criterion was considered in conjunction with other criteria (e.g., scree plot and the proportion of variance accounted for) [Reference Kaiser14]. This analysis used a cutoff on the eigenvalues of 0.65 when deciding how many components to retain and interpret.

  3. 3. The “proportion of variation accounted for” criterion. If the designated number of components do not account for at least 50% of the variance, then the analysis is aborted [Reference Bonizzoni, Gussoni and Agnelli15].

  4. 4. A given component contains variables with significant loadings, a loading of 0.30 being used as the cutoff point.

After determining the number of meaningful components to retain for the MCA, the team of mentoring experts independently interpreted the factors for all relevant results, discussed their interpretations collectively, and reached complete agreement on the interpretations of the factors and their alignment. Finally, confirmatory factor analysis (CFA) and internal consistency reliability analysis were performed to measure the construct validity and the reliability of the MCA with the components which were retained from the PCA. CFA was used to examine the fit of the measurement model and verify the factor structure of a measurement instrument. The use of CFA to conduct the construct validity of hypothesis-based testing instruments can add a level of statistical accuracy and develop condensed forms of an instrument or confirmation of its possible subdomains [Reference McArdle16]. The data were analyzed using Stata SE 16.0 for Mac (Stata Corp, College Station, TX, USA).

Results

Principal Component Analysis

The scree plot showed a sharp point of inflection after the first component (Fig. 1). Only three components had initial eigenvalues > 1, with values ranging from 1.16 to 12.20. Since the 26-item MCA was originally validated with six subscales, this study applied a cutoff on the eigenvalues of 0.65 to determine how many components should be retained to explain most items of the MCA. Considering the eigenvalue and the “proportion of variance accounted for” criterion, the first eight components were taken as the starting point for the analysis that explained most items of the MCA with 72% of variation.

Fig. 1. Scree plot of the eigenvalues of the factors. pca = principal component analysis.

Table 2 shows significant component loadings of the eight components with varimax rotation; 24 of the total 26 items were loaded into components. First, in measuring the MCA scale of maintaining effective communication, four of the six items were significantly loaded into one component, (1) active listening, (2) providing constructive feedback, (3) developing a trusting relationship, and (4) accommodating communication style. The item of (5) pursuing strategies to improve communication was not loaded into any components and the item of (6) coordinating with other mentors had high uniqueness which means that the item was loaded itself as a single component.

Table 2. Component loadings of the 8 components with varimax rotation (Blank if a loading < 0.30)

Second, items of aligning expectation scale and items of assessing understanding scale were mixed and loaded into three different components. The four items, (7) setting clear relationship expectations, (8) aligning expectations, (13) estimating mentee ability, and (14) enhancing mentee skills, were loaded into one component. Three items, (10) setting research goals, (11) developing strategies to meet goals, and (12) assessing mentee knowledge, were loaded into another component. The item of (9) considering differences that may impact expectations was loaded into the other component with two items of the addressing diversity scale, (20) accounting for biases and prejudices and (21) accounting for different backgrounds of mentors and mentees.

Third, the five items of the fostering independence scale were split up into two different components. The two items, (15) motivating mentees and (16) building confidence, were loaded into one component. Three items, (17) simulating creativity, (18) acknowledging mentees’ professional contributions, and (19) negotiating a path to independence, were loaded into the other component.

Lastly, four of the five items in the promoting professional development scale, (22) helping network effectively, (23) setting career goals, (24) helping establish a work/life balance, and (26) helping mentees acquire resources, were loaded into one component, while the item of (25) understanding impact as role model was not loaded into any components.

Confirmatory Factor Analysis and Reliability Analysis

Based on the results of the PCA, CFA, and Cronbach’s alpha analysis were performed to measure the construct validity and reliability of the MCA with eight components. In the CFA, we applied maximum likelihood estimation and principal components to assess how well the 24 items measured the eight components and applied the four goodness of fit statistics: chi-square, root mean square of approximation (RMSEA), comparative fit index (CFI), Tucker–Lewis index (TLI), and standardized root mean squared error (SRMR).

Two items, (5) employing strategies to improve communication and (25) understanding impact as role model, were excluded from the factor analysis and Cronbach’s alpha analysis, since these items were not significantly loaded into any components in the PCA. The single item which had high uniqueness, (6) coordinating with other mentors, was not included in the factor analysis and Cronbach’s alpha analysis since one item cannot measure the validity or reliability of a scale, though the item may have practical value.

Table 3 shows standardized factor loadings and Cronbach’s alpha scores for the seven components of 24-item MCA. At a minimum, the following indices should be reported and measured in combination: chi-square; RMSEA; CFI; TLI; and SRMR [Reference Kline17]. A seven-component structure was validated (χ 2 = 2085.970, p < 0.001, RMSEA = 0.079, CFI = 0.905, TLI = 0.886, SRMR = 0.042) and the hypothesized model of the seven components resulted in an acceptable fit to the data: RMSEA < 0.08; CFI > 0.80; TLI > 0.80; and SRMR < 0.05. Considering the sensitivity of the chi-square statistic to sample size, overall goodness-of-fit indices were applied to measure model adequacy. All the parameter estimates for each item were significant, with standardized factor loadings ranging from 0.47 to 1.00. The alpha coefficient for the seven components is from 0.77 to 0.86, suggesting that the items have relatively high internal consistency.

Table 3. Factor loadings and Cronbach’s alpha scores for the 24-item Mentoring Competency Assessment

Discussion

The results of this analysis demonstrate that the structure of the MCA has changed and needs recrafting. The first eight principal components were included to revalidate the MCA scale and identified which items were most strongly correlated with each component. Table 3 demonstrates the new structure of the MCA. Component 1 (Aligning Expectations) was strongly correlated with four items, (7) setting clear relationship expectations, (8) aligning expectations, (13) estimating mentee ability, and (14) enhancing mentee skills. Component 4 (Assessing Understanding) had larger positive associations with three items, (10) setting research goals, (11) developing strategies to meet goals, and (12) assessing mentee knowledge. Interestingly, the items from these two competencies were varied from the original validation structure (some aligning expectation items loading onto assessing understanding here and vice versa). In retrospect, this realignment is not surprising. In order to establish and align appropriate expectations with mentees, it is essential to have a clear understanding of their abilities so that you can enhance their skills. Likewise, in order to establish an appropriate research development plan, you should have a clear understanding of a mentee’s comprehension of key concepts.

Component 2 (Promoting Professional Development) had large positive associations with four items, (22) helping network effectively, (23) setting career goals, (24) helping establish a work/life balance, and (26) helping mentees acquire resources. These items remained the same as the original validation with the exception of the dropped item 25 on this competency (discussed below). Component 3 also held to its original structure, with four items (1) active listening, (2) providing constructive feedback, (3) developing a trusting relationship, and (4) accommodating communication style being strongly associated, and two items (5 and 6) being dropped (further discussed below).

Component 6 (Addressing Diversity) was strongly associated with three items, (9) considering differences may impact expectations, (20) accounting for biases and prejudices, and (21) accounting for different backgrounds of mentors and mentees. While item 9 was originally under aligning expectations, this item now loads under addressing diversity. This is a case where the curriculum content corresponds to a different competency. The authors had always intended to address diversity throughout the curriculum, which is why considering how personal and professional differences may influence expectations is one of the objectives of the Aligning Expectations session. The concept underlying this item does clearly align with addressing diversity.

Component 7 (Fostering Independence) had large positive associations with three items, (17) stimulating creativity, (18) acknowledging mentees’ professional contributions, and (19) negotiating a path to independence. This structure remains the same as the original analyses, though without items (15) motivating mentees and (16) building confidence, as these items strongly correlated with component 5 (Mentee Self-Efficacy). This component was not one of the competencies addressed in the curriculum when the original analysis was conducted, though Promoting Mentee Research Self-Efficacy has since been added as a core mentoring competency. Further, we now understand motivation and confidence to be central to mentee self-efficacy.

Component 8 (Navigating Mentoring Networks) is only correlated with one item, (6) coordinating with other mentors and was not one of the original competencies. This item was originally created to measure the dynamics within mentoring networks as at the time it was becoming more common to work in mentoring teams, and the mentees participating in the trial were in fact required to have multiple mentors as part of their career development awards. We anticipate that comentoring and mentor teams will become more common models and expanded in future curriculum modules.

This analysis also revealed that two items, (5) pursuing strategies to improve communication, and (25) understanding impact as a role model, did not load on any components with no associations with other items. We concluded that these measures are not capturing training content, or rather, not measuring something that was a part of the training. In reflection, curriculum developers agreed that these items are not directly covered in training. Though participants often report gains in these areas, it is likely a result of developing complementary skills. Further, mentors are asked to reflect on the impact of their own practices, which indirectly touches on modeling.

Future Directions and Recommendations

Based on this analysis to revalidate the MCA, we recommend dropping items 5, 6, 15, 16, and 25 (pursuing strategies to improve communication, coordinating with other mentors, motivating mentees, building confidence, and understanding impact as a role model). A condensed 21 items scale (now referred to as MCA-21) that loads onto 6 competencies should be used; it effectively measures mentoring skills in the competencies of maintaining effective communication, aligning expectations, fostering independence, addressing diversity, promoting professional development, and assessing understanding. Supplementary Table 1 shows the MCA-21.

An important note about the MCA-21 is that while the items have the majority of their variance loading onto the unit-scaled components, some of the variance can also be explained by other factors, thus demonstrating that there are cross-competency gains. This is a consideration for mentor training evaluation as some workshops may not include the full set of modules covering all targeted competencies, but rather just a few. Based upon our findings here, we recommend that the full revised MCA-21 can be used to capture skill gains despite the length or number of modules covered in the training.

The analysis suggests areas for future scale development. For example, our results suggest a new competency focused on helping mentees navigate mentoring networks. In the existing MCA, this competency is only assessed by one item; “coordinating with other mentors.” Another domain of interest is that of role modeling. These domains can be explored more fully by testing additional items built from existing network measures. To accomplish this work, we plan to use a similar approach to measures development and validation around the mentoring competencies of research self-efficacy and providing motivation [Reference Butz, Byars-Winston, Leverett, Branchaw and Pfund18,Reference Lewis, Martina and McDermott19].

Finally, the development and validation of measures such as the MCA-21 are important as we move toward the use of common measures across programs. Common measures allow the broader community to identify and examine the factors that matter in workforce development, such as mentorship, across large, diverse groups of mentors and mentees. Such efforts are being pursued by programs such as the National Research Mentoring Network and the CTSAs [Reference Welch, Tomoaia-Cotisel and Noubary20]. A national CTSA T32/TL1 survey has reported mentorship practices of 50 active CTSA hubs [Reference Sancheznieto, Sorkness and Attia21]. New NCATS Funding Opportunity Announcements have replaced the TL1 funding mechanism with three optional CTSA award components, including an NRSA predoctoral training grant (T32), an NRSA postdoctoral training grant (T32), and a Research Education Grants Program (R25). To be responsive to both training program faculty requirements and scored review criteria, CTSA programs must strengthen mentorship through the use of evidence-informed mentoring practices, evidence-informed mentor training, assessment of mentoring skills and behaviors, and monitoring of mentorship behaviors [Reference Sancheznieto, Sorkness and Attia21]. The common measure is being used widely and the MCA-21 described here could serve as an updated, validated, common measure.

Supplementary Material

To view supplementary material for this article, please visit https://doi.org/10.1017/cts.2022.381.

Acknowledgments

We would also like to thank all those who helped us implement the trainings assessed as part of this study and assisted with the data collection.

The research reported in this publication was supported by the National Institute of General Medical Sciences of the NIH under award number U54GM119023. The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH. Additional support was provided by the Institute for Clinical and Translational Research supported by the UW-Madison Clinical and Translational Science Award (CTSA) program, the National Center for Advancing Translational Sciences (NCATS), grant 1UL1TR002373.

Disclosures

There are no conflicts.

Ethical Approval

Ethical approval has been granted by the University of Wisconsin Institutional Review Board, protocol numbers 2016-0458, 2017-0026, 2017-1336.

References

National Academies of Sciences Engineering and Medicine. The Science of Effective Mentorship in STEMM. Washington DC: The National Academies Press, 2019.Google Scholar
Ng, YX, Koh, ZYK, Yap, HW, et al. Assessing mentoring: a scoping review of mentoring assessment tools in internal medicine between 1990 and 2019. PLOS ONE 2020; 15(5): e0232511. DOI 10.1371/journal.pone.0232511.CrossRefGoogle ScholarPubMed
Fleming, M, House, S, Hanson, VS, et al. The Mentoring Competency Assessment: validation of a new instrument to evaluate skills of research mentors. Academic Medicine 2013; 88(7): 10021008. DOI 10.1097/ACM.0b013e318295e298.CrossRefGoogle ScholarPubMed
Pfund, C, House, SC, Asquith, P, et al. Training mentors of Clinical and Translational Research scholars: a randomized controlled trial. Academic Medicine 2014; 89(5): 774782. DOI 10.1097/ACM.0000000000000218.CrossRefGoogle ScholarPubMed
Lau, C, Ford, J, Van Lieshout, RJ, Saperson, K, McConnell, M, McCabe, R. Developing mentoring competency: does a one session training workshop have impact? Academic Psychiatry 2016; 40(3): 429433. DOI 10.1007/s40596-016-0537-8.CrossRefGoogle ScholarPubMed
Johnson, MO, Gandhi, M. A mentor training program improves mentoring competency for researchers working with early-career investigators from underrepresented backgrounds. Advances in Health Sciences Education 2015; 20(3): 683689. DOI 10.1007/s10459-014-9555-z.CrossRefGoogle ScholarPubMed
Padek, M, Mir, N, Jacob, RR, et al. Training scholars in dissemination and implementation research for cancer prevention and control: a mentored approach. Implementation Science 2018; 13(1): 18. DOI 10.1186/s13012-018-0711-3.CrossRefGoogle ScholarPubMed
Stiltner, S, Kutz, M. The “Mentorship Competency Assessment of Research Mentors” applied to perceived effectiveness of athletic training clinical preceptors. Journal of Sports Medicine and Allied Health Sciences: Official Journal of the Ohio Athletic Trainers’ Association 2018; 4(1): 12.Google Scholar
Wiskur, B, Mickel, N, Williams, VN. Mentee Assessment of Mentoring Competencies at an Academic Health Sciences Center. Journal of Faculty Development 2020; 34(2): 3343.Google ScholarPubMed
Harker, K, O’Toole, E, Keshmiripour, S, McIntosh, M, Sassen, C. Mixed-methods assessment of a mentoring program. Journal of Library Administration 2019; 59(8): 873902. DOI 10.1080/01930826.2019.1661745.CrossRefGoogle Scholar
Orsini, J, Daroub, S, McAuslane, H, Stedman, N. Graduate student and faculty perceptions of mentoring competency. NACTA Journal 2019; 64(2): 6678.Google Scholar
Mickel, N, Wiskur, B, James, J, VanWagoner, T, Williams, VN. Assessing faculty capacity to build an effective mentoring network at an academic health sciences center. The Journal of Faculty Development 2018; 32(3): 3546.Google Scholar
Hatcher, L, O’Rourke, N. A Step-by-Step Approach to Using SAS for Factor Analysis and Structural Equation Modeling. SAS Institute, 2013.Google Scholar
Kaiser, HF. The application of electronic computers to factor analysis. Educational and Psychological Measurement 1960; 20(1): 141151. DOI 10.1177/001316446002000116.CrossRefGoogle Scholar
Bonizzoni, E, Gussoni, G, Agnelli, G, et al. The complexity of patients hospitalized in Internal Medicine wards evaluated by FADOI-COMPLIMED score(s). A hypothetical approach. PLOS ONE 2018; 13(4): e0195805. DOI 10.1371/journal.pone.0195805.CrossRefGoogle ScholarPubMed
McArdle, JJ. Current directions in structural factor analysis. Current Directions in Psychological Science 1996; 5(1): 1118.CrossRefGoogle Scholar
Kline, T. Psychological Testing: A Practical Approach to Design and Evaluation. Sage, 2005.CrossRefGoogle Scholar
Butz, A, Byars-Winston, A, Leverett, P, Branchaw, J, Pfund, C. Promoting STEM trainee research self-efficacy: A mentor training intervention. Understanding Interventions 2018; 9(1): 3730.Google Scholar
Lewis, V, Martina, CA, McDermott, MP, et al. A Randomized Controlled Trial of Mentoring Interventions for Underrepresented Minorities. Academic Medicine, 2015, http://europepmc.org/abstract/med/26717501.Google Scholar
Welch, LC, Tomoaia-Cotisel, A, Noubary, F, et al. Evaluation of initial progress to implement Common Metrics across the NIH Clinical and Translational Science Awards (CTSA) Consortium. Journal of Clinical and Translational Science 2021; 5(1): 704. DOI 10.1017/cts.2020.517.Google Scholar
Sancheznieto, F, Sorkness, CA, Attia, J, et al. Clinical and translational science award T32/TL1 training programs: program goals and mentorship practices. Journal of Clinical and Translational Science 2022; 6(1): e13. DOI 10.1017/cts.2021.884.CrossRefGoogle ScholarPubMed
Figure 0

Table 1. Characteristics of respondents

Figure 1

Fig. 1. Scree plot of the eigenvalues of the factors. pca = principal component analysis.

Figure 2

Table 2. Component loadings of the 8 components with varimax rotation (Blank if a loading < 0.30)

Figure 3

Table 3. Factor loadings and Cronbach’s alpha scores for the 24-item Mentoring Competency Assessment

Supplementary material: File

Hyun et al. supplementary material

Hyun et al. supplementary material

Download Hyun et al. supplementary material(File)
File 35.7 KB