Hostname: page-component-78c5997874-t5tsf Total loading time: 0 Render date: 2024-11-07T23:31:26.928Z Has data issue: false hasContentIssue false

How to make study documents clear and relevant: the impact of patient involvement

Published online by Cambridge University Press:  04 November 2021

Sagar Jilka*
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK and Division of Mental Health & Wellbeing, Warwick Medical School, University of Warwick, Coventry, UK
Georgie Hudson
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK
Sonja M. Jansli
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK
Esther Negbenose
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK; and Department of Psychology, South London and Maudsley NHS Foundation Trust, UK
Emma Wilson
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK; and Department of Psychology, South London and Maudsley NHS Foundation Trust, UK
Clarissa M. Odoi
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK; and Department of Psychology, South London and Maudsley NHS Foundation Trust, UK
Magano Mutepua
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK; and Department of Psychology, South London and Maudsley NHS Foundation Trust, UK
Til Wykes
Affiliation:
Institute of Psychiatry, Psychology & Neuroscience, King's College London, UK; and Department of Psychology, South London and Maudsley NHS Foundation Trust, UK
*
Correspondence: Sagar Jilka. Email: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Background

Patient and public involvement can improve study outcomes, but little data have been collected on why this might be. We investigated the impact of the Feasibility and Support to Timely Recruitment for Research (FAST-R) service, made up of trained patients and carers who review research documents at the beginning of the research pipeline.

Aims

To investigate the impact of the FAST-R service, and to provide researchers with guidelines to improve study documents.

Method

A mixed-methods design assessing changes and suggestions in documents submitted to the FAST-R service from 2011 to 2020. Quantitative measures were readability, word count, jargon words before and after review, the effects over time and if changes were implemented. We also asked eight reviewers to blindly select a pre- or post-review participant information sheet as their preferred version. Reviewers’ comments were analysed qualitatively via thematic analysis.

Results

After review, documents were longer and contained less jargon, but did not improve readability. Jargon and the number of suggested changes increased over time. Participant information sheets had the most suggested changes. Reviewers wanted clarity, better presentation and felt that documents lacked key information such as remuneration, risks involved and data management. Six out of eight reviewers preferred the post-review participant information sheet. FAST-R reviewers provided jargon words and phrases with alternatives for researchers to use.

Conclusions

Longer documents are acceptable if they are clear, with jargon explained or substituted. The highlighted barriers to true informed consent are not decreasing, although this study has suggestions for improving research document accessibility.

Type
Papers
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2021. Published by Cambridge University Press on behalf of the Royal College of Psychiatrists

It is important for researchers to provide clear participant information sheets to have true informed consent,Reference Ennis and Wykes1 and one way to achieve this is through patient and public involvement. Patient and public involvement is required for UK publicly funded health research, with researchers describing their involvement strategy.Reference Staley, Ashcroft, Doughty and Szmukler2,Reference Evans and Papoulias3 Patient and public involvement is defined as ‘research carried out with or by members of the public rather than to, about or for them’,4 where the public, patients and carers are active partners in research.Reference Barnes and Cotterell5 Traditional research models often confine patients to the end of the research ‘pipeline’, but ensuring collaboration across all stages can improve study outcomes, identify appropriate research questions and reduce ‘research waste’.Reference Trivedi and Wykes6Reference Ennis and Wykes9 The Feasibility and Support to Timely Recruitment for Research (FAST-R; https://www.maudsleybrc.nihr.ac.uk/patients-public/support-for-researchers/) service offers access to trained mental health patients and carers for improving participant facing documents. This service was set up by the Mental Health Research Network in London in 2011, and is now organised and funded by the National Institute for Health Research (NIHR) Maudsley Biomedical Research Centre. FAST-R is used in the early stages of the research cycle, before seeking funds or ethical approval. The service can speed up these processes by highlighting potential stumbling blocks; for example, by ensuring clear and accessible language in participant-facing documents and including information needed for genuinely informed consent.Reference Iliffe, McGrath and Mitchell10 Documents are considered by a group of reviewers with a facilitator, and are returned within 7 working days. Patient involvement has been contested despite the supports for co-production,Reference Faulkner, Carr, Gould, Khisa, Hafford-Letchfield and Cohen11,Reference Redman, Greenhalgh, Adedokun, Staniszewska and Denegri12 so it is essential to understand whether there is an impact. Several potential measures are unexplored that affect whether someone has informed consent, such as readability and jargon. The US Food and Drug Administration (FDA) recommend a readability grade of eight or lower (i.e. readable by someone aged 13 years).13 These quantitative measures do not tell the whole story, as valuable comments on the ways patient and public involvement benefits a study can be missed.Reference Rolfe, Ramsden, Banner and Graham14,Reference Crocker, Boylan, Bostock and Locock15 Therefore, in addition to quantitative measures, we explored FAST-R members’ views in detail, using qualitative methods.

Method

Design

This is a mixed-methods study auditing documents submitted to FAST-R before and after review, and investigating their accessibility (reading grade, jargon, length, etc.) by using quantitative analyses. All study documents were related to mental health research and consisted of a variety of study types and fields, such as clinical trials, cross-sectional investigations and qualitative studies (e.g. Martland et al,Reference Martland, Gaughran, Stubbs and Onwumere16 Greer at alReference Greer, Newbery, Cella and Wykes17). This was supplemented by a detailed qualitative analysis of patient/carer FAST-R member reviews found on the feedback forms and on the documents themselves. The authors selected the participant information sheets with at least 75% of changes implemented by the researcher. These eight studies were reviewed by eight FAST-R reviewers who each considered one study they had not seen before, with no information on which document was pre- or post-review. Each reviewer gave their preferred document and provided reasons for their choice. We extracted jargon words and phrases identified by the reviewers from the pre-review documents as a guide to how changes might be made in the future. This is a secondary data analysis, so no ethical approval was required because no participants were recruited and no reviewers were identifiable. The research was conducted in accordance with the Helsinki Declaration of 1975, as revised in 2008. See Fig. 1 for a visual representation of the study's methodology.

Fig. 1 The stages of the study's methodology, including the qualitative and quantitative analyses. FAST-R, Feasibility and Support to Timely Recruitment for Research.

FAST-R members

Members of the FAST-R service are recruited through a variety of sources, such as local mental health charities, referrals and The Maudsley Biomedical Research Centre's Consent for Contact programme.Reference Perera, Broadbent, Callard, Chang, Downs and Dutta18 All members are interviewed prior to joining the group and given information about the group and their role. Inclusion criteria for membership consists of lived experience of mental health service use and/or being a carer of a mental health patient, as well as some prior experience taking part in research studies. Members are paid and group demographics range in age, ethnicity and socioeconomic status; however, the majority of the reviewers are White British, middle-aged and university-educated.

Part one: quantitative assessment

Procedure

Data consisted of original and final versions of study documents submitted between July 2011 and July 2020.

Measures

We used the Flesch–Kincaid Grade Level measure,Reference Flesch19 which has been used extensively to assess healthcare information.Reference Ennis and Wykes1,Reference Grohol, Slimowicz and Granda20Reference Cochrane, Gregory and Wilson22 The readability score was extracted from Microsoft Word (version 2109 for Windows) before and after review. Higher scores represent more complicated text. Word count and the number of jargon words were calculated where words and phrases were classed as ‘jargon’ if reviewers said they were jargon, needed explaining or simplifying. We calculated changes by subtracting post from pre FAST-R scores so a negative reading grade means they became easier. We also counted the number of suggested changes and the number actually implemented in the final document.

We categorised the documents by type (e.g. participant information sheet, consent form) and investigated whether there were any changes over time.

Analysis

We characterised the documents using descriptive statistics. Readability scores, word count and jargon were subjected to paired samples t-tests to measure changes after FAST-R review. We investigated the relationship between reading grade, word count, jargon, the number of suggested changes and the percentage of changes implemented, with the year documents were submitted for review, using Pearson correlations to detect changes over time. We also report the differences between document types requiring and implementing changes. All analyses were carried out with SPSS version 26 for Windows.

Part two: qualitative assessment

Procedure

Tracked comments made on the original research documents, as well as comments for researchers on the FAST-R feedback forms, were the basis of all analyses. We selected participant information sheets with at least 75% of changes implemented by the researchers from our total sample. This consisted of eight studies, which we gave to eight FAST-R reviewers who reviewed one study each and selected their preference with reasons for their choice.

Analysis

Documents were analysed thematically,Reference Braun and Clarke23 and themes were inductively extracted by two researchers independently, using Pope et al's analysis framework.Reference Pope, Ziebland and Mays24 This is a five-stage process and involves: stage 1 involves familiarisation with raw data; stage 2 involves identifying a thematic framework; stage 3 involves indexing by applying the thematic framework to all the data by annotating the transcripts; stage 4 involves charting by rearranging the data according to the thematic framework; and stage 5 involves mapping and interpretation by defining concepts, mapping the range and nature of phenomena, and creating typologies.

Stages 1 and 2 were undertaken as a group through discussion to create a coding framework (‘framework category’). To minimise bias and maximise our inductive approach from an emic perspective,Reference Pike and Pike25 two patient researchers independently undertook inductive coding,Reference Thomas26 and categorised the codes into emerging themes and subthemes. Mapping and interpretation were repeated with a third researcher, a FAST-R facilitator, who reviewed the coding framework and made the final judgement on emerging themes and subthemes. Analysis was carried out with NVivo version 12 for Windows (QSR International, Melbourne, Australia; see https://www.qsrinternational.com/nvivo-qualitative-data-analysis-software/home).

For reviewers’ blind ratings, we report how many reviewers preferred the post-review participant information sheet, and summarise their feedback on the information sheets.

Results

Are there measurable differences in the documents after FAST-R review?

Over the 9 years of data collection, 99 documents had both the pre- and post-review. The most frequent were participant information sheets (n = 49) and consent forms (n = 29), but there was a range of document types (see Supplementary Table 1 available at https://doi.org/10.1192/bjo.2021.1040).

Accessibility

Table 1 summarises the outcomes before and after the FAST-R review.

Table 1 Reading grade, word count and jargon for all documents before and after FAST-R review

FAST-R, Feasibility and Support to Timely Recruitment for Research.

The final documents showed a significant reduction in the number of jargon words (before mean: 1.68 ± 8.23; after mean: 0.98 ± 5.44; t(98) = 2.37; P = 0.020; 95% CI 0.11–2.37), but documents became longer by 154 words (eight lines, or the equivalent of one paragraph) on average (before mean: 1199.4 ± 1116.6; after mean: 1353.3 ± 1336.4; t(98) = −4.64; P < 0.001; 95% CI −219.70 to −88.11). As patient-facing documents are important for consent, we separated them and found that participant information sheets became 12% longer after review, with an average increase of 275.41 words, and consent forms became 11% longer, but only by an average increase of 38 words.

Although there was no significant overall change in reading grade of the documents, 29.3% (n = 29) became easier to read (a negative grade), 12.1% (n = 12) did not change and 58.6% (n = 58) became more difficult to read. For participant information sheets, 20% (10 out of 49) improved, and for consent forms, 38% improved.

Most documents did not meet the recommended reading grade before FAST-R review (85%, n = 84), but six of these did meet the recommended reading grade after review (two consent forms, three participant information sheets and one lay summary). Of the 15 documents that initially met the recommended reading grade, three no longer met this criterion after review (two consent forms, one participant information sheet). So despite some improvements there were also detrimental changes.

Did documents improve over time?

There were few changes over time in the documents submitted for review, except that there was an increase in the amount of jargon (r (n = 99) = 0.20, P = 0.046), and, following review, the number of suggested changes also increased (r (n = 99) = 0.31, P = 0.002), suggesting a lack of improvement over time.

Do some documents produce more changes?

Nearly all documents required changes (97 out of 99), with an average of 14 suggested changes to participant information sheets (Table 2), which were one of the most edited documents where researchers implemented more changes. Consent forms had few suggested changes (n = 5), with half implemented, and this may be because the Health Regulations Authority (HRA) provide direct guidance in the form of bullet points for researchers to use (see Fig. 2 for an overview of differences between HRA and FAST-R reviewer advice).

Fig. 2 Comparison of FAST-R patient and carer guidance (right) with guidance from the Health Regulations Authority. FAST-R, Feasibility and Support to Timely Recruitment for Research; GDPR, General Data Protection Regulation; HRA, Health Regulations Authority.

Table 2 The number of suggested changes made by FAST-R reviewers, by document type and the percentage of changes implemented by researchers

FAST-R, Feasibility and Support to Timely Recruitment for Research.

What themes does FAST-R review produce for suggested changes?

With the inclusion of the feedback forms, we had 146 FAST-R documents. The analysis produced six categories: aspects valued by reviewers, issues around clarity, General Data Protection Regulation, language, study design and contact details, and presentation. A summary of the full framework with themes is shown in Table 3.

Table 3 Framework categories, and emerging themes with their subthemes

a. Occurrence refers to the number of documents coded to each theme, and percentage, relative to the framework category.

Aspects valued by reviewers

Reviewers noted when studies were ‘well designed and feasible’ and had real-life relevance (translational, e.g. ‘The reviewers feel that the study is worthwhile. Particularly as antipsychotics can lead to weight gain and other related physical problems’). Reviewers praised well-written documents so participants ‘could understand what they were being asked to do’ and contained substantial patient and public involvement. See quotes in Supplementary Table 2.

Clarity

Issues around a lack of clarity were common (see quotes in Supplementary Table 3), and several themes and subthemes emerged from this category.

Lack of information

Parts of documents lacked sufficient information, including how participants could benefit from the study (e.g. ‘the interaction from the education session and group walks’), compensation for involvement, study practicalities (e.g. ‘Reviewers were not sure what an eye tracker involved and felt that the procedure for this should be specified’), the study purpose (e.g. ‘How (the) research will ultimately benefit the population’), how participants would be kept safe (e.g. ‘what support is available for participants, should they become distressed’), side-effects and risks and how likely these are to occur (‘Reviewers hoped this risk could be quantified e.g. every one person in 100 who undergo this procedure’), and information on study team members.

Inclusion and exclusion criteria

Sometimes information on who could participate was lacking, which made it difficult for potential participants to know if they would be eligible. For example, if they were ‘allowed to be on medication’ or ‘involved in therapy’.

Terminology

Reviewers highlighted unclear terminology, such as jargon words and acronyms, and asked for further explanation on institutions or companies that were mentioned (‘The reviewers thought it would be good to spell out what NIHR stands for’). Issues with giving definitions for subjective issues were also highlighted (e.g. ‘recovery from anorexia’) and they suggested alternative words and phrases to explain jargon, as well as substituting negatively connotated words with more neutral terms (‘The reviewers felt that the terms “harmed” and “injured” could be interpreted negatively and suggested whether “affected” might offer an alternative?’).

Consistency (contradictions)

Reviewers highlighted contradictions and inconsistencies in the use of terminology and contradictory information on study procedures, e.g. its length, reimbursement, recruitment, people involved in the study and audio/video recordings.

General Data Protection Regulation

General Data Protection Regulation issues were central to many comments, and its themes are summarised in Supplementary Table 4. Reviewers particularly noted issues related to consent, such as the need for documents to be concise, as ‘this will make it more likely that participants will actually read it and be able to provide informed consent’, as well as transparency (e.g. ‘It is not mentioned that participants will be invited to attend a focus group or interview, or that a different protocol will be followed to gather information from service users and clinicians, and whether these will be audio recorded’) and how data will be used, storage methods, storage time and what will happen to data when participants withdraw from the study.

Language

Reviewers noted grammatical mistakes, misspelling and that ‘the tense should be consistent throughout’, as well as avoiding repetition (see Supplementary Table 5). Reviewers recommended using more inclusive phrasing (e.g. ‘In the gender section: the transgender option can be included’), as well as alternative phrasing of sentences in the documents that were difficult to read, unclear or could be offensive to the reader (e.g. ‘the term “diagnosis of schizophrenia” should be avoided as, from our experience, we find this can often cause offence, especially to those who do not believe they have this condition’).

Study design and contact details

Reviewers also commented on the study design and gave suggestions on contact details they felt were important to include (for a full breakdown see Supplementary Table 6). Reviewers queried ‘whether participants can ask to see a copy of the findings’, where participants would be recruited from and whether the amount of time allocated to different parts of the study process was realistic. Reviewers also provided suggestions, such as where to recruit participants and which contact details to include for community and crisis services, complaints departments and the research study team.

Presentation

Reviewers thought it was important that documents were aesthetically pleasing, and suggested using images, flowcharts, page numbers, and subheadings, as well as editing the text style and layout (e.g. formatting important sections in bold), breaking down paragraphs into bullet points and using checkboxes for demographic questions (see Supplementary Table 7). Researchers were recommended to ensure that the order of the text makes sense (‘We thought the section “Why are we doing the research” could be moved to below “What is the research about” as these sections complement each other’) and to avoid confusing or irrelevant information (‘We felt it would be more appropriate if the sentence starting “If you would like to take part … ” was removed to eliminate any confusion’).

Which participant information sheets did reviewers prefer?

Most (six out of eight) reviewers preferred the post-FAST-R reviewed participant information sheets, and these documents were the ones that had more changes implemented (average 92.5%).

Comments by reviewers explaining their preference reflected those outlined above, with reviewers noting that the post-FAST-R reviewed documents were more transparent overall and contained the essential information, such as details on the study process, and the risks and benefits of the study. The language and layout of the documents were also frequently mentioned, with the post-FAST-R reviewed participant information sheets being more concise, in terms of ‘layout, wording and in anticipating some of the concerns that the potential participant might have’, as well as containing less jargon, using friendlier language (e.g. replacing the phrase ‘what will happen to me if I take part?’ with ‘what will happen if I take part?’) and being better structured overall. Those who preferred the pre-review information sheets provided similar reasons, stating the pre-reviewed versions were more concise and got to the ‘crucial information quicker’, as well as having a more coherent structure, being more succinct and using a preferable font.

Tips for jargon busting

A list of jargon words/phrases along with alternatives were suggested by reviewers (see Table 4). Reviewers also highlighted other confusing terms, but were unable to provide alternatives (Table 5).

Table 4 Suggested alternative wordings by FAST-R reviewers

FAST-R, Feasibility and Support to Timely Recruitment for Research.

a. Indicates terms that were frequently changed by reviewers.

Table 5 Words and phrases flagged up by reviewers as confusing or needing further explanation

NIHR, National Institute for Health Research; MRC, Medical Research Council, PERT, XXX.

a. Relates to a specific medical study.

Discussion

This was a mixed-methods audit to understand how patient and public involvement (through the FAST-R service) affects the accessibility of research documents, and highlights how to improve them. Although jargon words decreased, the word count increased after FAST-R review, indicating that patients and carers do not mind slightly longer documents if they are clear, and with no jargon. We show that there are many words and phrases that researchers may struggle to put into lay terms and we provide alternative. Researchers may also benefit from resources like the Patient Information Forum (https://pifonline.org.uk/), which provides a glossary of terms. The increased word count and decreased jargon also suggests that researchers did take the reviewers’ advice, and we found that the participant information sheets – a crucial part of informed consent – are among those edited most after review. We also found that all documents were edited after review. Despite these changes, there was no improvement in readability. We identified issues considered important, jargon words and phrases, and provide alternatives to improve accessibility. Importantly researchers did not seem to improve the documents, as over time, there was an increase in the amount of jargon in documents submitted for review. The increased jargon over time may be because HRA or research and development teams are increasingly mandating specific terms, which are inaccessible to lay people. When directly comparing pre-FAST-R and post-FAST-R documents, 75% (six out of eight) of participants preferred the post-FAST-R versions. Comments made by participants mirrored those of the initial qualitative analysis, indicating that researchers should strive to find a balance between providing enough essential information (e.g. on data protection, study process and risks and benefits) and avoid unnecessary information. When researchers implemented suggested changes, those documents were preferred, indicating that there is value in taking FAST-R reviewers’ comments seriously. Taken together, researchers can use this information to make research documents more accessible in the future.

We found more suggested and implemented changes in participant information sheets than consent forms, probably because consent forms may provide less flexibility with wording. Study sponsors may provide specific wording but the accessibility of these ‘stock’ phrases should be reconsidered. The HRA also publishes guidance for preparing participant information sheets and consent forms, and most of their guidance is echoed by the FAST-R reviewers in this study. These include recommendations for formatting, including headings, using bullet points, putting important aspects in bold, using friendly/inviting language and using diagrams and pictures. The HRA and reviewers also similarly emphasised the importance of not using jargon, and transparency on participant payment and use of data. The HRA, however, also suggest considerations that reviewers do not think are important, like page and font size. Reviewers brought up problems overlooked by the HRA, such as clear inclusion and exclusion criteria, inclusive phrasing (including the gender option of transgender), avoiding repetition and considering the order of the text. This reinforces the importance of involving patients in the creation of these documents, as they can identify ways to make them more accessible that are overlooked by other bodies.

The FDA recommend a reading level of grade eight or lower to facilitate adequate understanding of health information.13 We found only 15 documents in this study (15.2%) complied with this before FAST-R review. The National Adult Literacy Survey revealed that about a quarter of American adults could not read or understand written materials above a fifth-grade level,Reference Liu, Lo and Hsieh27 and in the UK, around 15%, or 5.1 million adults, have literacy levels at or below those expected of an 11 year old.Reference Gilbert, Teravainen, Clark and Shaw28 In a readability assessment of 176 clinical research participant information leaflets and informed consent forms, several studies found few documents at this recommended readability grade.Reference Ennis and Wykes1,Reference O'Sullivan, Sukumar, Crowley, McAuliffe and Doran29 This highlights an urgent need to address accessibility issues, perhaps in the ethics guidance, so we ensure true informed consent.

Our FAST-R reviewers picked out similar issues that might prevent understanding by lay people, which replicates other studies,Reference Staley, Ashcroft, Doughty and Szmukler2,Reference Pilegaard and Ravn30 so this study emphasises the importance of these issues even for researchers who are not requesting patient involvement. It is important for everyone to consider the importance of true informed consent, rather than as an attempt to abide by enforced standards.

Limitations

There are limitations with using the Flesch–Kincaid Grade Level measure to assess readability. The tool measures semantic and syntactic difficulties, without considering that documents containing jargon terms may use other means to improve readability, such as visual aids.Reference Jindal and MacDermid31 Therefore, other measures may be more suitable to measure readability, such as the Clear Communication Index32 and the Suitability Assessment of Materials and future studies should aim to use these tools when assessing the impact of patient and public involvement.Reference Doak, Doak and Root33

Additionally, the training the reviewers receive may mean that they have different standards for research documents, as well as what constitutes jargon, compared with lay patients. Therefore their views may not be representative of other patients who are less experienced in participating in research. Future studies could investigate whether there are differences in this regard between these participants and lay patients.

In conclusion, we have highlighted common mistakes made by researchers in documents submitted for patient review by FAST-R. These issues have not changed over time, and one issue that has been made worse is the use of jargon, so we provided a list of jargon words and phrases pointed out by FAST-R and ways to avoid them. This work not only shows the influence of involving patients, as all but two documents had changes made to them, but also shows the benefits of involvement at the earliest stage. Only through this involvement will we have true participant informed consent.

Supplementary material

Supplementary material is available online at https://doi.org/10.1192/bjo.2021.1040.

Data availability

The data that support the findings of this study are available from the corresponding author, S.J., upon reasonable request.

Acknowledgements

We acknowledge the expert input of the Maudsley Biomedical Research Centre's FAST-R group, Anna Maria Fantini and Philip Sharma.

Author contributions

S.J. and T.W. designed the study. All authors carried out the literature search. S.J., C.M.O., E.W., M.M., G.H., S.M.J. and E.N. collected the data and contributed to data analysis. S.J., T.W., S.M.J., G.H. and E.N. interpreted the data and wrote the manuscript.

Funding

This paper represents independent research funded by the National Institute for Health Research (NIHR) Biomedical Research Centre at South London and Maudsley NHS Foundation Trust and King's College London (grant IS-BRC-1215-20018).

Declaration of interest

None.

References

Ennis, L, Wykes, T. Sense and readability: participant information sheets for research studies. Br J Psychiatry 2016; 208(2): 189–94.10.1192/bjp.bp.114.156687CrossRefGoogle ScholarPubMed
Staley, K, Ashcroft, J, Doughty, L, Szmukler, G. Making it clear and relevant: patients and carers add value to studies through research document reviews. Ment Health Soc Inclus 2016; 20(1): 3643.10.1108/MHSI-09-2015-0037CrossRefGoogle Scholar
Evans, J, Papoulias, S. Between funder requirements and ‘jobbing scientists’: the evolution of patient and public involvement in a mental health biomedical research centre - a qualitative study. Res Involv Engagem 2020; 6: 12.10.1186/s40900-020-00185-7CrossRefGoogle Scholar
National Institute for Health Research. I Want to Help with Research. National Institute for Health Research, 2021 (https://www.nihr.ac.uk/patients-carers-and-the-public/i-want-to-help-with-research/).Google Scholar
Barnes, M, Cotterell, P. Critical Perspectives on User Involvement. Policy Press, 2012.Google Scholar
Trivedi, P, Wykes, T. From passive subjects to equal partners: qualitative review of user involvement in research. Br J Psychiatry 2002; 181: 468–72.10.1192/bjp.181.6.468CrossRefGoogle ScholarPubMed
Callard, F, Rose, D, Wykes, T. Close to the bench as well as at the bedside: involving service users in all phases of translational research. Health Expect 2011; 15: 389400.10.1111/j.1369-7625.2011.00681.xCrossRefGoogle Scholar
Chalmers, I, Glasziou, P. Avoidable waste in the production and reporting of research evidence. Lancet 2009; 374(9683): 86–9.10.1016/S0140-6736(09)60329-9CrossRefGoogle ScholarPubMed
Ennis, L, Wykes, T. Impact of patient involvement in mental health research: longitudinal study. Br J Psychiatry 2013; 203(5): 381–6.10.1192/bjp.bp.112.119818CrossRefGoogle ScholarPubMed
Iliffe, S, McGrath, T, Mitchell, D. The impact of patient and public involvement in the work of the dementias & neurodegenerative diseases research network (DeNDRoN): case studies. Health Expect 2013; 16(4): 351–61.10.1111/j.1369-7625.2011.00728.xCrossRefGoogle Scholar
Faulkner, A, Carr, S, Gould, D, Khisa, C, Hafford-Letchfield, T, Cohen, R, et al. Dignity and respect’: an example of service user leadership and co-production in mental health research. Health Expect 2021; 24(suppl 1): 10–9.10.1111/hex.12963CrossRefGoogle ScholarPubMed
Redman, S, Greenhalgh, T, Adedokun, L, Staniszewska, S, Denegri, S. Co-production of knowledge: the future. BMJ 2021; 372: n434.10.1136/bmj.n434CrossRefGoogle ScholarPubMed
US Food and Drug Administration (FDA). Informed consent – Draft Guidance for IRBs, Clinical Investigators and Sponsors. FDA, 2014 (https://www.fda.gov/regulatory-information/search-fda-guidance-documents/informed-consent).Google Scholar
Rolfe, DE, Ramsden, VR, Banner, D, Graham, ID. Using qualitative health research methods to improve patient and public involvement and engagement in research. Res Involv Engagem 2018; 4: 49.10.1186/s40900-018-0129-8CrossRefGoogle ScholarPubMed
Crocker, JC, Boylan, AM, Bostock, J, Locock, L. Is it worth it? Patient and public views on the impact of their involvement in health research and its assessment: a UK-based qualitative interview study. Health Expect 2017; 20(3): 519–28.10.1111/hex.12479CrossRefGoogle ScholarPubMed
Martland, R, Gaughran, F, Stubbs, B, Onwumere, J. Perspectives on implementing HIIT interventions for service users in inpatient mental health settings: a qualitative study investigating patient, carer and staff attitudes. J Affect Disord 2021; 283: 198206.10.1016/j.jad.2021.01.065CrossRefGoogle ScholarPubMed
Greer, B, Newbery, K, Cella, M, Wykes, T. Predicting inpatient aggression in forensic services using remote monitoring technology: qualitative study of staff perspectives. J Med Internet Res 2019; 21(9): e15620.10.2196/15620CrossRefGoogle ScholarPubMed
Perera, G, Broadbent, M, Callard, F, Chang, CK, Downs, J, Dutta, R, et al. Cohort profile of the South London and Maudsley NHS Foundation Trust Biomedical Research Centre (SLaM BRC) case register: current status and recent enhancement of an electronic mental health record-derived data resource. BMJ Open 2016; 6(3): e008721.10.1136/bmjopen-2015-008721CrossRefGoogle ScholarPubMed
Flesch, R. A new readability yardstick. J Appl Psychol 1948; 32(3): 221–33.10.1037/h0057532CrossRefGoogle ScholarPubMed
Grohol, JM, Slimowicz, J, Granda, R. The quality of mental health information commonly searched for on the internet. Cyberpsychol Behav Soc Netw 2014; 17(4): 216–21.10.1089/cyber.2013.0258CrossRefGoogle ScholarPubMed
Williamson, JML, Martin, AG. Analysis of patient information leaflets provided by a district general hospital by the Flesch and Flesch–Kincaid method. Int J Clin Pract 2010; 63(13): 1824–31.10.1111/j.1742-1241.2010.02408.xCrossRefGoogle Scholar
Cochrane, ZR, Gregory, P, Wilson, A. Readability of consumer health information on the internet: a comparison of U.S. government–funded and commercially funded websites. J Health Commun 2012; 17(9): 1003–10.10.1080/10810730.2011.650823CrossRefGoogle Scholar
Braun, V, Clarke, V. Using thematic analysis in psychology. Qual Res Psychol 2006; 3: 77101.10.1191/1478088706qp063oaCrossRefGoogle Scholar
Pope, C, Ziebland, S, Mays, N. Analysing qualitative data. BMJ 2000; 320: 114.10.1136/bmj.320.7227.114CrossRefGoogle ScholarPubMed
Pike, KL. Etic and emic standpoints for the description of behavior. In Language in Relation to a Unified Theory of the Structure of Human Behavior (ed. Pike, K): 812. Mouton and Company, 1966.Google Scholar
Thomas, DR. A general inductive approach for analyzing qualitative evaluation data. Am J Eval 2006; 27(2): 237–46.10.1177/1098214005283748CrossRefGoogle Scholar
Liu, E, Lo, T-F, Hsieh, H-L. Literacy and health: evidence from the 2003 national assessment of adult literacy. J Appl Econ 2011; 8(1): 1735.Google Scholar
Gilbert, L, Teravainen, A, Clark, C, Shaw, S. Literacy and Life Expectancy - An Evidence Review Exploring the Link between Literacy and Life Expectancy in England through Health and Socioeconomic Factors. National Literacy Trust, 2018 (https://cdn.literacytrust.org.uk/media/documents/National_Literacy_Trust_-_Literacy_and_life_expectancy_report.pdf).Google Scholar
O'Sullivan, L, Sukumar, P, Crowley, R, McAuliffe, E, Doran, P. Readability and understandability of clinical research patient information leaflets and consent forms in Ireland and the UK: a retrospective quantitative analysis. BMJ Open 2020; 10: e037994.10.1136/bmjopen-2020-037994CrossRefGoogle ScholarPubMed
Pilegaard, M, Ravn, HB. Readability of patient information can be improved. Dan Med J 2012; 59(5): A4408.Google ScholarPubMed
Jindal, P, MacDermid, JC. Assessing reading levels of health information: uses and limitations of Flesch formula. Educ Health 2017; 30(1): 84.10.4103/1357-6283.210517CrossRefGoogle ScholarPubMed
Centers for Disease Control and Prevention (CDC). The CDC Clear Communication Index. CDC, 2021 (https://www.cdc.gov/ccindex/index.html).Google Scholar
Doak, CC, Doak, LG, Root, JH. Teaching Patients with Low Literacy Skills (2nd Ed.). J. B. Lippincott Company, 1996.Google Scholar
Figure 0

Fig. 1 The stages of the study's methodology, including the qualitative and quantitative analyses. FAST-R, Feasibility and Support to Timely Recruitment for Research.

Figure 1

Table 1 Reading grade, word count and jargon for all documents before and after FAST-R review

Figure 2

Fig. 2 Comparison of FAST-R patient and carer guidance (right) with guidance from the Health Regulations Authority. FAST-R, Feasibility and Support to Timely Recruitment for Research; GDPR, General Data Protection Regulation; HRA, Health Regulations Authority.

Figure 3

Table 2 The number of suggested changes made by FAST-R reviewers, by document type and the percentage of changes implemented by researchers

Figure 4

Table 3 Framework categories, and emerging themes with their subthemes

Figure 5

Table 4 Suggested alternative wordings by FAST-R reviewers

Figure 6

Table 5 Words and phrases flagged up by reviewers as confusing or needing further explanation

Supplementary material: File

Jilka et al. supplementary material

Jilka et al. supplementary material

Download Jilka et al. supplementary material(File)
File 39.4 KB
Submit a response

eLetters

No eLetters have been published for this article.