Journal Information
Vol. 99. Issue 5.
Pages 335-349 (01 November 2023)
Visits
866
Vol. 99. Issue 5.
Pages 335-349 (01 November 2023)
Special Article
Full text access
Reporting guidelines and a touch of critical appraisal
Guías para la elaboración de manuscritos y unas pinceladas de lectura crítica
Visits
866
Pilar Aizpurua Galdeanoa, Paz González Rodríguezb,
Corresponding author
, María Aparicio Rodrigoc, Nieves Balado Insunzad, Elena Pérez Gonzáleze, Juan Ruiz-Canela Cáceresf, Eduardo Ortega Páezg, en representación del Grupo de Trabajo de Pediatría Basada en la Evidencia de la AEPap y el Comité de Pediatría Basada en la Evidencia de la AEP
a Atención Primaria, San Sebastián, Spain
b Centro de Atención Primaria Barrio del Pilar, Madrid, Spain
c Centro de Atención Primaria Entrevías, Madrid, Spain
d Hospital Álvaro Cunqueiro, Vigo, Pontevedra, Spain
e Hospital Universitario Virgen Macarena, Sevilla, Spain
f Atención Primaria, Sevilla, Spain
g Atención Primaria, UGC Góngora, Distrito Granada Metropolitano, Granada, Spain
This item has received
Article information
Abstract
Full Text
Bibliography
Download PDF
Statistics
Figures (1)
Tables (6)
Table 1. Reporting guidelines for main study types.
Table 2. CONSORT 2010 checklist.
Table 3. TREND checklist.
Table 4. STROBE checklist.
Table 5. STARD 2015 checklist.
Table 6. PRISMA 2020 checklist.
Show moreShow less
Abstract

The biomedical research process must follow certain quality criteria in its design and development to ensure that the results are credible and reliable. Once completed, the time comes to write an article for publication. The article must present in sufficient detail, and in a clear and transparent manner, all the information on the research work that has been carried out. In this way, readers, after a critical reading of the published content, will be able to judge the validity and relevance of the study and, if they so wish, make use of the findings.

In order to improve the description of the research process for publication, a series of guidelines have been developed which, in a simple and structured way, guide authors in the preparation of a manuscript. They are presented in the form of a list, flowchart, or structured text, and are an invaluable aid when writing an article.

This article presents the reporting guidelines for the most common designs along with the corresponding checklists.

Keywords:
Medical writing/standards
Checklist
Epidemiologic research design
Resumen

El proceso de investigación biomédica debe seguir unos criterios de calidad en su diseño y elaboración, que garanticen que los resultados son creíbles y fiables. Una vez finalizado, llega el momento de escribir un artículo para su publicación. Este debe presentar con suficiente detalle, y de forma clara y transparente, toda la información del trabajo de investigación realizado. De esta forma, los lectores, tras una lectura crítica de lo publicado, podrán juzgar la validez y la relevancia del estudio, y si lo consideran, utilizar los hallazgos.

Con el objetivo de mejorar la descripción del proceso de investigación para su publicación, se han desarrollado una serie de guías que, de forma sencilla y estructurada, orientan a los autores a la hora de elaborar un manuscrito. Se presentan en forma de lista, diagrama de flujo, o texto estructurado, y son una ayuda inestimable a la hora de escribir un artículo.

Este artículo presenta las guías de elaboración de manuscritos de los diseños más habituales, con sus listas de verificación.

Palabras clave:
Escritura médica: estándares
Lista de verificación
Diseño de investigaciones epidemiológicas
Full Text
Background

In recent decades, there has been an exponential increase in the number of biomedical publications and, at the same time, concern has grown about the quality of their reporting. In the 1990s, a series of initiatives started to be developed with the aim of improving the quality of the content of scientific publications, at first focused on randomised controlled trials. These initiatives took shape in a series of guidelines that offered structured recommendations on the information that should be included in a paper depending on the study design.

The English term used to refer to them, reporting guidelines, has been translated to Spanish in different ways: “listas-guía”, ‘lista de directrices”, “directrices para la comunicación”, “directrices de presentación de informes”… This has given rise to some confusion as to what they are and what their purpose is, which is to improve reporting in published studies. They are also frequently referred to with the broad term of “checklists”, confusing them with other types of guidelines or tools whose purpose is to assess bias and the applicability of previously published reports. There are multiple tools for these purposes. One example would be the tools used by the authors of systematic reviews to assess the risk of bias of the primary studies included the review (Quadas-2,1 RoB-2,2 Newscastle-Otawa scale3…). There are also the tools intended for the readership, such as the user’s guides to the medical literature of the JAMA (Journal of the American Medical Association),4 the guidelines of the BMJ (British Medical Journal), based on the recommendations of the Evidence-Based Medicine Working Group, or the CASPe (Critical Appraisal Skills Programme Español) tools.5 The GRADE (Grading of Recommendations, Assessment, Development, and Evaluations) approach goes a step further for, in addition to the assessment of the quality of the evidence, grading the strength of recommendations in the development of clinical practice guidelines, systematic reviews or assessment of health care technologies.6

When it comes to reporting guidelines, which in the Spanish version of this article we called “guías para la elaboración de manuscritos” (based on their definition),7 their main goal is to guide authors in providing the best possible description of the methods used in each of the phases of research accurately and transparently. They are meant for manuscript authors, to be applied during the drafting of the paper, and also for journal editors and reviewers. It would be fair to say that the intent of these tools is to promote excellence in scientific reporting, regardless of the quality of the study. There can be studies that have an optimal design, implementation and analysis of results, but which are poorly reported because the published content is missing key information regarding the methods and results. This can create problems at a later stage in the assessment of study bias or applicability and in the inclusion of the study in a systematic review, all of which could have been easily prevented with a more detailed description.

The first reporting guideline, which is also the best known today, was the CONSORT Statement8 (CONsolidated Standards Of Reporting Trials, focusing on randomised trials). It was followed by others such as STARD9 (Standards for Reporting Diagnostic accuracy), PRISMA10 (Preferred Reporting Items for Systematic reviews and Meta-Analyses), STROBE11 (Strengthening the Reporting of OBservational studies in Epidemiology)… and extensions of these guidelines focusing on specific aspects within the type of study addressed by the guideline.

The proliferation of this type of guidelines spurred yet another initiative, EQUATOR (Enhancing the QUAlity and Transparency Of health Research, https://www.equator-network.org/).12 Active since 2006, its mission is to collect all valid reporting guidelines for any care setting, promote their application and assess the scientific literature on the subject (Table 1). It is the first attempt to coordinate efforts to address inadequate publication practices at a global scale. It encompasses researchers, medical journal editors and reviewers, guideline developers, research funding groups and other collaborators interested in improving the quality of scientific research and publications. In addition, the “Toolkit” section of the website13 offers practical help and resources to support professionals in different activities.

Table 1.

Reporting guidelines for main study types.

Randomised trials  CONSORT  Extensions 
Observational studies  STROBE  Extensions 
Systematic reviews  PRISMA  Extensions 
Study protocols  SPIRIT  PRISMA-P 
Diagnostic/prognostic studies  STARD  TRIPOD 
Case reports  CARE  Extensions 
Clinical practice guidelines  AGREE  RIGHT 
Nonrandomized controlled trials  TREND   
Qualitative research  SRQR  COREQ 
Animal pre-clinical studies  ARRIVE   
Quality improvement studies  SQUIRE  Extensions 
Economic evaluations  CHEERS   

Adapted from: https://www.equator-network.org/.

In this article, we present the guidelines that we consider most relevant based on the type of study and the frequency of their use for authors who seek to publish articles, peer reviewers and readers of scientific publications overall.

What should be included in a randomised trial report?

Well-designed and implemented randomised controlled trials (RCTs) are the gold standard for producing primary evidence to guide clinical decision-making regarding health care interventions.

Alarmed at the overwhelming evidence that the quality of RCT reporting tended to be far from optimal, a group of researchers and editors developed the Consolidated Standards of Reporting Trials (CONSORT) statement to improve the quality of reporting of RCTs for publication. The first version of the statement was published in 1996. Two revisions have been published since, the CONSORT 2001 and the CONSORT 2010 versions, with additional publication of several extensions. In 2022, very interesting extensions were published that delved in more detail on the reporting of intervention outcomes (The CONSORT-Outcomes 2022 Statement)14 and the adverse events (CONSORT Harms 2022 Statement).15

The CONSORT 2010 Statement, checklist, participant flow diagram and the indispensable explanation paper (CONSORT 2010 Explanation and Elaboration) are available at the EQUATOR8 website.16 The checklist comprises 25 items (some with subitems) that, due to their significant relevance, should always be included in any article reporting on a RCT. They are structured into 6 sections: Title and abstract, Introduction, Methods, Results, Discussion and Other information. The CONSORT Outcomes 2022 extension adds 17 more subitems to the CONSORT 2010 checklist, for greater detail in the Results (item 6a), Sample size (item 7a), Statistical methods (item 12a) and Ancillary analyses (item 18) sections.

Table 2 presents the CONSORT 2010 checklist, the Spanish translation of which is available in the Spanish version of this article.

Table 2.

CONSORT 2010 checklist.

Section/Topic  Item no.  Checklist item 
Title and abstract1a  Identification as a randomised trial in the title 
1b  Structured summary of trial design, methods, results, and conclusions (for specific guidance see CONSORT for abstracts) 
Introduction
Background and objectives2a  Scientific background and explanation of rationale 
2b  Specific objectives or hypotheses 
Methods
Trial design3a  Description of trial design (such as parallel, factorial) including allocation ratio 
3b  Important changes to methods after trial commencement (such as eligibility criteria), with reasons 
Participants4a  Eligibility criteria for participants 
4b  Settings and locations where the data were collected 
Interventions  The interventions for each group with sufficient details to allow replication, including how and when they were actually administered 
Results6a  Completely defined pre-specified primary and secondary outcome measures, including how and when they were assessed 
6b  Any changes to trial outcomes after the trial commenced, with reasons 
Sample size7a  How sample size was determined 
7b  When applicable, explanation of any interim analyses and stopping guidelines 
Randomisation
Sequence generation8a  Method used to generate the random allocation sequence 
8b  Type of randomisation; details of any restriction (such as blocking and block size) 
Allocation concealment mechanism  Mechanism used to implement the random allocation sequence (such as sequentially numbered containers), describing any steps taken to conceal the sequence until interventions were assigned 
Implementation  10  Who generated the random allocation sequence, who enrolled participants, and who assigned participants to interventions 
Blinding11a  If done, who was blinded after assignment to interventions (for example, participants, care providers, those assessing outcomes) and how 
11b  If relevant, description of the similarity of interventions 
Statistical methods12a  Statistical methods used to compare groups for primary and secondary outcomes 
12b  Methods for additional analyses, such as subgroup analyses and adjusted analyses 
Results
Participant flow (a diagram is strongly recommended)13a  For each group, the numbers of participants who were randomly assigned, received intended treatment, and were analysed for the primary outcome 
13b  For each group, losses and exclusions after randomisation, together with reasons 
Recruitment14a  Dates defining the periods of recruitment and follow-up 
14b  Why the trial ended or was stopped 
Baseline data  15  A table showing baseline demographic and clinical characteristics for each group 
Numbers analysed  16  For each group, number of participants (denominator) included in each analysis and whether the analysis was by original assigned groups 
Outcomes and estimation17a  For each primary and secondary outcome, results for each group, and the estimated effect size and its precision (such as 95% confidence interval) 
17b  For binary outcomes, presentation of both absolute and relative effect sizes is recommended 
Ancillary analyses  18  Results of any other analyses performed, including subgroup analyses and adjusted analyses, distinguishing pre-specified from exploratory 
Harms  19  All important harms or unintended effects in each group (for specific guidance see CONSORT for harms) 
Discussion
Limitations  20  Trial limitations, addressing sources of potential bias, imprecision, and, if relevant, multiplicity of analyses 
Generalisability  21  Generalisability (external validity, applicability) of the trial findings 
Interpretation  22  Interpretation consistent with results, balancing benefits and harms, and considering other relevant evidence 
Other information
Registration  23  Registration number and name of trial registry 
Protocol  24  Where the full trial protocol can be accessed, if available 
Funding  25  Sources of funding and other support (such as supply of drugs), role of funders 

Source/citation in spanish version: Cobos-Carbó A, Augustovski F. Declaración CONSORT 2010: actualización de la lista de comprobación para informar ensayos clínicos aleatorizados de grupos paralelos. Med Clin (Barc). 2011;137:213-5.

What should be included in a nonrandomised trial report?

Although a randomised controlled trial is the best possible method to demonstrate the efficacy of an intervention, it is not always feasible. We often encounter articles in the scientific literature reporting results obtained in nonrandomised clinical trials. These studies are conducted when carrying out a RCT would not be ethical, when the objective of the study is to analyse effectiveness in real-world clinical practice or the cost-effectiveness of a health care intervention or when the causal chain of the effect of the intervention is highly complex. Under these circumstances, it is appropriate to conduct nonrandomised trials, whose value has yet to be fully recognised. To be sure, RCTs offer the significant advantage of controlling biases or systematic errors that were not taken into account in the study design due to lack of previous awareness, but the lack of randomisation in nonrandomised trials can be compensated with Bayesian methods, used to quantify the degree to which the findings could be explained by potential selection biases.17

Many of the decisions made in our daily clinical practice or in public health policy are based on evidence from nonrandomised trials. Therefore, transparency, thoroughness and clarity are essential in reporting these studies. The Transparent Reporting of Evaluation with Non Randomized Designs (TREND) was developed to provide guidance on the subject.18,19

The first version of TREND was published in 2004 in the American Journal of Public Health17 on the initiative of the Center for Disease Control and Prevention (CDC)20 and several journal editors and representatives. The purpose was to provide recommendations for authors, reviewers and editors on the information that should be included in the reporting of nonrandomised studies. The TREND checklist comprises 22 items structured into 5 sections: Title, Introduction, Methods, Results and Discussion. Table 3 presents the most recent update of the checklist.

Table 3.

TREND checklist.

Paper Section/Topic  Item no.  Descriptor 
Title and Abstract
Title and Abstract1Information on how unit were allocated to interventions 
Structured abstract recommended 
Information on target population or study sample 
Introduction     
Background2Scientific background and explanation of rationale 
Theories used in designing behavioural interventions 
Methods     
Participants3Eligibility criteria for participants, including criteria at different levels in recruitment/sampling plan (eg, cities, clinics, subjects) 
Method of recruitment (eg, referral, self-selection), including the sampling method if a systematic sampling plan was implemented 
Recruitment setting 
Settings and locations where the data were collected 
Interventions4Details of the interventions intended for each study condition and how and when they were actually administered, specifically including: 
• Content: what was given? 
• Delivery method: how was the content given? 
• Unit of delivery: how were the subjects grouped during delivery? 
• Deliverer: who delivered the intervention? 
• Setting: where was the intervention delivered? 
• Exposure quantity and duration: how many sessions or episodes or events were intended to be delivered? How long were they intended to last? 
• Time span: how long was it intended to take to deliver the intervention to each unit? 
• Activities to increase compliance or adherence (eg, incentives) 
Objectives  Specific objectives and hypotheses 
Outcomes6Clearly defined primary and secondary outcome measures 
Methods used to collect data and any methods used to enhance the quality of measurements 
Information on validated instruments such as psychometric and biometric properties 
Sample size  How sample size was determined and, when applicable, explanation of any interim analyses and stopping rules 
Assignment method8Unit of assignment (the unit being assigned to study condition, eg, individual, group, community) 
Method used to assign units to study conditions, including details of any restriction (eg, blocking, stratification, minimization) 
Inclusion of aspects employed to help minimize potential bias induced due to non-randomization (eg, matching) 
Blinding (masking)  Whether or not participants, those administering the interventions, and those assessing the outcomes were blinded to study condition assignment; if so, statement regarding how the blinding was accomplished and how it was assessed. 
Unit of analysis10Description of the smallest unit that is being analysed to assess intervention effects (eg, individual, group, or community) 
If the unit of analysis differs from the unit of assignment, the analytical method used to account for this (eg, adjusting the standard error estimates by the design effect or using multilevel analysis) 
Statistical methods11Statistical methods used to compare study groups for primary methods outcome(s), including complex methods of correlated data 
Statistical methods used for additional analyses, such as a subgroup analyses and adjusted analysis 
Methods for imputing missing data, if used 
Statistical software or programs used 
Results
Participant flow12Flow of participants through each stage of the study: enrolment, assignment, allocation, and intervention exposure, follow-up, analysis (a diagram is strongly recommended) 
• Enrolment: the numbers of participants screened for eligibility, found to be eligible or not eligible, declined to be enrolled, and enrolled in the study 
• Assignment: the numbers of participants assigned to a study condition 
• Allocation and intervention exposure: the number of participants assigned to each study condition and the number of participants who received each intervention 
• Follow-up: the number of participants who completed the follow-up or did not complete the follow-up (i.e., lost to follow-up), by study condition 
• Analysis: the number of participants included in or excluded from the main analysis, by study condition 
Description of protocol deviations from study as planned, along with reasons 
Recruitment  13  Dates defining the periods of recruitment and follow-up 
Baseline data14Baseline demographic and clinical characteristics of participants in each study condition 
Baseline characteristics for each study condition relevant to specific disease prevention research 
Baseline comparisons of those lost to follow-up and those retained, overall and by study condition 
Comparison between study population at baseline and target population of interest 
Baseline equivalence  15  Data on study group equivalence at baseline and statistical methods used to control for baseline differences 
Numbers analysed16Number of participants (denominator) included in each analysis for each study condition, particularly when the denominators change for different outcomes; statement of the results in absolute numbers when feasible 
Indication of whether the analysis strategy was “intention to treat” or, if not, description of how non-compliers were treated in the analyses 
Outcomes and estimation17For each primary and secondary outcome, a summary of results for each estimation study condition, and the estimated effect size and a confidence interval to indicate the precision 
Inclusion of null and negative findings 
Inclusion of results from testing pre-specified causal pathways through which the intervention was intended to operate, if any 
Ancillary analyses  18  Summary of other analyses performed, including subgroup or restricted analyses, indicating which are pre-specified or exploratory 
Adverse events  19  Summary of all important adverse events or unintended effects in each study condition (including summary measures, effect size estimates, and confidence intervals) 
Discussion
Interpretation20Interpretation of the results, taking into account study hypotheses, sources of potential bias, imprecision of measures, multiplicative analyses, and other limitations or weaknesses of the study 
Discussion of results taking into account the mechanism by which the intervention was intended to work (causal pathways) or alternative mechanisms or explanations 
Discussion of the success of and barriers to implementing the intervention, fidelity of implementation 
Discussion of research, programmatic, or policy implications 
Generalizability  21  Generalizability (external validity) of the trial findings, taking into account the study population, the characteristics of the intervention, length of follow-up, incentives, compliance rates, specific sites/settings involved in the study, and other contextual issues 
Overall evidence  22  General interpretation of the results in the context of current evidence and current theory 

Source: TREND Statement checklist (pdf). Available at: https://www.cdc.gov/trendstatement/pdf/trendstatement_trend_checklist.pdf.

TREND is considered a complement to CONSORT (the guideline for randomised clinical trial reporting), and the two checklists have some overlapping elements and some distinct elements specific to each of them. The information presented in this section was updated by the CDC in September 2018 and by EQUATOR21 in November 2021.

What should be included in an observational study report?

One of the proposed definitions for epidemiology is “study of distribution and determinants of health-related states among specified populations and the application of that study to the control of health problems”. The aim of epidemiological studies is to answer questions regarding the frequency and trends in a given disease, which preventive interventions are most effective or the benefits of modifying certain diagnostic or therapeutic approaches, among others.

To answer these questions, the field of epidemiology has developed a methodology of its own that is based on observation. Thus, the design of an epidemiological study depends on the question that researchers ask regarding the population and the topic under study (disease, risk or protective factors, health care interventions, etc).22 For instance, the most appropriate design to study the prognosis of a condition would be a longitudinal observational study, while in aetiological research it would be a cohort or a case-control study.

Observational studies, by design, are intrinsically likely to suffer from systematic errors and biases that affect their internal validity, which in turn (in addition to other factors) may affect the external validity or generalizability of their results. Furthermore, articles reporting on observational studies frequently fail to present sufficiently clear and detailed information on the conducted research, which hampers the assessment of the evidence.

The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) Statement was developed with the aim of improving the reporting of observational studies, issuing a series of recommendations on the information that should be provided in a complete and accurate report of an observational study. The STROBE Statement was published in 2007 and it is available at the EQUATOR website11 or its own site.23

It provides checklists for the main study designs of analytical epidemiology: cohort, case–control, and cross-sectional studies. A combined checklist has also been published that can be used to any of the three designs (Table 4),24 as well as an article explaining each item in detail, offering a rationale and published examples (STROBE: explanation and elaboration).25

Table 4.

STROBE checklist.

  Item no.  Recommendation 
Title and abstract1(a) Indicate the study’s design with a commonly used term in the title or the abstract 
(b) Provide in the abstract an informative and balanced summary of what was done and what was found 
Introduction
Background/rationale  Explain the scientific background and rationale for the investigation being reported 
Objectives  State specific objectives, including any prespecified hypotheses 
Methods
Study design  Present key elements of study design early in the paper 
Setting  Describe the setting, locations, and relevant dates, including periods of recruitment, exposure, follow-up, and data collection 
Participants6(a) Cohort study—Give the eligibility criteria, and the sources and methods of selection of participants. Describe methods of follow-up 
Case-control study—Give the eligibility criteria, and the sources and methods of case ascertainment and control selection. Give the rationale for the choice of cases and controls 
Cross-sectional study—Give the eligibility criteria, and the sources and methods of selection of participants 
(b) Cohort study—For matched studies, give matching criteria and number of exposed and unexposed 
Case-control study—For matched studies, give matching criteria and the number of controls per case 
Variables  Clearly define all outcomes, exposures, predictors, potential confounders, and effect modifiers. Give diagnostic criteria, if applicable 
Data sources/measurement  8a  For each variable of interest, give sources of data and details of methods of assessment (measurement). Describe comparability of assessment methods if there is more than one group 
Bias  Describe any efforts to address potential sources of bias 
Study size  10  Explain how the study size was arrived at 
Quantitative variables  11  Explain how quantitative variables were handled in the analyses. If applicable, describe which groupings were chosen and why 
Statistical methods12(a) Describe all statistical methods, including those used to control for confounding 
(b) Describe any methods used to examine subgroups and interactions 
(c) Explain how missing data were addressed 
(d) Cohort study—If applicable, explain how loss to follow-up was addressed 
Case-control study—If applicable, explain how matching of cases and controls was addressed 
Cross-sectional study—If applicable, describe analytical methods taking account of sampling strategy 
(e) Describe any sensitivity analyses 
Results
Participants13a(a) Report numbers of individuals at each stage of study—eg numbers potentially eligible, examined for eligibility, confirmed eligible, included in the study, completing follow-up, and analysed 
(b) Give reasons for non-participation at each stage 
(c) Consider use of a flow diagram 
Descriptive data14a(a) Give characteristics of study participants (eg demographic, clinical, social) and information on exposures and potential confounders 
(b) Indicate number of participants with missing data for each variable of interest 
(c) Cohort study—Summarise follow-up time (eg, average and total amount) 
Outcome data15aCohort study—Report numbers of outcome events or summary measures over time 
Case-control study—Report numbers in each exposure category, or summary measures of exposure 
Cross-sectional study—Report numbers of outcome events or summary measures 
Main results16(a) Give unadjusted estimates and, if applicable, confounder-adjusted estimates and their precision (eg, 95% confidence interval). Make clear which confounders were adjusted for and why they were included 
(b) Report category boundaries when continuous variables were categorized 
(c) If relevant, consider translating estimates of relative risk into absolute risk for a meaningful time period 
Other analyses  17  Report other analyses done—eg analyses of subgroups and interactions, and sensitivity analyses 
Discussion
Key results  18  Summarise key results with reference to study objectives 
Limitations  19  Discuss limitations of the study, taking into account sources of potential bias or imprecision. Discuss both direction and magnitude of any potential bias 
Interpretation  20  Give a cautious overall interpretation of results considering objectives, limitations, multiplicity of analyses, results from similar studies, and other relevant evidence 
Generalizability  21  Discuss the generalisability (external validity) of the study results 
Other information
Funding  22  Give the source of funding and the role of the funders for the present study and, if applicable, for the original study on which the present article is based 

Source: Von Elm E, Altman DG, Egger M, Pocock S, Gotzsche PC, Vandenvroucke JP. Declaración de la iniciativa STROBE (Strengthening the Reporting of Observational studies in Epidemiology): directrices para la comunicación de estudios observacionales. Gac Sanit. 2008;22:144-50.

a

Give information separately for cases and controls in case-control studies and, if applicable, for exposed and unexposed groups in cohort and cross-sectional studies.

Each of the checklists comprises 22 items structured in 6 sections: Title and abstract, Introduction, Methods, Results, Discussion and Other information (Funding). In the combined checklist, 18 items are common to all designs and 4 are specific to each design.

What should be included in a diagnostic study report?

Reports of diagnostic accuracy studies should present all the key information readers may need be able to appraise the importance and credibility of their conclusions, in addition to a complete and rigorous description of their results. Only when all the necessary information is provided in the publication is it possible to assess the potential biases of the study and its generalizability at a later time. Yet, multiple reviews have evinced that diagnostic accuracy study reports frequently fail to report essential and adequate information on the selection of participants or the study design or the full results. In addition, the conclusions regarding diagnostic accuracy may be overly optimistic.26

In 2003, the Standards for Reporting Diagnostic Accuracy (STARD) statement was published following the example of the CONSORT statement. This guideline was published in prominent general and specialised medical journals. It included a 25-item checklist, a flow diagram and an explanatory document clarifying the meaning and rationale of each of the items and a short summary of the available evidence they were based on.27 Its publication prompted several editorials recommending its acceptance by journal editors and review committees.

In 2015, the statement was updated to include new evidence on the sources of bias and variability and to facilitate its use.

Table 5 presents the 30 items of the STARD 2015 checklist, of which the translation to Spanish is available in the Spanish version of this article. The items are structured into the usual sections found in original articles: Introduction, Methods, Results and Discussion, in addition to one section at the start (Title and abstract) and another at the end (Other information).

Table 5.

STARD 2015 checklist.

Section and topic  No.  Item 
Title or abstract
  Identification as a study of diagnostic accuracy using at least one measure of accuracy (such as sensitivity, specificity, predictive values, or AUC) 
Abstract
  Structured summary of study design, methods, results, and conclusions (for specific guidance, see STARD for Abstracts) 
Introduction
  Scientific and clinical background, including the intended use and clinical role of the index test 
  Study objectives and hypotheses 
Methods
Study design  Whether data collection was planned before the index test and reference standard were performed (prospective study) or after (retrospective study) 
ParticipantsEligibility criteria 
On what basis potentially eligible participants were identified (such as symptoms, results from previous tests, inclusion in registry) 
Where and when potentially eligible participants were identified (setting, location and dates) 
Whether participants formed a consecutive, random or convenience series 
Test methods10a  Index test, in sufficient detail to allow replication 
10b  Reference standard, in sufficient detail to allow replication 
11  Rationale for choosing the reference standard (if alternatives exist) 
12a  Definition of and rationale for test positivity cut-offs or result categories of the index test, distinguishing pre-specified from exploratory 
12b  Definition of and rationale for test positivity cut-offs or result categories of the reference standard, distinguishing pre-specified from exploratory 
13a  Whether clinical information and reference standard results were available to the performers/readers of the index test 
13b  Whether clinical information and index test results were available to the assessors of the reference standard 
Analysis14  Methods for estimating or comparing measures of diagnostic accuracy 
15  How indeterminate index test or reference standard results were handled 
16  How missing data on the index test and reference standard were handled 
17  Any analyses of variability in diagnostic accuracy, distinguishing pre-specified from exploratory 
18  Intended sample size and how it was determined 
Results
Participants19  Flow of participants, using a diagram 
20  Baseline demographic and clinical characteristics of participants 
21a  Distribution of severity of disease in those with the target condition 
21b  Distribution of alternative diagnoses in those without the target condition 
22  Time interval and any clinical interventions between index test and reference standard 
Test results23  Cross tabulation of the index test results (or their distribution) by the results of the reference standard 
24  Estimates of diagnostic accuracy and their precision (such as 95% confidence intervals) 
25  Any adverse events from performing the index test or the reference standard 
Discussion
  26  Study limitations, including sources of potential bias, statistical uncertainty, and generalisability 
  27  Implications for practice, including the intended use and clinical role of the index test 
Other information
  28  Registration number and name of registry 
  29  Where the full study protocol can be accessed 
  30  Sources of funding and other support; role of funders 

Source: STARD 2015 checklist. Available at: https://www.equator-network.org/wp-content/uploads/2015/03/STARD-2015-checklist.pdf.

Among the explanation and elaboration papers of the STARD 2015 statement, available at the EQUATOR website,12 the reader can find the original article by Cohen et al.28 with commented examples and a detailed explanation of each of the 30 items in the checklist. This information is provided to facilitate the writing of diagnostic accuracy reports and ensure they include all the relevant information. It is also meant to help journal reviewers and editors as well as readers to ensure that submitted and published articles are sufficiently detailed.

What should be included in a systematic review article?

The systematic review (SR) and meta-analysis is a type of study that synthesises the available evidence on a specific topic. The methodology used in this type of studies follows a clear, standardised and reproducible protocol to ensure the quality, consistency and transparency of the review process.29 When reporting results, it is very important to describe the development process in full and in detail, including why it was done, how and what was found. The PRISMA 2020 Statement provides recommendations to ensure that this happens, reflecting the most recent advances in methods to identify, select, appraise, and synthesise studies eligible for inclusion in systematic reviews.

All the information on PRISMA 2020 can be found in its own website,30 including translations to Spanish of some of the documents. Both its website and the EQUATOR website31 offer access to the PRISMA 2020 Statement, the 27-item checklist, the flow diagram template and, for those interested in more in-depth information, a more thorough and detailed version of the checklist and the PRISMA 2020 explanation and elaboration paper.32

The 27 items of the verification checklists are structured into 7 sections: Title, Abstract, Introduction, Methods, Results, Discussion and Other information (including financial support).

Table 6 presents the PRISMA 2020 checklist33 and Fig. 1, a template PRISMA flow diagram.

Table 6.

PRISMA 2020 checklist.

Section and topic  Item no.  Checklist item 
Title
Title  Identify the report as a systematic review. 
Abstract
Structured abstract  See the PRISMA 2020 for Abstracts checklist. 
Introduction
Rationale  Describe the rationale for the review in the context of existing knowledge. 
Objectives  Provide an explicit statement of the objective(s) or question(s) the review addresses. 
Methods
Eligibility criteria  Specify the inclusion and exclusion criteria for the review and how studies were grouped for the syntheses. 
Information sources  Specify all databases, registers, websites, organisations, reference lists and other sources searched or consulted to identify studies. Specify the date when each source was last searched or consulted. 
Search strategy  Present the full search strategies for all databases, registers and websites, including any filters and limits used. 
Selection process  Specify the methods used to decide whether a study met the inclusion criteria of the review, including how many reviewers screened each record and each report retrieved, whether they worked independently, and if applicable, details of automation tools used in the process. 
Data collection process  Specify the methods used to collect data from reports, including how many reviewers collected data from each report, whether they worked independently, any processes for obtaining or confirming data from study investigators, and if applicable, details of automation tools used in the process. 
Data items10a  List and define all outcomes for which data were sought. Specify whether all results that were compatible with each outcome domain in each study were sought (eg for all measures, time points, analyses), and if not, the methods used to decide which results to collect. 
10b  List and define all other variables for which data were sought (eg participant and intervention characteristics, funding sources). Describe any assumptions made about any missing or unclear information. 
Study risk of bias assessment  11  Specify the methods used to assess risk of bias in the included studies, including details of the tool(s) used, how many reviewers assessed each study and whether they worked independently, and if applicable, details of automation tools used in the process. 
Effect measures  12  Specify for each outcome the effect measure(s) (eg, risk ratio, mean difference) used in the synthesis or presentation of results. 
Synthesis methods13a  Describe the processes used to decide which studies were eligible for each synthesis (eg, tabulating the study intervention characteristics and comparing against the planned groups for each synthesis (item #5)). 
13b  Describe any methods required to prepare the data for presentation or synthesis, such as handling of missing summary statistics, or data conversions. 
13c  Describe any methods used to tabulate or visually display results of individual studies and syntheses. 
13d  Describe any methods used to synthesize results and provide a rationale for the choice(s). If meta-analysis was performed, describe the model(s), method(s) to identify the presence and extent of statistical heterogeneity, and software package(s) used. 
13e  Describe any methods used to explore possible causes of heterogeneity among study results (eg, subgroup analysis, meta-regression). 
13f  Describe any sensitivity analyses conducted to assess robustness of the synthesized results. 
Reporting bias assessment  14  Describe any methods used to assess risk of bias due to missing results in a synthesis (arising from reporting biases). 
Certainty assessment  15  Describe any methods used to assess certainty (or confidence) in the body of evidence for an outcome. 
Results
Study selection16a  Describe the results of the search and selection process, from the number of records identified in the search to the number of studies included in the review, ideally using a flow diagram. 
16b  Cite studies that might appear to meet the inclusion criteria, but which were excluded, and explain why they were excluded. 
Study characteristics  17  Cite each included study and present its characteristics. 
Risk of bias in studies  18  Present assessments of risk of bias for each included study. 
Results of individual studies  19  For all outcomes, present, for each study: (a) summary statistics for each group (where appropriate) and (b) an effect estimate and its precision (eg, confidence/credible interval), ideally using structured tables or plots. 
Results of syntheses20a  For each synthesis, briefly summarise the characteristics and risk of bias among contributing studies. 
20b  Present results of all statistical syntheses conducted. If meta-analysis was done, present for each the summary estimate and its precision (eg confidence/credible interval) and measures of statistical heterogeneity. If comparing groups, describe the direction of the effect. 
20c  Present results of all investigations of possible causes of heterogeneity among study results. 
20d  Present results of all sensitivity analyses conducted to assess the robustness of the synthesized results. 
Reporting biases  21  Present assessments of risk of bias due to missing results (arising from reporting biases) for each synthesis assessed. 
Certainty of evidence  22  Present assessments of certainty (or confidence) in the body of evidence for each outcome assessed. 
Discussion
Discussion23a  Provide a general interpretation of the results in the context of other evidence. 
23b  Discuss any limitations of the evidence included in the review. 
23c  Discuss any limitations of the review processes used. 
23d  Discuss implications of the results for practice, policy, and future research. 
Other information
Registration and protocol24a  Provide registration information for the review, including register name and registration number, or state that the review was not registered. 
24b  Indicate where the review protocol can be accessed, or state that a protocol was not prepared. 
24c  Describe and explain any amendments to information provided at registration or in the protocol. 
Support  25  Describe sources of financial or non-financial support for the review, and the role of the funders or sponsors in the review. 
Competing interests  26  Declare any competing interests of review authors. 
Availability of data, code and other materials  27  Report which of the following are publicly available and where they can be found: template data collection forms; data extracted from included studies; data used for all analyses; analytic code; any other materials used in the review. 

Source: Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. Declaración PRISMA 2020: una guía actualizada para la publicación de revisiones sistemáticas. Rev Esp Cardiol. 2021;74:790-9.

Figure 1.

PRISMA 2020 flow diagram template.

Source: Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ. 2021;372:n71.

(0.61MB).
How to read a published article

The process of publishing a scientific paper culminates when it reaches the target readership. Readers, however, should not be mere uncritical recipients of the information at their disposal. If the article has been written following the recommendations of the applicable reporting guideline, the reader will have sufficient information for, upon reading it critically, appraise the potential for bias in the evidence and whether the results can be applied locally.

Multiple tools are available to support the critical appraisal of published reports. The Critical Appraisal Skills Programme (CASPe)5 offers workshops on critical appraisal skills and tools to aid critical appraisal in Spanish, available at the website https://redcaspe.org/materiales/. At this site, the reader can find checklists composed of relevant questions that can guide the reading of reports for most of the study designs addressed in this article. These questions have only 3 possible answers: “yes”, “no” and “can’t tell”. A “can’t tell” answer usually means that the necessary information to answer the question is not available in the report. In some instances, it is due to deficiencies in the study methodology, but in others the problem stems from deficient reporting. The widespread application of reporting guidelines would reduce uncertainty and facilitate the appraisal of research publications.

With this article, the Working Group of Evidence-Based Paediatrics encourages future authors to use reporting guidelines to report their findings in a complete and structured manner. This will facilitate the subsequent publishing process of the article in addition to its critical reading by the audience.

Conflicts of interest

The authors have no conflicts of interest to disclose.

References
[1]
Bristol U of. QUADAS-2 [Internet]. University of Bristol [accessed 30 May 2023]. Available from: https://www.bristol.ac.uk/population-health-sciences/projects/quadas/quadas-2/.
[2]
RoB 2: A revised Cochrane risk-of-bias tool for randomized trials | Cochrane Bias [Internet] [accessed 30 May 2023]. Available from: https://methods.cochrane.org/bias/resources/rob-2-revised-cochrane-riskbias-tool-randomized-trials.
[3]
Ottawa Hospital Research Institute [Internet] [accessed 30 May 2023]. Available from: https://www.ohri.ca/programs/clinical_epidemiology/oxford.asp.
[4]
A.D. Oxman, D.L. Sackett, G.H. Guyatt, G. Browman, D. Cook, H. Gerstein, et al.
Users’ guides to the medical literature: I. How to get started.
JAMA., 270 (1993), pp. 2093-2095
[5]
Materiales – Redcaspe [Internet] [accessed 30 May 2023]. Available from: https://redcaspe.org/materiales/.
[6]
What is GRADE? | BMJ Best Practice [Internet] [accessed 30 May 2023]. Available from: https://bestpractice.bmj.com/info/toolkit/learn-ebm/what-is-grade/.
[7]
What is a reporting guideline? [Internet] [accessed 30 May 2023]. Available from: https://www.equator-network.org/about-us/what-is-a-reporting-guideline/.
[8]
CONSORT 2010 Statement: Updated guidelines for reporting parallel group randomised trials | EQUATOR Network [Internet] [accessed 30 May 2023]. Available from: https://www.equator-network.org/reporting-guidelines/consort/.
[9]
STARD 2015: An Updated List of Essential Items for Reporting Diagnostic Accuracy Studies [Internet] [accessed 22 May 2023]. Available from: https://www.equator-network.org/reporting-guidelines/stard/.
[10]
The PRISMA 2020 statement: An updated guideline for reporting systematic reviews | EQUATOR Network [Internet] [accessed 30 May 2023]. Available from: https://www.equator-network.org/reporting-guidelines/prisma/.
[11]
The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) Statement: guidelines for reporting observational studies [Internet] [accessed 30 May 2023]. Available from: https://www.equator-network.org/reporting-guidelines/strobe/.
[12]
EQUATOR Network | Enhancing the QUAlity and Transparency Of Health Research [Internet] [accessed 26 May 2023]. Available from: https://www.equator-network.org/.
[13]
How to write a great research paper using reporting guidelines [Internet] [accessed 25 May 2023]. Available from: https://www.equator-network.org/toolkits/writing-research/.
[14]
N.J. Butcher, A. Monsour, E.J. Mew, A.W. Chan, D. Moher, E. Mayo-Wilson, et al.
Guidelines for reporting outcomes in trial reports: the CONSORT-outcomes 2022 extension.
JAMA., 328 (2022), pp. 2252-2264
[15]
D.R. Junqueira, L. Zorzela, S. Golder, Y. Loke, J.J. Gagnier, S.A. Julious, et al.
CONSORT Harms 2022 statement, explanation, and elaboration: updated guideline for the reporting of harms in randomised trials.
BMJ., 381 (2023),
[16]
D. Moher, S. Hopewell, K.F. Schulz, V. Montori, P.C. Gøtzsche, P.J. Devereaux, et al.
CONSORT 2010 explanation and elaboration: updated guidelines for reporting parallel group randomised trials.
BMJ., 340 (2010), pp. c869
[17]
D.C. Des Jarlais, C. Lyles, N. Crepaz, TREND Group.
Improving the reporting quality of nonrandomized evaluations of behavioral and public health interventions: the TREND statement.
Am J Public Health., 94 (2004), pp. 361-366
[18]
Listas guía de comprobación de intervenciones no aleatorizadas: declaración TREND.
[19]
C. Vallvé, M. Artés, E. Cobo.
Estudios de intervención no aleatorizados (TREND).
Med Clín., 125 (2005), pp. 38-42
[20]
TREND statement | CDC [Internet]. 2019 [accessed 25 May 2023]. Available from: https://www.cdc.gov/trendstatement/index.html.
[21]
Improving the reporting quality of nonrandomized evaluations of behavioral and public health interventions: the TREND statement | EQUATOR Network [Internet] [accessed 30 May 2023]. Available from: https://www.equator-network.org/reporting-guidelines/improving-the-reporting-quality-of-nonrandomized-evaluations-of-behavioral-and-public-health-interventions-the-trend-statement/.
[22]
M. Molina Arias, C. Ochoa Sangrador.
Tipos de estudios epidemiológicos.
[23]
STROBE [Internet]. STROBE [accessed 4 Jun 2023]. Available from: https://www.strobe-statement.org/.
[24]
E. von Elm, D.G. Altman, M. Egger, S.J. Pocock, P.C. Gøtzsche, J.P. Vandenbroucke.
Declaración de la Iniciativa STROBE (Strengthening the Reporting of Observational studies in Epidemiology): directrices para la comunicación de estudios observacionales.
Gaceta Sanitaria., 22 (2008), pp. 144-150
[25]
J.P. Vandenbroucke, E. von Elm, D.G. Altman, P.C. Gøtzsche, C.D. Mulrow, S.J. Pocock, et al.
Strengthening the reporting of observational studies in epidemiology (STROBE): explanation and elaboration.
[26]
D.G. Altman, P.M. Bossuyt.
Estudios de precisión diagnóstica (STARD) y pronóstica (REMARK).
Med Clin (Barc)., 125 (2005), pp. 49-55
[27]
P.M. Bossuyt, J.B. Reitsma, D.E. Bruns, C.A. Gatsonis, P.P. Glasziou, L.M. Irwig, et al.
The STARD statement for reporting studies of diagnostic accuracy: explanation and elaboration.
[28]
J.F. Cohen, D.A. Korevaar, D.G. Altman, D.E. Bruns, C.A. Gatsonis, L. Hooft, et al.
STARD 2015 guidelines for reporting diagnostic accuracy studies: explanation and elaboration.
BMJ Open., 6 (2016),
[29]
Guirao-Goris JA. Utilidad y tipos de revisión de literatura. 9 (2) 2015. [Internet] [accessed 15 May 2023]. Available from: https://doi.org/10.4321/S1988-348X2015000200002.
[30]
PRISMA transparent reporting of systematic reviews and metaanalyses [Internet] [accessed 28 May 2023]. Available from: http://www.prisma-statement.org/.
[31]
The PRISMA 2020 statement: An updated guideline for reporting systematic reviews | EQUATOR Network [Internet] [accessed 2 Jun 2023]. Available from: https://www.equator-network.org/reporting-guidelines/prisma/.
[32]
M.J. Page, D. Moher, P.M. Bossuyt, I. Boutron, T.C. Hoffmann, C.D. Mulrow, et al.
PRISMA 2020 explanation and elaboration: updated guidance and exemplars for reporting systematic reviews.
BMJ., 372 (2021), pp. n160
[33]
M.J. Page, J.E. McKenzie, P.M. Bossuyt, I. Boutron, T.C. Hoffmann, C.D. Mulrow, et al.
Declaración PRISMA 2020: una guía actualizada para la publicación de revisiones sistemáticas.
Rev Esp Cardiol., 74 (2021), pp. 790-799
Copyright © 2023. Asociación Española de Pediatría
Idiomas
Anales de Pediatría (English Edition)
Article options
Tools
es en

¿Es usted profesional sanitario apto para prescribir o dispensar medicamentos?

Are you a health professional able to prescribe or dispense drugs?