Many randomized controlled trials (RCTs) are biased and difficult to reproduce due to methodological flaws and poor reporting. There is increasing attention for responsible research practices and implementation of reporting guidelines, but whether these efforts have improved the methodological quality of RCTs (e.g., lower risk of bias) is unknown. We, therefore, mapped risk-of-bias trends over time in RCT publications in relation to journal and author characteristics. Meta-information of 176,620 RCTs published between 1966 and 2018 was extracted. The risk-of-bias probability (random sequence generation, allocation concealment, blinding of patients/personnel, and blinding of outcome assessment) was assessed using a risk-of-bias machine learning tool. This tool was simultaneously validated using 63,327 human risk-of-bias assessments obtained from 17,394 RCTs evaluated in the Cochrane Database of Systematic Reviews (CDSR). Moreover, RCT registration and CONSORT Statement reporting were assessed using automated searches. Publication characteristics included the number of authors, journal impact factor (JIF), and medical discipline. The annual number of published RCTs substantially increased over 4 decades, accompanied by increases in authors (5.2 to 7.8) and institutions (2.9 to 4.8). The risk of bias remained present in most RCTs but decreased over time for allocation concealment (63% to 51%), random sequence generation (57% to 36%), and blinding of outcome assessment (58% to 52%). Trial registration (37% to 47%) and the use of the CONSORT Statement (1% to 20%) also rapidly increased. In journals with a higher impact factor (>10), the risk of bias was consistently lower with higher levels of RCT registration and the use of the CONSORT Statement. Automated risk-of-bias predictions had accuracies above 70% for allocation concealment (70.7%), random sequence generation (72.1%), and blinding of patients/personnel (79.8%), but not for blinding of outcome assessment (62.7%). In conclusion, the likelihood of bias in RCTs has generally decreased over the last decades. This optimistic trend may be driven by increased knowledge augmented by mandatory trial registration and more stringent reporting guidelines and journal requirements. Nevertheless, relatively high probabilities of bias remain, particularly in journals with lower impact factors. This emphasizes that further improvement of RCT registration, conduct, and reporting is still urgently needed.
Predicting adverse outcomes in patients with peripheral arterial disease (PAD) is a complex task owing to the heterogeneity in patient and disease characteristics. This systematic review aimed to identify prognostic factors and prognostic models to predict mortality outcomes in patients with PAD Fontaine stage I - III or Rutherford category 0 - 4.
Current cardiovascular risk assessment in people living with HIV is based on general risk assessment tools; however, whether these tools can be applied in sub-Saharan African populations has been questioned.The study aimed to assess cardiovascular risk classification of common cardiovascular disease (CVD) risk prediction models compared to the Data Collection on Adverse Events of Anti-HIV Drugs (D:A:D) 2010 and 2016 models in people living with HIV.Cardiovascular disease risk was estimated by Framingham Cardiovascular and Heart Disease (FHS-CVD, FHS-CHD), Atherosclerotic Cardiovascular Disease (ASCVD) and D:A:D 2010 and 2016 risk prediction models for HIV-infected participants of the Ndlovu Cohort Study, Limpopo, rural South Africa. Participants were classified to be at low (< 10%), moderate (10% - 20%), or high-risk (> 20%) of CVD within 10 years for general CVD and five years for D:A:D models. Kappa statistics were used to determine agreement between CVD risk prediction models. Subgroup analysis was performed according to age.The analysis comprised 735 HIV-infected individuals, predominantly women (56.7%), average age 43.9 (8.8) years. The median predicted CVD risk for D:A:D 2010 and FHS-CVD was 4% and for ASCVD and FHS-CHD models, 3%. For the D:A:D 2016 risk prediction model, the figure was 5%. High 10-year CVD risk was predicted for 2.9%, 0.5%, 0.7%, 3.1% and 6.6% of the study participants by FHS-CVD, FHS-CHD, ASCVD, and D:A:D 2010 and 2016. Kappa statistics ranged from 0.34 for ASCVD to 0.60 for FHS-CVD as compared to the D:A:D 2010 risk prediction model.Overall, predicted CVD risk is low in this population. Compared to D:A:D 2010, CVD risk estimated by the FHS-CVD model showed similar overall results for risk classification. With the exception of the D:A:D model, all other risk prediction models classified fewer people to be at high estimated CVD risk. Prospective studies are needed to develop and validate CVD risk algorithms in people living with HIV in sub-Saharan Africa.
ObjectivesIn biomedical research, spin is the overinterpretation of findings, and it is a growing concern. To date, the presence of spin has not been evaluated in prognostic model research in oncology, including studies developing and validating models for individualized risk prediction.Study Design and SettingWe conducted a systematic review, searching MEDLINE and EMBASE for oncology-related studies that developed and validated a prognostic model using machine learning published between 1st January, 2019, and 5th September, 2019. We used existing spin frameworks and described areas of highly suggestive spin practices.ResultsWe included 62 publications (including 152 developed models; 37 validated models). Reporting was inconsistent between methods and the results in 27% of studies due to additional analysis and selective reporting. Thirty-two studies (out of 36 applicable studies) reported comparisons between developed models in their discussion and predominantly used discrimination measures to support their claims (78%). Thirty-five studies (56%) used an overly strong or leading word in their title, abstract, results, discussion, or conclusion.ConclusionThe potential for spin needs to be considered when reading, interpreting, and using studies that developed and validated prognostic models in oncology. Researchers should carefully report their prognostic model research using words that reflect their actual results and strength of evidence.
As complete reporting is essential to judge the validity and applicability of multivariable prediction models, a guideline for the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) was introduced. We assessed the completeness of reporting of prediction model studies published just before the introduction of the TRIPOD statement, to refine and tailor its implementation strategy. Within each of 37 clinical domains, 10 journals with the highest journal impact factor were selected. A PubMed search was performed to identify prediction model studies published before the launch of TRIPOD in these journals (May 2014). Eligible publications reported on the development or external validation of a multivariable prediction model (either diagnostic or prognostic) or on the incremental value of adding a predictor to an existing model. We included 146 publications (84% prognostic), from which we assessed 170 models: 73 (43%) on model development, 43 (25%) on external validation, 33 (19%) on incremental value, and 21 (12%) on combined development and external validation of the same model. Overall, publications adhered to a median of 44% (25th–75th percentile 35–52%) of TRIPOD items, with 44% (35–53%) for prognostic and 41% (34–48%) for diagnostic models. TRIPOD items that were completely reported for less than 25% of the models concerned abstract (2%), title (5%), blinding of predictor assessment (6%), comparison of development and validation data (11%), model updating (14%), model performance (14%), model specification (17%), characteristics of participants (21%), model performance measures (methods) (21%), and model-building procedures (24%). Most often reported were TRIPOD items regarding overall interpretation (96%), source of data (95%), and risk groups (90%). More than half of the items considered essential for transparent reporting were not fully addressed in publications of multivariable prediction model studies. Essential information for using a model in individual risk prediction, i.e. model specifications and model performance, was incomplete for more than 80% of the models. Items that require improved reporting are title, abstract, and model-building procedures, as they are crucial for identification and external validation of prediction models.