Refine
Has Fulltext
- no (3)
Document Type
- Article (2)
- Working Paper (1)
Is part of the Bibliography
- yes (3)
Keywords
- Comprehensibility (1)
- Health literacy (1)
- Internet (1)
- Patient information (1)
- Readability (1)
- conomics (1)
- open science (1)
- political science (1)
- replication (1)
- reproduction (1)
Institute
Besides their function as one of the main contact points, websites of hospitals serve as medical information portals. All patients should be able to understand medical information texts; regardless of their literacy skills and educational level. Online texts should thus have an appropriate structure to ease their comprehension.
Patient information texts on every German nonuniversity ENT hospital website (n = 125) were systematically analysed. For ten different ENT topics a representative medical information text was extracted from each website. Using objective text parameters and five established readability indices, the texts were analysed in terms of their readability and structure. Furthermore, we stratified the analysis in relation to the hospital organisation system and geographical region in Germany.
Texts from 142 internet sites could be used for the definite analysis. On average, texts consisted of 15 sentences and 237 words. Readability indices congruously showed that the analysed texts could generally only be understood by a well-educated or even academic reader.
The majority of patient information texts on German hospital websites are difficult to understand for most patients. In order to fulfil their goal of adequately informing the general population about disease, therapeutic options and the particular focal points of the clinic, a revision of most medical texts on the websites of German ENT hospitals is recommended.
Collinearity a review of methods to deal with it and a simulation study evaluating their performance
(2013)
Collinearity refers to the non independence of predictor variables, usually in a regression-type analysis. It is a common feature of any descriptive ecological data set and can be a problem for parameter estimation because it inflates the variance of regression parameters and hence potentially leads to the wrong identification of relevant predictors in a statistical model. Collinearity is a severe problem when a model is trained on data from one region or time, and predicted to another with a different or unknown structure of collinearity. To demonstrate the reach of the problem of collinearity in ecology, we show how relationships among predictors differ between biomes, change over spatial scales and through time. Across disciplines, different approaches to addressing collinearity problems have been developed, ranging from clustering of predictors, threshold-based pre-selection, through latent variable methods, to shrinkage and regularisation. Using simulated data with five predictor-response relationships of increasing complexity and eight levels of collinearity we compared ways to address collinearity with standard multiple regression and machine-learning approaches. We assessed the performance of each approach by testing its impact on prediction to new data. In the extreme, we tested whether the methods were able to identify the true underlying relationship in a training dataset with strong collinearity by evaluating its performance on a test dataset without any collinearity. We found that methods specifically designed for collinearity, such as latent variable methods and tree based models, did not outperform the traditional GLM and threshold-based pre-selection. Our results highlight the value of GLM in combination with penalised methods (particularly ridge) and threshold-based pre-selection when omitted variables are considered in the final interpretation. However, all approaches tested yielded degraded predictions under change in collinearity structure and the folk lore'-thresholds of correlation coefficients between predictor variables of |r| >0.7 was an appropriate indicator for when collinearity begins to severely distort model estimation and subsequent prediction. The use of ecological understanding of the system in pre-analysis variable selection and the choice of the least sensitive statistical approaches reduce the problems of collinearity, but cannot ultimately solve them.
This study pushes our understanding of research reliability by reproducing and replicating claims from 110 papers in leading economic and political science journals. The analysis involves computational reproducibility checks and robustness assessments. It reveals several patterns. First, we uncover a high rate of fully computationally reproducible results (over 85%). Second, excluding minor issues like missing packages or broken pathways, we uncover coding errors for about 25% of studies, with some studies containing multiple errors. Third, we test the robustness of the results to 5,511 re-analyses. We find a robustness reproducibility of about 70%. Robustness reproducibility rates are relatively higher for re-analyses that introduce new data and lower for re-analyses that change the sample or the definition of the dependent variable. Fourth, 52% of re-analysis effect size estimates are smaller than the original published estimates and the average statistical significance of a re-analysis is 77% of the original. Lastly, we rely on six teams of researchers working independently to answer eight additional research questions on the determinants of robustness reproducibility. Most teams find a negative relationship between replicators' experience and reproducibility, while finding no relationship between reproducibility and the provision of intermediate or even raw data combined with the necessary cleaning codes.