Show simple item record

dc.contributor.authorClarke, Mike
dc.date.accessioned2010-03-12T15:36:07Z
dc.date.available2010-03-12T15:36:07Z
dc.date.issued2009
dc.identifier.citationCan you believe what you read in the papers? 2009, 10:55notTrialsen
dc.identifier.issn1745-6215
dc.identifier.pmid19607671
dc.identifier.doi10.1186/1745-6215-10-55
dc.identifier.urihttp://hdl.handle.net/10147/94169
dc.description.abstractThe number of reports of clinical trials grows by hundreds every week. However, this does not mean that people making decisions about healthcare are finding it easier to obtain reliable knowledge for these decisions. Some of the information is unreliable. Systematic reviews are helping to resolve this by bringing together the research on a topic, appraising and summarising it. But the quality of these reviews depends greatly on the quality of the studies, and this usually means the quality of their reports. If there are fundamental flaws within a study, such as the use of inappropriate 'randomisation' techniques in the context of reviews of the effects of interventions, the reviewers will not be able to fix these. Worse still, if they are not aware of underlying flaws, they might make incorrect judgements about the quality of the research in their review. A study by Wu and colleagues of 'randomised trials' from China provides a reminder of the cautious approach needed by users of scientific articles. They contacted the authors of more than 2000 research articles, which purported to be reports of randomised trials; and concluded that ten of every 11 studies claiming to be a randomised trial probably did not use random allocation. Better education of researchers, peer reviewers and editors about what is, and is not, a properly randomised trial is needed; along with better reporting of the details for how participants were allocated to the different interventions. Systematic reviewers must be cautious in making assumptions about the conduct of trials based on simple phrases about the trial methodology, rather than a full description of the methods actually used. It's not that you can't believe anything that you read in the papers, just that you cannot believe everything.
dc.language.isoenen
dc.subject.meshClinical Trials as Topic
dc.subject.meshHealth Policy
dc.subject.meshHumans
dc.subject.meshInformation Dissemination
dc.subject.meshNewspapers
dc.subject.meshPeer Review, Research
dc.titleCan you believe what you read in the papers?en
dc.contributor.departmentSchool of Nursing and Midwifery, Trinity College Dublin, Dublin 2, Ireland. mclarke@cochrane.co.uken
dc.identifier.journalTrialsen
refterms.dateFOA2018-09-03T10:37:36Z
html.description.abstractThe number of reports of clinical trials grows by hundreds every week. However, this does not mean that people making decisions about healthcare are finding it easier to obtain reliable knowledge for these decisions. Some of the information is unreliable. Systematic reviews are helping to resolve this by bringing together the research on a topic, appraising and summarising it. But the quality of these reviews depends greatly on the quality of the studies, and this usually means the quality of their reports. If there are fundamental flaws within a study, such as the use of inappropriate 'randomisation' techniques in the context of reviews of the effects of interventions, the reviewers will not be able to fix these. Worse still, if they are not aware of underlying flaws, they might make incorrect judgements about the quality of the research in their review. A study by Wu and colleagues of 'randomised trials' from China provides a reminder of the cautious approach needed by users of scientific articles. They contacted the authors of more than 2000 research articles, which purported to be reports of randomised trials; and concluded that ten of every 11 studies claiming to be a randomised trial probably did not use random allocation. Better education of researchers, peer reviewers and editors about what is, and is not, a properly randomised trial is needed; along with better reporting of the details for how participants were allocated to the different interventions. Systematic reviewers must be cautious in making assumptions about the conduct of trials based on simple phrases about the trial methodology, rather than a full description of the methods actually used. It's not that you can't believe anything that you read in the papers, just that you cannot believe everything.


Files in this item

Thumbnail
Name:
19607671.pdf
Size:
178.5Kb
Format:
PDF

This item appears in the following Collection(s)

Show simple item record