Next Lesson - Patient Safety and Quality Control
Contents
Abstract
Evidence based practice occurs when evidence from systematic research is used as a basis for healthcare decisions to improve the quality of care given. It is commonly done through systematic reviews (overview of primary studies done using high quality methods) and meta-analysis (synthesis of the results of two or more primary studies that addressed the same hypothesis in the same way). Forest plots are the diagrammatic representation of meta-analysis, with the diamond demonstrating the pooled estimate and pooled confidence interval.
The fixed effect and random effect models exist to assess the variation between studies. Fixed effect states that all studies are aiming for one ‘true’ value, whereas the random effects model states the ‘true’ value is the mean of all the study data.
Publication bias is an issue with meta-analysis, because the selection of studies to include is biased against smaller studies that show no effector have ‘unfavourable’ results. This is tested by using a funnel plot.
Core
Evidence based practice involves the integration of individual clinical expertise (from the healthcare worker in question) with the best available external evidence from systematic research. Evidence based practice attempts to reduce the number of healthcare workers continuing to use ineffective practices or failing to take on practices that have been proved to be effective.
There are two main types of evidence:
- Systematic Review – an overview of primary studies done using explicit and reproducible methods ( method is clear and reproducible)
- Meta-analysis – a quantitative synthesis of the results of two or more primary studies that addressed the same hypothesis in the same way.
This means that a meta-analysis looks at a smaller pool of studies than a systematic review, because a meta-analysis needs the studies to be really similar, whereas a systematic review can be more broad (ie have greater clinical heterogenicity).
Problems getting evidence into practice
- Evidence exists but healthcare professionals don’t know how to access it or choose not to use it.
- Organisational systems or resources are not available to implement the change in advice following new evidence.
- Patients might not want the recommended treatment.
- NICE guidelines offering one treatment as a first choice can lead to reluctance to fund other things.
Systematic reviews are a type of literature review that collects together and analyses multiple studies around the same topic. They attempt to provide a complete summary of the current research. Systematic reviews may use meta – analyses as part of their reasoning.
They exist to: avoid or reduce bias, highlight gaps in research, address clinical uncertainty (if studies on the same topic come to different conclusions), and to assess the quality of studies and their results.
They are useful to clinicians because: they have less bias than individual studies, reduce the delay between research and implementation (because all the research is collated), save clinicians from having to investigate the research themselves, and to increase quality control and certainty.
However, it is important to note that not all systematic reviews are of equal quality. Therefore a poorly conducted systematic review can be just as harmful as a poorly conducted study. Furthermore, if a systematic review uses poor quality studies to create its’ conclusions than those conclusions will themselves be of poor quality. The phrase ‘ rubbish in, rubbish out’ sums this issue up.
Criticisms of Systematic Reviews
- It can be difficult to maintain a constant state of all studies being collated into a systematic review.
- Doctors have reduced autonomy because they have to follow the systematic reviews.
Meta-analyses are a collation of similar studies that have been collected to form one conclusion. They facilitate the collection of a large amount of study data and reduce problems of small sample size. They also allow the researchers to quantify the size of an effect and the confidence intervals surrounding this value as a pooled estimate. Basically, doing a meta-analysis means that the researchers look at all the data on one topic, work out the average result, work out the degree of uncertainty on this (confidence interval) and then can release it to the public.
Meta-analysis also have formal protocols specifying many things before they start:
- The complete set of studies to be included (important to reduce bias if more information is released later)
- Identification of the common variable between the studies
- Analysis types looking for sources of variation (like checking if one study only included men)
Forest plots are the diagrammatic form of a meta-analysis.
Diagram of a Forest Plot. All values and studies and fictional.
SimpleMed Original by Maddie Swannack
- A study’s individual odds ratio is indicated by a square. The size of the square is representative of how much weight is given to that study. In general the amount of weight given to a study is determined by the confidence interval of the study and the sample size (smaller confidence interval/ larger sample size = more weight = bigger square).
- A study’s confidence interval is represented by the horizontal lines out the side of each box.
- The y-axis represents the null hypothesis (at 1 because the odds ratios are used).
- The x-axis represents the odds ratio. For continuous outcomes, mean difference may be used instead of odds ratio, making the y axis (null hypothesis) 0 rather than 1.
- The diamond represents the pooled estimate of the odds ratio of the studies. The centre of the diamond (represented by the dotted line) represents the pooled odds ratio. The width of the diamond represents the pooled confidence interval.
- Heterogenicity between studies (studies too different)
- Fixed-effect model vs random effect model comes into play
- Sub-group analysis needs to be carried out to analyse variation
- Variable quality of studies - not all are carried out to the same standard, particularly variable between countries and based on budget-size
- Publication bias in selection of the studies - funding/moral issues that could lead to some studies remaining unpublished if they do not show the correlation they were designed to show
Variation is assessed through two methods: the fixed effect model and the random effect model.
Fixed effect model – assumes that all the studies are estimating exactly the same true effect, so all the studies are aiming for the same ‘true effect’ value. Random error from each study result is then worked out to a fixed point. No 2 studies can be exactly the same or have the same populations so the exact same true effect is a sizeable assumption in this model.
Random effect model – assumes that the studies are estimating similar (but not the same) ‘true effect’ values. As a result the ‘true value’ is the mean of these values. More widely accepted model.
Diagram showing the differences between the fixed effect (when all studies are aiming for one true effect) and random effects (when studies are aiming for their own effect value, and the true effect is a mean of these) models.
SimpleMed Original by Maddie Swannack
These different models can produce different values for things:
- The final estimate for odds ratio is often similar with both methods.
- The confidence interval is often wider with the random effects model.
- The weighting of studies is more equal in the random effects model (more weight is given to smaller studies).
Random effects modelling accounts for variation but cannot explain it. Sub-group analysis (separating the group of participants into two groups to look if that characteristic has an effect) is used to help explain the heterogenicity, which can be useful in explaining the effects of a certain intervention or exposure.
There are two main types:
- Study characteristics – year of publication (is a later study different), length of follow up (different conclusion if follow up is twice as long), % female participants (different conclusion if all female).
- Participant profile – all data is analysed separately by type of participant eg male vs female, adult vs child, white vs non-white
Issues with Studies of Variable Study Quality
Variable quality can be due to:
- Poor study design (this is easy to assess when looking at study’s results)
- Poorly designed protocols
- Poor protocol implementation (this is hard to assess after the study is done).
Some studies are more prone to bias than others:
- Randomised control trials (these are the least susceptible)
- Non-randomised control trials
- Cohort studies
- Case-control studies (these are the most susceptible to bias)
There are two possible approaches to combatting these:
- Define a basic quality standard and only accept studies that meet this criteria (ie the Cochrane Reviews only include randomised control trials).
- Score each study for quality and then:
- Incorporate quality score in the weighting process so higher quality studies have greater influence on the pooled estimate.
- Use sub-group analysis to investigate the differences eg results from poor quality studies vs results from higher quality studies.
There are many things to look at when assessing the quality of randomised control trials:
- Allocation methods (eg randomisation)
- Blinding and outcome assessment
- Patient drop out (ideally under 10%)
- Intention to treat analysis (see next article)
More than one assessor should be used when looking at the quality of studies, and they should be blinded to the results of the study, but this can be difficult as the assessors often have seen the study results before they begin.
Studies with statistically significant results or ‘favourable’ results are more likely to be published than those without, and this occurs especially with smaller studies. This means that any systematic review or meta-analysis can be flawed by this bias as some data is simply not available.
This can be combated through using a funnel plot and through checking the meta-analysis protocol for choosing the studies to include (they should check for unpublished studies on the same topic). The plot is funnel shaped as studies with larger sample sizes have smaller observed effects.
Diagram showing two funnel plots, the left side being unbiased, and the right being biased. If a funnel plot is symmetrical (it has no areas within the funnel shape that have no studies in, or all the funnel is filled easily with studies) there is no evidence of publication bias. If a funnel plot has gaps in the funnel shape (as shown by the blue circle), it shows that some studies have not been published because they show an ‘unfavourable’ and are smaller studies, therefore there is publication bias.
SimpleMed Original by Maddie Swannack
Edited by: Dr. Ben Appleby
Reviewed by: Dr. Marcus Judge
- 5226