dataset
stringclasses
1 value
id
stringlengths
1
5
messages
listlengths
2
2
query-laysum
6668
[ { "role": "user", "content": "Abstract: Background\nChildhood vaccination is one of the most effective ways to prevent serious illnesses and deaths in children. However, worldwide, many children do not receive all recommended vaccinations, for several potential reasons. Vaccines might be unavailable, or parents may experience difficulties in accessing vaccination services; for instance, because of poor quality health services, distance from a health facility, or lack of money. Some parents may not accept available vaccines and vaccination services.\nOur understanding of what influences parents’ views and practices around childhood vaccination, and why some parents may not accept vaccines for their children, is still limited.\nThis synthesis links to Cochrane Reviews of the effectiveness of interventions to improve coverage or uptake of childhood vaccination.\nObjectives\n- Explore parents’ and informal caregivers’ views and practices regarding routine childhood vaccination, and the factors influencing acceptance, hesitancy, or nonacceptance of routine childhood vaccination.\n- Develop a conceptual understanding of what and how different factors reduce parental acceptance of routine childhood vaccination.\n- Explore how the findings of this review can enhance our understanding of the related Cochrane Reviews of intervention effectiveness.\nSearch methods\nWe searched MEDLINE, Embase, CINAHL, and three other databases for eligible studies from 1974 to June 2020.\nSelection criteria\nWe included studies that: utilised qualitative methods for data collection and analysis; focused on parents’ or caregivers’ views, practices, acceptance, hesitancy, or refusal of routine vaccination for children aged up to six years; and were from any setting globally where childhood vaccination is provided.\nData collection and analysis\nWe used a pre-specified sampling frame to sample from eligible studies, aiming to capture studies that were conceptually rich, relevant to the review's phenomenon of interest, from diverse geographical settings, and from a range of income-level settings. We extracted contextual and methodological data from each sampled study. We used a meta-ethnographic approach to analyse and synthesise the evidence. We assessed methodological limitations using a list of criteria used in previous Cochrane Reviews and originally based on the Critical Appraisal Skills Programme quality assessment tool for qualitative studies. We used the GRADE-CERQual (Confidence in the Evidence from Reviews of Qualitative research) approach to assess our confidence in each finding. We integrated the findings of this review with those from relevant Cochrane Reviews of intervention effectiveness. We did this by mapping whether the underlying theories or components of trial interventions included in those reviews related to or targeted the overarching factors influencing parental views and practices regarding routine childhood vaccination identified by this review.\nMain results\nWe included 145 studies in the review and sampled 27 of these for our analysis. Six studies were conducted in Africa, seven in the Americas, four in South-East Asia, nine in Europe, and one in the Western Pacific. Studies included urban and rural settings, and high-, middle-, and low-income settings.\nMany complex factors were found to influence parents’ vaccination views and practices, which we divided into four themes.\nFirstly, parents’ vaccination ideas and practices may be influenced by their broader ideas and practices surrounding health and illness generally, and specifically with regards to their children, and their perceptions of the role of vaccination within this context. Secondly, many parents’ vaccination ideas and practices were influenced by the vaccination ideas and practices of the people they mix with socially. At the same time, shared vaccination ideas and practices helped some parents establish social relationships, which in turn strengthened their views and practices around vaccination. Thirdly, parents' vaccination ideas and practices may be influenced by wider political issues and concerns, and particularly their trust (or distrust) in those associated with vaccination programmes. Finally, parents' vaccination ideas and practices may be influenced by their access to and experiences of vaccination services and their frontline healthcare workers.\nWe developed two concepts for understanding possible pathways to reduced acceptance of childhood vaccination.\nThe first concept, ‘neoliberal logic’, suggests that many parents, particularly from high-income countries, understood health and healthcare decisions as matters of individual risk, choice, and responsibility. Some parents experienced this understanding as in conflict with vaccination programmes, which emphasise generalised risk and population health. This perceived conflict led some parents to be less accepting of vaccination for their children.\nThe second concept, ‘social exclusion’, suggests that some parents, particularly from low- and middle-income countries, were less accepting of childhood vaccination due to their experiences of social exclusion. Social exclusion may damage trustful relationships between government and the public, generate feelings of isolation and resentment, and give rise to demotivation in the face of public services that are poor quality and difficult to access. These factors in turn led some parents who were socially excluded to distrust vaccination, to refuse vaccination as a form of resistance or a way to bring about change, or to avoid vaccination due to the time, costs, and distress it creates.\nMany of the overarching factors our review identified as influencing parents' vaccination views and practices were underrepresented in the interventions tested in the four related Cochrane Reviews of intervention effectiveness.\nAuthors' conclusions\nOur review has revealed that parents’ views and practices regarding childhood vaccination are complex and dynamic social processes that reflect multiple webs of influence, meaning, and logic. We have provided a theorised understanding of the social processes contributing to vaccination acceptance (or not), thereby complementing but also extending more individualistic models of vaccination acceptance. Successful development of interventions to promote acceptance and uptake of childhood vaccination will require an understanding of, and then tailoring to, the specific factors influencing vaccination views and practices of the group(s) in the target setting. The themes and concepts developed through our review could serve as a basis for gaining this understanding, and subsequent development of interventions that are potentially more aligned with the norms, expectations, and concerns of target users.\n\nGiven the provided abstract, please respond to the following query: \"How up-to-date is this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched for studies published before 3 June 2020." } ]
query-laysum
6669
[ { "role": "user", "content": "Abstract: Background\nTobacco use is highly prevalent amongst people living with HIV/AIDS (PLWHA) and has a substantial impact on morbidity and mortality.\nObjectives\nTo assess the effectiveness of interventions to motivate and assist tobacco use cessation for people living with HIV/AIDS (PLWHA), and to evaluate the risks of any harms associated with those interventions.\nSearch methods\nWe searched the Cochrane Tobacco Addiction Group's Specialised Register, Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, EMBASE, and PsycINFO in June 2015. We also searched EThOS, ProQuest, four clinical trial registries, reference lists of articles, and searched for conference abstracts using Web of Science and handsearched speciality conference databases.\nSelection criteria\nControlled trials of behavioural or pharmacological interventions for tobacco cessation for PLWHA.\nData collection and analysis\nTwo review authors independently extracted all data using a standardised electronic data collection form. They extracted data on the nature of the intervention, participants, and proportion achieving abstinence and they contacted study authors to obtain missing information. We collected data on long-term (greater than or equal to six months) and short-term (less than six months) outcomes. Where appropriate, we performed meta-analysis and estimated the pooled effects using the Mantel-Haenszel fixed-effect method. Two authors independently assessed and reported the risk of bias according to prespecified criteria.\nMain results\nWe identified 14 studies relevant to this review, of which we included 12 in a meta-analysis (n = 2087). All studies provided an intervention combining behavioural support and pharmacotherapy, and in most studies this was compared to a less intensive control, typically comprising a brief behavioural intervention plus pharmacotherapy.\nThere was moderate quality evidence from six studies for the long-term abstinence outcome, which showed no evidence of effect for more intense cessation interventions: (risk ratio (RR) 1.00, 95% confidence interval (CI) 0.72 to 1.39) with no evidence of heterogeneity (I2 = 0%). The pooled long-term abstinence was 8% in both intervention and control conditions. There was very low quality evidence from 11 studies that more intense tobacco cessation interventions were effective in achieving short-term abstinence (RR 1.51, 95% CI 1.15 to 2.00); there was moderate heterogeneity (I2 = 42%). Abstinence in the control group at short-term follow-up was 8% (n = 67/848) and in the intervention group was 13% (n = 118/937). The effect of tailoring the intervention for PLWHA was unclear. We further investigated the effect of intensity of behavioural intervention via number of sessions and total duration of contact. We failed to detect evidence of a difference in effect according to either measure of intensity, although there were few studies in each subgroup. It was not possible to perform the planned analysis of adverse events or HIV outcomes since these were not reported in more than one study.\nAuthors' conclusions\nThere is moderate quality evidence that combined tobacco cessation interventions provide similar outcomes to controls in PLWHA in the long-term. There is very low quality evidence that combined tobacco cessation interventions were effective in helping PLWHA achieve short-term abstinence. Despite this, tobacco cessation interventions should be offered to PLWHA, since even non-sustained periods of abstinence have proven benefits. Further large, well designed studies of cessation interventions for PLWHA are needed.\n\nGiven the provided abstract, please respond to the following query: \"Methods\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We reviewed the available evidence from trials to help PLWHA stop using tobacco. This evidence is correct up to June 2015. We conducted analyses of whether people were able to successfully quit tobacco use in the long-term (six months and over) and short-term (measured at less than six months)." } ]
query-laysum
6670
[ { "role": "user", "content": "Abstract: Background\nPeripheral arterial occlusive disease (PAOD) is a common cause of morbidity and mortality due to cardiovascular disease in the general population. Although numerous treatments have been adopted for patients at different disease stages, no option other than amputation is available for patients presenting with critical limb ischaemia (CLI) unsuitable for rescue or reconstructive intervention. In this regard, prostanoids have been proposed as a therapeutic alternative, with the aim of increasing blood supply to the limb with occluded arteries through their vasodilatory, antithrombotic, and anti-inflammatory effects. This is an update of a review first published in 2010.\nObjectives\nTo determine the effectiveness and safety of prostanoids in patients with CLI unsuitable for rescue or reconstructive intervention.\nSearch methods\nFor this update, the Cochrane Vascular Information Specialist searched the Specialised Register (January 2017) and the Cochrane Central Register of Controlled Trials (CENTRAL; 2017, Issue 1). In addition, we searched trials registries (January 2017) and contacted pharmaceutical manufacturers, in our efforts to identify unpublished data and ongoing trials.\nSelection criteria\nRandomised controlled trials describing the efficacy and safety of prostanoids compared with placebo or other pharmacological control treatments for patients presenting with CLI without chance of rescue or reconstructive intervention.\nData collection and analysis\nTwo review authors independently selected trials, assessed trials for eligibility and methodological quality, and extracted data. We resolved disagreements by consensus or by consultation with a third review author.\nMain results\nFor this update, 15 additional studies fulfilled selection criteria. We included in this review 33 randomised controlled trials with 4477 participants; 21 compared different prostanoids versus placebo, seven compared prostanoids versus other agents, and five conducted head-to-head comparisons using two different prostanoids.\nWe found low-quality evidence that suggests no clear difference in the incidence of cardiovascular mortality between patients receiving prostanoids and those given placebo (risk ratio (RR) 0.81, 95% confidence interval (CI) 0.41 to 1.58). We found high-quality evidence showing that prostanoids have no effect on the incidence of total amputations when compared with placebo (RR 0.97, 95% CI 0.86 to 1.09). Adverse events were more frequent with prostanoids than with placebo (RR 2.11, 95% CI 1.79 to 2.50; moderate-quality evidence). The most commonly reported adverse events were headache, nausea, vomiting, diarrhoea, flushing, and hypotension. We found moderate-quality evidence showing that prostanoids reduced rest-pain (RR 1.30, 95% CI 1.06 to 1.59) and promoted ulcer healing (RR 1.24, 95% CI 1.04 to 1.48) when compared with placebo, although these small beneficial effects were diluted when we performed a sensitivity analysis that excluded studies at high risk of bias. Additionally, we found evidence of low to very low quality suggesting the effects of prostanoids versus other active agents or versus other prostanoids because studies conducting these comparisons were few and we judged them to be at high risk of bias. None of the included studies assessed quality of life.\nAuthors' conclusions\nWe found high-quality evidence showing that prostanoids have no effect on the incidence of total amputations when compared against placebo. Moderate-quality evidence showed small beneficial effects of prostanoids for rest-pain relief and ulcer healing when compared with placebo. Additionally, moderate-quality evidence showed a greater incidence of adverse effects with the use of prostanoids, and low-quality evidence suggests that prostanoids have no effect on cardiovascular mortality when compared with placebo. None of the included studies reported quality of life measurements. The balance between benefits and harms associated with use of prostanoids in patients with critical limb ischaemia with no chance of reconstructive intervention is uncertain; therefore careful assessment of therapeutic alternatives should be considered. Main reasons for downgrading the quality of evidence were high risk of attrition bias and imprecision of effect estimates.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found that, when compared with placebo, prostanoids provided a small beneficial effect by alleviating pain in the leg at rest and improving ulcer healing. Prostanoids did not reduce deaths or the need for an amputation. We found that no studies evaluated the quality of life of people with this condition. We found insufficient evidence to compare effects of prostanoids against those of other medications or other prostanoids.\nOur findings suggest that taking prostanoids does cause harm. When 1000 patients are treated with prostanoids, on average 674 (572 to 798) will experience adverse events, compared with 319 given placebo. Adverse events usually include nausea, vomiting, diarrhoea, headache, dizziness, and flushing. More severe adverse events include low blood pressure, chest pain, and abnormalities in heart rhythm." } ]
query-laysum
6671
[ { "role": "user", "content": "Abstract: Background\nShort-stay units are hospital units that provide short-term care for selected patients. Studies have indicated that short-stay units might reduce admission rates, time of hospital stays, hospital readmissions and expenditure without compromising the quality of care. Short-stay units are often defined by a target patient category, a target function, and a target time frame. Hypothetically, short-stay units could be established as part of any department, but this review focuses on short-stay units that provide care for participants with internal medicine diseases and conditions.\nObjectives\nTo assess beneficial and harmful effects of short-stay unit hospitalisation compared with usual care in people with internal medicine diseases and conditions.\nSearch methods\nWe searched CENTRAL, MEDLINE, Embase, three other databases and two trials registers up to 13 December 2017 together with reference checking, citation searching and contact with study authors to identify additional studies. We also searched several grey literature sources and performed a forward citation search for included studies.\nSelection criteria\nWe included randomised trials and cluster-randomised trials, comparing hospitalisation in a short-stay unit with usual care (hospitalisation in a traditional hospital ward or other services). We defined a short-stay unit to be a hospital ward where the targeted length of stay in hospital for patients was five days or less. We included both multipurpose and specialised short-stay units. Participants were adults admitted to hospital with an internal medicine disease or condition. We excluded surgical, obstetric, psychiatric, gynaecological, and ambulatory participants. Trials were included irrespective of publication status, date, and language.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. Two review authors independently extracted data and assessed the risk of bias of each included trial. We measured intervention effect sizes by meta-analyses for two primary outcomes, mortality and serious adverse events, and one secondary outcome, hospital readmission. We narratively reported the following important outcomes: quality of life, activities of daily living, non-serious adverse events, and costs. We used risk ratio differences of 15% for mortality and of 20% for serious adverse events for minimal relevant clinical consideration. We rated the certainty of the evidence and the strength of recommendations of the outcomes using the GRADE approach.\nMain results\nWe included 19 records reporting on 14 randomised trials with a total of 2872 participants. One trial was ongoing. Thirteen trials evaluated short-stay unit hospitalisation for six specific conditions (acute decompensated heart failure (one trial), asthma (one trial), atrial fibrillation (one trial), chest pain (seven trials), syncope (two trials), and transient ischaemic attack (one trial)) and one trial investigated participants presenting with miscellaneous internal medicine disease and conditions. The components of the intervention differed among the trials as dictated by the trial participants' condition. All included trials were at high risk of bias.\nThe certainty of the evidence for all outcomes was very low. Consequently, it is uncertain whether hospitalisation in short-stay units compared with usual care reduces mortality (risk ratio (RR) 0.73, 95% confidence interval (CI) 0.47 to 1.15) 5 trials (seven additional trials reporting on 1299 participants reported no deaths in either group)); serious adverse events (RR 0.95, 95% CI 0.59 to 1.54; 7 trials (one additional trial with 108 participants reported no serious adverse events in either group)), and hospital readmission (RR 0.80, 95% CI 0.54 to 1.19, 8 trials (one additional trial with 424 participants did not report results for participants)). There was not enough information to confirm or refute that short-stay unit hospitalisation had relevant effects on quality of life, activities of daily living, non-serious adverse events, and costs.\nAuthors' conclusions\nOverall, the quantity and the certainty of the evidence was very low. Consequently, it is uncertain whether there are any beneficial or harmful effects of short-stay unit hospitalisation for adults with internal medicine diseases and conditions - more trials comparing the effects of short-stay units with usual care are needed. Such trials ought to be conducted with low risk of bias and low risks of random errors to improve the overall confidence in the evidence.\n\nGiven the provided abstract, please respond to the following query: \"Key messages\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We are unsure about the effect of short-stay unit hospitalisation for adults with internal medicine diseases and conditions compared to usual care. The evidence was uncertain for several important reasons; including not having enough data, differences among participants and co-interventions, and problems with the methods that the trials used that could have led to false results. We need more high-quality trials to test the impact of short-stay unit hospitalisation on individual patients and costs." } ]
query-laysum
6672
[ { "role": "user", "content": "Abstract: Background\nPressure ulcers (also known as pressure injuries, pressure sores and bed sores) are localised injuries to the skin or underlying soft tissue, or both, caused by unrelieved pressure, shear or friction. Specific kinds of beds, overlays and mattresses are widely used with the aim of preventing and treating pressure ulcers.\nObjectives\nTo summarise evidence from Cochrane Reviews that assess the effects of beds, overlays and mattresses on reducing the incidence of pressure ulcers and on increasing pressure ulcer healing in any setting and population.\nTo assess the relative effects of different types of beds, overlays and mattresses for reducing the incidence of pressure ulcers and increasing pressure ulcer healing in any setting and population.\nTo cumulatively rank the different treatment options of beds, overlays and mattresses in order of their effectiveness in pressure ulcer prevention and treatment.\nMethods\nIn July 2020, we searched the Cochrane Library. Cochrane Reviews reporting the effectiveness of beds, mattresses or overlays for preventing or treating pressure ulcers were eligible for inclusion in this overview. Two review authors independently screened search results and undertook data extraction and risk of bias assessment using the ROBIS tool. We summarised the reported evidence in an overview of reviews. Where possible, we included the randomised controlled trials from each included review in network meta-analyses. We assessed the relative effectiveness of beds, overlays and mattresses for preventing or treating pressure ulcers and their probabilities of being, comparably, the most effective treatment. We assessed the certainty of the evidence using the GRADE approach.\nMain results\nWe include six Cochrane Reviews in this overview of reviews, all at low or unclear risk of bias.\nPressure ulcer prevention: four reviews (of 68 studies with 18,174 participants) report direct evidence for 27 pairwise comparisons between 12 types of support surface on the following outcomes: pressure ulcer incidence, time to pressure ulcer incidence, patient comfort response, adverse event rates, health-related quality of life, and cost-effectiveness. Here we focus on outcomes with some evidence at a minimum of low certainty.\n(1) Pressure ulcer incidence: our overview includes direct evidence for 27 comparisons that mostly (19/27) have very low-certainty evidence concerning reduction of pressure ulcer risk. We included 40 studies (12,517 participants; 1298 participants with new ulcers) in a network meta-analysis involving 13 types of intervention. Data informing the network are sparse and this, together with the high risk of bias in most studies informing the network, means most network contrasts (64/78) yield evidence of very low certainty. There is low-certainty evidence that, compared with foam surfaces (reference treatment), reactive air surfaces (e.g. static air overlays) (risk ratio (RR) 0.46, 95% confidence interval (CI) 0.29 to 0.75), alternating pressure (active) air surfaces (e.g. alternating pressure air mattresses, large-celled ripple mattresses) (RR 0.63, 95% CI 0.42 to 0.93), and reactive gel surfaces (e.g. gel pads used on operating tables) (RR 0.47, 95% CI 0.22 to 1.01) may reduce pressure ulcer incidence. The ranking of treatments in terms of effectiveness is also of very low certainty for all interventions. It is unclear which treatment is best for preventing ulceration.\n(2) Time to pressure ulcer incidence: four reviews had direct evidence on this outcome for seven comparisons. We included 10 studies (7211 participants; 699 participants with new ulcers) evaluating six interventions in a network meta-analysis. Again, data from most network contrasts (13/15) are of very low certainty. There is low-certainty evidence that, compared with foam surfaces (reference treatment), reactive air surfaces may reduce the hazard of developing new pressure ulcers (hazard ratio (HR) 0.20, 95% CI 0.04 to 1.05). The ranking of all support surfaces for preventing pressure ulcers in terms of time to healing is uncertain.\n(3) Cost-effectiveness: this overview includes direct evidence for three comparisons. For preventing pressure ulcers, alternating pressure air surfaces are probably more cost-effective than foam surfaces (moderate-certainty evidence).\nPressure ulcer treatment: two reviews (of 12 studies with 972 participants) report direct evidence for five comparisons on: complete pressure ulcer healing, time to complete pressure ulcer healing, patient comfort response, adverse event rates, and cost-effectiveness. Here we focus on outcomes with some evidence at a minimum of low certainty.\n(1) Complete pressure ulcer healing: our overview includes direct evidence for five comparisons. There is uncertainty about the relative effects of beds, overlays and mattresses on ulcer healing. The corresponding network meta-analysis (with four studies, 397 participants) had only three direct contrasts and a total of six network contrasts. Again, most network contrasts (5/6) have very low-certainty evidence. There was low-certainty evidence that more people with pressure ulcers may heal completely using reactive air surfaces than using foam surfaces (RR 1.32, 95% CI 0.96 to 1.80). We are uncertain which surfaces have the highest probability of being the most effective (all very low-certainty evidence).\n(2) Time to complete pressure ulcer healing: this overview includes direct evidence for one comparison: people using reactive air surfaces may be more likely to have healed pressure ulcers compared with those using foam surfaces in long-term care settings (HR 2.66, 95% CI 1.34 to 5.17; low-certainty evidence).\n(3) Cost-effectiveness: this overview includes direct evidence for one comparison: compared with foam surfaces, reactive air surfaces may cost an extra 26 US dollars for every ulcer-free day in the first year of use in long-term care settings (low-certainty evidence).\nAuthors' conclusions\nCompared with foam surfaces, reactive air surfaces may reduce pressure ulcer risk and may increase complete ulcer healing. Compared with foam surfaces, alternating pressure air surfaces may reduce pressure ulcer risk and are probably more cost-effective in preventing pressure ulcers. Compared with foam surfaces, reactive gel surfaces may reduce pressure ulcer risk, particularly for people in operating rooms and long-term care settings. There are uncertainties for the relative effectiveness of other support surfaces for preventing and treating pressure ulcers, and their efficacy ranking.\nMore high-quality research is required; for example, for the comparison of reactive air surfaces with alternating pressure air surfaces. Future studies should consider time-to-event outcomes and be designed to minimise any risk of bias.\n\nGiven the provided abstract, please respond to the following query: \"What did we do?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched for Cochrane Reviews that summarised the results of all available carefully designed studies (controlled trials) evaluating different beds, mattresses and overlays in preventing and treating pressure ulcers. A Cochrane Review provides a high level of evidence on the effectiveness of healthcare interventions. We summarised the results of these reviews in a single document (called an overview of reviews).\nWe also collected studies included in these reviews and compared all available treatments at the same time in a single analysis (called network meta-analysis). We then summarised these results, and rated our confidence in the evidence, based on factors such as study methods and sizes." } ]
query-laysum
6673
[ { "role": "user", "content": "Abstract: Background\nPrimary care is an important setting in which to treat tobacco addiction. However, the rates at which providers address smoking cessation and the success of that support vary. Strategies can be implemented to improve and increase the delivery of smoking cessation support (e.g. through provider training), and to increase the amount and breadth of support given to people who smoke (e.g. through additional counseling or tailored printed materials).\nObjectives\nTo assess the effectiveness of strategies intended to increase the success of smoking cessation interventions in primary care settings.\nTo assess whether any effect that these interventions have on smoking cessation may be due to increased implementation by healthcare providers.\nSearch methods\nWe searched the Cochrane Tobacco Addiction Group's Specialized Register, the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, Embase, and trial registries to 10 September 2020.\nSelection criteria\nWe included randomized controlled trials (RCTs) and cluster-RCTs (cRCTs) carried out in primary care, including non-pregnant adults. Studies investigated a strategy or strategies to improve the implementation or success of smoking cessation treatment in primary care. These strategies could include interventions designed to increase or enhance the quality of existing support, or smoking cessation interventions offered in addition to standard care (adjunctive interventions). Intervention strategies had to be tested in addition to and in comparison with standard care, or in addition to other active intervention strategies if the effect of an individual strategy could be isolated. Standard care typically incorporates physician-delivered brief behavioral support, and an offer of smoking cessation medication, but differs across studies. Studies had to measure smoking abstinence at six months' follow-up or longer.\nData collection and analysis\nWe followed standard Cochrane methods. Our primary outcome - smoking abstinence - was measured using the most rigorous intention-to-treat definition available. We also extracted outcome data for quit attempts, and the following markers of healthcare provider performance: asking about smoking status; advising on cessation; assessment of participant readiness to quit; assisting with cessation; arranging follow-up for smoking participants. Where more than one study investigated the same strategy or set of strategies, and measured the same outcome, we conducted meta-analyses using Mantel-Haenszel random-effects methods to generate pooled risk ratios (RRs) and 95% confidence intervals (CIs).\nMain results\nWe included 81 RCTs and cRCTs, involving 112,159 participants. Fourteen were rated at low risk of bias, 44 at high risk, and the remainder at unclear risk.\nWe identified moderate-certainty evidence, limited by inconsistency, that the provision of adjunctive counseling by a health professional other than the physician (RR 1.31, 95% CI 1.10 to 1.55; I2 = 44%; 22 studies, 18,150 participants), and provision of cost-free medications (RR 1.36, 95% CI 1.05 to 1.76; I2 = 63%; 10 studies,7560 participants) increased smoking quit rates in primary care. There was also moderate-certainty evidence, limited by risk of bias, that the addition of tailored print materials to standard smoking cessation treatment increased the number of people who had successfully stopped smoking at six months' follow-up or more (RR 1.29, 95% CI 1.04 to 1.59; I2 = 37%; 6 studies, 15,978 participants).\nThere was no clear evidence that providing participants who smoked with biomedical risk feedback increased their likelihood of quitting (RR 1.07, 95% CI 0.81 to 1.41; I2 = 40%; 7 studies, 3491 participants), or that provider smoking cessation training (RR 1.10, 95% CI 0.85 to 1.41; I2 = 66%; 7 studies, 13,685 participants) or provider incentives (RR 1.14, 95% CI 0.97 to 1.34; I2 = 0%; 2 studies, 2454 participants) increased smoking abstinence rates. However, in assessing the former two strategies we judged the evidence to be of low certainty and in assessing the latter strategies it was of very low certainty. We downgraded the evidence due to imprecision, inconsistency and risk of bias across these comparisons. There was some indication that provider training increased the delivery of smoking cessation support, along with the provision of adjunctive counseling and cost-free medications. However, our secondary outcomes were not measured consistently, and in many cases analyses were subject to substantial statistical heterogeneity, imprecision, or both, making it difficult to draw conclusions.\nThirty-four studies investigated multicomponent interventions to improve smoking cessation rates. There was substantial variation in the combinations of strategies tested, and the resulting individual study effect estimates, precluding meta-analyses in most cases. Meta-analyses provided some evidence that adjunctive counseling combined with either cost-free medications or provider training enhanced quit rates when compared with standard care alone. However, analyses were limited by small numbers of events, high statistical heterogeneity, and studies at high risk of bias. Analyses looking at the effects of combining provider training with flow sheets to aid physician decision-making, and with outreach facilitation, found no clear evidence that these combinations increased quit rates; however, analyses were limited by imprecision, and there was some indication that these approaches did improve some forms of provider implementation.\nAuthors' conclusions\nThere is moderate-certainty evidence that providing adjunctive counseling by an allied health professional, cost-free smoking cessation medications, and tailored printed materials as part of smoking cessation support in primary care can increase the number of people who achieve smoking cessation. There is no clear evidence that providing participants with biomedical risk feedback, or primary care providers with training or incentives to provide smoking cessation support enhance quit rates. However, we rated this evidence as of low or very low certainty, and so conclusions are likely to change as further evidence becomes available. Most of the studies in this review evaluated smoking cessation interventions that had already been extensively tested in the general population. Further studies should assess strategies designed to optimize the delivery of those interventions already known to be effective within the primary care setting. Such studies should be cluster-randomized to account for the implications of implementation in this particular setting. Due to substantial variation between studies in this review, identifying optimal characteristics of multicomponent interventions to improve the delivery of smoking cessation treatment was challenging. Future research could use component network meta-analysis to investigate this further.\n\nGiven the provided abstract, please respond to the following query: \"What did we do?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched for studies that looked at ways to improve standard stop-smoking support within primary care, and where the treatments people received were decided at random.\nWe wanted to find out:\n- how many people were asked about their smoking and provided with advice and support;\n- how many people tried to quit smoking; and\n- how many people stopped smoking for at least six months.\nWe included evidence published to 10th September 2020." } ]
query-laysum
6674
[ { "role": "user", "content": "Abstract: Background\nRepetitive task training (RTT) involves the active practice of task-specific motor activities and is a component of current therapy approaches in stroke rehabilitation.\nObjectives\nPrimary objective: To determine if RTT improves upper limb function/reach and lower limb function/balance in adults after stroke.\n\nSecondary objectives: 1) To determine the effect of RTT on secondary outcome measures including activities of daily living, global motor function, quality of life/health status and adverse events. 2) To determine the factors that could influence primary and secondary outcome measures, including the effect of 'dose' of task practice; type of task (whole therapy, mixed or single task); timing of the intervention and type of intervention.\nSearch methods\nWe searched the Cochrane Stroke Group Trials Register (4 March 2016); the Cochrane Central Register of Controlled Trials (CENTRAL) (the Cochrane Library 2016, Issue 5: 1 October 2006 to 24 June 2016); MEDLINE (1 October 2006 to 8 March 2016); Embase (1 October 2006 to 8 March 2016); CINAHL (2006 to 23 June 2016); AMED (2006 to 21 June 2016) and SPORTSDiscus (2006 to 21 June 2016).\nSelection criteria\nRandomised/quasi-randomised trials in adults after stroke, where the intervention was an active motor sequence performed repetitively within a single training session, aimed towards a clear functional goal.\nData collection and analysis\nTwo review authors independently screened abstracts, extracted data and appraised trials. We determined the quality of evidence within each study and outcome group using the Cochrane 'Risk of bias' tool and GRADE (Grades of Recommendation, Assessment, Development and Evaluation) criteria. We did not assess follow-up outcome data using GRADE. We contacted trial authors for additional information.\nMain results\nWe included 33 trials with 36 intervention-control pairs and 1853 participants. The risk of bias present in many studies was unclear due to poor reporting; the evidence has therefore been rated 'moderate' or 'low' when using the GRADE system.\nThere is low-quality evidence that RTT improves arm function (standardised mean difference (SMD) 0.25, 95% confidence interval (CI) 0.01 to 0.49; 11 studies, number of participants analysed = 749), hand function (SMD 0.25, 95% CI 0.00 to 0.51; eight studies, number of participants analysed = 619), and lower limb functional measures (SMD 0.29, 95% CI 0.10 to 0.48; five trials, number of participants analysed = 419).\nThere is moderate-quality evidence that RTT improves walking distance (mean difference (MD) 34.80, 95% CI 18.19 to 51.41; nine studies, number of participants analysed = 610) and functional ambulation (SMD 0.35, 95% CI 0.04 to 0.66; eight studies, number of participants analysed = 525). We found significant differences between groups for both upper-limb (SMD 0.92, 95% CI 0.58 to 1.26; three studies, number of participants analysed = 153) and lower-limb (SMD 0.34, 95% CI 0.16 to 0.52; eight studies, number of participants analysed = 471) outcomes up to six months post treatment but not after six months. Effects were not modified by intervention type, dosage of task practice or time since stroke for upper or lower limb. There was insufficient evidence to be certain about the risk of adverse events.\nAuthors' conclusions\nThere is low- to moderate-quality evidence that RTT improves upper and lower limb function; improvements were sustained up to six months post treatment. Further research should focus on the type and amount of training, including ways of measuring the number of repetitions actually performed by participants. The definition of RTT will need revisiting prior to further updates of this review in order to ensure it remains clinically meaningful and distinguishable from other interventions.\n\nGiven the provided abstract, please respond to the following query: \"Review question\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "What are the effects of repeated practice of functional tasks on recovery after stroke when compared with usual care or placebo treatments?" } ]
query-laysum
6675
[ { "role": "user", "content": "Abstract: Background\nSupplemental oxygen is frequently administered to patients with acute respiratory distress syndrome (ARDS), including ARDS secondary to viral illness such as coronavirus disease 19 (COVID-19). An up-to-date understanding of how best to target this therapy (e.g. arterial partial pressure of oxygen (PaO2) or peripheral oxygen saturation (SpO2) aim) in these patients is urgently required.\nObjectives\nTo address how oxygen therapy should be targeted in adults with ARDS (particularly ARDS secondary to COVID-19 or other respiratory viruses) and requiring mechanical ventilation in an intensive care unit, and the impact oxygen therapy has on mortality, days ventilated, days of catecholamine use, requirement for renal replacement therapy, and quality of life.\nSearch methods\nWe searched the Cochrane COVID-19 Study Register, CENTRAL, MEDLINE, and Embase from inception to 15 May 2020 for ongoing or completed randomized controlled trials (RCTs).\nSelection criteria\nTwo review authors independently assessed all records in accordance with standard Cochrane methodology for study selection.\nWe included RCTs comparing supplemental oxygen administration (i.e. different target PaO2 or SpO2 ranges) in adults with ARDS and receiving mechanical ventilation in an intensive care setting. We excluded studies exploring oxygen administration in patients with different underlying diagnoses or those receiving non-invasive ventilation, high-flow nasal oxygen, or oxygen via facemask.\nData collection and analysis\nOne review author performed data extraction, which a second review author checked. We assessed risk of bias in included studies using the Cochrane 'Risk of bias' tool. We used the GRADE approach to judge the certainty of the evidence for the following outcomes; mortality at longest follow-up, days ventilated, days of catecholamine use, and requirement for renal replacement therapy.\nMain results\nWe identified one completed RCT evaluating oxygen targets in patients with ARDS receiving mechanical ventilation in an intensive care setting. The study randomized 205 mechanically ventilated patients with ARDS to either conservative (PaO2 55 to 70 mmHg, or SpO2 88% to 92%) or liberal (PaO2 90 to 105 mmHg, or SpO2≥ 96%) oxygen therapy for seven days.\nOverall risk of bias was high (due to lack of blinding, small numbers of participants, and the trial stopping prematurely), and we assessed the certainty of the evidence as very low. The available data suggested that mortality at 90 days may be higher in those participants receiving a lower oxygen target (odds ratio (OR) 1.83, 95% confidence interval (CI) 1.03 to 3.27). There was no evidence of a difference between the lower and higher target groups in mean number of days ventilated (14.0, 95% CI 10.0 to 18.0 versus 14.5, 95% CI 11.8 to 17.1); number of days of catecholamine use (8.0, 95% CI 5.5 to 10.5 versus 7.2, 95% CI 5.9 to 8.4); or participants receiving renal replacement therapy (13.7%, 95% CI 5.8% to 21.6% versus 12.0%, 95% CI 5.0% to 19.1%). Quality of life was not reported.\nAuthors' conclusions\nWe are very uncertain as to whether a higher or lower oxygen target is more beneficial in patients with ARDS and receiving mechanical ventilation in an intensive care setting. We identified only one RCT with a total of 205 participants exploring this question, and rated the risk of bias as high and the certainty of the findings as very low. Further well-conducted studies are urgently needed to increase the certainty of the findings reported here. This review should be updated when more evidence is available.\n\nGiven the provided abstract, please respond to the following query: \"Certainty of evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Our certainty (confidence) in these findings is very low as data were only available from one study that had only a small number of participants, and was stopped earlier than anticipated because of safety concerns. We are therefore unable to definitively say whether giving more or less oxygen to ARDS patients is helpful." } ]
query-laysum
6676
[ { "role": "user", "content": "Abstract: Background\nCommunication about end of life (EoL) and EoL care is critically important for providing quality care as people approach death. Such communication is often complex and involves many people (patients, family members, carers, health professionals). How best to communicate with people in the period approaching death is not known, but is an important question for quality of care at EoL worldwide. This review fills a gap in the evidence on interpersonal communication (between people and health professionals) in the last year of life, focusing on interventions to improve interpersonal communication and patient, family member and carer outcomes.\nObjectives\nTo assess the effects of interventions designed to improve verbal interpersonal communication about EoL care between health practitioners and people affected by EoL.\nSearch methods\nWe searched CENTRAL, MEDLINE, Embase, PsycINFO, and CINAHL from inception to July 2018, without language or date restrictions. We contacted authors of included studies and experts and searched reference lists to identify relevant papers. We searched grey literature sources, conference proceedings, and clinical trials registries in September 2019. Database searches were re-run in June 2021 and potentially relevant studies listed as awaiting classification or ongoing.\nSelection criteria\nThis review assessed the effects of interventions, evaluated in randomised and quasi-randomised trials, intended to enhance interpersonal communication about EoL care between patients expected to die within 12 months, their family members and carers, and health practitioners involved in their care. Patients of any age from birth, in any setting or care context (e.g. acute catastrophic injury, chronic illness), and all health professionals involved in their care were eligible. All communication interventions were eligible, as long as they included interpersonal interaction(s) between patients and family members or carers and health professionals. Interventions could be simple or complex, with one or more communication aims (e.g. to inform, skill, engage, support). Effects were sought on outcomes for patients, family and carers, health professionals and health systems, including adverse (unintended) effects.\nTo ensure this review's focus was maintained on interpersonal communication in the last 12 months of life,  we excluded studies that addressed specific decisions, shared or otherwise, and the tools involved in such decision-making. We also excluded studies focused on advance care planning (ACP) reporting ACP uptake or completion as the primary outcome. Finally, we excluded studies of communication skills training for health professionals unless patient outcomes were reported as primary outcomes.\nData collection and analysis\nStandard Cochrane methods were used, including dual review author study selection, data extraction and quality assessment of the included studies.\nMain results\nEight trials were included. All assessed intervention effects compared with usual care. Certainty of the evidence was low or very low. All outcomes were downgraded for indirectness based on the review’s purpose, and many were downgraded for imprecision and/or inconsistency. Certainty was not commonly downgraded for methodological limitations.\nA summary of the review's findings is as follows.\nKnowledge and understanding (four studies, low-certainty evidence; one study without usable data): interventions to improve communication (e.g. question prompt list, with or without patient and physician training) may have little or no effect on knowledge of illness and prognosis, or information needs and preferences, although studies were small and measures used varied across trials.\nEvaluation of the communication (six studies measuring several constructs (communication quality, patient-centredness, involvement preferences, doctor-patient relationship, satisfaction with consultation), most low-certainty evidence): across constructs there may be minimal or no effects of interventions to improve EoL communication, and there is uncertainty about effects of interventions such as a patient-specific feedback sheet on quality of communication.\nDiscussions of EoL or EoL care (six studies measuring selected outcomes, low- or very low-certainty evidence): a family conference intervention may increase duration of EoL discussions in an intensive care unit (ICU) setting, while use of a structured serious illness conversation guide may lead to earlier discussions of EoL and EoL care (each assessed by one study). We are uncertain about effects on occurrence of discussions and question asking in consultations, and there may be little or no effect on content of communication in consultations.\nAdverse outcomes or unintended effects (limited evidence): there is insufficient evidence to determine whether there are adverse outcomes associated with communication interventions  (e.g. question prompt list, family conference, structured discussions) for EoL and EoL care. Patient and/or carer anxiety was reported by three studies, but judged as confounded. No other unintended consequences, or worsening of desired outcomes, were reported.\nPatient/carer quality of life (four studies, low-certainty evidence; two without useable data): interventions to improve communication may have little or no effect on quality of life.\nHealth practitioner outcomes (three studies, low-certainty evidence; two without usable data): interventions to improve communication may have little or no effect on health practitioner outcomes (satisfaction with communication during consultation; one study); effects on other outcomes (knowledge, preparedness to communicate) are unknown.\nHealth systems impacts: communication interventions (e.g. structured EoL conversations) may have little or no effect on carer or clinician ratings of quality of EoL care (satisfaction with care, symptom management, comfort assessment, quality of care) (three studies, low-certainty evidence), or on patients' self-rated care and illness, or numbers of care goals met (one study, low-certainty evidence). Communication interventions (e.g. question prompt list alone or with nurse-led communication skills training) may slightly increase mean consultation length (two studies), but other health service impacts (e.g. hospital admissions) are unclear.\nAuthors' conclusions\nFindings of this review are inconclusive for practice. Future research might contribute meaningfully by seeking to fill gaps for populations not yet studied in trials; and to develop responsive outcome measures with which to better assess the effects of communication on the range of people involved in EoL communication episodes. Mixed methods and/or qualitative research may contribute usefully to better understand the complex interplay between different parties involved in communication, and to inform development of more effective interventions and appropriate outcome measures. Co-design of such interventions and outcomes, involving the full range of people affected by EoL communication and care, should be a key underpinning principle for future research in this area.\n\nGiven the provided abstract, please respond to the following query: \"Key messages\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We did not find enough good quality evidence to be able to say which ways of communicating about end of life (EoL) are best for the people involved. One study of a family conference intervention found that communication interventions might increase the length of EoL discussions between families and health professionals in some situations, and one other found that an intervention which used a structured conversation guide might lead to earlier discussions between patients, carers, and health professionals about EoL and EoL care. We did not find any evidence of harmful or negative effects of communication interventions, and we are uncertain about effects on outcomes like knowledge or quality of EoL care." } ]
query-laysum
6677
[ { "role": "user", "content": "Abstract: Background\nDiabetic retinopathy is a complication of diabetes in which high blood sugar levels damage the blood vessels in the retina. Sometimes new blood vessels grow in the retina, and these can have harmful effects; this is known as proliferative diabetic retinopathy. Laser photocoagulation is an intervention that is commonly used to treat diabetic retinopathy, in which light energy is applied to the retina with the aim of stopping the growth and development of new blood vessels, and thereby preserving vision.\nObjectives\nTo assess the effects of laser photocoagulation for diabetic retinopathy compared to no treatment or deferred treatment.\nSearch methods\nWe searched CENTRAL (which contains the Cochrane Eyes and Vision Group Trials Register) (2014, Issue 5), Ovid MEDLINE, Ovid MEDLINE In-Process and Other Non-Indexed Citations, Ovid MEDLINE Daily, Ovid OLDMEDLINE (January 1946 to June 2014), EMBASE (January 1980 to June 2014), the metaRegister of Controlled Trials (mRCT) (www.controlled-trials.com), ClinicalTrials.gov (www.clinicaltrials.gov) and the World Health Organization (WHO) International Clinical Trials Registry Platform (ICTRP) (www.who.int/ictrp/search/en). We did not use any date or language restrictions in the electronic searches for trials. We last searched the electronic databases on 3 June 2014.\nSelection criteria\nWe included randomised controlled trials (RCTs) where people (or eyes) with diabetic retinopathy were randomly allocated to laser photocoagulation or no treatment or deferred treatment. We excluded trials of lasers that are no longer in routine use. Our primary outcome was the proportion of people who lost 15 or more letters (3 lines) of best-corrected visual acuity (BCVA) as measured on a logMAR chart at 12 months. We also looked at longer-term follow-up of the primary outcome at two to five years. Secondary outcomes included mean best corrected distance visual acuity, severe visual loss, mean near visual acuity, progression of diabetic retinopathy, quality of life, pain, loss of driving licence, vitreous haemorrhage and retinal detachment.\nData collection and analysis\nWe used standard methods as expected by the Cochrane Collaboration. Two review authors selected studies and extracted data.\nMain results\nWe identified a large number of trials of laser photocoagulation of diabetic retinopathy (n = 83) but only five of these studies were eligible for inclusion in the review, i.e. they compared laser photocoagulation with currently available lasers to no (or deferred) treatment. Three studies were conducted in the USA, one study in the UK and one study in Japan. A total of 4786 people (9503 eyes) were included in these studies. The majority of participants in four of these trials were people with proliferative diabetic retinopathy; one trial recruited mainly people with non-proliferative retinopathy. Four of the studies evaluated panretinal photocoagulation with argon laser and one study investigated selective photocoagulation of non-perfusion areas. Three studies compared laser treatment to no treatment and two studies compared laser treatment to deferred laser treatment. All studies were at risk of performance bias because the treatment and control were different and no study attempted to produce a sham treatment. Three studies were considered to be at risk of attrition bias.\nAt 12 months there was little difference between eyes that received laser photocoagulation and those allocated to no treatment (or deferred treatment), in terms of loss of 15 or more letters of visual acuity (risk ratio (RR) 0.99, 95% confidence interval (CI) 0.89 to 1.11; 8926 eyes; 2 RCTs, low quality evidence). Longer term follow-up did not show a consistent pattern, but one study found a 20% reduction in risk of loss of 15 or more letters of visual acuity at five years with laser treatment. Treatment with laser reduced the risk of severe visual loss by over 50% at 12 months (RR 0.46, 95% CI 0.24 to 0.86; 9276 eyes; 4 RCTs, moderate quality evidence). There was a beneficial effect on progression of diabetic retinopathy with treated eyes experiencing a 50% reduction in risk of progression of diabetic retinopathy (RR 0.49, 95% CI 0.37 to 0.64; 8331 eyes; 4 RCTs, low quality evidence) and a similar reduction in risk of vitreous haemorrhage (RR 0.56, 95% CI 0.37 to 0.85; 224 eyes; 2 RCTs, low quality evidence).\nNone of the studies reported near visual acuity or patient-relevant outcomes such as quality of life, pain, loss of driving licence or adverse effects such as retinal detachment.\nWe did not plan any subgroup analyses, but there was a difference in baseline risk in participants with non-proliferative retinopathy compared to those with proliferative retinopathy. With the small number of included studies we could not do a formal subgroup analysis comparing effect in proliferative and non-proliferative retinopathy.\nAuthors' conclusions\nThis review provides evidence that laser photocoagulation is beneficial in treating proliferative diabetic retinopathy. We judged the evidence to be moderate or low, depending on the outcome. This is partly related to reporting of trials conducted many years ago, after which panretinal photocoagulation has become the mainstay of treatment of proliferative diabetic retinopathy.\nFuture Cochrane Reviews on variations in the laser treatment protocol are planned. Future research on laser photocoagulation should investigate the combination of laser photocoagulation with newer treatments such as anti-vascular endothelial growth factors (anti-VEGFs).\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found that moderate vision loss at 12 months was similar in eyes treated with laser and eyes that were not treated, but similar assessments made at a later date showed that eyes treated with laser were less likely to have suffered moderate vision loss. Treatment with laser reduced the risk of severe visual loss by over 50% at 12 months. There was a similar effect on the progression of DR. None of the studies reported patient-relevant outcomes such as pain or loss of driving licence." } ]
query-laysum
6678
[ { "role": "user", "content": "Abstract: Background\nThe World Health Organization (WHO) in 2015 stated atovaquone-proguanil can be used in travellers, and is an option in malaria-endemic areas in combination with artesunate, as an alternative treatment where first-line artemisinin-based combination therapy (ACT) is not available or effective. This review is an update of a Cochrane Review undertaken in 2005.\nObjectives\nTo assess the efficacy and safety of atovaquone-proguanil (alone and in combination with artemisinin drugs) versus other antimalarial drugs for treating uncomplicated Plasmodium falciparum malaria in adults and children.\nSearch methods\nThe date of the last trial search was 30 January 2020. Search locations for published trials included the Cochrane Infectious Diseases Group Specialized Register, CENTRAL, MEDLINE, Embase, and LILACS. To include recently published and unpublished trials, we also searched ClinicalTrials.gov, the metaRegister of Controlled Trials and the WHO International Clinical Trials Registry Platform Search Portal.\nSelection criteria\nRandomized controlled trials (RCTs) reporting efficacy and safety data for atovaquone-proguanil or atovaquone-proguanil with a partner drug compared with at least one other antimalarial drug for treating uncomplicated Plasmodium falciparum infection.\nData collection and analysis\nFor this update, two review authors re-extracted data and assessed certainty of evidence. We meta-analyzed data to calculate risk ratios (RRs) with 95% confidence intervals (CI) for treatment failures between comparisons, and for safety outcomes between and across comparisons. Outcome measures include unadjusted treatment failures and polymerase chain reaction (PCR)-adjusted treatment failures. PCR adjustment differentiates new infection from recrudescent infection.\nMain results\nSeventeen RCTs met our inclusion criteria providing 4763 adults and children from Africa, South-America, and South-East Asia. Eight trials reported PCR-adjusted data to distinguish between new and recrudescent infection during the follow-up period. In this abstract, we report only the comparisons against the three WHO-recommended antimalarials which were included within these trials.\nThere were two comparisons with artemether-lumefantrine, one trial from 2008 in Ethiopia with 60 participants had two failures with atovaquone-proguanil compared to none with artemether-lumefantrine (PCR-adjusted treatment failures at day 28). A second trial from 2012 in Colombia with 208 participants had one failure in each arm (PCR-adjusted treatment failures at day 42).\nThere was only one comparison with artesunate-amodiaquine from a 2014 trial conducted in Cameroon. There were six failures with atovaquone-proguanil at day 28 and two with artesunate-amodiaquine (PCR-adjusted treatment failures at day 28: 9.4% with atovaquone-proguanil compared to 2.9% with artesunate-amodiaquine; RR 3.19, 95% CI 0.67 to 15.22; 1 RCT, 132 participants; low-certainty evidence), although there was a similar number of PCR-unadjusted treatment failures (9 (14.1%) with atovaquone-proguanil and 8 (11.8%) with artesunate-amodiaquine; RR 1.20, 95% CI 0.49 to 2.91; 1 RCT, 132 participants; low-certainty evidence).\nThere were two comparisons with artesunate-mefloquine from a 2012 trial in Colombia and a 2002 trial in Thailand where there are high levels of multi-resistant malaria. There were similar numbers of PCR-adjusted treatment failures between groups at day 42 (2.7% with atovaquone-proguanil compared to 2.4% with artesunate-mefloquine; RR 1.15, 95% CI 0.57 to 2.34; 2 RCTs, 1168 participants; high-certainty evidence). There were also similar PCR-unadjusted treatment failures between groups (5.3% with atovaquone-proguanil compared to 6.6% with artesunate-mefloquine; RR 0.8, 95% CI 0.5 to 1.3; 1 RCT, 1063 participants; low-certainty evidence).\nWhen atovaquone-proguanil was combined with artesunate, there were fewer treatment failures with and without PCR-adjustment at day 28 (PCR-adjusted treatment failures at day 28: 2.16% with atovaquone-proguanil compared to no failures with artesunate-atovaquone-proguanil; RR 5.14, 95% CI 0.61 to 43.52; 2 RCTs, 375 participants, low-certainty evidence) and day 42 (PCR-adjusted treatment failures at day 42: 3.82% with atovaquone-proguanil compared to 2.05% with artesunate-atovaquone-proguanil (RR 1.84, 95% CI 0.95 to 3.56; 2 RCTs, 1258 participants, moderate-certainty evidence). In the 2002 trial in Thailand, there were fewer treatment failures in the artesunate-atovaquone-proguanil group compared to the atovaquone-proguanil group at day 42 with PCR-adjustment.\nWhilst there were some small differences in which adverse events were more frequent in the atovaquone-proguanil groups compared to comparator drugs, there were no recurrent associations to suggest that atovaquone-proguanil is strongly associated with any specific adverse event.\nAuthors' conclusions\nAtovaquone-proguanil was effective against uncomplicated P falciparum malaria, although in some instances treatment failure rates were between 5% and 10%. The addition of artesunate to atovaquone-proguanil may reduce treatment failure rates. Artesunate-atovaquone-proguanil and the development of parasite resistance may represent an area for further research.\n\nGiven the provided abstract, please respond to the following query: \"Key messages\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Atovaquone-proguanil is as effective for treating uncomplicated Plasmodium falciparum malaria as artesunate-mefloquine. It may be less effective than artemether-lumefantrine, artesunate-amodiaquine, and artesunate-atovaquone-proguanil, though more robust evidence is needed to confirm this. Side effects seem similar with atovaquone-proguanil." } ]
query-laysum
6679
[ { "role": "user", "content": "Abstract: Background\nThis review supersedes the original Cochrane review first published in 2008 (Huertas-Ceballos 2008).\nBetween 4% and 25% of school-aged children complain of recurrent abdominal pain (RAP) severe enough to interfere with their daily activities. No organic cause for this pain can be found on physical examination or investigation for the majority of such children. Although many children are managed by reassurance and simple measures, a large range of psychosocial interventions involving cognitive and behavioural components have been recommended.\nObjectives\nTo determine the effectiveness of psychosocial interventions for reducing pain in school-aged children with RAP.\nSearch methods\nIn June 2016 we searched CENTRAL, MEDLINE, Embase, eight other databases, and two trials registers. We also searched the references of identified studies and relevant reviews.\nSelection criteria\nRandomised controlled trials comparing psychosocial therapies with usual care, active control, or wait-list control for children and adolescents (aged 5 to 18 years) with RAP or an abdominal pain-related functional gastrointestinal disorder defined by the Rome III criteria were eligible for inclusion.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. Five review authors independently selected studies, assessed them for risk of bias, and extracted relevant data. We also assessed the quality of the evidence using the GRADE approach.\nMain results\nThis review includes 18 randomised controlled trials (14 new to this version), reported in 26 papers, involving 928 children and adolescents with RAP between the ages of 6 and 18 years. The interventions were classified into four types of psychosocial therapy: cognitive behavioural therapy (CBT), hypnotherapy (including guided imagery), yoga, and written self-disclosure. The studies were carried out in the USA, Australia, Canada, the Netherlands, Germany, and Brazil. The majority of the studies were small and short term; only two studies included more than 100 participants, and only five studies had follow-up assessments beyond six months. Small sample sizes and the degree of assessed risk of performance and detection bias in many studies led to the overall quality of the evidence being rated as low to very low for all outcomes.\nFor CBT compared to control, we found evidence of treatment success postintervention (odds ratio (OR) 5.67, 95% confidence interval (CI) 1.18 to 27.32; Z = 2.16; P = 0.03; 4 studies; 175 children; very low-quality evidence), but no evidence of treatment success at medium-term follow-up (OR 3.08, 95% CI 0.93 to 10.16; Z = 1.85; P = 0.06; 3 studies; 139 children; low-quality evidence) or long-term follow-up (OR 1.29, 95% CI 0.50 to 3.33; Z = 0.53; P = 0.60; 2 studies; 120 children; low-quality evidence). We found no evidence of effects of intervention on pain intensity scores measured postintervention (standardised mean difference (SMD) -0.33, 95% CI -0.74 to 0.08; 7 studies; 405 children; low-quality evidence), or at medium-term follow-up (SMD -0.32, 95% CI -0.85 to 0.20; 4 studies; 301 children; low-quality evidence).\nFor hypnotherapy (including studies of guided imagery) compared to control, we found evidence of greater treatment success postintervention (OR 6.78, 95% CI 2.41 to 19.07; Z = 3.63; P = 0.0003; 4 studies; 146 children; low-quality evidence) as well as reductions in pain intensity (SMD -1.01, 95% CI -1.41 to -0.61; Z = 4.97; P < 0.00001; 4 studies; 146 children; low-quality evidence) and pain frequency (SMD -1.28, 95% CI -1.84 to -0.72; Z = 4.48; P < 0.00001; 4 studies; 146 children; low-quality evidence). The only study of long-term effect reported continued benefit of hypnotherapy compared to usual care after five years, with 68% reporting treatment success compared to 20% of controls (P = 0.005).\nFor yoga therapy compared to control, we found no evidence of effectiveness on pain intensity reduction postintervention (SMD -0.31, 95% CI -0.67 to 0.05; Z = 1.69; P = 0.09; 3 studies; 122 children; low-quality evidence).\nThe single study of written self-disclosure therapy reported no benefit for pain.\nThere was no evidence of effect from the pooled analyses for any type of intervention on the secondary outcomes of school performance, social or psychological functioning, and quality of daily life.\nThere were no adverse effects for any of the interventions reported.\nAuthors' conclusions\nThe data from trials to date provide some evidence for beneficial effects of CBT and hypnotherapy in reducing pain in the short term in children and adolescents presenting with RAP. There was no evidence for the effectiveness of yoga therapy or written self-disclosure therapy. There were insufficient data to explore effects of treatment by RAP subtype.\nHigher-quality, longer-duration trials are needed to fully investigate the effectiveness of psychosocial interventions. Identifying the active components of the interventions and establishing whether benefits are sustained in the long term are areas of priority. Future research studies would benefit from employing active control groups to help minimise potential bias from wait-list control designs and to help account for therapist and intervention time.\n\nGiven the provided abstract, please respond to the following query: \"Conclusion\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Cognitive behavioural therapy and hypnotherapy warrant consideration by clinicians as part of the management strategy for children with recurrent abdominal pain. The overall quality of the evidence was low to very low. More high-quality research is needed to evaluate the particular aspects of the therapies that are effective and to establish whether benefits are maintained over time." } ]
query-laysum
6680
[ { "role": "user", "content": "Abstract: Background\nSomatoform disorders are characterised by chronic, medically unexplained physical symptoms (MUPS). Although different medications are part of treatment routines for people with somatoform disorders in clinics and private practices, there exists no systematic review or meta-analysis on the efficacy and tolerability of these medications. We aimed to synthesise to improve optimal treatment decisions.\nObjectives\nTo assess the effects of pharmacological interventions for somatoform disorders (specifically somatisation disorder, undifferentiated somatoform disorder, somatoform autonomic dysfunction, and pain disorder) in adults.\nSearch methods\nWe searched the Cochrane Depression, Anxiety and Neurosis Review Group's Specialised Register (CCDANCTR) (to 17 January 2014). This register includes relevant randomised controlled trials (RCTs) from The Cochrane Library (all years), MEDLINE (1950 to date), EMBASE (1974 to date), and PsycINFO (1967 to date). To identify ongoing trials, we searched ClinicalTrials.gov, Current Controlled Trials metaRegister, the World Health Organization International Clinical Trials Registry Platform, and the Chinese Clinical Trials Registry. For grey literature, we searched ProQuest Dissertation & Theses Database, OpenGrey, and BIOSIS Previews. We handsearched conference proceedings and reference lists of potentially relevant papers and systematic reviews and contacted experts in the field.\nSelection criteria\nWe selected RCTs or cluster RCTs of pharmacological interventions versus placebo, treatment as usual, another medication, or a combination of different medications for somatoform disorders in adults. We included people fulfilling standardised diagnostic criteria for somatisation disorder, undifferentiated somatoform disorder, somatoform autonomic dysfunction, or somatoform pain disorder.\nData collection and analysis\nOne review author and one research assistant independently extracted data and assessed risk of bias. Primary outcomes included the severity of MUPS on a continuous measure, and acceptability of treatment.\nMain results\nWe included 26 RCTs (33 reports), with 2159 participants, in the review. They examined the efficacy of different types of antidepressants, the combination of an antidepressant and an antipsychotic, antipsychotics alone, or natural products (NPs). The duration of the studies ranged between two and 12 weeks.\nOne meta-analysis of placebo-controlled studies showed no clear evidence of a significant difference between tricyclic antidepressants (TCAs) and placebo for the outcome severity of MUPS (SMD -0.13; 95% CI -0.39 to 0.13; 2 studies, 239 participants; I2 = 2%; low-quality evidence). For new-generation antidepressants (NGAs), there was very low-quality evidence showing they were effective in reducing the severity of MUPS (SMD -0.91; 95% CI -1.36 to -0.46; 3 studies, 243 participants; I2 = 63%). For NPs there was low-quality evidence that they were effective in reducing the severity of MUPS (SMD -0.74; 95% CI -0.97 to -0.51; 2 studies, 322 participants; I2 = 0%).\nOne meta-analysis showed no clear evidence of a difference between TCAs and NGAs for severity of MUPS (SMD -0.16; 95% CI -0.55 to 0.23; 3 studies, 177 participants; I2 = 42%; low-quality evidence). There was also no difference between NGAs and other NGAs for severity of MUPS (SMD -0.16; 95% CI -0.45 to 0.14; 4 studies, 182 participants; I2 = 0%).\nFinally, one meta-analysis comparing selective serotonin reuptake inhibitors (SSRIs) with a combination of SSRIs and antipsychotics showed low-quality evidence in favour of combined treatment for severity of MUPS (SMD 0.77; 95% CI 0.32 to 1.22; 2 studies, 107 participants; I2 = 23%).\nDifferences regarding the acceptability of the treatment (rate of all-cause drop-outs) were neither found between NGAs and placebo (RR 1.01, 95% CI 0.64 to 1.61; 2 studies, 163 participants; I2 = 0%; low-quality evidence) or NPs and placebo (RR 0.85, 95% CI 0.40 to 1.78; 3 studies, 506 participants; I2 = 0%; low-quality evidence); nor between TCAs and other medication (RR 1.48, 95% CI 0.59 to 3.72; 8 studies, 556 participants; I2 =14%; low-quality evidence); nor between antidepressants and the combination of an antidepressant and an antipsychotic (RR 0.80, 95% CI 0.25 to 2.52; 2 studies, 118 participants; I2 = 0%; low-quality evidence). Percental attrition rates due to adverse effects were high in all antidepressant treatments (0% to 32%), but low for NPs (0% to 1.7%).\nThe risk of bias was high in many domains across studies. Seventeen trials (65.4%) gave no information about random sequence generation and only two (7.7%) provided information about allocation concealment. Eighteen studies (69.2%) revealed a high or unclear risk in blinding participants and study personnel; 23 studies had high risk of bias relating to blinding assessors. For the comparison NGA versus placebo, there was relatively high imprecision and heterogeneity due to one outlier study. Although we identified 26 studies, each comparison only contained a few studies and small numbers of participants so the results were imprecise.\nAuthors' conclusions\nThe current review found very low-quality evidence for NGAs and low-quality evidence for NPs being effective in treating somatoform symptoms in adults when compared with placebo. There was some evidence that different classes of antidepressants did not differ in efficacy; however, this was limited and of low to very low quality. These results had serious shortcomings such as the high risk of bias, strong heterogeneity in the data, and small sample sizes. Furthermore, the significant effects of antidepressant treatment have to be balanced against the relatively high rates of adverse effects. Adverse effects produced by medication can have amplifying effects on symptom perceptions, particularly in people focusing on somatic symptoms without medical causes. We can only draw conclusions about short-term efficacy of the pharmacological interventions because no trial included follow-up assessments. For each of the comparisons where there were available data on acceptability rates (NGAs versus placebo, NPs versus placebo, TCAs versus other medication, and antidepressants versus a combination of an antidepressant and an antipsychotic), no clear differences between the intervention and comparator were found.\nFuture high-quality research should be carried out to determine the effectiveness of medications other than antidepressants, to compare antidepressants more thoroughly, and to follow-up participants over longer periods (the longest follow up was just 12 weeks). Another idea for future research would be to include other outcomes such as functional impairment or dysfunctional behaviours and cognitions as well as the classical outcomes such as symptom severity, depression, or anxiety.\n\nGiven the provided abstract, please respond to the following query: \"Who may be interested in this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "- People with long-term unexplained physical symptoms (somatoform disorders) and their family and friends.\n- Professionals working with people with somatoform disorders.\n- Professionals working in chronic pain services.\n- General practitioners." } ]
query-laysum
6681
[ { "role": "user", "content": "Abstract: Background\nAge-related macular degeneration (AMD) is a progressive, late-onset disorder of the macula affecting central vision. It is the leading cause of blindness in people over 65 years in industrialized countries. Recent epidemiologic, genetic, and pathological evidence has shown that AMD shares a number of risk factors with atherosclerosis, leading to the hypothesis that statins may exert protective effects in AMD.\nObjectives\nThe objective of this review was to examine the effectiveness of statins compared with other treatments, no treatment, or placebo in delaying the onset and progression of AMD.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL) (which contains the Cochrane Eyes and Vision Trials Register) (2016, Issue 3), Ovid MEDLINE, Ovid MEDLINE In-Process and Other Non-Indexed Citations, Ovid MEDLINE Daily, Ovid OLDMEDLINE (January 1946 to March 2016), EMBASE (January 1980 to March 2016), Latin American and Caribbean Health Sciences Literature Database (LILACS) (January 1982 to March 2016), PubMed (January 1946 to March 2016), the metaRegister of Controlled Trials (mRCT) (www.controlled-trials.com) (last searched 5 June 2014), ClinicalTrials.gov (www.clinicaltrials.gov), and the WHO International Clinical Trials Registry Platform (ICTRP) (www.who.int/ictrp/search/en). We did not use any date or language restrictions in the electronic searches for trials. We last searched the electronic databases on 31 March 2016.\nSelection criteria\nWe included randomized controlled trials (RCTs) and quasi-randomized trials that compared statins with other treatments, no treatment, or placebo in people who were diagnosed as having the early stages of AMD.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. Two review authors independently evaluated the search results against the selection criteria, abstracted data, and assessed risk of bias. We did not perform meta-analysis due to heterogeneity in the interventions and outcomes between the included studies.\nMain results\nTwo RCTs with a total of 144 participants met the selection criteria. Both trials compared simvastatin versus placebo in older people (older than 50 or 60 years) with high risk of developing AMD (drusen present on examination). Overall, we judged the quality of the evidence to be low, as we downgraded all outcomes due to limitations in the designs of the trials and insufficient outcome reporting. The larger trial, with 114 participants, was conducted in Australia and used a higher dose (40 mg daily) of simvastatin for three years. Participants and study personnel in this trial were adequately masked, however data were missing for 30% of participants at three years' follow-up. The smaller trial, with 30 participants, was conducted in Italy and used a lower dose (20 mg) of simvastatin for three months. This trial reported insufficient details to assess the risk of bias.\nNeither trial reported data for change in visual acuity. Low-quality evidence from the smaller trial, with 30 participants, did not show a statistically significant difference between the simvastatin and placebo groups in visual acuity values at three months of treatment (decimal visual acuity 0.21 ± 0.56 in simvastatin group and 0.19 ± 0.40 in placebo group) or 45 days after the completion of treatment (decimal visual acuity 0.20 ± 0.50 in simvastatin group and 0.19 ± 0.48 in placebo group). The lack of a difference in visual acuity was not explained by lens or retina status, which remained unchanged during and after the treatment period for both groups.\nPreliminary analyses of 42 participants who had completed 12 months' follow-up in the larger trial did not show a statistically significant difference between simvastatin and the placebo groups for visual acuity, drusen score, or visual function (effect estimates and confidence intervals were not available). Complete data for these outcomes at three years' follow-up were not reported. At three years, low-quality evidence showed an effect of simvastatin in slowing progression of AMD compared with placebo to be uncertain (odds ratio 0.51, 95% confidence interval 0.23 to 1.09).\nOne trial did not report adverse outcomes. The second trial reported no difference between groups in terms of adverse events such as death, muscle aches, and acute hepatitis.\nAuthors' conclusions\nEvidence from currently available RCTs is insufficient to conclude that statins have a role in preventing or delaying the onset or progression of AMD.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "AMD is a progressive disease of the macula (area in the back of the eye affecting central vision). AMD is the leading cause of blindness in people over 65 years in industrialized countries. Studies have shown that some of the factors that may lead to heart disease and strokes are the same as those that may lead to AMD. Statins are a type of drug that aims to lower blood cholesterol levels. As statins are very effective in preventing strokes, it is possible that they are also protective for AMD." } ]
query-laysum
6682
[ { "role": "user", "content": "Abstract: Background\nFast-track cardiac care is a complex intervention involving several components of care during cardiac anaesthesia and in the postoperative period, with the ultimate aim of early extubation after surgery, to reduce length of stay in the intensive care unit and in the hospital. Safe and effective fast-track cardiac care may reduce hospital costs. This is an update of a Cochrane review first published in 2003, updated in 2012 and updated now in 2016.\nObjectives\nTo determine the safety and effectiveness of fast-track cardiac care compared with conventional (not fast-track) care in adult patients undergoing cardiac surgery. Fast-track cardiac care intervention includes administration of low-dose opioid-based general anaesthesia or use of a time-directed extubation protocol, or both.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL; 2015, Issue 5), MEDLINE (January 2012 to May 2015), Embase (January 2012 to May 2015), the Cumulative Index to Nursing and Allied Health Literature (CINAHL; January 2012 to May 2015) and the Institute for Scientific Information (ISI) Web of Science (January 2012 to May 2015), along with reference lists of articles, to identify additional trials. We applied no language restrictions.\nSelection criteria\nWe included all randomized controlled trials of adult cardiac surgical patients (coronary artery bypass grafts, aortic valve replacement, mitral valve replacement) that compared fast-track cardiac care and conventional (not fast-track) care groups. We focused on the following fast-track interventions, which were designed for early extubation after surgery: administration of low-dose opioid-based general anaesthesia during cardiac surgery and use of a time-directed extubation protocol after surgery. The primary outcome was risk of mortality. Secondary outcomes included postoperative complications, reintubation within 24 hours of surgery, time to extubation, length of stay in the intensive care unit and in the hospital, quality of life after surgery and hospital costs.\nData collection and analysis\nTwo review authors independently assessed trial quality and extracted study data. We contacted study authors for additional information. We calculated a Peto odds ratio (OR) for risk of mortality and used a random-effects model to report risk ratio (RR), mean difference (MD) and 95% confidence intervals (95% CIs) for all secondary outcomes.\nMain results\nWe included 28 trials (4438 participants) in the updated review. We considered most participants to be at low to moderate risk of death after surgery. We assessed two studies as having low risk of bias and 11 studies high risk of bias. Investigators reported no differences in risk of mortality within the first year after surgery between low-dose versus high-dose opioid-based general anaesthesia groups (OR 0.53, 95% CI 0.25 to 1.12; eight trials, 1994 participants, low level of evidence) and between a time-directed extubation protocol versus usual care (OR 0.80, 95% CI 0.45 to 1.45; 10 trials, 1802 participants, low level of evidence).\nResearchers noted no significant differences between low-dose and high-dose opioid-based anaesthesia groups in the following postoperative complications: myocardial infarction (RR 0.98, 95% CI 0.48 to 1.99; eight trials, 1683 participants, low level of evidence), stroke (RR 1.17, 95% CI 0.36 to 3.78; five trials, 562 participants, low level of evidence) and tracheal reintubation (RR 1.77, 95% CI 0.38 to 8.27; five trials, 594 participants, low level of evidence).\nComparisons with usual care revealed no significant differences in the risk of postoperative complications associated with a time-directed extubation protocol: myocardial infarction (RR 0.59, 95% CI 0.27 to 1.31; eight trials, 1378 participants, low level of evidence), stroke (RR 0.85, 95% CI 0.33 to 2.16; 11 trials, 1646 participants, low level of evidence) and tracheal reintubation (RR 1.34, 95% CI 0.74 to 2.41; 12 trials, 1261 participants, low level of evidence).\nAlthough levels of heterogeneity were high, low-dose opioid anaesthesia was associated with reduced time to extubation (reduction of 4.3 to 10.5 hours, 14 trials, 2486 participants, low level of evidence) and length of stay in the intensive care unit (reduction of 0.4 to 7.0 hours, 12 trials, 1394 participants, low level of evidence). Use of a time-directed extubation protocol was associated with reduced time to extubation (reduction of 3.7 to 8.8 hours, 16 trials, 2024 participants, low level of evidence) and length of stay in the intensive care unit (reduction of 3.9 to 10.5 hours, 13 trials, 1888 participants, low level of evidence). However, these two fast-track care interventions were not associated with reduced total length of stay in the hospital (low level of evidence).\nAuthors' conclusions\nLow-dose opioid-based general anaesthesia and time-directed extubation protocols for fast-track interventions have risks of mortality and major postoperative complications similar to those of conventional (not fast-track) care, and therefore appear to be safe for use in patients considered to be at low to moderate risk. These fast-track interventions reduced time to extubation and shortened length of stay in the intensive care unit but did not reduce length of stay in the hospital.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found 28 relevant randomized controlled studies, conducted between 1994 and 2015. Most of the 4438 adults who participated in these studies were undergoing first-time elective coronary artery graft bypass or valve replacement surgery, or both. They were at low to moderate risk of death after surgery. Eighteen studies examined the use of low-dose opioid-based general anaesthesia. Sixteen studies assessed how effective the protocols were in guiding staff to remove the tube that provided breathing support within eight hours after surgery." } ]
query-laysum
6683
[ { "role": "user", "content": "Abstract: Background\nPeople undergoing major vascular surgery have an increased risk of postoperative cardiac complications. Beta-adrenergic blockers represent an important and established pharmacological intervention in the prevention of cardiac complications in people with coronary artery disease. It has been proposed that this class of drugs may reduce the risk of perioperative cardiac complications in people undergoing major non-cardiac vascular surgery.\nObjectives\nTo review the efficacy and safety of perioperative beta-adrenergic blockade in reducing cardiac or all-cause mortality, myocardial infarction, and other cardiovascular safety outcomes in people undergoing major non-cardiac vascular surgery.\nSearch methods\nThe Cochrane Peripheral Vascular Diseases Group Trials Search Co-ordinator searched the Specialised Register (January 2014) and the Cochrane Central Register of Controlled Trials (CENTRAL; 2013, Issue 12). We searched trials databases and checked reference lists of relevant articles.\nSelection criteria\nWe included prospective, randomised controlled trials of perioperative beta-adrenergic blockade of people over 18 years of age undergoing non-cardiac vascular surgery.\nData collection and analysis\nTwo review authors independently performed study selection and data extraction. We resolved disagreements through discussion. We performed meta-analysis using a fixed-effect model with odds ratios (ORs) and 95% confidence intervals (CIs).\nMain results\nWe included two studies in this review, both of which were double-blind, randomised controlled trials comparing perioperative beta-adrenergic blockade (metoprolol) with placebo, on cardiovascular outcomes in people undergoing major non-cardiac vascular surgery. We included 599 participants receiving beta-adrenergic blockers (301 participants) or placebo (298 participants). The overall quality of studies was good. However, one study did not report random sequence generation or allocation concealment techniques, indicating possible selection bias, and the other study did not report outcome assessor blinding and was possibly underpowered. It should be noted that several of the outcomes were only reported in a single study and neither of the studies reported on vascular patency/graft occlusion, which reduces the quality of evidence to moderate. There was no evidence that perioperative beta-adrenergic blockade reduced all-cause mortality (OR 0.62, 95% CI 0.03 to 15.02), cardiovascular mortality (OR 0.34, 95% CI 0.01 to 8.32), non-fatal myocardial infarction (OR 0.83, 95% CI 0.46 to 1.49; P value = 0.53), arrhythmia (OR 0.70, 95% CI 0.26 to 1.88), heart failure (OR 1.71, 95% CI 0.40 to 7.23), stroke (OR 2.67, 95% CI 0.11 to 67.08), composite cardiovascular events (OR 0.87, 95% CI 0.55 to 1.39; P value = 0.57) or re-hospitalisation at 30 days (OR 0.86, 95% CI 0.48 to 1.52). However, there was strong evidence that beta-adrenergic blockers increased the odds of intra-operative bradycardia (OR 4.97, 95% CI 3.22 to 7.65; P value < 0.00001) and intra-operative hypotension (OR 1.84, 95% CI 1.31 to 2.59; P value = 0.0005).\nAuthors' conclusions\nThis meta-analysis currently offers no clear evidence that perioperative beta-adrenergic blockade reduces postoperative cardiac morbidity and mortality in people undergoing major non-cardiac vascular surgery. There is evidence that intra-operative bradycardia and hypotension are more likely in people taking perioperative beta-adrenergic blockers, which should be weighed with any benefit.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We identified two studies that evaluated beta-blockers giving during surgery (perioperatively) in people undergoing major non-cardiac vascular surgery, with follow-up data on cardiovascular outcomes. A total of 599 participants were randomised to receive beta-blockers (301 participants) or placebo (298 participants). Both studies were double-blind (neither participants nor surgeon were aware of the treatment), randomised controlled trials evaluating the beta-blocker, metoprolol." } ]
query-laysum
6684
[ { "role": "user", "content": "Abstract: Background\nLower-limb running injuries are common. Running shoes have been proposed as one means of reducing injury risk. However, there is uncertainty as to how effective running shoes are for the prevention of injury. It is also unclear how the effects of different characteristics of running shoes prevent injury.\nObjectives\nTo assess the effects (benefits and harms) of running shoes for preventing lower-limb running injuries in adult runners.\nSearch methods\nWe searched the following databases: CENTRAL, MEDLINE, Embase, AMED, CINAHL Plus and SPORTDiscus plus trial registers WHO ICTRP and ClinicalTrials.gov. We also searched additional sources for published and unpublished trials. The date of the search was June 2021.\nSelection criteria\nWe included randomised controlled trials (RCTs) and quasi-RCTs involving runners or military personnel in basic training that either compared a) a running shoe with a non-running shoe; b) different types of running shoes (minimalist, neutral/cushioned, motion control, stability, soft midsole, hard midsole); or c) footwear recommended and selected on foot posture versus footwear not recommended and not selected on foot posture for preventing lower-limb running injuries. Our primary outcomes were number of people sustaining a lower-limb running injury and number of lower-limb running injuries. Our secondary outcomes were number of runners who failed to return to running or their previous level of running, runner satisfaction with footwear, adverse events other than musculoskeletal injuries, and number of runners requiring hospital admission or surgery, or both, for musculoskeletal injury or adverse event.\nData collection and analysis\nTwo review authors independently assessed study eligibility and performed data extraction and risk of bias assessment. The certainty of the included evidence was assessed using GRADE methodology.\nMain results\nWe included 12 trials in the analysis which included a total of 11,240 participants, in trials that lasted from 6 to 26 weeks and were carried out in North America, Europe, Australia and South Africa. Most of the evidence was low or very low certainty as it was not possible to blind runners to their allocated running shoe, there was variation in the definition of an injury and characteristics of footwear, and there were too few studies for most comparisons.\nWe did not find any trials that compared running shoes with non-running shoes.\nNeutral/cushioned versus minimalist (5 studies, 766 participants)\nNeutral/cushioned shoes may make little or no difference to the number of runners sustaining a lower-limb running injuries when compared with minimalist shoes (low-certainty evidence) (risk ratio (RR) 0.77, 95% confidence interval (CI) 0.59 to 1.01).\nOne trial reported that 67% and 92% of runners were satisfied with their neutral/cushioned or minimalist running shoes, respectively (RR 0.73, 95% CI 0.47 to 1.12). Another trial reported mean satisfaction scores ranged from 4.0 to 4.3 in the neutral/ cushioned group and 3.6 to 3.9 in the minimalist running shoe group out of a total of 5. Hence neutral/cushioned running shoes may make little or no difference to runner satisfaction with footwear (low-certainty evidence).\nMotion control versus neutral / cushioned (2 studies, 421 participants)\nIt is uncertain whether or not motion control shoes reduce the number of runners sustaining a lower-limb running injuries when compared with neutral / cushioned shoes because the quality of the evidence has been assessed as very low certainty (RR 0.92, 95% CI 0.30 to 2.81).\nSoft midsole versus hard midsole (2 studies, 1095 participants)\nSoft midsole shoes may make little or no difference to the number of runners sustaining a lower-limb running injuries when compared with hard midsole shoes (low-certainty of evidence) (RR 0.82, 95% CI 0.61 to 1.10).\nStability versus neutral / cushioned (1 study, 57 participants)\nIt is uncertain whether or not stability shoes reduce the number of runners sustaining a lower-limb running injuries when compared with neutral/cushioned shoes because the quality of the evidence has been assessed as very low certainty (RR 0.49, 95% CI 0.18 to 1.31).\nMotion control versus stability (1 study, 56 participants)\nIt is uncertain whether or not motion control shoes reduce the number of runners sustaining a lower-limb running injuries when compared with stability shoes because the quality of the evidence has been assessed as very low certainty (RR 3.47, 95% CI 1.43 to 8.40).\nRunning shoes prescribed and selected on foot posture (3 studies, 7203 participants)\nThere was no evidence that running shoes prescribed based on static foot posture reduced the number of injuries compared with those who received a shoe not prescribed based on foot posture in military recruits (Rate Ratio 1.03, 95% CI 0.94 to 1.13). Subgroup analysis confirmed these findings were consistent between males and females. Therefore, prescribing running shoes and selecting on foot posture probably makes little or no difference to lower-limb running injuries (moderate-certainty evidence).\nData were not available for all other review outcomes.\nAuthors' conclusions\nMost evidence demonstrates no reduction in lower-limb running injuries in adults when comparing different types of running shoes. Overall, the certainty of the evidence determining whether different types of running shoes influence running injury rates was very low to low, and as such we are uncertain as to the true effects of different types of running shoes upon injury rates.\nThere is no evidence that prescribing footwear based on foot type reduces running-related lower-limb injures in adults. The evidence for this comparison was rated as moderate and as such we can have more certainty when interpreting these findings. However, all three trials included in this comparison used military populations and as such the findings may differ in recreational runners.\nFuture researchers should develop a consensus definition of running shoe design to help standardise classification. The definition of a running injury should also be used consistently and confirmed via health practitioners. More researchers should consider a RCT design to increase the evidence in this area. Lastly, future work should look to explore the influence of different types or running shoes upon injury rates in specific subgroups.\n\nGiven the provided abstract, please respond to the following query: \"How up to date is this evidence?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The evidence is up to date to June 2021." } ]
query-laysum
6685
[ { "role": "user", "content": "Abstract: Background\nChordoma is a rare primary bone tumour with a high propensity for local recurrence. Surgical resection is the mainstay of treatment, but complete resection is often morbid due to tumour location. Similarly, the dose of radiotherapy (RT) that surrounding healthy organs can tolerate is frequently below that required to provide effective tumour control. Therefore, clinicians have investigated different radiation delivery techniques, often in combination with surgery, aimed to improve the therapeutic ratio.\nObjectives\nTo assess the effects and toxicity of proton and photon adjuvant radiotherapy (RT) in people with biopsy-confirmed chordoma.\nSearch methods\nWe searched CENTRAL (2021, Issue 4); MEDLINE Ovid (1946 to April 2021); Embase Ovid (1980 to April 2021) and online registers of clinical trials, and abstracts of scientific meetings up until April 2021.\nSelection criteria\nWe included adults with pathologically confirmed primary chordoma, who were irradiated with curative intent, with protons or photons in the form of fractionated RT, SRS (stereotactic radiosurgery), SBRT (stereotactic body radiotherapy), or IMRT (intensity modulated radiation therapy). We limited analysis to studies that included outcomes of participants treated with both protons and photons.\nData collection and analysis\nThe primary outcomes were local control, mortality, recurrence, and treatment-related toxicity. We followed current standard Cochrane methodological procedures for data extraction, management, and analysis. We used the ROBINS-I tool to assess risk of bias, and GRADE to assess the certainty of the evidence.\nMain results\nWe included six observational studies with 187 adult participants. We judged all studies to be at high risk of bias. Four studies were included in meta-analysis.\nWe are uncertain if proton compared to photon therapy worsens or has no effect on local control (hazard ratio (HR) 5.34, 95% confidence interval (CI) 0.66 to 43.43; 2 observational studies, 39 participants; very low-certainty evidence).\nMedian survival time ranged between 45.5 months and 66 months. We are uncertain if proton compared to photon therapy reduces or has no effect on mortality (HR 0.44, 95% CI 0.13 to 1.57; 4 observational studies, 65 participants; very low-certainty evidence).\nMedian recurrence-free survival ranged between 3 and 10 years. We are uncertain whether proton compared to photon therapy reduces or has no effect on recurrence (HR 0.34, 95% CI 0.10 to 1.17; 4 observational studies, 94 participants; very low-certainty evidence).\nOne study assessed treatment-related toxicity and reported that four participants on proton therapy developed radiation-induced necrosis in the temporal bone, radiation-induced damage to the brainstem, and chronic mastoiditis; one participant on photon therapy developed hearing loss, worsening of the seventh cranial nerve paresis, and ulcerative keratitis (risk ratio (RR) 1.28, 95% CI 0.17 to 9.86; 1 observational study, 33 participants; very low-certainty evidence). There is no evidence that protons led to reduced toxicity.\nThere is very low-certainty evidence to show an advantage for proton therapy in comparison to photon therapy with respect to local control, mortality, recurrence, and treatment related toxicity.\nAuthors' conclusions\nThere is a lack of published evidence to confirm a clinical difference in effect with either proton or photon therapy for the treatment of chordoma. As radiation techniques evolve, multi-institutional data should be collected prospectively and published, to help identify persons that would most benefit from the available radiation treatment techniques.\n\nGiven the provided abstract, please respond to the following query: \"Authors' conclusions\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "From the currently available evidence, we do not know if proton therapy, used for adults with chordoma, is associated with clinically appreciable benefits and acceptable harms.\nThe existing research on this question is solely of non-randomised design, and is underpowered, with substantial imprecision. Any estimates of effect based on the existing insufficient evidence is very uncertain and is likely to change with future research." } ]
query-laysum
6686
[ { "role": "user", "content": "Abstract: Background\nTo limit long-term sick leave and associated consequences, insurers, healthcare providers and employers provide programmes to facilitate disabled people's return to work. These programmes include a variety of coordinated and individualised interventions. Despite the increasing popularity of such programmes, their benefits remain uncertain. We conducted a systematic review to determine the long-term effectiveness of return-to-work coordination programmes compared to usual practice in workers at risk for long-term disability.\nObjectives\nTo assess the effects of return-to-work coordination programmes versus usual practice for workers on sick leave or disability.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL; 2016, Issue 11), MEDLINE, Embase, CINAHL and PsycINFO up to 1 November 2016.\nSelection criteria\nWe included randomised controlled trials (RCTs) that enrolled workers absent from work for at least four weeks and randomly assigned them to return-to-work coordination programmes or usual practice.\nData collection and analysis\nTwo review authors independently screened titles, abstracts and full-text articles for study eligibility; extracted data; and assessed risk of bias from eligible trials. We contacted authors for additional data where required. We conducted random-effects meta-analyses and used the GRADE approach to rate the quality of the evidence.\nMain results\nWe identified 14 studies from nine countries that enrolled 12,568 workers. Eleven studies focused on musculoskeletal problems, two on mental health and one on both. Most studies (11 of 14) followed workers 12 months or longer. Risk of bias was low in 10 and high in 4 studies, but findings were not sensitive to their exclusion.\nWe found no benefits for return-to-work coordination programmes on return-to-work outcomes.\nFor short-term follow-up of six months, we found no effect on time to return to work (hazard ratio (HR) 1.32, 95% confidence interval (CI) 0.93 to 1.88, low-quality evidence), cumulative sickness absence (mean difference (MD) −16.18 work days per year, 95% CI −32.42 to 0.06, moderate-quality evidence), the proportion of participants at work at end of the follow-up (risk ratio (RR) 1.06, 95% CI 0.86 to 1.30, low-quality evidence) or on the proportion of participants who had ever returned to work, that is, regardless of whether they had remained at work until last follow-up (RR 0.87, 95% CI 0.63 to 1.19, very low-quality evidence).\nFor long-term follow-up of 12 months, we found no effect on time to return to work (HR 1.25, 95% CI 0.95 to 1.66, low-quality evidence), cumulative sickness absence (MD −14.84 work days per year, 95% CI −38.56 to 8.88, low-quality evidence), the proportion of participants at work at end of the follow-up (RR 1.06, 95% CI 0.99 to 1.15, low-quality evidence) or on the proportion of participants who had ever returned to work (RR 1.03, 95% CI 0.97 to 1.09, moderate-quality evidence).\nFor very long-term follow-up of longer than 12 months, we found no effect on time to return to work (HR 0.93, 95% CI 0.74 to 1.17, low-quality evidence), cumulative sickness absence (MD 7.00 work days per year, 95% CI −15.17 to 29.17, moderate-quality evidence), the proportion of participants at work at end of the follow-up (RR 0.94, 95% CI 0.82 to 1.07, low-quality evidence) or on the proportion of participants who had ever returned to work (RR 0.95, 95% CI 0.88 to 1.02, low-quality evidence).\nWe found only small benefits for return-to-work coordination programmes on patient-reported outcomes. All differences were below the minimal clinically important difference (MID).\nAuthors' conclusions\nOffering return-to-work coordination programmes for workers on sick leave for at least four weeks results in no benefits when compared to usual practice. We found no significant differences for the outcomes time to return to work, cumulative sickness absence, the proportion of participants at work at end of the follow-up or the proportion of participants who had ever returned to work at short-term, long-term or very long-term follow-up. For patient-reported outcomes, we found only marginal effects below the MID. The quality of the evidence ranged from very low to moderate across all outcomes.\n\nGiven the provided abstract, please respond to the following query: \"What is the aim of this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The aim of this review was to compare the effect of return-to-work coordination programmes to usual practice in workers on sick leave or disability." } ]
query-laysum
6687
[ { "role": "user", "content": "Abstract: Background\nRegional analgesia is more effective than conventional analgesia for controlling pain and may facilitate rehabilitation after large joint replacement in the short term. It remains unclear if regional anaesthesia improves functional outcomes after joint replacement beyond three months after surgery.\nObjectives\nTo assess the effects of regional anaesthesia and analgesia on long-term functional outcomes 3, 6 and 12 months after elective major joint (knee, shoulder and hip) replacement surgery.\nSearch methods\nWe performed an electronic search of several databases (CENTRAL, MEDLINE, EMBASE, CINAHL), and handsearched reference lists and conference abstracts. We updated our search in June 2015.\nSelection criteria\nWe included randomized controlled trials (RCTs) comparing regional analgesia versus conventional analgesia in patients undergoing total shoulder, hip or knee replacement. We included studies that reported a functional outcome with a follow-up of at least three months after surgery.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. We contacted study authors for additional information.\nMain results\nWe included six studies with 350 participants followed for at least three months. All of these studies enrolled participants undergoing total knee replacement. Studies were at least partially blinded. Three studies had a high risk of performance bias and one a high risk of attrition bias, but the risk of bias was otherwise unclear or low.\nOnly one study assessed joint function using a global score. Due to heterogeneity in outcome and reporting, we could only pool three out of six RCTs, with range of motion assessed at three months after surgery used as a surrogate for joint function. All studies had a high risk of detection bias. Using the random-effects model, there was no statistically significant difference between the experimental and control groups (mean difference 3.99 degrees, 95% confidence interval (CI) − 2.23 to 10.21; P value = 0.21, 3 studies, 140 participants, very low quality evidence).\nWe did not perform further analyses because immediate adverse effects were not part of the explicit outcomes of any of these typically small studies, and long-term adverse events after regional anaesthesia are rare.\nNone of the included studies elicited or reported long-term adverse effects like persistent nerve damage.\nAuthors' conclusions\nMore high-quality studies are needed to establish the effects of regional analgesia on function after major joint replacement, as well as on the risk of adverse events (falls).\n\nGiven the provided abstract, please respond to the following query: \"Quality of the evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We deemed the quality of the evidence to be very low, as the included studies were not considered to be of high calibre and included few participants." } ]
query-laysum
6688
[ { "role": "user", "content": "Abstract: Background\nThis is an updated version of the original Cochrane Review published in Issue 10, 2013.\nExtramammary Paget's disease is a rare form of superficial skin cancer. The most common site of involvement is the vulva. It is seen mainly in postmenopausal white women. Paget's disease of the vulva often spreads in an occult fashion, with margins extending beyond the apparent edges of the lesion. There is a range of interventions from surgical to non-invasive techniques or treatments. The challenges of interventions are to remove or treat disease that may not be visible, without overtreatment and with minimisation of morbidity from radical surgery. There is little consensus regarding treatment. Surgery, by default, is the most common treatment, but it is challenging to excise the disease adequately, and recurrence is common, leading to repeated operations, and destruction of anatomy. Alternative treatments of photodynamic therapy, laser therapy, radiotherapy, topical treatments or even chemotherapy have been mooted, and it is important to evaluate the available evidence. It is essential to assess whether newer cell-specific treatments, such as photodynamic therapy and imiquimod, can reduce the need for radical surgery.\nObjectives\nTo evaluate the benefits and harms of different treatment modalities for the management of Paget's disease of the vulva.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE (via Ovid) and Embase (via Ovid) up to 8 May 2018. We also searched registers of clinical trials, abstracts of scientific meetings and reference lists of review articles.\nSelection criteria\nWe searched for randomised controlled trials (RCTs) and well-designed non-randomised studies that compared different interventions in women with Paget's disease of the vulva,\nData collection and analysis\nTwo review authors independently assessed whether potentially relevant studies met the inclusion criteria. We found no trials and, therefore, analysed no data.\nMain results\nThe search for the original version of the review identified 635 unique references. We found 31 references (which reported on 30 studies) in full text after inspection of titles and abstracts, but we excluded them all as they did not meet the inclusion criteria. However, we have included a comprehensive narrative account of studies where we identified an analysis of more than 10 women, as this forms the only evidence base in this rare disease. Surgery continues to be the mainstay of treatment in the current literature, with other treatments limited to case reports or treatment of inoperable or recurrent disease.\nThis update between September 2013 and May 2018 identified 35 new studies. None of these met the inclusion criteria. There was only one prospective study of 5% imiquimod in recurrent Paget's disease of the vulva, which although of good quality only included eight women.\nAuthors' conclusions\nSince the last version of the review was published there are many more cases in the literature reporting a clinical response to 5% imiquimod cream. There is one prospective study of eight women treated with 5% imiquimod for recurrent Paget's disease of the vulva, and one prospective trial of 20 women was due to be reported. This increasing evidence for the safety and efficacy of 5% imiquimod will be helpful for women and clinicians alike. Ideally, a multicentre RCT of reasonable size is needed, but ongoing publications of high-quality non-randomised prospective studies will enhance the current available literature.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Extramammary Paget’s disease of the vulva is a rare type of superficial skin cancer. It is most common in postmenopausal white women. It is an intraepithelial (layer of cells that forms the surface or lining of an organ) condition that can present as white and red scaly areas on the vulva that may be itchy and painful. The diagnosis is made by examination and tissue sampling. Abnormal cells often extend outside the clinically abnormal area, so some studies suggest frozen section at time of surgery, where a pathologist can give a rapid report of small biopsies to say whether the skin is involved with Paget's or not. Other treatments include: topical medication, such as imiquimod (self-applied cream); radiotherapy; chemotherapy; photodynamic therapy (form of phototherapy using light-sensitive compounds that are exposed selectively to light, whereupon they become toxic to targeted cancerous and other diseased cells); laser therapy; or a combination of these approaches. The challenges of interventions are to remove or treat disease that may not be visible, without overtreatment. Avoiding the long-term complications of radical surgery, such as pain and scarring, a feeling of mutilation and loss of femininity, is very important to women. Surgery is still the most common treatment, but it is challenging to remove the disease completely, and recurrence is common, leading to repeated operations and mutilation of the vulva. The aim of this review was to evaluate the benefits and harms of different treatments for Paget's disease of the vulva." } ]
query-laysum
6689
[ { "role": "user", "content": "Abstract: Background\nBone marrow failure disorders include a heterogenous group of disorders, of which myelodysplastic syndrome (MDS), forms the largest subgroup. MDS is predominantly a disease of the elderly, with many elderly people managed conservatively with regular allogeneic red blood cell (RBC) transfusions to treat their anaemia. However, RBC transfusions are not without risk. Despite regular transfusions playing a central role in treating such patients, the optimal RBC transfusion strategy (restrictive versus liberal) is currently unclear.\nObjectives\nTo assess the efficacy and safety of a restrictive versus liberal red blood cell transfusion strategy for patients with myelodysplasia, acquired aplastic anaemia, and other inherited bone marrow failure disorders.\nSearch methods\nWe searched for randomised controlled trials (RCTs) in the Cochrane Central Register of Controlled Trials (CENTRAL) (The Cochrane Library 2015, Issue 4), Ovid MEDLINE (from 1946), Ovid EMBASE (from 1974), EBSCO CINAHL (from 1937), the Transfusion Evidence Library (from 1980) and ongoing trial databases to 26th May 2015.\nSelection criteria\nRCTs including patients with long-term bone marrow failure disorders that require allogeneic blood transfusion, who are not being actively treated with a haematopoietic stem cell transplant, or intensive chemotherapy.\nData collection and analysis\nWe used standard Cochrane review methodology. One author initially screened all references, and excluded any that were clearly irrelevant or duplicates. Two authors then independently screened all abstracts of articles, identified by the review search strategy, for relevancy. Two authors independently assessed the full text of all potentially relevant articles for eligibility, completed the data extraction and assessed the studies for risk of bias using The Cochrane Collaboration’s ’Risk of bias’ tool.\nMain results\nWe included one trial (13 participants) and identified three ongoing trials that assess RBC transfusion strategies in people with MDS.\nThe quality of the evidence was very low across different outcomes according to GRADE methodology.\nThe one included study randomised participants to a restrictive [haemoglobin (Hb) transfusion trigger < 72 g/L, 8 participants] or liberal [Hb trigger < 96 g/L, 5 participants] transfusion policy. There was insufficient evidence to determine a difference in all-cause mortality (1 RCT; 13 participants; RR 0.13, 95% CI 0.01 to 2.32; very low quality evidence). There was insufficient evidence to determine a difference in the number of red blood cell transfusions (1 RCT; 13 participants; 1.8 units per patient per month in the liberal group, compared to 0.8 in the restrictive arm, no standard deviation was reported; very low quality evidence). There were no anaemia-related complications reported (cardiac failure) and no reported effect on activity levels (no statistics provided). The study did not report: mortality due to bleeding/infection/transfusion reactions or iron overload, quality of life, frequency and length of hospital admissions, serious infections (requiring admission to hospital), or serious bleeding (e.g. WHO/CTCAE grade 3 (or equivalent) or above).\nAuthors' conclusions\nThis review indicates that there is currently a lack of evidence for the recommendation of a particular transfusion strategy for bone marrow failure patients undergoing supportive treatment only. The one RCT included in this review was only published as an abstract and contained only 13 participants. Further randomised trials with robust methodology are required to develop the optimal transfusion strategy for such patients, particularly as the incidence of the main group of bone marrow failure disorders, MDS, rises with an ageing population.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The bone marrow is where many types of blood cells are produced. Red blood cells are necessary to bring oxygen to all parts of the body, white blood cells fight against infection, and platelets in the blood help to form clots and prevent bleeding. Bone marrow failure can have different causes and can happen at birth or later in life, and may result in too few of any or all of the three types of blood cells in the body. Too few red blood cells causes a low haemoglobin level, called anaemia, which may result in poor delivery of oxygen to the body. This can cause shortness of breath, tiredness and has a significant impact on quality of life.\nRegular red blood cell transfusions are used to support patients and improve their quality of life when treatments that might cure the condition are not an option. However, there are certain risks involved with regular use of red blood cell transfusions, for example, transfusion reactions, transfusion-transmitted infections and iron overload.\nCurrently it is not clear if the best transfusion policy is a restrictive or liberal one." } ]
query-laysum
6690
[ { "role": "user", "content": "Abstract: Background\nMyasthenia is a condition in which neuromuscular transmission is affected by antibodies against neuromuscular junction components (autoimmune myasthenia gravis, MG; and neonatal myasthenia gravis, NMG) or by defects in genes for neuromuscular junction proteins (congenital myasthenic syndromes, CMSs). Clinically, some individuals seem to benefit from treatment with ephedrine, but its effects and adverse effects have not been systematically evaluated.\nObjectives\nTo assess the effects and adverse effects of ephedrine in people with autoimmune MG, transient neonatal MG, and the congenital myasthenic syndromes.\nSearch methods\nOn 17 November 2014, we searched the Cochrane Neuromuscular Disease Group Specialized Register, the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE and EMBASE. We also searched reference lists of articles, conference proceedings of relevant conferences, and prospective trial registers. In addition, we contacted manufacturers and researchers in the field.\nSelection criteria\nWe considered randomised controlled trials (RCTs) and quasi-RCTs comparing ephedrine as a single or add-on treatment with any other active treatment, placebo, or no treatment in adults or children with autoimmune MG, NMG, or CMSs.\nData collection and analysis\nTwo review authors independently assessed study design and quality, and extracted data. We contacted study authors for additional information. We collected information on adverse effects from included articles, and contacted authors.\nMain results\nWe found no RCTs or quasi-RCTs, and therefore could not establish the effect of ephedrine on MG, NMG and CMSs. We describe the results of 53 non-randomised studies narratively in the Discussion section, including observations of endurance, muscle strength and quality of life. Effects may differ depending on the type of myasthenia. Thirty-seven studies were in participants with CMS, five in participants with MG, and in 11 the precise form of myasthenia was unknown. We found no studies for NMG. Reported adverse effects included tachycardia, sleep disturbances, nervousness, and withdrawal symptoms.\nAuthors' conclusions\nThere was no evidence available from RCTs or quasi-RCTs, but some observations from non-randomised studies are available. There is a need for more evidence from suitable forms of prospective RCTs, such as series of n-of-one RCTs, that use appropriate and validated outcome measures.\n\nGiven the provided abstract, please respond to the following query: \"Review question\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We reviewed the evidence about the effect of ephedrine in adults and children with myasthenia gravis (MG), neonatal myasthenia and the congenital myasthenic syndromes (CMSs)." } ]
query-laysum
6691
[ { "role": "user", "content": "Abstract: Background\nInfective endocarditis is a microbial infection of the endocardial surface of the heart. Antibiotics are the cornerstone of treatment, but due to the differences in presentation, populations affected, and the wide variety of micro-organisms that can be responsible, their use is not standardised. This is an update of a review previously published in 2016.\nObjectives\nTo assess the existing evidence about the clinical benefits and harms of different antibiotics regimens used to treat people with infective endocarditis.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, Embase Classic and Embase, LILACS, CINAHL, and the Conference Proceedings Citation Index - Science on 6 January 2020. We also searched three trials registers and handsearched the reference lists of included papers. We applied no language restrictions.\nSelection criteria\nWe included randomised controlled trials (RCTs) assessing the effects of antibiotic regimens for treating definitive infective endocarditis diagnosed according to modified Duke's criteria. We considered all-cause mortality, cure rates, and adverse events as the primary outcomes. We excluded people with possible infective endocarditis and pregnant women.\nData collection and analysis\nTwo review authors independently performed study selection, 'Risk of bias' assessment, and data extraction in duplicate. We constructed 'Summary of findings' tables and used GRADE methodology to assess the quality of the evidence. We described the included studies narratively.\nMain results\nSix small RCTs involving 1143 allocated/632 analysed participants met the inclusion criteria of this first update. The included trials had a high risk of bias. Three trials were sponsored by drug companies. Due to heterogeneity in outcome definitions and different antibiotics used data could not be pooled.\nThe included trials compared miscellaneous antibiotic schedules having uncertain effects for all of the prespecified outcomes in this review. Evidence was either low or very low quality due to high risk of bias and very low number of events and small sample size.\nThe results for all-cause mortality were as follows: one trial compared quinolone (levofloxacin) plus standard treatment (antistaphylococcal penicillin (cloxacillin or dicloxacillin), aminoglycoside (tobramycin or netilmicin), and rifampicin) versus standard treatment alone and reported 8/31 (26%) with levofloxacin plus standard treatment versus 9/39 (23%) with standard treatment alone; risk ratio (RR) 1.12, 95% confidence interval (CI) 0.49 to 2.56. One trial compared fosfomycin plus imipenem 3/4 (75%) versus vancomycin 0/4 (0%) (RR 7.00, 95% CI 0.47 to 103.27), and one trial compared partial oral treatment 7/201 (3.5%) versus conventional intravenous treatment 13/199 (6.53%) (RR 0.53, 95% CI 0.22 to 1.31).\nThe results for rates of cure with or without surgery were as follows: one trial compared daptomycin versus low-dose gentamicin plus an antistaphylococcal penicillin (nafcillin, oxacillin, or flucloxacillin) or vancomycin and reported 9/28 (32.1%) with daptomycin versus 9/25 (36%) with low-dose gentamicin plus antistaphylococcal penicillin or vancomycin; RR 0.89, 95% CI 0.42 to 1.89. One trial compared glycopeptide (vancomycin or teicoplanin) plus gentamicin with cloxacillin plus gentamicin (13/23 (56%) versus 11/11 (100%); RR 0.59, 95% CI 0.40 to 0.85). One trial compared ceftriaxone plus gentamicin versus ceftriaxone alone (15/34 (44%) versus 21/33 (64%); RR 0.69, 95% CI 0.44 to 1.10), and one trial compared fosfomycin plus imipenem versus vancomycin (1/4 (25%) versus 2/4 (50%); RR 0.50, 95% CI 0.07 to 3.55).\nThe included trials reported adverse events, the need for cardiac surgical interventions, and rates of uncontrolled infection, congestive heart failure, relapse of endocarditis, and septic emboli, and found no conclusive differences between groups (very low-quality evidence). No trials assessed quality of life.\nAuthors' conclusions\nThis first update confirms the findings of the original version of the review. Limited and low to very low-quality evidence suggests that the comparative effects of different antibiotic regimens in terms of cure rates or other relevant clinical outcomes are uncertain. The conclusions of this updated Cochrane Review were based on few RCTs with a high risk of bias. Accordingly, current evidence does not support or reject any regimen of antibiotic therapy for the treatment of infective endocarditis.\n\nGiven the provided abstract, please respond to the following query: \"Quality of evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The confidence in the results of this review is low to very low. The included studies had limitations in the way they were designed and performed, and three studies were sponsored by the manufacturer of the medicine that was assessed. Moreover, the limited number of people included in the studies led to uncertain results. Larger studies are required to provide more information about the best antibiotic regimens to treat people with infective endocarditis." } ]
query-laysum
6692
[ { "role": "user", "content": "Abstract: Background\nMagnetic seizure therapy (MST) is a potential alternative to electroconvulsive therapy (ECT). Reports to date on use of MST for patients with treatment-resistant depression (TRD) are limited.\nObjectives\nTo evaluate the effects of MST in comparison with sham-MST, antidepressant, and other forms of electric or magnetic treatment for adults with TRD.\nSearch methods\nIn March 2020, we searched a wide range of international electronic sources for published, unpublished, and ongoing studies. We handsearched the reference lists of all included studies and relevant systematic reviews and conference proceedings of the Annual Meeting of the American College of Neuropsychopharmacology (ACNP), the Annual Scientific Convention and Meeting, and the Annual Meeting of the European College of Neuropsychopharmacology (ECNP) to identify additional studies.\nSelection criteria\nAll randomised clinical trials (RCTs) focused on MST for adults with TRD.\nData collection and analysis\nTwo review authors extracted data independently. For binary outcomes, we calculated risk ratios (RRs) and 95% confidence intervals (CIs). For continuous data, we estimated mean differences (MDs) between groups and 95% CIs. We employed a random-effects model for analyses. We assessed risk of bias for included studies and created a 'Summary of findings' table using the GRADE approach. Our main outcomes of interest were symptom severity, cognitive function, suicide, quality of life, social functioning, dropout for any reason, serious adverse events, and adverse events that led to discontinuation of treatment.\nMain results\nWe included three studies (65 participants) comparing MST with ECT. Two studies reported depressive symptoms with the Hamilton Rating Scale for Depression (HAMD). However, in one study, the data were skewed and there was an imbalance in baseline characteristics. Analysis of these two studies showed no clear differences in depressive symptoms between treatment groups (MD 0.71, 95% CI -2.23 to 3.65; 2 studies, 40 participants; very low-certainty evidence). Two studies investigated multiple domains of cognitive function. However most of the outcomes were not measured by validated neuropsychological tests, and many of the data suffered from unbalanced baseline and skewed distribution. Analysis of immediate memory performance measured by the Wechsler Memory Scale showed no clear differences between treatment groups (MD 0.40, 95% CI -4.16 to 4.96; 1 study, 20 participants; very low-certainty evidence). Analysis of delayed memory performance measured by the Wechsler Memory Scale also showed no clear differences between treatment groups (MD 2.57, 95% CI -2.39 to 7.53; 1 study, 20 participants; very low-certainty evidence). Only one study reported quality of life, but the data were skewed and baseline data were unbalanced across groups. Analysis of quality of life showed no clear differences between treatment groups (MD 14.86, 95% CI -42.26 to 71.98; 1 study, 20 participants; very low-certainty evidence). Only one study reported dropout and adverse events that led to discontinuation of treatment. Analysis of reported data showed no clear differences between treatment groups for this outcome (RR 1.38, 95% CI 0.28 to 6.91; 1 study, 25 participants; very low-certainty evidence). Adverse events occurred in only two participants who received ECT (worsening of preexisting coronary heart disease and a cognitive adverse effect). None of the included studies reported outcomes on suicide and social functioning. No RCTs comparing MST with other treatments were identified.\nAuthors' conclusions\nEvidence regarding effects of MST on patients with TRD is currently insufficient. Our analyses of available data did not reveal clearly different effects between MST and ECT. We are uncertain about these findings because of risk of bias and imprecision of estimates. Large, long, well-designed, and well-reported trials are needed to further examine the effects of MST.\n\nGiven the provided abstract, please respond to the following query: \"Review question\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Whether magnetic seizure therapy (MST) is effective and acceptable to treat treatment-resistant depression (TRD)." } ]
query-laysum
6693
[ { "role": "user", "content": "Abstract: Background\nThis is an updated version of the original Cochrane Review published in Issue 12, 2016. This review is one in a series of Cochrane Reviews investigating pair-wise monotherapy comparisons.\nEpilepsy is a common neurological condition in which abnormal electrical discharges from the brain cause recurrent unprovoked seizures. It is believed that with effective drug treatment, up to 70% of individuals with active epilepsy have the potential to become seizure-free and go into long-term remission shortly after starting drug therapy, the majority of which may be able to achieve remission with a single antiepileptic drug (AED).\nThe correct choice of first-line AED for individuals with newly diagnosed seizures is of great importance and should be based on the highest-quality evidence available regarding the potential benefits and harms of various treatments for an individual.\nTopiramate and carbamazepine are commonly used AEDs. Performing a synthesis of the evidence from existing trials will increase the precision of results of outcomes relating to efficacy and tolerability, and may help inform a choice between the two drugs.\nObjectives\nTo review the time to treatment failure, remission and first seizure with topiramate compared with carbamazepine when used as monotherapy in people with focal onset seizures (simple or complex focal and secondarily generalised), or generalised onset tonic-clonic seizures (with or without other generalised seizure types).\nSearch methods\nFor the latest update we searched the Cochrane Register of Studies (CRS Web), which includes the Cochrane Epilepsy Group Specialized Register and the Cochrane Central Register of Controlled Trials (CENTRAL); MEDLINE (Ovid); ClinicalTrials.gov; and the WHO International Clinical Trials Registry Platform (ICTRP) to 22 May 2018. We imposed no language restrictions. We also contacted pharmaceutical companies and trial investigators.\nSelection criteria\nRandomised controlled trials (RCTs) comparing monotherapy with either topiramate or carbamazepine in children or adults with focal onset seizures or generalised onset tonic-clonic seizures (with or without other generalised seizure types).\nData collection and analysis\nThis was an individual participant data (IPD), review. Our primary outcome was time to treatment failure. Our secondary outcomes were time to first seizure post-randomisation, time to six-month remission, time to 12-month remission, and incidence of adverse events. We used Cox proportional hazards regression models to obtain trial-specific estimates of hazard ratios (HRs), with 95% confidence intervals (CIs), using the generic inverse variance method to obtain the overall pooled HR and 95% CI.\nMain results\nIPD were available for 1151 of 1239 eligible individuals from two of three eligible studies (93% of the potential data). A small proportion of individuals recruited into these trials had 'unclassified seizures;' for analysis purposes, these individuals are grouped with those with generalised onset seizures. For remission outcomes, a HR < 1 indicated an advantage for carbamazepine, and for first seizure and treatment failure outcomes, a HR < 1 indicated an advantage for topiramate.\nThe main overall results for the primary outcome, time to treatment failure, given as pooled HR adjusted for seizure type were: time to failure for any reason related to treatment 1.16 (95% CI 0.97 to 1.38); time to failure due to adverse events 1.02 (95% CI 0.82 to 1.27); and time to failure due to lack of efficacy 1.46 (95% CI 1.08 to 1.98). Overall results for secondary outcomes were time to first seizure 1.11 (95% CI 0.96 to 1.29); and time to six-month remission 0.88 (0.76 to 1.01). There were no statistically significant differences between the drugs. A statistically significant advantage for carbamazepine was shown for time to 12-month remission: 0.84 (95% CI 0.71 to 0.99).\nThe results of this review are applicable mainly to individuals with focal onset seizures; 81% of individuals included within the analysis experienced seizures of this type at baseline. For individuals with focal onset seizures, a statistically significant advantage for carbamazepine was shown for time to failure for any reason related to treatment (HR 1.21, 95% CI 1.01 to 1.46), time to treatment failure due to lack of efficacy (HR 1.47, 95% CI 1.07 to 2.02), and time to 12-month remission (HR 0.82, 95% CI 0.69 to 0.99). There was no statistically significant difference between topiramate and carbamazepine for 'time to first seizure' and 'time to six-month remission'.\nEvidence for individuals with generalised tonic-clonic seizures (9% of participants contributing to the analysis), and unclassified seizure types (10% of participants contributing to the analysis) was very limited; no statistically significant differences were found but CIs were wide; therefore we cannot exclude an advantage to either drug, or a difference between drugs.\nThe most commonly reported adverse events with both drugs were drowsiness or fatigue, \"pins and needles\" (tingling sensation), headache, gastrointestinal disturbance and anxiety or depression. The rate of adverse events was similar across the two drugs.\nWe judged the methodological quality of the included trials generally to be good; however, there was some evidence that the open-label design of the larger of the two trials may have influenced the treatment failure rate within the trial. Hence, we judged the certainty of the evidence for treatment failure to be moderate for individuals with focal onset seizures and low for individuals with generalised onset seizures. For efficacy outcomes (first seizure, remission), we judged the certainty of evidence from this review to be high for individuals with focal onset seizures and moderate for individuals with generalised onset or unclassified seizures.\nAuthors' conclusions\nFor individuals with focal onset seizures, there is moderate-certainty evidence that carbamazepine is less likely to be withdrawn and high-certainty evidence that 12-month remission will be achieved earlier than with topiramate. We did not find any differences between the drugs in terms of the other outcomes measured in the review and for individuals with generalised tonic-clonic seizures or unclassified epilepsy; however, we encourage caution in the interpretation of results including small numbers of participants with these seizure types.\nFuture trials should be designed to the highest quality possible and take into consideration masking, choice of population, classification of seizure type, duration of follow-up, choice of outcomes and analysis, and presentation of results.\n\nGiven the provided abstract, please respond to the following query: \"Results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Most (81%) of the people included in the two trials experienced focal seizures, so the results of this review apply mainly to people with this seizure type. Many of the remaining 19% of people experienced a seizure type which was difficult to classify as focal or generalised (unclassified seizures). Considering only people with focal seizures, the results showed that those taking carbamazepine were more likely to take their treatment for longer and to achieve a remission of 12 months duration earlier than those taking topiramate. No differences were found between the drugs in individuals with generalised onset or unclassified epilepsy.\nThe most common side effects reported by the participants during the trials were fatigue, 'pins and needles' (tingling sensation), headache, gastrointestinal problems and anxiety or depression. These side effects were reported a similar number of times by people taking topiramate or carbamazepine." } ]
query-laysum
6694
[ { "role": "user", "content": "Abstract: Background\nSymptoms of autism spectrum disorder (ASD) have been associated, in part, with the dysfunction of N-methyl-D-aspartate (NMDA) glutamate receptors at excitatory synapses and glutamate abnormalities. Medications related to glutamatergic neurotransmission, such as D-cycloserine - which is a partial agonist of the NMDA glutamate receptor - are potential treatment options for the core features of ASD. However, the potential effect of D-cycloserine on the social and communication skills deficits of individuals with ASD has not been thoroughly explored and no systematic reviews of the evidence have been conducted.\nObjectives\nTo assess the efficacy and adverse effects of D-cycloserine compared with placebo for social and communication skills in individuals with ASD.\nSearch methods\nIn November 2020, we searched CENTRAL, MEDLINE, Embase, six other databases and two trials registers. We also searched the reference lists of relevant publications and contacted the authors of the included study, Minshawi 2016, to identify any additional studies. In addition, we contacted pharmaceutical companies, searched manufacturers' websites and sources of reports of adverse events.\nSelection criteria\nAll randomised controlled trials (RCTs) of any duration and dose of D-cycloserine, with or without adjunct treatment, compared to placebo in individuals with ASD.\nData collection and analysis\nTwo review authors independently selected studies for inclusion, extracted relevant data, assessed the risk of bias, graded the certainty of the evidence using the GRADE approach, and analysed and evaluated the data. We provide a narrative report of the findings as only one study is included in this review.\nMain results\nWe included a single RCT (Minshawi 2016) funded by the United States Department of Defense. It was conducted at two sites in the USA: Indiana University School of Medicine and Cincinnati Children's Hospital Medical Centre. In the included study, 67 children with ASD aged between 5 and 11 years were randomised to receive either 10 weeks (10 doses) of (50 mg) D-cycloserine plus social skills training, or placebo plus social skills training. Randomisation was carried out 1:1 between D-cycloserine and placebo arms, and outcome measures were recorded at one-week post-treatment. The 'risk of bias' assessment for the included study was low for five domains and unclear for two domains.\nThe study (67 participants) reported low certainty evidence of little to no difference between the two groups for all outcomes measured at one week post-treatment: social interaction impairment (mean difference (MD) 3.61 (assessed with the Social Responsiveness Scale), 95% confidence interval (CI) -5.60 to 12.82); social communication impairment (MD -1.08 (measured using the inappropriate speech subscale of the Aberrant Behavior Checklist (ABC)), 95% CI -2.34 to 0.18); restricted, repetitive, stereotyped patterns of behaviour (MD 0.12 (measured by the ABC stereotypy subscale), 95% CI -1.71 to 1.95); serious adverse events (risk ratio (RR) 1.11, 95% CI 0.94 to 1.31); non-core symptoms of ASD (RR 0.97 (measured by the Clinical Global Impression-Improvement scale), 95% CI 0.49 to 1.93); and tolerability of D-cycloserine (RR 0.32 (assessed by the number of dropouts), 95% CI 0.01 to 7.68).\nAuthors' conclusions\nWe are unable to conclude with certainty whether D-cycloserine is effective for individuals with ASD. This review included low certainty data from only one study with methodological issues and imprecision. The added value of this review compared to the included study is we assessed the risk of bias and evaluated the certainty of evidence using the GRADE approach. Moreover, if we find new trials in future updates of this review, we could potentially pool the data, which may either strengthen or decrease the evidence for our findings.\n\nGiven the provided abstract, please respond to the following query: \"Key messages\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "There appears to be no clear difference between D-cycloserine plus social skills training and social skills training alone on social and communication skills in individuals with ASD. However, we are uncertain about these results." } ]
query-laysum
6695
[ { "role": "user", "content": "Abstract: Background\nShared decision making is an important component of patient-centred care. It is a set of communication and evidence-based practice skills that elicits patients' expectations, clarifies any misperceptions and discusses the best available evidence for benefits and harms of treatment. Acute respiratory infections (ARIs) are one of the most common reasons for consulting in primary care and obtaining prescriptions for antibiotics. However, antibiotics offer few benefits for ARIs, and their excessive use contributes to antibiotic resistance – an evolving public health crisis. Greater explicit consideration of the benefit-harm trade-off within shared decision making may reduce antibiotic prescribing for ARIs in primary care.\nObjectives\nTo assess whether interventions that aim to facilitate shared decision making increase or reduce antibiotic prescribing for ARIs in primary care.\nSearch methods\nWe searched CENTRAL (2014, Issue 11), MEDLINE (1946 to November week 3, 2014), EMBASE (2010 to December 2014) and Web of Science (1985 to December 2014). We searched for other published, unpublished or ongoing trials by searching bibliographies of published articles, personal communication with key trial authors and content experts, and by searching trial registries at the National Institutes of Health and the World Health Organization.\nSelection criteria\nRandomised controlled trials (RCTs) (individual level or cluster-randomised), which evaluated the effectiveness of interventions that promote shared decision making (as the focus or a component of the intervention) about antibiotic prescribing for ARIs in primary care.\nData collection and analysis\nTwo review authors independently extracted and collected data. Antibiotic prescribing was the primary outcome, and secondary outcomes included clinically important adverse endpoints (e.g. re-consultations, hospital admissions, mortality) and process measures (e.g. patient satisfaction). We assessed the risk of bias of all included trials and the quality of evidence. We contacted trial authors to obtain missing information where available.\nMain results\nWe identified 10 published reports of nine original RCTs (one report was a long-term follow-up of the original trial) in over 1100 primary care doctors and around 492,000 patients.\nThe main risk of bias came from participants in most studies knowing whether they had received the intervention or not, and we downgraded the rating of the quality of evidence because of this.\nWe meta-analysed data using a random-effects model on the primary and key secondary outcomes and formally assessed heterogeneity. Remaining outcomes are presented narratively.\nThere is moderate quality evidence that interventions that aim to facilitate shared decision making reduce antibiotic use for ARIs in primary care (immediately after or within six weeks of the consultation), compared with usual care, from 47% to 29%: risk ratio (RR) 0.61, 95% confidence interval (CI) 0.55 to 0.68. Reduction in antibiotic prescribing occurred without an increase in patient-initiated re-consultations (RR 0.87, 95% CI 0.74 to 1.03, moderate quality evidence) or a decrease in patient satisfaction with the consultation (OR 0.86, 95% CI 0.57 to 1.30, low quality evidence). There were insufficient data to assess the effects of the intervention on sustained reduction in antibiotic prescribing, adverse clinical outcomes (such as hospital admission, incidence of pneumonia and mortality), or measures of patient and caregiver involvement in shared decision making (such as satisfaction with the consultation; regret or conflict with the decision made; or treatment compliance following the decision). No studies assessed antibiotic resistance in colonising or infective organisms.\nAuthors' conclusions\nInterventions that aim to facilitate shared decision making reduce antibiotic prescribing in primary care in the short term. Effects on longer-term rates of prescribing are uncertain and more evidence is needed to determine how any sustained reduction in antibiotic prescribing affects hospital admission, pneumonia and death.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Shared decision making enables health decisions to be made jointly by a clinician and patient. The decision making occurs after the options and their benefits and harms have been discussed together with the patient's values and preferences.\nAcute respiratory infections (such as an acute cough, middle ear infection or sore throat) are one of the most common reasons to see a health professional, and antibiotics are commonly prescribed despite good evidence that they have little benefit for these conditions. Any decision to prescribe an antibiotic should be balanced by any benefits against the risk of common harms (such as rash and stomach upset) and the contribution to antibiotic resistance - now a major threat to human health.\nShared decision making provides an ideal opportunity within a primary care consultation for greater consideration about the trade-off between benefit and harm of antibiotics for acute respiratory illnesses. Antibiotic prescribing may decrease as a result." } ]
query-laysum
6696
[ { "role": "user", "content": "Abstract: Background\nMalignant melanoma is a form of skin cancer associated with significant mortality once it has spread beyond the skin. Melanoma in situ (MIS) is the earliest histologically recognisable stage of malignant melanoma and represents a precursor of invasive melanoma. Lentigo maligna (LM) represents a subtype of pre-invasive intraepidermal melanoma associated specifically with chronic exposure to ultraviolet (UV) radiation. Over the past two decades, the incidence of MIS has increased significantly, even more than the invasive counterpart. There are several treatment options for MIS, but no consensus exists on the best therapeutic management of this condition.\nObjectives\nTo assess the effects of all available interventions, surgical and non-surgical, for the treatment of melanoma in situ, including LM.\nSearch methods\nWe searched the following databases up to November 2014: the Cochrane Skin Group Specialised Register, CENTRAL in The Cochrane Library (2014, Issue 10), MEDLINE (from 1946), Embase (from 1974), LILACS (from 1982), African Index Medicus (from inception), IndeMED of India (from inception), and Index Medicus for the South-East Asia Region (IMSEAR) (from inception). We scanned the references of included and excluded studies for further references to relevant trials and searched five trials registries. We checked the abstracts of major dermatology and oncology conference proceedings, and we shared our lists of included and excluded studies with industry contacts and other experts in the field of melanoma to try to identify further relevant trials.\nSelection criteria\nWe included randomised controlled trials (RCT) on the management of MIS, including LM, that compared any intervention to placebo or active treatment. We included individuals, irrespective of age and sex, diagnosed with MIS, including LM, based on histological examination.\nData collection and analysis\nTwo authors independently evaluated possible studies for inclusion; extracted data from the included study using a standard data extraction form modified for our review; assessed risk of bias; and analysed data on efficacy, safety, and tolerability. They resolved any disagreements by discussion with a third author. We collected adverse effects information from included studies.\nMain results\nOur search identified only 1 study eligible for inclusion (and 1 ongoing study in active recruitment stage), which was a single centre, open label, parallel group, 2-arm RCT with 90 participants, who had 91 histologically proven LM lesions.\nForty-four participants, with 44 LM lesions, were treated with imiquimod 5% cream 5 days per week plus tazarotene 0.1% gel 2 days/week for 3 months, and 46 participants, with 47 LM lesions, were treated with imiquimod 5% cream 5 days per week for 3 months. Two months after cessation of topical treatment, the initial tumour footprint was excised using 2 mm margins via a staged excision. This study was open label, and analysis was not intention-to-treat, leading to a high risk of incomplete outcome data.\nOur primary outcome 'Histological or clinical complete response' was measured at 5 months in 29/44 participants (66%) treated with imiquimod plus tazarotene (combination therapy) and 27/46 participants (59%) treated with imiquimod (monotherapy). The difference was not statistically significant (risk ratio (RR) 1.12, 95% confidence interval (CI) 0.81 to 1.55, P value = 0.48).\nWith regard to our secondary outcomes on recurrence and inflammation, after a mean follow up of 42 months, no local recurrences were observed among complete responders. Difference in overall inflammation score between the 2 groups was significant (mean difference (MD) 0.6, 95% CI 0.2 to 1, P value = 0.004), with the mean overall inflammation score being significantly higher in the combination group.\nThe study authors did not clearly report on side-effects. Because of adverse effects, there was a dropout rate of 6/44 participants (13.7%) in the combination group compared with 1/46 (2.2%) in the imiquimod monotherapy group (due to excessive inflammation) before the cessation of topical treatment (first 3 months), but this was not statistically significant (RR 6.27, 95% CI 0.79 to 50.02, P value = 0.08).\nAuthors' conclusions\nThere is a lack of high-quality evidence for the treatment of MIS and LM.\nFor the treatment of MIS, we found no RCTs of surgical interventions aiming to optimise margin control (square method, perimeter technique, 'slow Mohs', staged radial sections, staged \"mapped\" excisions, or Mohs micrographic surgery), which are the most widely used interventions recommended as first-line therapy. The use of non-surgical interventions in selected cases (patients with contraindications to surgical interventions) may be effective and may be considered preferable for experienced providers and under close and adequate follow up.\nFor the treatment of LM, we found no RCTs of surgical interventions, which remain the most widely used and recommended available treatment. The use of non-surgical interventions, such as imiquimod, as monotherapy may be effective and may be considered in selected cases where surgical procedures are contraindicated and used preferentially by experienced providers under close and adequate follow up. The use of topical therapies, such as 5-fluorouracil and imiquimod, as neoadjuvant therapies warrants further investigation. There is insufficient evidence to support or refute the addition of tazarotene to imiquimod as adjuvant therapy; the current evidence suggests that it can increase topical inflammatory response and withdrawal of participants because of treatment-related side-effects.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We identified one randomised controlled clinical trial (RCT) of the non-surgical treatment of 90 people with LM: 44 were treated with imiquimod cream plus tazarotene gel, and 46 were treated with imiquimod cream for 3 months; these interventions were followed by staged excision after 2 months. We did not find any RCTs of surgical treatments." } ]
query-laysum
6697
[ { "role": "user", "content": "Abstract: Background\nThe World Health Organization (WHO) End TB Strategy stresses universal access to drug susceptibility testing (DST). DST determines whether Mycobacterium tuberculosis bacteria are susceptible or resistant to drugs. Xpert MTB/XDR is a rapid nucleic acid amplification test for detection of tuberculosis and drug resistance in one test suitable for use in peripheral and intermediate level laboratories. In specimens where tuberculosis is detected by Xpert MTB/XDR, Xpert MTB/XDR can also detect resistance to isoniazid, fluoroquinolones, ethionamide, and amikacin.\nObjectives\nTo assess the diagnostic accuracy of Xpert MTB/XDR for pulmonary tuberculosis in people with presumptive pulmonary tuberculosis (having signs and symptoms suggestive of tuberculosis, including cough, fever, weight loss, night sweats).\nTo assess the diagnostic accuracy of Xpert MTB/XDR for resistance to isoniazid, fluoroquinolones, ethionamide, and amikacin in people with tuberculosis detected by Xpert MTB/XDR, irrespective of rifampicin resistance (whether or not rifampicin resistance status was known) and with known rifampicin resistance.\nSearch methods\nWe searched multiple databases to 23 September 2021. We limited searches to 2015 onwards as Xpert MTB/XDR was launched in 2020.\nSelection criteria\nDiagnostic accuracy studies using sputum in adults with presumptive or confirmed pulmonary tuberculosis. Reference standards were culture (pulmonary tuberculosis detection); phenotypic DST (pDST), genotypic DST (gDST),composite (pDST and gDST) (drug resistance detection).\nData collection and analysis\nTwo review authors independently reviewed reports for eligibility and extracted data using a standardized form. For multicentre studies, we anticipated variability in the type and frequency of mutations associated with resistance to a given drug at the different centres and considered each centre as an independent study cohort for quality assessment and analysis. We assessed methodological quality with QUADAS-2, judging risk of bias separately for each target condition and reference standard. For pulmonary tuberculosis detection, owing to heterogeneity in participant characteristics and observed specificity estimates, we reported a range of sensitivity and specificity estimates and did not perform a meta-analysis. For drug resistance detection, we performed meta-analyses by reference standard using bivariate random-effects models. Using GRADE, we assessed certainty of evidence of Xpert MTB/XDR accuracy for detection of resistance to isoniazid and fluoroquinolones in people irrespective of rifampicin resistance and to ethionamide and amikacin in people with known rifampicin resistance, reflecting real-world situations. We used pDST, except for ethionamide resistance where we considered gDST a better reference standard.\nMain results\nWe included two multicentre studies from high multidrug-resistant/rifampicin-resistant tuberculosis burden countries, reporting on six independent study cohorts, involving 1228 participants for pulmonary tuberculosis detection and 1141 participants for drug resistance detection. The proportion of participants with rifampicin resistance in the two studies was 47.9% and 80.9%. For tuberculosis detection, we judged high risk of bias for patient selection owing to selective recruitment. For ethionamide resistance detection, we judged high risk of bias for the reference standard, both pDST and gDST, though we considered gDST a better reference standard.\nPulmonary tuberculosis detection\n- Xpert MTB/XDR sensitivity range, 98.3% (96.1 to 99.5) to 98.9% (96.2 to 99.9) and specificity range, 22.5% (14.3 to 32.6) to 100.0% (86.3 to 100.0); median prevalence of pulmonary tuberculosis 91.3%, (interquartile range, 89.3% to 91.8%), (2 studies; 1 study reported on 2 cohorts, 1228 participants; very low-certainty evidence, sensitivity and specificity).\nDrug resistance detection\nPeople irrespective of rifampicin resistance\n- Isoniazid resistance: Xpert MTB/XDR summary sensitivity and specificity (95% confidence interval (CI)) were 94.2% (87.5 to 97.4) and 98.5% (92.6 to 99.7) against pDST, (6 cohorts, 1083 participants, moderate-certainty evidence, sensitivity and specificity).\n- Fluoroquinolone resistance: Xpert MTB/XDR summary sensitivity and specificity were 93.2% (88.1 to 96.2) and 98.0% (90.8 to 99.6) against pDST, (6 cohorts, 1021 participants; high-certainty evidence, sensitivity; moderate-certainty evidence, specificity).\nPeople with known rifampicin resistance\n- Ethionamide resistance: Xpert MTB/XDR summary sensitivity and specificity were 98.0% (74.2 to 99.9) and 99.7% (83.5 to 100.0) against gDST, (4 cohorts, 434 participants; very low-certainty evidence, sensitivity and specificity).\n- Amikacin resistance: Xpert MTB/XDR summary sensitivity and specificity were 86.1% (75.0 to 92.7) and 98.9% (93.0 to 99.8) against pDST, (4 cohorts, 490 participants; low-certainty evidence, sensitivity; high-certainty evidence, specificity).\nOf 1000 people with pulmonary tuberculosis, detected as tuberculosis by Xpert MTB/XDR:\n- where 50 have isoniazid resistance, 61 would have an Xpert MTB/XDR result indicating isoniazid resistance: of these, 14/61 (23%) would not have isoniazid resistance (FP); 939 (of 1000 people) would have a result indicating the absence of isoniazid resistance: of these, 3/939 (0%) would have isoniazid resistance (FN).\n- where 50 have fluoroquinolone resistance, 66 would have an Xpert MTB/XDR result indicating fluoroquinolone resistance: of these, 19/66 (29%) would not have fluoroquinolone resistance (FP); 934 would have a result indicating the absence of fluoroquinolone resistance: of these, 3/934 (0%) would have fluoroquinolone resistance (FN).\n- where 300 have ethionamide resistance, 296 would have an Xpert MTB/XDR result indicating ethionamide resistance: of these, 2/296 (1%) would not have ethionamide resistance (FP); 704 would have a result indicating the absence of ethionamide resistance: of these, 6/704 (1%) would have ethionamide resistance (FN).\n- where 135 have amikacin resistance, 126 would have an Xpert MTB/XDR result indicating amikacin resistance: of these, 10/126 (8%) would not have amikacin resistance (FP); 874 would have a result indicating the absence of amikacin resistance: of these, 19/874 (2%) would have amikacin resistance (FN).\nAuthors' conclusions\nReview findings suggest that, in people determined by Xpert MTB/XDR to be tuberculosis-positive, Xpert MTB/XDR provides accurate results for detection of isoniazid and fluoroquinolone resistance and can assist with selection of an optimised treatment regimen. Given that Xpert MTB/XDR targets a limited number of resistance variants in specific genes, the test may perform differently in different settings. Findings in this review should be interpreted with caution. Sensitivity for detection of ethionamide resistance was based only on Xpert MTB/XDR detection of mutations in the inhA promoter region, a known limitation. High risk of bias limits our confidence in Xpert MTB/XDR accuracy for pulmonary tuberculosis.\nXpert MTB/XDR's impact will depend on its ability to detect tuberculosis (required for DST), prevalence of resistance to a given drug, health care infrastructure, and access to other tests.\n\nGiven the provided abstract, please respond to the following query: \"How up-to-date is this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched for studies up to 23 September 2021. Searches were limited to 2015 onwards as Xpert MTB/XDR was launched in July 2020." } ]
query-laysum
6698
[ { "role": "user", "content": "Abstract: Background\nThe vicious cycle hypothesis for bronchiectasis predicts that bacterial colonisation of the respiratory tract perpetuates inflammatory change. This damages the mucociliary escalator, preventing bacterial clearance and allowing persistence of pro-inflammatory mediators. Conventional treatment with physiotherapy and intermittent antibiotics is believed to improve the condition of people with bronchiectasis, although no conclusive data show that these interventions influence the natural history of the condition. Various strategies have been tried to interrupt this cycle of infection and inflammation, including prolonging antibiotic treatment with the goal of allowing the airway mucosa to heal.\nObjectives\nTo determine the benefits of prolonged antibiotic therapy in the treatment of patients with bronchiectasis.\nSearch methods\nWe searched the Cochrane Airways Group Trials Register and reference lists of identified articles. Searches were current as of February 2014.\nSelection criteria\nRandomised trials examining the use of prolonged antibiotic therapy (for four or more weeks) in the treatment of bronchiectasis compared with placebo or usual care.\nData collection and analysis\nTwo review authors independently assessed trial quality and extracted data. We contacted study authors to ask for missing information.\nMain results\nEighteen trials met the inclusion criteria, randomly assigning a total of 1157 participants. Antibiotics were given for between four weeks and 83 weeks. Limited meta-analysis was possible because of the diversity of outcomes reported in these trials. Based on the number of participants with at least one exacerbation, the meta-analysis showed significant effects in favour of the intervention (odds ratio (OR) 0.31, 95% confidence interval (CI) 0.19 to 0.52; P value < 0.00001), with events occurring in 271 per 1000 people in the intervention arm (95% CI 126 to 385) and in 546 per 1000 in the control population, based on evidence of moderate quality. A non-statistically significant reduction in hospitalisation favoured the use of prolonged antibiotics with a moderate quality grade of supporting evidence (37 per 1000 in the intervention arm (95% CI 13 to 96) and 87 per 1000 in control (OR 0.40, 95% CI 0.14 to 1.11; P value = 0.08). Drug resistance developed in 36 of 220 participants taking antibiotics compared with 10 of 211 participants given placebo or standard therapy (OR 3.48, 95% CI 1.20 to 10.07; P value = 0.02), translating to natural frequencies of 155 per 1000 in the intervention arm (95% CI 59 to 346) and 50 per 1000 in the control arm. The intervention was well tolerated with no overall significant difference in withdrawal between treatment and placebo groups (OR 0.91, 95% CI 0.56 to 1.49). Diarrhoea was commonly reported as an adverse event, particularly with an oral intervention.\nAuthors' conclusions\nAvailable evidence shows benefit associated with use of prolonged antibiotics in the treatment of patients with bronchiectasis, at least halving the odds of exacerbation (with 275 fewer exacerbations per every 1000 people treated in the antibiotic arm compared with the control arm) and hospitalisation (50 fewer hospitalisations per 1000 people in the antibiotic arm compared with the control arm). However, the risk of emerging drug resistance is increased more than threefold. This review is limited by diversity of trials and by evidence of moderate to low quality. Further randomised controlled trials with adequate power and standardised end points are required.\n\nGiven the provided abstract, please respond to the following query: \"Conclusion\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Prolonged antibiotic therapy in bronchiectasis provides benefit, especially in reducing the risk of future exacerbations and hospitalisations. Antibiotics are well tolerated by participants without significant differences in overall adverse effects (e.g. intolerance, chest symptoms, fatigue, fever, palpitations).\nHowever, antibiotic resistance is a matter of concern, particularly for patients with drug allergies, which further limit their future treatment." } ]
query-laysum
6699
[ { "role": "user", "content": "Abstract: Background\nInvasive aspergillosis (IA) is a life-threatening opportunistic mycosis that occurs in some people with a compromised immune system. The serum galactomannan enzyme-linked immunosorbent assay (ELISA) rapidly gained widespread acceptance as part of the diagnostic work-up of a patient suspected of IA. Due to its non-invasive nature, it can be used as a routine screening test. The ELISA can also be performed on bronchoalveolar lavage (BAL), allowing sampling of the immediate vicinity of the infection. The invasive nature of acquiring BAL, however, changes the role of the galactomannan test significantly, for example by precluding its use as a routine screening test.\nObjectives\nTo assess the diagnostic accuracy of galactomannan detection in BAL for the diagnosis of IA in people who are immunocompromised, at different cut-off values for test positivity, in accordance with the Cochrane Diagnostic Test Accuracy Handbook.\nSearch methods\nWe searched three bibliographic databases including MEDLINE on 9 September 2016 for aspergillosis and galactomannan as text words and subject headings where appropriate. We checked reference lists of included studies for additional studies.\nSelection criteria\nWe included cohort studies that examined the accuracy of BAL galactomannan for the diagnosis of IA in immunocompromised patients if they used the European Organization for Research and Treatment of Cancer/Invasive Fungal Infections Cooperative Group and the National Institute of Allergy and Infectious Diseases Mycoses Study Group (EORTC/MSG) classification as reference standard.\nData collection and analysis\nTwo review authors assessed study quality and extracted data. Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) was used for quality assessment.\nMain results\nWe included 17 studies in our review. All studies except one had a high risk of bias in two or more domains. The diagnostic performance of an optical density index (ODI) of 0.5 as cut-off value was reported in 12 studies (with 1123 patients). The estimated sensitivity was 0.88 (95% confidence interval (CI) 0.75 to 1.00) and specificity 0.81 (95% CI 0.71 to 0.91). The performance of an ODI of 1.0 as cut-off value could be determined in 11 studies (with 648 patients). The sensitivity was 0.78 (95% CI 0.61 to 0.95) and specificity 0.93 (95% CI 0.87 to 0.98). At a cut-off ODI of 1.5 or higher, the heterogeneity in specificity decreased significantly and was invariably >90%.\nAuthors' conclusions\nThe optimal cut-off value depends on the local incidence and clinical pathway. At a prevalence of 12% a hypothetical population of 1000 patients will consist of 120 patients with IA. At a cut-off value of 0.5 14 patients with IA will be missed and there will be 167 patients incorrectly diagnosed with IA. If we use the test at a cut-off value of 1.0, we will miss 26 patients with IA. And there will be 62 patients incorrectly diagnosed with invasive aspergillosis. The populations and results were very heterogeneous. Therefore, interpretation and extrapolation of these results has to be performed with caution. A test result of 1.5 ODI or higher appears a strong indicator of IA.\n\nGiven the provided abstract, please respond to the following query: \"What is the aim of this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The aim of this review was to determine the error rate of galactomannan testing on BAL. To this end, we searched for studies on this subject. The galactomannan test results in an 'optical density index' (ODI), not a yes or no answer. The studies differed with respect to the cut-off value above which a result was considered positive. Most considered a result above an ODI of 0.5 or 1.0 as positive." } ]
query-laysum
6700
[ { "role": "user", "content": "Abstract: Background\nChronic hepatitis B is a liver disease associated with high morbidity and mortality. Chronic hepatitis B requires long-term management aiming to reduce the risks of hepatocellular inflammatory necrosis, liver fibrosis, decompensated liver cirrhosis, liver failure, and liver cancer, as well as to improve health-related quality of life. Acupuncture is being used to decrease discomfort and improve immune function in people with chronic hepatitis B. However, the benefits and harms of acupuncture still need to be established in a rigorous way.\nObjectives\nTo assess the benefits and harms of acupuncture versus no intervention or sham acupuncture in people with chronic hepatitis B.\nSearch methods\nWe undertook electronic searches of the Cochrane Hepato-Biliary Group Controlled Trials Register, CENTRAL, MEDLINE, Embase, LILACS, Science Citation Index Expanded, Conference Proceedings Citation Index - Science, China National Knowledge Infrastructure (CNKI), Chongqing VIP (CQVIP), Wanfang Data, and SinoMed to 1 March 2019. We also searched the World Health Organization International Clinical Trials Registry Platform (www.who.int/ictrp), ClinicalTrials.gov (www.clinicaltrials.gov/), and the Chinese Clinical Trial Registry (ChiCTR) for ongoing or unpublished trials until 1 March 2019.\nSelection criteria\nWe included randomised clinical trials, irrespective of publication status, language, and blinding, comparing acupuncture versus no intervention or sham acupuncture in people with chronic hepatitis B. We included participants of any sex and age, diagnosed with chronic hepatitis B as defined by the trialists or according to guidelines. We allowed co-interventions when the co-interventions were administered equally to all intervention groups.\nData collection and analysis\nReview authors in pairs individually retrieved data from reports and through correspondence with investigators. Primary outcomes were all-cause mortality, proportion of participants with one or more serious adverse events, and health-related quality of life. Secondary outcomes were hepatitis B-related mortality, hepatitis B-related morbidity, and adverse events considered not to be serious. We presented the pooled results as risk ratios (RRs) with 95% confidence intervals (CIs). We assessed the risks of bias using risk of bias domains with predefined definitions. We put more weight on the estimate closest to zero effect when results with fixed-effect and random-effects models differed. We evaluated the certainty of evidence using GRADE.\nMain results\nWe included eight randomised clinical trials with 555 randomised participants. All included trials compared acupuncture versus no intervention. These trials assessed heterogeneous acupuncture interventions. All trials used heterogeneous co-interventions applied equally in the compared groups. Seven trials included participants with chronic hepatitis B, and one trial included participants with chronic hepatitis B with comorbid tuberculosis. All trials were assessed at overall high risk of bias, and the certainty of evidence for all outcomes was very low due to high risk of bias for each outcome, imprecision of results (the confidence intervals were wide), and publication bias (small sample size of the trials, and all trials were conducted in China). Additionally, 79 trials lacked the necessary methodological information to ensure their inclusion in our review.\nNone of the included trials aim to assess all-cause mortality, serious adverse events, health-related quality of life, hepatitis B-related mortality, and hepatitis B-related morbidity. We are uncertain whether acupuncture, compared with no intervention, has an effect regarding adverse events considered not to be serious (RR 0.67, 95% CI 0.43 to 1.06; I² = 0%; 3 trials; 203 participants; very low-certainty evidence) or detectable hepatitis B e-antigen (HBeAg) (RR 0.64, 95% CI 0.11 to 3.68; I² = 98%; 2 trials; 158 participants; very low-certainty evidence). Acupuncture showed a reduction in detectable hepatitis B virus (HBV) DNA (a non-validated surrogate outcome; RR 0.45, 95% CI 0.27 to 0.74; 1 trial, 58 participants; very low-certainty evidence). We are uncertain whether acupuncture has an effect regarding the remaining separately reported adverse events considered not to be serious.\nThree of the eight included trials received academic funding from government or hospital. None of the remaining five trials reported information on funding.\nAuthors' conclusions\nThe clinical effects of acupuncture for chronic hepatitis B remain unknown. The included trials lacked data on all-cause mortality, health-related quality of life, serious adverse events, hepatitis-B related mortality, and hepatitis-B related morbidity. The vast number of excluded trials lacked clear descriptions of their design and conduct. Whether acupuncture influences adverse events considered not to be serious is uncertain. It remains unclear if acupuncture affects HBeAg, and if it is associated with reduction in detectable HBV DNA. Based on available data from only one or two small trials on adverse events considered not to be serious and on the surrogate outcomes HBeAg and HBV DNA, the certainty of evidence is very low. In view of the wide usage of acupuncture, any conclusion that one might try to draw in the future should be based on data on patient and clinically relevant outcomes, assessed in large, high-quality randomised sham-controlled trials with homogeneous groups of participants and transparent funding.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Chronic hepatitis B infection has a substantial economic, psychological, and life impact on people with chronic hepatitis B and their families. Acupuncture has been used in treating people with chronic hepatitis B, as it is believed that it decreases discomfort and improves immune function in people with this disease. However, the benefits and harms of acupuncture have never been established in systematic reviews in a rigorous way." } ]
query-laysum
6701
[ { "role": "user", "content": "Abstract: Background\nVenous thromboembolism (VTE) is a leading cause of morbidity and mortality in hospitalized patients. While numerous randomized controlled trials (RCTs) have shown that the appropriate use of thromboprophylaxis in hospitalized patients at risk for VTE is safe, effective, and cost-effective, thromboprophylaxis remains underused or inappropriately used. Our previous review suggested that system-wide interventions, such as education, alerts, and multifaceted interventions were more effective at improving the prescribing of thromboprophylaxis than relying on individual providers’ behaviors. However, 47 of the 55 included studies in our previous review were observational in design. Thus, an update to our systematic review, focused on the higher level of evidence of RCTs only, was warranted.\nObjectives\nTo assess the effects of system-wide interventions designed to increase the implementation of thromboprophylaxis and decrease the incidence of VTE in hospitalized adult medical and surgical patients at risk for VTE, focusing on RCTs only.\nSearch methods\nOur research librarian conducted a systematic literature search of MEDLINE Ovid, and subsequently translated it to CENTRAL, PubMed, Embase Ovid, BIOSIS Previews Ovid, CINAHL, Web of Science, the Database of Abstracts of Reviews of Effects (DARE; in the Cochrane Library), NHS Economic Evaluation Database (EED; in the Cochrane Library), LILACS, and clinicaltrials.gov from inception to 7 January 2017. We also screened reference lists of relevant review articles. We identified 12,920 potentially relevant records.\nSelection criteria\nWe included all types of RCTs, with random or quasi-random methods of allocation of interventions, which either randomized individuals (e.g. parallel group, cross-over, or factorial design RCTs), or groups of individuals (cluster RCTs (CRTs)), which aimed to increase the use of prophylaxis or appropriate prophylaxis, or decrease the occurrence of VTE in hospitalized adult patients. We excluded observational studies, studies in which the intervention was simply distribution of published guidelines, and studies whose interventions were not clearly described. Studies could be in any language.\nData collection and analysis\nWe collected data on the following outcomes: the number of participants who received prophylaxis or appropriate prophylaxis (as defined by study authors), the occurrence of any VTE (symptomatic or asymptomatic), mortality, and safety outcomes, such as bleeding. We categorized the interventions into alerts (computer or human alerts), multifaceted interventions (combination of interventions that could include an alert component), educational interventions (e.g. grand rounds, courses), and preprinted orders (written predefined orders completed by the physician on paper or electronically). We meta-analyzed data across RCTs using a random-effects model. For CRTs, we pooled effect estimates (risk difference (RD) and risk ratio (RR), with 95% confidence interval (CI), adjusted for clustering, when possible. We pooled results if three or more trials were available for a particular intervention. We assessed the certainty of the evidence according to the GRADE approach.\nMain results\nFrom the 12,920 records identified by our search, we included 13 RCTs (N = 35,997 participants) in our qualitative analysis and 11 RCTs (N = 33,207 participants) in our meta-analyses.\nPrimary outcome: Alerts were associated with an increase in the proportion of participants who received prophylaxis (RD 21%, 95% CI 15% to 27%; three studies; 5057 participants; I² = 75%; low-certainty evidence). The substantial statistical heterogeneity may be in part explained by patient types, type of hospital, and type of alert. Subgroup analyses were not feasible due to the small number of studies included in the meta-analysis.\nMultifaceted interventions were associated with a small increase in the proportion of participants who received prophylaxis (cluster-adjusted RD 4%, 95% CI 2% to 6%; five studies; 9198 participants; I² = 0%; moderate-certainty evidence). Multifaceted interventions with an alert component were found to be more effective than multifaceted interventions that did not include an alert, although there were not enough studies to conduct a pooled analysis.\nSecondary outcomes: Alerts were associated with an increase in the proportion of participants who received appropriate prophylaxis (RD 16%, 95% CI 12% to 20%; three studies; 1820 participants; I² = 0; moderate-certainty evidence). Alerts were also associated with a reduction in the rate of symptomatic VTE at three months (RR 64%, 95% CI 47% to 86%; three studies; 5353 participants; I² = 15%; low-certainty evidence). Computer alerts were associated with a reduction in the rate of symptomatic VTE, although there were not enough studies to pool computer alerts and human alerts results separately.\nAuthors' conclusions\nWe reviewed RCTs that implemented a variety of system-wide strategies aimed at improving thromboprophylaxis in hospitalized patients. We found increased prescription of prophylaxis associated with alerts and multifaceted interventions, and increased prescription of appropriate prophylaxis associated with alerts. While multifaceted interventions were found to be less effective than alerts, a multifaceted intervention with an alert was more effective than one without an alert. Alerts, particularly computer alerts, were associated with a reduction in symptomatic VTE at three months, although there were not enough studies to pool computer alerts and human alerts results separately.\nOur analysis was underpowered to assess the effect on mortality and safety outcomes, such as bleeding.\nThe incomplete reporting of relevant study design features did not allow complete assessment of the certainty of the evidence. However, the certainty of the evidence for improvement in outcomes was judged to be better than for our previous review (low- to moderate-certainty evidence, compared to very low-certainty evidence for most outcomes). The results of our updated review will help physicians, hospital administrators, and policy makers make practical decisions about adopting specific system-wide measures to improve prescription of thromboprophylaxis, and ultimately prevent VTE in hospitalized patients.\n\nGiven the provided abstract, please respond to the following query: \"What was studied in this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Blood clots that occur in the leg veins (deep vein thrombosis) or in the lung circulation (pulmonary embolism) are together known as venous thromboembolism (VTE). VTE is a potential complication for patients who have been hospitalized for medical or surgical reasons. These complications lengthen hospital stay and are a leading cause of death and long-term disability. Risk factors for VTE include hospitalization for surgical or medical illness, cancer, trauma or immobilization, medications, such as oral contraceptives or hormone replacement therapy, and pregnancy or postpartum. Other risk factors are older age, obesity, previous blood clots, and family history of blood clots.\nThromboprophylaxis involves the administration of small doses of anticoagulant (i.e. blood thinning) medications, such as heparin, low molecular weight heparin, or oral blood thinners, or the application of physical measures, such as graduated compression stockings or sequential compression devices. In the USA, thromboprophylaxis has been ranked as the number one strategy to improve patient safety in hospitals, and interventions to improve the implementation of thromboprophylaxis were recently ranked as a top-10 patient safety strategy that demanded action." } ]
query-laysum
6702
[ { "role": "user", "content": "Abstract: Background\nBonding of orthodontic brackets to teeth is important to enable effective and efficient treatment with fixed appliances. The problem is bracket failure during treatment which increases operator chairside time and lengthens treatment time. A prolonged treatment is likely to increase the oral health risks of orthodontic treatment with fixed appliances one of which is irreversible enamel decalcification. This is an update of the Cochrane Review first published in 2003. A new full search was conducted on 26 September 2017 but no new studies were identified. We have only updated the search methods section in this new version. The conclusions of this Cochrane Review remain the same.\nObjectives\nTo evaluate the effects of different orthodontic adhesives for bonding.\nSearch methods\nCochrane Oral Health's Information Specialist searched the following databases: Cochrane Oral Health's Trials Register (to 26 September 2017), the Cochrane Central Register of Controlled Trials (CENTRAL; 2017, Issue 8) in the Cochrane Library (searched 26 September 2017), MEDLINE Ovid (1946 to 26 September 2017), and Embase Ovid (1980 to 26 September 2017). The US National Institutes of Health Ongoing Trials Register (ClinicalTrials.gov) and the World Health Organization International Clinical Trials Registry Platform were searched for ongoing trials. No restrictions were placed on the language or date of publication when searching the electronic databases.\nSelection criteria\nTrials were selected if they met the following criteria: randomised controlled trials (RCTs) and controlled clinical trials (CCTs) comparing two different adhesive groups. Participants were patients with fixed orthodontic appliances. The interventions were adhesives that bonded stainless steel brackets to all teeth except the molars. The primary outcome was debond or bracket failure.\nData collection and analysis\nData were recorded on decalcification as a secondary outcome, if present. Information regarding methods, participants, interventions, outcome measures and results were extracted in duplicate by pairs of review authors. Since the data were not presented in a form that was amenable to meta-analysis, the results of the review are presented in narrative form only.\nMain results\nThree trials satisfied the inclusion criteria. A chemical cured composite was compared with a light cured composite (one trial), a conventional glass ionomer cement (one trial) and a polyacid-modified resin composite (compomer) (one trial). The quality of the trial reports was generally poor.\nAuthors' conclusions\nThere is no clear evidence on which to make a clinical decision of the type of orthodontic adhesive to use.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "There is no clear evidence on which to make a clinical decision of the type of orthodontic adhesive to use." } ]
query-laysum
6703
[ { "role": "user", "content": "Abstract: Background\nDipyrone (metamizole) is a nonsteroidal anti-inflammatory drug used in some countries to treat pain (postoperative, colic, cancer, and migraine); it is banned in other countries because of an association with life-threatening blood disorders. This review replaces a 2010 Cochrane review that has been withdrawn.\nObjectives\nTo assess the analgesic efficacy and associated adverse events of single dose dipyrone for moderate to severe acute postoperative pain using methods that permit comparison with other analgesics evaluated in standardised trials using almost identical methods and outcomes.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, EMBASE, and LILACS to 11 August 2015; the Oxford Pain Relief Database; two clinical trial registries; and the reference lists of articles.\nSelection criteria\nWe included randomised, double-blind, placebo-controlled trials of single dose dipyrone for relief of established moderate to severe postoperative pain in adults. We accepted oral, rectal, intramuscular, and intravenous routes of administration.\nData collection and analysis\nTwo review authors independently considered studies for inclusion in the review, assessed risk of bias, and extracted data. We used summed total pain relief or pain intensity difference (TOTPAR or SPID) over four to six hours to calculate the number of participants achieving at least 50% pain relief. From derived results, we calculated the risk ratio and number needed to treat for an additional beneficial outcome (NNT), with 95% confidence intervals (CI), for one participant to experience at least 50% pain relief over four to six hours compared to placebo. We looked at use of rescue medication and time to use of rescue medication as additional measures of efficacy. We also looked for information on adverse events and withdrawals.\nMain results\nWe included eight studies, involving 809 participants, comparing oral dipyrone 500 mg (143 participants), oral dipyrone 1000 mg (57 participants), and intramuscular dipyrone 2000 mg (35 participants) with placebo (236 participants). In addition to placebo, all studies used active controls (ibuprofen, paracetamol, aspirin, flurbiprofen, ketoprofen; 338 participants). Seven studies used the oral route of administration, and one study used the intramuscular route. The mean age ranged from 23 to 62 years. Six studies included both men and women, and two studies included only women. All the studies were small, but were otherwise of moderate to good quality.\nOver 70% of participants experienced our primary outcome of at least 50% pain relief over four to six hours with oral dipyrone 500 mg compared to 30% with placebo (five studies, 288 participants; NNT 2.4 (95% CI 1.8 to 3.1)) (moderate quality evidence). There were insufficient data to assess other doses or routes of administration of dipyrone.\nFewer participants needed rescue medication within four to six hours with dipyrone 500 mg than with placebo (7% with dipyrone versus 34% with placebo; four studies, 248 participants) (low quality evidence).\nThe data on numbers of participants experiencing any adverse event was inconsistently reported and no analysis was possible. No serious adverse events or adverse event withdrawals were reported (very low quality evidence).\nThere were too few data to compare dipyrone directly with other active treatments.\nAuthors' conclusions\nBased on very limited information, a single dose of dipyrone 500 mg provides good pain relief to about 70% of people treated, compared to about 30% with placebo. For every five people given dipyrone 500 mg, two people would experience this level of pain relief over four to six hours who would not have done with placebo, and fewer people would need rescue medication.\nWe were unable to compare dipyrone directly with other active treatments, or to assess the effects of different doses or routes of administration, or the number of participants experiencing adverse events, because of insufficient data and inadequate reporting.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched medical databases for studies of dipyrone used to treat pain following surgery in adults and compared with placebo (a pretend treatment). The medicines could be given by mouth, into a vein, into a muscle, or into the rectum. The evidence is current to 11 August 2015. We found eight studies, involving 809 participants treated with dipyrone, placebo, and various other painkillers. The studies were all small, but otherwise of moderate to good quality." } ]
query-laysum
6704
[ { "role": "user", "content": "Abstract: Background\nPain is a common symptom with cancer, and 30% to 50% of all people with cancer will experience moderate to severe pain that can have a major negative impact on their quality of life. Non-opioid drugs are commonly used to treat mild to moderate cancer pain, and are recommended for this purpose in the WHO cancer pain treatment ladder, either alone or in combination with opioids.\nA previous Cochrane review that examined the evidence for nonsteroidal anti-inflammatory drugs (NSAIDs) or paracetamol, alone or combined with opioids, for cancer pain was withdrawn in 2015 because it was out of date; the date of the last search was 2005. This review, and another on NSAIDs, updates the evidence.\nObjectives\nTo assess the efficacy of oral paracetamol (acetaminophen) for cancer pain in adults and children, and the adverse events reported during its use in clinical trials.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, and Embase from inception to March 2017, together with reference lists of retrieved papers and reviews, and two online study registries.\nSelection criteria\nWe included randomised, double-blind, studies of five days' duration or longer, comparing paracetamol alone with placebo, or paracetamol in combination with an opioid compared with the same dose of the opioid alone, for cancer pain of any intensity. Single-blind and open studies were also eligible for inclusion. The minimum study size was 25 participants per treatment arm at the initial randomisation.\nData collection and analysis\nTwo review authors independently searched for studies, extracted efficacy and adverse event data, and examined issues of study quality and potential bias. We did not carry out any pooled analyses. We assessed the quality of the evidence using GRADE and created a 'Summary of findings' table.\nMain results\nThree studies in adults satisfied the inclusion criteria, lasting up to one week; 122 participants were randomised initially, and 95 completed treatment. We found no studies in children. One study was parallel-group, and two had a cross-over design. All used paracetamol as an add-on to established treatment with strong opioids (median daily morphine equivalent doses of 60 mg, 70 mg, and 225 mg, with some participants taking several hundred mg of oral morphine equivalents daily). Other non-paracetamol medication included non-steroidal anti-inflammatory drugs (NSAIDs), tricyclic antidepressants, or neuroleptics. All studies were at high risk of bias for incomplete outcome data and small size; none was unequivocally at low risk of bias.\nNone of the studies reported any of our primary outcomes: participants with pain reduction of at least 50%, and at least 30%, from baseline; participants with pain no worse than mild at the end of the treatment period; participants with Patient Global Impression of Change (PGIC) of much improved or very much improved (or equivalent wording). What pain reports there were indicated no difference between paracetamol and placebo when added to another treatment. There was no convincing evidence of paracetamol being different from placebo with regards to quality of life, use of rescue medication, or participant satisfaction or preference. Measures of harm (serious adverse events, other adverse events, and withdrawal due to lack of efficacy) were inconsistently reported and provided no clear evidence of difference.\nOur GRADE assessment of evidence quality was very low for all outcomes, because studies were at high risk of bias from several sources.\nAuthors' conclusions\nThere is no high-quality evidence to support or refute the use of paracetamol alone or in combination with opioids for the first two steps of the three-step WHO cancer pain ladder. It is not clear whether any additional analgesic benefit of paracetamol could be detected in the available studies, in view of the doses of opioids used.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "One person in two or three who gets cancer will suffer from pain that becomes moderate or severe in intensity. The pain tends to get worse as the cancer progresses. In 1986, the World Health Organization recommended taking morphine-like drugs (opioids) for moderate to severe pain from cancer, and non-opioid drugs like paracetamol, alone for mild to moderate pain, or alongside opioids in people with moderate to severe pain." } ]
query-laysum
6705
[ { "role": "user", "content": "Abstract: Background\nAt least one-third of community-dwelling people over 65 years of age fall each year. Exercises that target balance, gait and muscle strength have been found to prevent falls in these people. An up-to-date synthesis of the evidence is important given the major long-term consequences associated with falls and fall-related injuries\nObjectives\nTo assess the effects (benefits and harms) of exercise interventions for preventing falls in older people living in the community.\nSearch methods\nWe searched CENTRAL, MEDLINE, Embase, three other databases and two trial registers up to 2 May 2018, together with reference checking and contact with study authors to identify additional studies.\nSelection criteria\nWe included randomised controlled trials (RCTs) evaluating the effects of any form of exercise as a single intervention on falls in people aged 60+ years living in the community. We excluded trials focused on particular conditions, such as stroke.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. Our primary outcome was rate of falls.\nMain results\nWe included 108 RCTs with 23,407 participants living in the community in 25 countries. There were nine cluster-RCTs. On average, participants were 76 years old and 77% were women. Most trials had unclear or high risk of bias for one or more items. Results from four trials focusing on people who had been recently discharged from hospital and from comparisons of different exercises are not described here.\nExercise (all types) versus control\nEighty-one trials (19,684 participants) compared exercise (all types) with control intervention (one not thought to reduce falls). Exercise reduces the rate of falls by 23% (rate ratio (RaR) 0.77, 95% confidence interval (CI) 0.71 to 0.83; 12,981 participants, 59 studies; high-certainty evidence). Based on an illustrative risk of 850 falls in 1000 people followed over one year (data based on control group risk data from the 59 studies), this equates to 195 (95% CI 144 to 246) fewer falls in the exercise group. Exercise also reduces the number of people experiencing one or more falls by 15% (risk ratio (RR) 0.85, 95% CI 0.81 to 0.89; 13,518 participants, 63 studies; high-certainty evidence). Based on an illustrative risk of 480 fallers in 1000 people followed over one year (data based on control group risk data from the 63 studies), this equates to 72 (95% CI 52 to 91) fewer fallers in the exercise group. Subgroup analyses showed no evidence of a difference in effect on both falls outcomes according to whether trials selected participants at increased risk of falling or not.\nThe findings for other outcomes are less certain, reflecting in part the relatively low number of studies and participants. Exercise may reduce the number of people experiencing one or more fall-related fractures (RR 0.73, 95% CI 0.56 to 0.95; 4047 participants, 10 studies; low-certainty evidence) and the number of people experiencing one or more falls requiring medical attention (RR 0.61, 95% CI 0.47 to 0.79; 1019 participants, 5 studies; low-certainty evidence). The effect of exercise on the number of people who experience one or more falls requiring hospital admission is unclear (RR 0.78, 95% CI 0.51 to 1.18; 1705 participants, 2 studies, very low-certainty evidence). Exercise may make little important difference to health-related quality of life: conversion of the pooled result (standardised mean difference (SMD) -0.03, 95% CI -0.10 to 0.04; 3172 participants, 15 studies; low-certainty evidence) to the EQ-5D and SF-36 scores showed the respective 95% CIs were much smaller than minimally important differences for both scales.\nAdverse events were reported to some degree in 27 trials (6019 participants) but were monitored closely in both exercise and control groups in only one trial. Fourteen trials reported no adverse events. Aside from two serious adverse events (one pelvic stress fracture and one inguinal hernia surgery) reported in one trial, the remainder were non-serious adverse events, primarily of a musculoskeletal nature. There was a median of three events (range 1 to 26) in the exercise groups.\nDifferent exercise types versus control\nDifferent forms of exercise had different impacts on falls (test for subgroup differences, rate of falls: P = 0.004, I² = 71%). Compared with control, balance and functional exercises reduce the rate of falls by 24% (RaR 0.76, 95% CI 0.70 to 0.81; 7920 participants, 39 studies; high-certainty evidence) and the number of people experiencing one or more falls by 13% (RR 0.87, 95% CI 0.82 to 0.91; 8288 participants, 37 studies; high-certainty evidence). Multiple types of exercise (most commonly balance and functional exercises plus resistance exercises) probably reduce the rate of falls by 34% (RaR 0.66, 95% CI 0.50 to 0.88; 1374 participants, 11 studies; moderate-certainty evidence) and the number of people experiencing one or more falls by 22% (RR 0.78, 95% CI 0.64 to 0.96; 1623 participants, 17 studies; moderate-certainty evidence). Tai Chi may reduce the rate of falls by 19% (RaR 0.81, 95% CI 0.67 to 0.99; 2655 participants, 7 studies; low-certainty evidence) as well as reducing the number of people who experience falls by 20% (RR 0.80, 95% CI 0.70 to 0.91; 2677 participants, 8 studies; high-certainty evidence). We are uncertain of the effects of programmes that are primarily resistance training, or dance or walking programmes on the rate of falls and the number of people who experience falls. No trials compared flexibility or endurance exercise versus control.\nAuthors' conclusions\nExercise programmes reduce the rate of falls and the number of people experiencing falls in older people living in the community (high-certainty evidence). The effects of such exercise programmes are uncertain for other non-falls outcomes. Where reported, adverse events were predominantly non-serious.\nExercise programmes that reduce falls primarily involve balance and functional exercises, while programmes that probably reduce falls include multiple exercise categories (typically balance and functional exercises plus resistance exercises). Tai Chi may also prevent falls but we are uncertain of the effect of resistance exercise (without balance and functional exercises), dance, or walking on the rate of falls.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "At least one-third of community-dwelling people over 65 years of age fall each year. Exercises that target balance, gait and muscle strength have previously been found to prevent falls in these people." } ]
query-laysum
6706
[ { "role": "user", "content": "Abstract: Background\nEpidermal growth factor receptor (EGFR) inhibitors prevent cell growth and have shown benefit in the treatment of metastatic colorectal cancer, whether used as single agents or in combination with chemotherapy. Clear benefit has been shown in trials of EGFR monoclonal antibodies (EGFR MAb) but not EGFR tyrosine kinase inhibitors (EGFR TKI). However, there is ongoing debate as to which patient populations gain maximum benefit from EGFR inhibition and where they should be used in the metastatic colorectal cancer treatment paradigm to maximise efficacy and minimise toxicity.\nObjectives\nTo determine the efficacy, safety profile, and potential harms of EGFR inhibitors in the treatment of people with metastatic colorectal cancer when given alone, in combination with chemotherapy, or with other biological agents.\nThe primary outcome of interest was progression-free survival; secondary outcomes included overall survival, tumour response rate, quality of life, and adverse events.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), the Cochrane Library, Issue 9, 2016; Ovid MEDLINE (from 1950); and Ovid Embase (from 1974) on 9 September 2016; and ClinicalTrials.gov and the World Health Organization International Clinical Trials Registry Platform (WHO ICTRP) on 14 March 2017. We also searched proceedings from the major oncology conferences ESMO, ASCO, and ASCO GI from 2012 to December 2016. We further scanned reference lists from eligible publications and contacted corresponding authors for trials for further information where needed.\nSelection criteria\nWe included randomised controlled trials on participants with metastatic colorectal cancer comparing: 1) the combination of EGFR MAb and 'standard therapy' (whether chemotherapy or best supportive care) to standard therapy alone, 2) the combination of EGFR TKI and standard therapy to standard therapy alone, 3) the combination of EGFR inhibitor (whether MAb or TKI) and standard therapy to another EGFR inhibitor (or the same inhibitor with a different dosing regimen) and standard therapy, or 4) the combination of EGFR inhibitor (whether MAb or TKI), anti-angiogenic therapy, and standard therapy to anti-angiogenic therapy and standard therapy alone.\nData collection and analysis\nWe used standard methodological procedures defined by Cochrane. Summary statistics for the endpoints used hazard ratios (HR) with 95% confidence intervals (CI) for overall survival and progression-free survival, and odds ratios (OR) for response rate (RR) and toxicity. Subgroup analyses were performed by Kirsten rat sarcoma viral oncogene homolog (KRAS) and neuroblastoma RAS viral (V-Ras) oncogene homolog (NRAS) status - firstly by status of KRAS exon 2 testing (mutant or wild type) and also by status of extended KRAS/NRAS testing (any mutation present or wild type).\nMain results\nWe identified 33 randomised controlled trials for analysis (15,025 participants), including trials of both EGFR MAb and EGFR TKI. Looking across studies, significant risk of bias was present, particularly with regard to the risk of selection bias (15/33 unclear risk, 1/33 high risk), performance bias (9/33 unclear risk, 9/33 high risk), and detection bias (7/33 unclear risk, 11/33 high risk).\nThe addition of EGFR MAb to standard therapy in the KRAS exon 2 wild-type population improves progression-free survival (HR 0.70, 95% CI 0.60 to 0.82; high-quality evidence), overall survival (HR 0.88, 95% CI 0.80 to 0.98; high-quality evidence), and response rate (OR 2.41, 95% CI 1.70 to 3.41; high-quality evidence). We noted evidence of significant statistical heterogeneity in all three of these analyses (progression-free survival: I2 = 76%; overall survival: I2 = 40%; and response rate: I2 = 77%), likely due to pooling of studies investigating EGFR MAb use in different lines of therapy. Rates of overall grade 3 to 4 toxicity, diarrhoea, and rash were increased (moderate-quality evidence for all three outcomes), but there was no evidence for increased rates of neutropenia.\nFor the extended RAS wild-type population (no mutations in KRAS or NRAS), addition of EGFR MAb improved progression-free survival (HR 0.60, 95% CI 0.48 to 0.75; moderate-quality evidence) and overall survival (HR 0.77, 95% CI 0.67 to 0.88; high-quality evidence). Response rate was also improved (OR 4.28, 95% CI 2.61 to 7.03; moderate-quality evidence). We noted significant statistical heterogeneity in the progression-free survival analysis (I2 = 61%), likely due to the pooling of studies combining EGFR MAb with chemotherapy with monotherapy studies.\nWe observed no evidence of a statistically significant difference when EGFR MAb was compared to bevacizumab, in progression-free survival (HR 1.02, 95% CI 0.93 to 1.12; high quality evidence) or overall survival (HR 0.84, 95% CI 0.70 to 1.01; moderate-quality evidence). We noted significant statistical heterogeneity in the overall survival analysis (I2 = 51%), likely due to the pooling of first-line and second-line studies.\nThe addition of EGFR TKI to standard therapy in molecularly unselected participants did not show benefit in limited data sets (meta-analysis not performed). The addition of EGFR MAb to bevacizumab plus chemotherapy in people with KRAS exon 2 wild-type metastatic colorectal cancer did not improve progression-free survival (HR 1.04, 95% CI 0.83 to 1.29; very low quality evidence), overall survival (HR 1.00, 95% CI 0.69 to 1.47; low-quality evidence), or response rate (OR 1.20, 95% CI 0.67 to 2.12; very low-quality evidence) but increased toxicity (OR 2.57, 95% CI 1.45 to 4.57; low-quality evidence). We noted significant between-study heterogeneity in most analyses.\nScant information on quality of life was reported in the identified studies.\nAuthors' conclusions\nThe addition of EGFR MAb to either chemotherapy or best supportive care improves progression-free survival (moderate- to high-quality evidence), overall survival (high-quality evidence), and tumour response rate (moderate- to high-quality evidence), but may increase toxicity in people with KRAS exon 2 wild-type or extended RAS wild-type metastatic colorectal cancer (moderate-quality evidence). The addition of EGFR TKI to standard therapy does not improve clinical outcomes. EGFR MAb combined with bevacizumab is of no clinical value (very low-quality evidence). Future studies should focus on optimal sequencing and predictive biomarkers and collect quality of life data.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Cancer of the colon or rectum that has spread to other organs (metastatic colorectal cancer) is a commonly occurring disease that usually cannot be surgically removed. The main treatment is chemotherapy, targeted therapy (such as EGFR inhibitors, the subject of this review), or both. The epidermal growth factor receptor (EGFR) is a protein found on cells that plays a vital role in promoting cell growth. Monoclonal antibodies are molecules developed to attach to a particular protein in order to enhance or decrease action at that protein site. EGFR monoclonal antibodies (EGFR MAb), such as cetuximab and panitumumab, specifically target and block EGFR, which stops cancer cell growth. Research has shown that people with mutations (gene changes) in KRAS (a gene related to EGFR) may not benefit from these drugs ('KRAS mutant'), but those without mutations ('KRAS wild type') do benefit. Recent research also suggests that people with mutations in another related gene (NRAS) may not benefit from these drugs either – that is, patients need to have no mutation in either KRAS or NRAS (otherwise known as 'extended RAS wild type').\nAnother type of EGFR-blocking drug known as tyrosine kinase inhibitor (EGFR TKI) (e.g. erlotinib and gefitinib) is effective in the treatment of lung cancer with EGFR mutations, but its benefit in colorectal cancer is unclear." } ]
query-laysum
6707
[ { "role": "user", "content": "Abstract: Background\nThe prevalence of opiate use among pregnant women can range from 1% to 2% to as high as 21%. Just in the United States alone, among pregnant women with hospital delivery, a fourfold increase in opioid use is reported from 1999 to 2014 (Haight 2018). Heroin crosses the placenta, and pregnant, opiate-dependent women experience a six-fold increase in maternal obstetric complications such as low birth weight, toxaemia, third trimester bleeding, malpresentation, puerperal morbidity, fetal distress and meconium aspiration. Neonatal complications include narcotic withdrawal, postnatal growth deficiency, microcephaly, neuro-behavioural problems, increased neonatal mortality and a 74-fold increase in sudden infant death syndrome. This is an updated version of the original Cochrane Review first published in 2008 and last updated in 2013.\nObjectives\nTo assess the effectiveness of any maintenance treatment alone or in combination with a psychosocial intervention compared to no intervention, other pharmacological intervention or psychosocial interventions alone for child health status, neonatal mortality, retaining pregnant women in treatment, and reducing the use of substances.\nSearch methods\nWe updated our searches of the following databases to February 2020: the Cochrane Drugs and Alcohol Group Specialised Register, CENTRAL, MEDLINE, Embase, PsycINFO, CINAHL, and Web of Science. We also searched two trials registers and checked the reference lists of included studies for further references to relevant randomised controlled trials (RCTs).\nSelection criteria\nRandomised controlled trials which assessed the efficacy of any pharmacological maintenance treatment for opiate-dependent pregnant women.\nData collection and analysis\nWe used the standard methodological procedures expected by Cochrane.\nMain results\nWe found four trials with 271 pregnant women. Three compared methadone with buprenorphine and one methadone with oral slow-release morphine. Three out of four studies had adequate allocation concealment and were double-blind. The major flaw in the included studies was attrition bias: three out of four had a high dropout rate (30% to 40%), and this was unbalanced between groups.\nMethadone versus buprenorphine:\nThere was probably no evidence of a difference in the dropout rate from treatment (risk ratio (RR) 0.66, 95% confidence interval (CI) 0.37 to 1.20, three studies, 223 participants, moderate-quality evidence). There may be no evidence of a difference in the use of primary substances between methadone and buprenorphine (RR 1.81, 95% CI 0.70 to 4.68, two studies, 151 participants, low-quality evidence). Birth weight may be higher in the buprenorphine group in the two trials that reported data MD;-530.00 g, 95%CI -662.78 to -397.22 (one study, 19 particpants) and MD: -215.00 g, 95%CI -238.93 to -191.07 (one study, 131 participants) although the results could not be pooled due to very high heterogeneity (very low-quality of evidence). The third study reported that there was no evidence of a difference. We found there may be no evidence of a difference in the APGAR score (MD: 0.00, 95% CI -0.03 to 0.03, two studies,163 participants, low-quality evidence). Many measures were used in the studies to assess neonatal abstinence syndrome. The number of newborns treated for neonatal abstinence syndrome, which is the most critical outcome, may not differ between groups (RR 1.19, 95% CI 0.87 to1.63, three studies, 166 participants, low-quality evidence).\nOnly one study which compared methadone with buprenorphine reported side effects. We found there may be no evidence of a difference in the number of mothers with serious adverse events (AEs) (RR 1.69, 95% CI 0.75 to 3.83, 175 participants, low-quality evidence) and we found there may be no difference in the numbers of newborns with serious AEs (RR 4.77, 95% CI 0.59, 38.49,131 participants, low-quality evidence).\nMethadone versus slow-release morphine:\nThere were no dropouts in either treatment group. Oral slow-release morphine may be superior to methadone for abstinence from heroin use during pregnancy (RR 2.40, 95% CI 1.00 to 5.77, one study, 48 participants, low-quality evidence).\nIn the comparison between methadone and slow-release morphine, no side effects were reported for the mother. In contrast, one child in the methadone group had central apnoea, and one child in the morphine group had obstructive apnoea (low-quality evidence).\nAuthors' conclusions\nMethadone and buprenorphine may be similar in efficacy and safety for the treatment of opioid-dependent pregnant women and their babies. There is not enough evidence to make conclusions for the comparison between methadone and slow-release morphine. Overall, the body of evidence is too small to make firm conclusions about the equivalence of the treatments compared. There is still a need for randomised controlled trials of adequate sample size comparing different maintenance treatments.\n\nGiven the provided abstract, please respond to the following query: \"Quality of evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "In the comparison of methadone with buprenorphine, the quality of the evidence ranged from moderate to very low because of inconsistency in the results of the studies for some outcomes, high rates of participants who dropped out from the studies, and small sample sizes of the included studies. In the comparison of methadone with slow-release morphine, the quality of the evidence was low because of the small sample size of the study." } ]
query-laysum
6708
[ { "role": "user", "content": "Abstract: Background\nAdequate nutrients early in life promote cognitive development and are critical for proper growth and functioning. The effect of individual nutrients consumed through food is often not the same as consuming the same nutrients in supplementary form due to 'food synergy', the biological and chemical interrelations that occur between nutrients. Animal-source foods, such as eggs, meat, fish, and dairy, are energy dense and contain multiple micronutrients and essential fatty acids with high bioavailability. The benefits of animal-source foods may include higher food synergy relative to fortified foods as well as decreasing dependence on external suppliers of fortified foods.\nObjectives\nTo assess the effectiveness of animal-source foods compared to any other feeding interventions or no intervention in improving growth and developmental outcomes in children aged 6 to 59 months.\nSearch methods\nWe searched CENTRAL, MEDLINE, Embase, CINAHL, 18 other databases, and three trials registers up to August 2018. We also contacted authors and known experts in the field for assistance in identifying ongoing or unpublished data, and searched the reference lists of included studies and reviews, and websites of relevant organizations, for other studies that may not have been captured by our electronic searches.\nSelection criteria\nWe included randomized controlled trials and quasi-randomized controlled trials of any duration, where children between 5 months and 59 months (6 years) of age were provided with an animal-source food (e.g. consumption of milk, meat, or eggs), prepared with any cooking method, compared with any intervention or no intervention.\nData collection and analysis\nTwo review authors independently assessed trial eligibility using prespecified criteria, extracted data, assessed risk of bias, and graded the quality of the evidence using the GRADE approach.\nMain results\nStudy characteristics\nWe included 6 studies that analyzed data from 3036 children aged 5 to 50 months. The studies were conducted in China, the Democratic Republic of Congo, Ecuador, Guatemala, Pakistan, the USA, and Zambia, and lasted between 5 and 12 months. Three studies were funded, in part, by government entities; one study was supported by a nonprofit organization. Two studies did not report a funding source.\nThree studies compared the effects of feeding an animal-source food with a fortified (iron or iron and zinc), or unfortified cereal; two used a control group with no intervention; one compared a meat-based diet to a dairy-based diet. The types of animal-source foods tested included yogurt, eggs, cheese, lyophilized (freeze-dried) beef product, ground and frozen pork, puréed and jarred beef with gravy or pork, and powdered whey protein.\nWe judged four studies to be at unclear risk of bias overall; three studies because they were funded by an industry with a plausible interest in the outcome of the intervention; and one study because there was insufficient information to assess five of the seven bias 'Risk of bias' domains. We judged two of the six studies to be at high risk of bias overall; one study because there was significant baseline imbalance in length-for-age z scores (LAZ) between groups and evidence of selective reporting; the other study because there there was both a significant baseline imbalance in LAZ and weight-for-age z scores (WAZ) between groups, and a large-scale social media campaign that may have influenced care received at home in the control group.\nKey results\nAnimal-source foods versus cereal-based foods or no intervention\nFive studies (2972 children) measured change in linear growth with either height-for-age z scores (HAZ) or LAZ. Three studies (592 children) reported a significant increase in HAZ and LAZ in the intervention group compared to the control group. Two studies (2380 children) reported a decline in LAZ in both groups. In one study (1062 children) there was no difference between the groups in the rate of decline; in the other (1318 children) the decrease in LAZ was significantly smaller in the intervention group.\nFive studies (2972 children) measured weight gain using WAZ. Three studies (592 children) reported a significant increase in WAZ in the intervention group compared to the control group. In two studies (2380 children), WAZ decreased in both groups. In one of these studies (1318 children), the decrease in the intervention group was significantly smaller than in the control group. In the other study (1062 children), there was no difference between the groups.\nThree studies (1612 children) reported impacts on all-cause morbidity, but metrics were inconsistent between studies. One study with yogurt (402 children) reported a significant reduction in duration and incidence of diarrhea and upper respiratory infections in the intervention group. One study with eggs (148 children) reported a significant increase in the incidence of diarrhea in the intervention group, but this may have been due to cultural associations with eggs and gastrointestional problems. There were no other significant differences in fever, respiratory infections, or skin conditions between groups. The third study (1062 children) found no differences between intervention and control groups across morbidity measures.\nNo studies reported data on anemia.\nMeat-based diet versus dairy-based diet\nOne study (64 children) measured change in LAZ and WAZ in infants fed either a meat-based diet or dairy-based diet. There was a significant increase in LAZ among infants consuming the meat-based diet and a significant decrease in LAZ among infants consuming a dairy-based diet. WAZ increased in both groups, with no significant difference between groups.\nThe study did not assess all-cause morbidity or anemia.\nQuality of the evidence\nWe rated the quality of the evidence as very low overall due to baseline imbalances between intervention and control groups, high heterogeneity in meta-analysis, and imprecision due to wide confidence intervals and inconsistent direction of effects. We have little confidence in the results; further research is likely to change the estimate of magnitude and direction of treatment effect.\nAuthors' conclusions\nGiven the limited quality of the evidence, we are uncertain of the effects of the provision of animal-source food versus cereal products or no intervention on the growth or development of children. More adequately powered trials with deliberately selected animal-source foods are needed.\n\nGiven the provided abstract, please respond to the following query: \"What is the aim of this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We reviewed the evidence about the effect of animal-source foods on the growth and development of children between 6 and 59 months of age." } ]
query-laysum
6709
[ { "role": "user", "content": "Abstract: Background\nHistorically, whole brain radiation therapy (WBRT) has been the main treatment for brain metastases. Stereotactic radiosurgery (SRS) delivers high-dose focused radiation and is being increasingly utilized to treat brain metastases. The benefit of adding SRS to WBRT is unclear. This is an updated version of the original Cochrane Review published in Issue 9, 2012.\nObjectives\nTo assess the efficacy of WBRT plus SRS versus WBRT alone in the treatment of adults with brain metastases.\nSearch methods\nFor the original review, in 2009 we searched the following electronic databases: CENTRAL, MEDLINE, Embase, and CancerLit in order to identify trials for inclusion in this review. For the first update the searches were updated in May 2012.\nFor this update, in May 2017 we searched CENTRAL, MEDLINE, and Embase in order to identify trials for inclusion in the review.\nSelection criteria\nWe restricted the review to randomized controlled trials (RCTs) that compared use of WBRT plus SRS versus WBRT alone for upfront treatment of adults with newly diagnosed metastases (single or multiple) in the brain resulting from any primary, extracranial cancer.\nData collection and analysis\nWe used the generic inverse variance method, random-effects model in Review Manager 5 for the meta-analysis.\nMain results\nWe identified three studies and one abstract for inclusion but we could only include two studies, with a total of 358 participants in a meta-analysis. This found no difference in overall survival (OS) between the WBRT plus SRS and WBRT alone groups (hazard ratio (HR) 0.82, 95% confidence interval (CI) 0.65 to 1.02; 2 studies, 358 participants; moderate-quality evidence). For participants with one brain metastasis median survival was significantly longer in the WBRT plus SRS group (6.5 months) versus WBRT group (4.9 months; P = 0.04). Participants in the WBRT plus SRS group had decreased local failure compared to participants who received WBRT alone (HR 0.27, 95% CI 0.14 to 0.52; 2 studies, 129 participants; moderate-quality evidence). Furthermore, we observed an improvement in performance status scores and decrease in steroid use in the WBRT plus SRS group (risk ratio (RR) 0.64 CI 0.42 to 0.97; 1 study, 118 participants; low-quality evidence). Unchanged or improved Karnofsky Performance Scale (KPS) at six months was seen in 43% of participants in the combined therapy group versus only 28% in the WBRT-alone group (RR 0.78 CI 0.61 to 1.00; P value = 0.05; 1 study, 118 participants; low-quality evidence). Overall, risk of bias in the included studies was unclear.\nAuthors' conclusions\nSince the last version of this review we have identified one new study that met the inclusion criteria. However, due to a lack of data from this study we were not able to include it in a meta-analysis. Given the unclear risk of bias in the included studies, the results of this analysis have to be interpreted with caution. In our analysis of all included participants, SRS plus WBRT did not show a survival benefit over WBRT alone. However, performance status and local control were significantly better in the SRS plus WBRT group. Furthermore, significantly longer OS was reported in the combined treatment group for recursive partitioning analysis (RPA) Class I patients as well as patients with single metastasis. Most of our outcomes of interest were graded as moderate-quality evidence according to the GRADE criteria and the risk of bias in the majority of included studies was mostly unclear.\n\nGiven the provided abstract, please respond to the following query: \"What are the conclusions?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Most of our conclusions are based on the results of one large trial with unclear risk of bias and therefore, we cautiously make the following remarks: we found that when radiosurgery was added to WBRT, there was no evidence to suggest that people lived any longer than if they had WBRT alone, except for people with only one brain metastasis (who may live longer if they receive the combination treatment). People having combination treatment also seemed to function better in daily life, their treated tumors were associated with having less chance of growing back, and they had to take less steroid medication. The side effects of combined therapy and WBRT alone were similar." } ]
query-laysum
6710
[ { "role": "user", "content": "Abstract: Background\nAlveolar osteitis (dry socket) is a complication of dental extractions more often involving mandibular molar teeth. It is associated with severe pain developing 2 to 3 days postoperatively with or without halitosis, a socket that may be partially or totally devoid of a blood clot, and increased postoperative visits. This is an update of the Cochrane Review first published in 2012.\nObjectives\nTo assess the effects of local interventions used for the prevention and treatment of alveolar osteitis (dry socket) following tooth extraction.\nSearch methods\nAn Information Specialist searched four bibliographic databases up to 28 September 2021 and used additional search methods to identify published, unpublished, and ongoing studies.\nSelection criteria\nWe included randomised controlled trials of adults over 18 years of age who were having permanent teeth extracted or who had developed dry socket postextraction. We included studies with any type of local intervention used for the prevention or treatment of dry socket, compared to a different local intervention, placebo or no treatment. We excluded studies reporting on systemic use of antibiotics or the use of surgical techniques because these interventions are evaluated in separate Cochrane Reviews.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. We followed Cochrane statistical guidelines and reported dichotomous outcomes as risk ratios (RR) and calculated 95% confidence intervals (CI) using random-effects models. For some of the split-mouth studies with sparse data, it was not possible to calculate RR so we calculated the exact odds ratio (OR) instead. We used GRADE to assess the certainty of the body of evidence.\nMain results\nWe included 49 trials with 6771 participants; 39 trials (with 6219 participants) investigated prevention of dry socket and 10 studies (with 552 participants) looked at the treatment of dry socket. 16 studies were at high risk of bias, 30 studies at unclear risk of bias, and 3 studies at low risk of bias.\nChlorhexidine in the prevention of dry socket\nWhen compared to placebo, rinsing with chlorhexidine mouthrinses (0.12% and 0.2% concentrations) both before and 24 hours after extraction(s) substantially reduced the risk of developing dry socket with an OR of 0.38 (95% CI 0.25 to 0.58; P < 0.00001; 6 trials, 1547 participants; moderate-certainty evidence). The prevalence of dry socket varies from 1% to 5% in routine dental extractions to upwards of 30% in surgically extracted third molars. The number of patients needed to be treated (NNT) with chlorhexidine rinse to prevent one patient having dry socket was 162 (95% CI 155 to 240), 33 (95% CI 27 to 49), and 7 (95% CI 5 to 10) for control prevalence of dry socket 0.01, 0.05, and 0.30 respectively.\nCompared to placebo, placing chlorhexidine gel intrasocket after extractions reduced the odds of developing a dry socket by 58% with an OR of 0.44 (95% CI 0.27 to 0.71; P = 0.0008; 7 trials, 753 participants; moderate-certainty evidence). The NNT with chlorhexidine gel (0.2%) to prevent one patient developing dry socket was 180 (95% CI 137 to 347), 37 (95% CI 28 to 72), and 7 (95% CI 5 to 15) for control prevalence of dry socket of 0.01, 0.05, and 0.30 respectively.\nCompared to chlorhexidine rinse (0.12%), placing chlorhexidine gel (0.2%) intrasocket after extractions was not superior in reducing the risk of dry socket (RR 0.74, 95% CI 0.46 to 1.20; P = 0.22; 2 trials, 383 participants; low-certainty evidence).\nThe present review found some evidence for the association of minor adverse reactions with use of 0.12%, 0.2% chlorhexidine mouthrinses (alteration in taste, staining of teeth, stomatitis) though most studies were not designed explicitly to detect the presence of hypersensitivity reactions to mouthwash as part of the study protocol. No adverse events were reported in relation to the use of 0.2% chlorhexidine gel placed directly into a socket.\nPlatelet rich plasma in the prevention of dry socket\nCompared to placebo, placing platelet rich plasma after extractions was not superior in reducing the risk of having a dry socket (RR 0.51, 95% CI 0.19 to 1.33; P = 0.17; 2 studies, 127 participants; very low-certainty evidence).\nA further 21 intrasocket interventions to prevent dry socket were each evaluated in single studies, and there is insufficient evidence to determine their effects.\nZinc oxide eugenol versus Alvogyl in the treatment of dry socket\nTwo studies, with 80 participants, showed that Alvogyl (old formulation) is more effective than zinc oxide eugenol at reducing pain at day 7 (mean difference (MD) -1.40, 95% CI -1.75 to -1.04; P < 0.00001; 2 studies, 80 participants; very low-certainty evidence)\nA further nine interventions for the treatment of dry socket were evaluated in single studies, providing insufficient evidence to determine their effects.\nAuthors' conclusions\nTooth extractions are generally undertaken by dentists for a variety of reasons, however, all but five studies included in the present review included participants undergoing extraction of third molars, most of which were undertaken by oral surgeons. There is moderate-certainty evidence that rinsing with chlorhexidine (0.12% and 0.2%) or placing chlorhexidine gel (0.2%) in the sockets of extracted teeth, probably results in a reduction in dry socket. There was insufficient evidence to determine the effects of the other 21 preventative interventions each evaluated in single studies. There was limited evidence of very low certainty that Alvogyl (old formulation) may reduce pain at day 7 in patients with dry socket when compared to zinc oxide eugenol.\n\nGiven the provided abstract, please respond to the following query: \"What did we find?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We identified 49 trials; 39 trials (6219 participants) investigated prevention of dry socket and 10 trials (552 participants) investigated the treatment of dry socket.\nWe found that:\n- rinsing both before and after tooth extraction (commencing 24 hours after extraction) with chlorhexidine gluconate rinse (at 0.12% and 0.2% strength) probably results in a reduction in dry socket;\n- placing chlorhexidine gel (0.2% strength) in the socket of an extracted tooth probably results in a reduction in dry socket;\n- chlorhexidine rinses and gels are equally effective at reducing dry socket, but the evidence for this comparison is very uncertain;\n- chlorhexidine rinses produced some minor unwanted effects; chlorhexidine intrasocket gels appeared to produce no unwanted effects;\n- there was a small amount of evidence of very low certainty from two studies that Alvogyl (old formulation) may reduce pain at day 7 in patients with dry socket when compared to zinc oxide eugenol. This evidence relates to the old formulation of Alvogyl which is no longer available. It should be noted that the formulation of Alvogyl has changed, it is now called Alveogyl.\nTooth extractions are generally undertaken by dentists for a variety of reasons, however, all but five studies included in the present review included participants undergoing extraction of third molars, most of which were undertaken by oral surgeons." } ]
query-laysum
6711
[ { "role": "user", "content": "Abstract: Background\nHaving nurses take on tasks that are typically conducted by doctors (doctor-nurse substitution, a form of 'task-shifting') may help to address doctor shortages and reduce doctors' workload and human resource costs. A Cochrane Review of effectiveness studies suggested that nurse-led care probably leads to similar healthcare outcomes as care delivered by doctors. This finding highlights the need to explore the factors that affect the implementation of strategies to substitute doctors with nurses in primary care. In our qualitative evidence synthesis (QES), we focused on studies of nurses taking on tasks that are typically conducted by doctors working in primary care, including substituting doctors with nurses or expanding nurses' roles.\nObjectives\n(1) To identify factors influencing implementation of interventions to substitute doctors with nurses in primary care. (2) To explore how our synthesis findings related to, and helped to explain, the findings of the Cochrane intervention review of the effectiveness of substituting doctors with nurses. (3) To identify hypotheses for subgroup analyses for future updates of the Cochrane intervention review.\nSearch methods\nWe searched CINAHL and PubMed, contacted experts in the field, scanned the reference lists of relevant studies and conducted forward citation searches for key articles in the Social Science Citation Index and Science Citation Index databases, and 'related article' searches in PubMed.\nSelection criteria\nWe constructed a maximum variation sample (exploring variables such as country level of development, aspects of care covered and the types of participants) from studies that had collected and analysed qualitative data related to the factors influencing implementation of doctor-nurse substitution and the expansion of nurses' tasks in community or primary care worldwide. We included perspectives of doctors, nurses, patients and their families/carers, policymakers, programme managers, other health workers and any others directly involved in or affected by the substitution. We excluded studies that collected data using qualitative methods but did not analyse the data qualitatively.\nData collection and analysis\nWe identified factors influencing implementation of doctor-nurse substitution strategies using a framework thematic synthesis approach. Two review authors independently assessed the methodological strengths and limitations of included studies using a modified Critical Appraisal Skills Programme (CASP) tool. We assessed confidence in the evidence for the QES findings using the GRADE-CERQual approach. We integrated our findings with the evidence from the effectiveness review of doctor-nurse substitution using a matrix model. Finally, we identified hypotheses for subgroup analyses for updates of the review of effectiveness.\nMain results\nWe included 66 studies (69 papers), 11 from low- or middle-income countries and 55 from high-income countries. These studies found several factors that appeared to influence the implementation of doctor-nurse substitution strategies. The following factors were based on findings that we assessed as moderate or high confidence.\nPatients in many studies knew little about nurses' roles and the difference between nurse-led and doctor-led care. They also had mixed views about the type of tasks that nurses should deliver. They preferred doctors when the tasks were more 'medical' but accepted nurses for preventive care and follow-ups. Doctors in most studies also preferred that nurses performed only 'non-medical' tasks. Nurses were comfortable with, and believed they were competent to deliver a wide range of tasks, but particularly emphasised tasks that were more health promotive/preventive in nature.\nPatients in most studies thought that nurses were more easily accessible than doctors. Doctors and nurses also saw nurse-doctor substitution and collaboration as a way of increasing people's access to care, and improving the quality and continuity of care.\nNurses thought that close doctor-nurse relationships and doctor's trust in and acceptance of nurses was important for shaping their roles. But nurses working alone sometimes found it difficult to communicate with doctors.\nNurses felt they had gained new skills when taking on new tasks. But nurses wanted more and better training. They thought this would increase their skills, job satisfaction and motivation, and would make them more independent.\nNurses taking on doctors' tasks saw this as an opportunity to develop personally, to gain more respect and to improve the quality of care they could offer to patients. Better working conditions and financial incentives also motivated nurses to take on new tasks. Doctors valued collaborating with nurses when this reduced their own workload.\nDoctors and nurses pointed to the importance of having access to resources, such as enough staff, equipment and supplies; good referral systems; experienced leaders; clear roles; and adequate training and supervision. But they often had problems with these issues. They also pointed to the huge number of documents they needed to complete when tasks were moved from doctors to nurses.\nAuthors' conclusions\nPatients, doctors and nurses may accept the use of nurses to deliver services that are usually delivered by doctors. But this is likely to depend on the type of services. Nurses taking on extra tasks want respect and collaboration from doctors; as well as proper resources; good referral systems; experienced leaders; clear roles; and adequate incentives, training and supervision. However, these needs are not always met.\n\nGiven the provided abstract, please respond to the following query: \"What was the aim of the review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "In this Cochrane Review of qualitative studies ('qualitative evidence synthesis'), we explored peoples' views and experiences of moving tasks from doctors to nurses in primary healthcare. We collected relevant studies on this topic and included 66 studies (69 papers).\nThis synthesis links to another Cochrane Review that assesses the effectiveness of moving tasks from doctors to nurses in primary care." } ]
query-laysum
6712
[ { "role": "user", "content": "Abstract: Background\nPrimary congenital glaucoma (PCG) is an optic neuropathy with high intraocular pressure (IOP) that manifests within the first few years of a child's life and is not associated with other systemic or ocular abnormalities. PCG results in considerable morbidity even in high-income countries.\nObjectives\nTo compare the effectiveness and safety of different surgical techniques for PCG.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL) (2020, Issue 4); Ovid MEDLINE; Embase.com; PubMed; metaRegister of Controlled Trials (mRCT) (last searched 23 June 2014); ClinicalTrials.gov; and the World Health Organization (WHO) International Clinical Trials Registry Platform (ICTRP). We did not use any date or language restrictions in the electronic search. We last searched the electronic databases on 27 April 2020.\nSelection criteria\nWe included randomized controlled trials (RCTs) and quasi-RCTs comparing different surgical interventions in children under five years of age with PCG.\nData collection and analysis\nWe used standard Cochrane methodology.\nMain results\nWe included 16 trials (13 RCTs and three quasi-RCTs) with 587 eyes in 446 children. Eleven (69%) trials were conducted in Egypt and the Middle East, three in India, and two in the USA. All included trials involved children younger than five years of age, with follow-up ranging from six to 80 months.\nThe interventions compared varied across trials. Three trials (on 68 children) compared combined trabeculotomy and trabeculectomy (CTT) with trabeculotomy. Meta-analysis of these trials suggests there may be little to no evidence of a difference between groups in mean IOP (mean difference (MD) 0.27 mmHg, 95% confidence interval (CI) −0.74 to 1.29; 88 eyes; 2 studies) and surgical success (risk ratio (RR) 1.01, 95% CI 0.90 to 1.14; 102 eyes; 3 studies) at one year postoperatively. We assessed the certainty of evidence as very low for these outcomes, downgrading for risk of bias (-1) and imprecision (-2). Hyphema was the most common adverse outcome in both groups (no meta-analysis due to considerable heterogeneity; I2 = 83%).\nTwo trials (on 39 children) compared viscotrabeculotomy to conventional trabeculotomy. Meta-analysis of 42 eyes suggests there is no evidence of between groups difference in mean IOP (MD −1.64, 95% CI −5.94 to 2.66) and surgical success (RR 1.11, 95% CI 0.70 to 1.78) at six months postoperatively. We assessed the certainty of evidence as very low, downgrading for risk of bias and imprecision due to small sample size. Hyphema was the most common adverse outcome (38% in viscotrabeculotomy and 28% in conventional trabeculotomy), with no evidence of difference difference (RR 1.33, 95% CI 0.63 to 2.83).\nTwo trials (on 95 children) compared microcatheter-assisted 360-degree circumferential trabeculotomy to conventional trabeculotomy. Meta-analysis of two trials suggests that mean IOP may be lower in the microcatheter group at six months (MD −2.44, 95% CI −3.69 to −1.19; 100 eyes) and at 12 months (MD −1.77, 95% CI −2.92 to −0.63; 99 eyes); and surgical success was more likely to be achieved in the microcatheter group compared to the conventional trabeculotomy group (RR 1.59, 95% CI 1.14 to 2.21; 60 eyes; 1 trial at 6 months; RR 1.54, 95% CI 1.20 to 1.97; 99 eyes; 2 trials at 12 months). We assessed the certainty of evidence for these outcomes as moderate due to small sample size. Hyphema was the most common adverse outcome (40% in the microcatheter group and 17% in the conventional trabeculotomy group), with greater likelihood of occurring in the microcatheter group (RR 2.25, 95% CI 1.25 to 4.04); the evidence was of moderate certainty due to small sample size (−1).\nOf the nine remaining trials, no two trials compared the same two surgical interventions: one trial compared CTT versus CTT with sclerectomy; three trials compared various suturing techniques and adjuvant use including mitomycin C, collagen implant in CTT; one trial compared CTT versus Ahmed valve implant in previously failed surgeries; one trial compared CTT with trabeculectomy; one trial compared trabeculotomy to goniotomy; and two trials compared different types of goniotomy. No trials reported quality of life or economic data.\nMany of the included trials had limitations in study design, implementation, and reporting, therefore the reliability and applicability of the evidence remains unclear.\nAuthors' conclusions\nThe evidence suggests that there may be little to no evidence of difference between CTT and routine conventional trabeculotomy, or between viscotrabeculotomy and routine conventional trabeculotomy. A 360-degree circumferential trabeculotomy may show greater surgical success than conventional trabeculotomy. Considering the rarity of the disease, future research would benefit from a multicenter, possibly international trial, involving parents of children with PCG and with a follow-up of at least one year.\n\nGiven the provided abstract, please respond to the following query: \"How up-to-date is this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The evidence in this Cochrane Review is current to 27 April 2020." } ]
query-laysum
6713
[ { "role": "user", "content": "Abstract: Background\nCaries (dental decay) is a disease of the hard tissues of the teeth caused by an imbalance, over time, in the interactions between cariogenic bacteria in dental plaque and fermentable carbohydrates (mainly sugars). Regular toothbrushing with fluoride toothpaste is the principal non-professional intervention to prevent caries, but the caries-preventive effect varies according to different concentrations of fluoride in toothpaste, with higher concentrations associated with increased caries control. Toothpastes with higher fluoride concentration increases the risk of fluorosis (enamel defects) in developing teeth. This is an update of the Cochrane Review first published in 2010.\nObjectives\nTo determine and compare the effects of toothpastes of different fluoride concentrations (parts per million (ppm)) in preventing dental caries in children, adolescents, and adults.\nSearch methods\nCochrane Oral Health's Information Specialist searched the following databases: Cochrane Oral Health's Trials Register (to 15 August 2018); the Cochrane Central Register of Controlled Trials (CENTRAL; 2018, Issue 7) in the Cochrane Library (searched 15 August 2018); MEDLINE Ovid (1946 to 15 August 2018); and Embase Ovid (1980 to 15 August 2018). The US National Institutes of Health Ongoing Trials Register (ClinicalTrials.gov) and the World Health Organization International Clinical Trials Registry Platform were searched for ongoing trials (15 August 2018). No restrictions were placed on the language or date of publication when searching the electronic databases.\nSelection criteria\nRandomised controlled trials that compared toothbrushing with fluoride toothpaste with toothbrushing with a non-fluoride toothpaste or toothpaste of a different fluoride concentration, with a follow-up period of at least 1 year. The primary outcome was caries increment measured by the change from baseline in the decayed, (missing), and filled surfaces or teeth index in all permanent or primary teeth (D(M)FS/T or d(m)fs/t).\nData collection and analysis\nTwo members of the review team, independently and in duplicate, undertook the selection of studies, data extraction, and risk of bias assessment. We graded the certainty of the evidence through discussion and consensus. The primary effect measure was the mean difference (MD) or standardised mean difference (SMD) caries increment. Where it was appropriate to pool data, we used random-effects pairwise or network meta-analysis.\nMain results\nWe included 96 studies published between 1955 and 2014 in this updated review. Seven studies with 11,356 randomised participants (7047 evaluated) reported the effects of fluoride toothpaste up to 1500 ppm on the primary dentition; one study with 2500 randomised participants (2008 evaluated) reported the effects of 1450 ppm fluoride toothpaste on the primary and permanent dentition; 85 studies with 48,804 randomised participants (40,066 evaluated) reported the effects of toothpaste up to 2400 ppm on the immature permanent dentition; and three studies with 2675 randomised participants (2162 evaluated) reported the effects of up to 1100 ppm fluoride toothpaste on the mature permanent dentition. Follow-up in most studies was 36 months.\nIn the primary dentition of young children, 1500 ppm fluoride toothpaste reduces caries increment when compared with non-fluoride toothpaste (MD -1.86 dfs, 95% confidence interval (CI) -2.51 to -1.21; 998 participants, one study, moderate-certainty evidence); the caries-preventive effects for the head-to-head comparison of 1055 ppm versus 550 ppm fluoride toothpaste are similar (MD -0.05, dmfs, 95% CI -0.38 to 0.28; 1958 participants, two studies, moderate-certainty evidence), but toothbrushing with 1450 ppm fluoride toothpaste slightly reduces decayed, missing, filled teeth (dmft) increment when compared with 440 ppm fluoride toothpaste (MD -0.34, dmft, 95%CI -0.59 to -0.09; 2362 participants, one study, moderate-certainty evidence). The certainty of the remaining evidence for this comparison was judged to be low.\nWe included 81 studies in the network meta-analysis of D(M)FS increment in the permanent dentition of children and adolescents. The network included 21 different comparisons of seven fluoride concentrations. The certainty of the evidence was judged to be low with the following exceptions: there was high- and moderate-certainty evidence that 1000 to 1250 ppm or 1450 to 1500 ppm fluoride toothpaste reduces caries increments when compared with non-fluoride toothpaste (SMD -0.28, 95% CI -0.32 to -0.25, 55 studies; and SMD -0.36, 95% CI -0.43 to -0.29, four studies); there was moderate-certainty evidence that 1450 to 1500 ppm fluoride toothpaste slightly reduces caries increments when compared to 1000 to 1250 ppm (SMD -0.08, 95% CI -0.14 to -0.01, 10 studies); and moderate-certainty evidence that the caries increments are similar for 1700 to 2200 ppm and 2400 to 2800 ppm fluoride toothpaste when compared to 1450 to 1500 ppm (SMD 0.04, 95% CI -0.07 to 0.15, indirect evidence only; SMD -0.05, 95% CI -0.14 to 0.05, two studies).\nIn the adult permanent dentition, 1000 or 1100 ppm fluoride toothpaste reduces DMFS increment when compared with non-fluoride toothpaste in adults of all ages (MD -0.53, 95% CI -1.02 to -0.04; 2162 participants, three studies, moderate-certainty evidence). The evidence for DMFT was low certainty.\nOnly a minority of studies assessed adverse effects of toothpaste. When reported, effects such as soft tissue damage and tooth staining were minimal.\nAuthors' conclusions\nThis Cochrane Review supports the benefits of using fluoride toothpaste in preventing caries when compared to non-fluoride toothpaste. Evidence for the effects of different fluoride concentrations is more limited, but a dose-response effect was observed for D(M)FS in children and adolescents. For many comparisons of different concentrations the caries-preventive effects and our confidence in these effect estimates are uncertain and could be challenged by further research. The choice of fluoride toothpaste concentration for young children should be balanced against the risk of fluorosis.\n\nGiven the provided abstract, please respond to the following query: \"Authors conclusions\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "There are benefits of using fluoride toothpaste at certain strengths to prevent tooth decay when compared with non-fluoride toothpaste. The stronger the fluoride concentration, the more decay is prevented. For many of the comparisons of different strengths of toothpaste, the findings are uncertain and could be challenged by further research. The choice of fluoride toothpaste for young children should be balanced against the risk of fluorosis." } ]
query-laysum
6714
[ { "role": "user", "content": "Abstract: Background\nNutritional labelling is advocated as a means to promote healthier food purchasing and consumption, including lower energy intake. Internationally, many different nutritional labelling schemes have been introduced. There is no consensus on whether such labelling is effective in promoting healthier behaviour.\nObjectives\nTo assess the impact of nutritional labelling for food and non-alcoholic drinks on purchasing and consumption of healthier items. Our secondary objective was to explore possible effect moderators of nutritional labelling on purchasing and consumption.\nSearch methods\nWe searched 13 electronic databases including CENTRAL, MEDLINE and Embase to 26 April 2017. We also handsearched references and citations and sought unpublished studies through websites and trials registries.\nSelection criteria\nEligible studies: were randomised or quasi-randomised controlled trials (RCTs/Q-RCTs), controlled before-and-after studies, or interrupted time series (ITS) studies; compared a labelled product (with information on nutrients or energy) with the same product without a nutritional label; assessed objectively measured purchasing or consumption of foods or non-alcoholic drinks in real-world or laboratory settings.\nData collection and analysis\nTwo authors independently selected studies for inclusion and extracted study data. We applied the Cochrane 'Risk of bias' tool and GRADE to assess the quality of evidence. We pooled studies that evaluated similar interventions and outcomes using a random-effects meta-analysis, and we synthesised data from other studies in a narrative summary.\nMain results\nWe included 28 studies, comprising 17 RCTs, 5 Q-RCTs and 6 ITS studies. Most (21/28) took place in the USA, and 19 took place in university settings, 14 of which mainly involved university students or staff. Most (20/28) studies assessed the impact of labelling on menus or menu boards, or nutritional labelling placed on, or adjacent to, a range of foods or drinks from which participants could choose. Eight studies provided participants with only one labelled food or drink option (in which labelling was present on a container or packaging, adjacent to the food or on a display board) and measured the amount consumed. The most frequently assessed labelling type was energy (i.e. calorie) information (12/28).\nEleven studies assessed the impact of nutritional labelling on purchasing food or drink options in real-world settings, including purchases from vending machines (one cluster-RCT), grocery stores (one ITS), or restaurants, cafeterias or coffee shops (three RCTs, one Q-RCT and five ITS). Findings on vending machines and grocery stores were not interpretable, and were rated as very low quality. A meta-analysis of the three RCTs, all of which assessed energy labelling on menus in restaurants, demonstrated a statistically significant reduction of 47 kcal in energy purchased (MD −46.72 kcal, 95% CI −78.35, −15.10, N = 1877). Assuming an average meal of 600 kcal, energy labelling on menus would reduce energy purchased per meal by 7.8% (95% CI 2.5% to 13.1%). The quality of the evidence for these three studies was rated as low, so our confidence in the effect estimate is limited and may change with further studies. Of the remaining six studies, only two (both ITS studies involving energy labels on menus or menus boards in a coffee shop or cafeteria) were at low risk of bias, and their results support the meta-analysis. The results of the other four studies which were conducted in a restaurant, cafeterias (2 studies) or a coffee shop, were not clearly reported and were at high risk of bias.\nSeventeen studies assessed the impact of nutritional labels on consumption in artificial settings or scenarios (henceforth referred to as laboratory studies or settings). Of these, eight (all RCTs) assessed the effect of labels on menus or placed on a range of food options. A meta-analysis of these studies did not conclusively demonstrate a reduction in energy consumed during a meal (MD −50 kcal, 95% CI −104.41, 3.88, N = 1705). We rated the quality of the evidence as low, so our confidence in the effect estimate is limited and may change with further studies.\nSix laboratory studies (four RCTs and two Q-RCTs) assessed the impact of labelling a single food or drink option (such as chocolate, pasta or soft drinks) on energy consumed during a snack or meal. A meta-analysis of these studies did not demonstrate a statistically significant difference in energy (kcal) consumed (SMD 0.05, 95% CI −0.17 to 0.27, N = 732). However, the confidence intervals were wide, suggesting uncertainty in the true effect size. We rated the quality of the evidence as low, so our confidence in the effect estimate is limited and may change with further studies.\nThere was no evidence that nutritional labelling had the unintended harm of increasing energy purchased or consumed. Indirect evidence came from five laboratory studies that involved mislabelling single nutrient content (i.e. placing low energy or low fat labels on high-energy foods) during a snack or meal. A meta-analysis of these studies did not demonstrate a statistically significant increase in energy (kcal) consumed (SMD 0.19, 95% CI −0.14to 0.51, N = 718). The effect was small and the confidence intervals wide, suggesting uncertainty in the true effect size. We rated the quality of the evidence from these studies as very low, providing very little confidence in the effect estimate.\nAuthors' conclusions\nFindings from a small body of low-quality evidence suggest that nutritional labelling comprising energy information on menus may reduce energy purchased in restaurants. The evidence assessing the impact on consumption of energy information on menus or on a range of food options in laboratory settings suggests a similar effect to that observed for purchasing, although the evidence is less definite and also of low quality.\nAccordingly, and in the absence of observed harms, we tentatively suggest that nutritional labelling on menus in restaurants could be used as part of a wider set of measures to tackle obesity. Additional high-quality research in real-world settings is needed to enable more certain conclusions.\nFurther high-quality research is also needed to address the dearth of evidence from grocery stores and vending machines and to assess potential moderators of the intervention effect, including socioeconomic status.\n\nGiven the provided abstract, please respond to the following query: \"What is the aim of this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "This review investigated whether nutritional labels (i.e. labels providing information about nutritional content) persuade people to buy or consume different (healthy) kinds of food. We searched for all available evidence to answer this question and found 28 studies." } ]
query-laysum
6715
[ { "role": "user", "content": "Abstract: Background\nEvery year, an estimated one million children and young adolescents become ill with tuberculosis, and around 226,000 of those children die. Xpert MTB/RIF Ultra (Xpert Ultra) is a molecular World Health Organization (WHO)-recommended rapid diagnostic test that simultaneously detects Mycobacterium tuberculosis complex and rifampicin resistance. We previously published a Cochrane Review 'Xpert MTB/RIF and Xpert MTB/RIF Ultra assays for tuberculosis disease and rifampicin resistance in children'. The current review updates evidence on the diagnostic accuracy of Xpert Ultra in children presumed to have tuberculosis disease. Parts of this review update informed the 2022 WHO updated guidance on management of tuberculosis in children and adolescents.\nObjectives\nTo assess the diagnostic accuracy of Xpert Ultra for detecting: pulmonary tuberculosis, tuberculous meningitis, lymph node tuberculosis, and rifampicin resistance, in children with presumed tuberculosis.\nSecondary objectives\nTo investigate potential sources of heterogeneity in accuracy estimates. For detection of tuberculosis, we considered age, comorbidity (HIV, severe pneumonia, and severe malnutrition), and specimen type as potential sources.\nTo summarize the frequency of Xpert Ultra trace results.\nSearch methods\nWe searched the Cochrane Infectious Diseases Group Specialized Register, MEDLINE, Embase, three other databases, and three trial registers without language restrictions to 9 March 2021.\nSelection criteria\nCross-sectional and cohort studies and randomized trials that evaluated Xpert Ultra in HIV-positive and HIV-negative children under 15 years of age. We included ongoing studies that helped us address the review objectives. We included studies evaluating sputum, gastric, stool, or nasopharyngeal specimens (pulmonary tuberculosis), cerebrospinal fluid (tuberculous meningitis), and fine needle aspirate or surgical biopsy tissue (lymph node tuberculosis). For detecting tuberculosis, reference standards were microbiological (culture) or composite reference standard; for stool, we also included Xpert Ultra performed on a routine respiratory specimen. For detecting rifampicin resistance, reference standards were drug susceptibility testing or MTBDRplus.\nData collection and analysis\nTwo review authors independently extracted data and, using QUADAS-2, assessed methodological quality judging risk of bias separately for each target condition and reference standard. For each target condition, we used the bivariate model to estimate summary sensitivity and specificity with 95% confidence intervals (CIs). We stratified all analyses by type of reference standard. We summarized the frequency of Xpert Ultra trace results; trace represents detection of a very low quantity of Mycobacterium tuberculosis DNA. We assessed certainty of evidence using GRADE.\nMain results\nWe identified 14 studies (11 new studies since the previous review). For detection of pulmonary tuberculosis, 335 data sets (25,937 participants) were available for analysis. We did not identify any studies that evaluated Xpert Ultra accuracy for tuberculous meningitis or lymph node tuberculosis. Three studies evaluated Xpert Ultra for detection of rifampicin resistance. Ten studies (71%) took place in countries with a high tuberculosis burden based on WHO classification. Overall, risk of bias was low.\nDetection of pulmonary tuberculosis\nSputum, 5 studies\nXpert Ultra summary sensitivity verified by culture was 75.3% (95% CI 64.3 to 83.8; 127 participants; high-certainty evidence), and specificity was 97.1% (95% CI 94.7 to 98.5; 1054 participants; high-certainty evidence).\nGastric aspirate, 7 studies\nXpert Ultra summary sensitivity verified by culture was 70.4% (95% CI 53.9 to 82.9; 120 participants; moderate-certainty evidence), and specificity was 94.1% (95% CI 84.8 to 97.8; 870 participants; moderate-certainty evidence).\nStool, 6 studies\nXpert Ultra summary sensitivity verified by culture was 56.1% (95% CI 39.1 to 71.7; 200 participants; moderate-certainty evidence), and specificity was 98.0% (95% CI 93.3 to 99.4; 1232 participants; high certainty-evidence).\nNasopharyngeal aspirate, 4 studies\nXpert Ultra summary sensitivity verified by culture was 43.7% (95% CI 26.7 to 62.2; 46 participants; very low-certainty evidence), and specificity was 97.5% (95% CI 93.6 to 99.0; 489 participants; high-certainty evidence).\nXpert Ultra sensitivity was lower against a composite than a culture reference standard for all specimen types other than nasopharyngeal aspirate, while specificity was similar against both reference standards.\nInterpretation of results\nIn theory, for a population of 1000 children:\n• where 100 have pulmonary tuberculosis in sputum (by culture):\n- 101 would be Xpert Ultra-positive, and of these, 26 (26%) would not have pulmonary tuberculosis (false positive); and\n- 899 would be Xpert Ultra-negative, and of these, 25 (3%) would have tuberculosis (false negative).\n• where 100 have pulmonary tuberculosis in gastric aspirate (by culture):\n- 123 would be Xpert Ultra-positive, and of these, 53 (43%) would not have pulmonary tuberculosis (false positive); and\n- 877 would be Xpert Ultra-negative, and of these, 30 (3%) would have tuberculosis (false negative).\n• where 100 have pulmonary tuberculosis in stool (by culture):\n- 74 would be Xpert Ultra-positive, and of these, 18 (24%) would not have pulmonary tuberculosis (false positive); and\n- 926 would be Xpert Ultra-negative, and of these, 44 (5%) would have tuberculosis (false negative).\n• where 100 have pulmonary tuberculosis in nasopharyngeal aspirate (by culture):\n- 66 would be Xpert Ultra-positive, and of these, 22 (33%) would not have pulmonary tuberculosis (false positive); and\n- 934 would be Xpert Ultra-negative, and of these, 56 (6%) would have tuberculosis (false negative).\nDetection of rifampicin resistance\nXpert Ultra sensitivity was 100% (3 studies, 3 participants; very low-certainty evidence), and specificity range was 97% to 100% (3 studies, 128 participants; low-certainty evidence).\nTrace results\nXpert Ultra trace results, regarded as positive in children by WHO standards, were common. Xpert Ultra specificity remained high in children, despite the frequency of trace results.\nAuthors' conclusions\nWe found Xpert Ultra sensitivity to vary by specimen type, with sputum having the highest sensitivity, followed by gastric aspirate and stool. Nasopharyngeal aspirate had the lowest sensitivity. Xpert Ultra specificity was high against both microbiological and composite reference standards. However, the evidence base is still limited, and findings may be imprecise and vary by study setting. Although we found Xpert Ultra accurate for detection of rifampicin resistance, results were based on a very small number of studies that included only three children with rifampicin resistance. Therefore, findings should be interpreted with caution. Our findings provide support for the use of Xpert Ultra as an initial rapid molecular diagnostic in children being evaluated for tuberculosis.\n\nGiven the provided abstract, please respond to the following query: \"What children do the results of this review apply to?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Children and young adolescents (birth to 14 years) who are HIV-positive or HIV-negative, with signs or symptoms of pulmonary tuberculosis. The results also apply to children with severe pneumonia or malnutrition and tuberculosis symptoms. In this review, we did not identify any studies that evaluated Xpert Ultra accuracy for tuberculous meningitis or lymph node tuberculosis." } ]
query-laysum
6716
[ { "role": "user", "content": "Abstract: Background\nPhysical restraints (PR), such as bedrails and belts in chairs or beds, are commonly used for older people receiving long-term care, despite clear evidence for the lack of effectiveness and safety, and widespread recommendations that their use should be avoided. This systematic review of the efficacy and safety of interventions to prevent and reduce the use of physical restraints outside hospital settings, i.e. in care homes and the community, updates our previous review published in 2011.\nObjectives\nTo evaluate the effects of interventions to prevent and reduce the use of physical restraints for older people who require long-term care (either at home or in residential care facilities)\nSearch methods\nWe searched ALOIS, the Cochrane Dementia and Cognitive Improvement Group's register, MEDLINE (Ovid Sp), Embase (Ovid SP), PsycINFO (Ovid SP), CINAHL (EBSCOhost), Web of Science Core Collection (ISI Web of Science), LILACS (BIREME), ClinicalTrials.gov and the World Health Organization's meta-register, the International Clinical Trials Registry Portal, on 3 August 2022.\nSelection criteria\nWe included randomised controlled trials (RCTs) and controlled clinical trials (CCTs) that investigated the effects of interventions intended to prevent or reduce the use of physical restraints in older people who require long-term care. Studies conducted in residential care institutions or in the community, including patients' homes, were eligible for inclusion. We assigned all included interventions to categories based on their mechanisms and components.\nData collection and analysis\nTwo review authors independently selected the publications for inclusion, extracted study data, and assessed the risk of bias of all included studies. Primary outcomes were the number or proportion of people with at least one physical restraint, and serious adverse events related to PR use, such as death or serious injuries. We performed meta-analyses if necessary data were available. If meta-analyses were not feasible, we reported results narratively. We used GRADE methods to describe the certainty of the evidence.\nMain results\nWe identified six new studies and included 11 studies with 19,003 participants in this review update. All studies were conducted in long-term residential care facilities. Ten studies were RCTs and one study a CCT. All studies included people with dementia. The mean age of the participants was approximately 85 years.\nFour studies investigated organisational interventions aiming to implement a least-restraint policy; six studies investigated simple educational interventions; and one study tested an intervention that provided staff with information about residents' fall risk. The control groups received usual care only in most studies although, in two studies, additional information materials about physical restraint reduction were provided.\nWe judged the risk of selection bias to be high or unclear in eight studies. Risk of reporting bias was high in one study and unclear in eight studies.\nThe organisational interventions intended to promote a least-restraint policy included a variety of components, such as education of staff, training of 'champions' of low-restraint practice, and components which aimed to facilitate a change in institutional policies and culture of care. We found moderate-certainty evidence that organisational interventions aimed at implementation of a least-restraint policy probably lead to a reduction in the number of residents with at least one use of PR (RR 0.86, 95% CI 0.78 to 0.94; 3849 participants, 4 studies) and a large reduction in the number of residents with at least one use of a belt for restraint (RR 0.54, 95% CI 0.40 to 0.73; 2711 participants, 3 studies). No adverse events occurred in the one study which reported this outcome. There was evidence from one study that organisational interventions probably reduce the duration of physical restraint use. We found that the interventions may have little or no effect on the number of falls or fall-related injuries (low-certainty evidence) and probably have little or no effect on the number of prescribed psychotropic medications (moderate-certainty evidence). One study found that organisational interventions result in little or no difference in quality of life (high-certainty evidence) and another study found that they may make little or no difference to agitation (low-certainty evidence).\nThe simple educational interventions were intended to increase knowledge and change staff attitudes towards PR. As well as providing education, some interventions included further components to support change, such as ward-based guidance. We found pronounced between-group baseline imbalances in PR prevalence in some of the studies, which might have occurred because of the small number of clusters in the intervention and control groups. One study did not assess bedrails, which is the most commonly used method of restraint in nursing homes. Regarding the number of residents with at least one restraint, the results were inconsistent. We found very-low certainty evidence and we are uncertain about the effects of simple educational interventions on the number of residents with PR. None of the studies assessed or reported any serious adverse events. We found moderate-certainty evidence that simple educational interventions probably result in little or no difference in restraint intensity and may have little or no effect on falls, fall-related injuries, or agitation (low-certainty evidence each). Based on very low-certainty evidence we are uncertain about the effects of simple educational interventions on the number of participants with a prescription of at least one psychotropic medication.\nOne study investigated an intervention that provided information about residents' fall risk to the nursing staff. We found low-certainty evidence that providing information about residents' fall risk may result in little or no difference in the mean number of PR or the number of falls. The study did not assess overall adverse events.\nAuthors' conclusions\nOrganisational interventions aimed to implement a least-restraint policy probably reduce the number of residents with at least one PR and probably largely reduce the number of residents with at least one belt. We are uncertain whether simple educational interventions reduce the use of physical restraints, and interventions providing information about residents' fall risk may result in little to no difference in the use of physical restraints. These results apply to long-term care institutions; we found no studies from community settings.\n\nGiven the provided abstract, please respond to the following query: \"What did we want to find out?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We wanted to know which interventions are most effective for preventing or reducing the use of PR for older people receiving long-term care either in care institutions or at home. Interventions for preventing and reducing the use of PR typically include education and training for nursing staff and may also include changes to policies and the way care is organised." } ]
query-laysum
6717
[ { "role": "user", "content": "Abstract: Background\nAmong people who have suffered a traumatic brain injury, increased intracranial pressure continues to be a major cause of early death; it is estimated that about 11 people per 100 with traumatic brain injury die.\nIndomethacin (also known as indometacin) is a powerful cerebral vasoconstrictor that can reduce intracranial pressure and, ultimately, restore cerebral perfusion and oxygenation. Thus, indomethacin may improve the recovery of a person with traumatic brain injury.\nObjectives\nTo assess the effects of indomethacin for adults with severe traumatic brain injury.\nSearch methods\nWe ran the searches from inception to 23 August 2019. We searched the Cochrane Central Register of Controlled Trials (CENTRAL; 2019, Issue 8) in the Cochrane Library, Ovid MEDLINE, Ovid Embase, CINAHL Plus (EBSCO), four other databases, and clinical trials registries. We also screened reference lists and conference abstracts, and contacted experts in the field.\nSelection criteria\nOur search criteria included randomised controlled trials (RCTs) that compared indomethacin with any control in adults presenting with severe traumatic brain injury associated with elevated intracranial pressure, with no previous decompressive surgery.\nData collection and analysis\nTwo review authors independently decided on the selection of the studies. We followed standard Cochrane methods.\nMain results\nWe identified no eligible studies for this review, either completed or ongoing.\nAuthors' conclusions\nWe found no studies, either completed or ongoing, that assessed the effects of indomethacin in controlling intracranial hypertension secondary to severe traumatic brain injury. Thus, we cannot draw any conclusions about the effects of indomethacin on intracranial pressure, mortality rates, quality of life, disability or adverse effects.\nThis absence of evidence should not be interpreted as evidence of no effect for indomethacin in controlling intracranial hypertension secondary to severe traumatic brain injury. It means that we have not identified eligible research for this review.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found no trials, either completed or ongoing, that answered our review question. Thus, this review cannot draw any conclusions about the effects of indomethacin on raised intracranial pressure, mortality rates, quality of life, disability or adverse effects in adults." } ]
query-laysum
6718
[ { "role": "user", "content": "Abstract: Background\nNicotine receptor partial agonists may help people to stop smoking by a combination of maintaining moderate levels of dopamine to counteract withdrawal symptoms (acting as an agonist) and reducing smoking satisfaction (acting as an antagonist). This is an update of a Cochrane Review first published in 2007.\nObjectives\nTo assess the effectiveness of nicotine receptor partial agonists, including varenicline and cytisine, for smoking cessation.\nSearch methods\nWe searched the Cochrane Tobacco Addiction Group's Specialised Register in April 2022 for trials, using relevant terms in the title or abstract, or as keywords. The register is compiled from searches of CENTRAL, MEDLINE, Embase, and PsycINFO.\nSelection criteria\nWe included randomised controlled trials that compared the treatment drug with placebo, another smoking cessation drug, e-cigarettes, or no medication. We excluded trials that did not report a minimum follow-up period of six months from baseline.\nData collection and analysis\nWe followed standard Cochrane methods. Our main outcome was abstinence from smoking at longest follow-up using the most rigorous definition of abstinence, preferring biochemically validated rates where reported. We pooled risk ratios (RRs), using the Mantel-Haenszel fixed-effect model. We also reported the number of people reporting serious adverse events (SAEs).\nMain results\nWe included 75 trials of 45,049 people; 45 were new for this update. We rated 22 at low risk of bias, 18 at high risk, and 35 at unclear risk.\nWe found moderate-certainty evidence (limited by heterogeneity) that cytisine helps more people to quit smoking than placebo (RR 1.30, 95% confidence interval (CI) 1.15 to 1.47; I2 = 83%; 4 studies, 4623 participants), and no evidence of a difference in the number reporting SAEs (RR 1.04, 95% CI 0.78 to 1.37; I2 = 0%; 3 studies, 3781 participants; low-certainty evidence). SAE evidence was limited by imprecision. We found no data on neuropsychiatric or cardiac SAEs.\nWe found high-certainty evidence that varenicline helps more people to quit than placebo (RR 2.32, 95% CI 2.15 to 2.51; I2 = 60%, 41 studies, 17,395 participants), and moderate-certainty evidence that people taking varenicline are more likely to report SAEs than those not taking it (RR 1.23, 95% CI 1.01 to 1.48; I2 = 0%; 26 studies, 14,356 participants). While point estimates suggested increased risk of cardiac SAEs (RR 1.20, 95% CI 0.79 to 1.84; I2 = 0%; 18 studies, 7151 participants; low-certainty evidence), and decreased risk of neuropsychiatric SAEs (RR 0.89, 95% CI 0.61 to 1.29; I2 = 0%; 22 studies, 7846 participants; low-certainty evidence), in both cases evidence was limited by imprecision, and confidence intervals were compatible with both benefit and harm.\nPooled results from studies that randomised people to receive cytisine or varenicline found no clear evidence of difference in quit rates (RR 1.00, 95% CI 0.79 to 1.26; I2 = 65%; 2 studies, 2131 participants; low-certainty evidence) and reported SAEs (RR 0.67, 95% CI 0.44 to 1.03; I2 = 45%; 2 studies, 2017 participants; low-certainty evidence). However, the evidence was limited by imprecision, and confidence intervals incorporated the potential for benefit from either cytisine or varenicline. We found no data on neuropsychiatric or cardiac SAEs.\nWe found high-certainty evidence that varenicline helps more people to quit than bupropion (RR 1.36, 95% CI 1.25 to 1.49; I2 = 0%; 9 studies, 7560 participants), and no clear evidence of difference in rates of SAEs (RR 0.89, 95% CI 0.61 to 1.31; I2 = 0%; 5 studies, 5317 participants), neuropsychiatric SAEs (RR 1.05, 95% CI 0.16 to 7.04; I2 = 10%; 2 studies, 866 participants), or cardiac SAEs (RR 3.17, 95% CI 0.33 to 30.18; I2 = 0%; 2 studies, 866 participants). Evidence of harms was of low certainty, limited by imprecision.\nWe found high-certainty evidence that varenicline helps more people to quit than a single form of nicotine replacement therapy (NRT) (RR 1.25, 95% CI 1.14 to 1.37; I2 = 28%; 11 studies, 7572 participants), and low-certainty evidence, limited by imprecision, of fewer reported SAEs (RR 0.70, 95% CI 0.50 to 0.99; I2 = 24%; 6 studies, 6535 participants). We found no data on neuropsychiatric or cardiac SAEs.\nWe found no clear evidence of a difference in quit rates between varenicline and dual-form NRT (RR 1.02, 95% CI 0.87 to 1.20; I2 = 0%; 5 studies, 2344 participants; low-certainty evidence, downgraded because of imprecision). While pooled point estimates suggested increased risk of SAEs (RR 2.15, 95% CI 0.49 to 9.46; I2 = 0%; 4 studies, 1852 participants) and neuropsychiatric SAEs (RR 4.69, 95% CI 0.23 to 96.50; I2 not estimable as events only in 1 study; 2 studies, 764 participants), and reduced risk of cardiac SAEs (RR 0.32, 95% CI 0.01 to 7.88; I2 not estimable as events only in 1 study; 2 studies, 819 participants), in all three cases evidence was of low certainty and confidence intervals were very wide, encompassing both substantial harm and benefit.\nAuthors' conclusions\nCytisine and varenicline both help more people to quit smoking than placebo or no medication. Varenicline is more effective at helping people to quit smoking than bupropion, or a single form of NRT, and may be as or more effective than dual-form NRT. People taking varenicline are probably more likely to experience SAEs than those not taking it, and while there may be increased risk of cardiac SAEs and decreased risk of neuropsychiatric SAEs, evidence was compatible with both benefit and harm. Cytisine may lead to fewer people reporting SAEs than varenicline. Based on studies that directly compared cytisine and varenicline, there may be no difference or a benefit from either medication for quitting smoking.\nFuture trials should test the effectiveness and safety of cytisine compared with varenicline and other pharmacotherapies, and should also test variations in dose and duration. There is limited benefit to be gained from more trials testing the effect of standard-dose varenicline compared with placebo for smoking cessation. Further trials on varenicline should test variations in dose and duration, and compare varenicline with e-cigarettes for smoking cessation.\n\nGiven the provided abstract, please respond to the following query: \"What are 'nicotine receptor partial agonists'?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Smoking tobacco is extremely bad for people’s health. For people who smoke, quitting is the best thing they can do to improve their health. Many people find it difficult to quit smoking. Nicotine receptor partial agonists (NRPAs) are a type of medication used to help people to stop smoking. They help to reduce the withdrawal symptoms people experience when they stop smoking, like cravings and unpleasant mood changes. They also reduce the pleasure people usually experience when they smoke. The most widely-available treatment in this drug type is varenicline. Cytisine is another, similar medication. They may cause unwanted effects such as feeling sick (nausea) and other stomach problems, difficulties sleeping, abnormal dreams, and headache. They may also lead to potentially serious unwanted effects, such as suicidal thoughts, heart problems and raised blood pressure." } ]
query-laysum
6719
[ { "role": "user", "content": "Abstract: Background\nCaries is one of the most prevalent, preventable conditions worldwide. A wide variety of management options are available at different thresholds of disease, ranging from non-operative preventive strategies such as improved oral hygiene, reduced sugar diet, and application of topical fluoride, to minimally invasive treatments for early lesions which are limited to enamel, through to selective removal and restoration for extensive lesions. The cornerstone of caries detection is a visual and tactile dental examination, however, an increasing array of methods of caries lesion detection have been proposed that could potentially support traditional methods of detection and diagnosis. Earlier identification of disease could afford patients the opportunity of less invasive treatment with less destruction of tooth tissue, reduce the need for treatment with aerosol-generating procedures, and potentially result in a reduced cost of care to the patient and to healthcare services.\nObjectives\nOur primary objective was to determine the diagnostic accuracy of different electrical conductance devices for the detection and diagnosis of non-cavitated coronal dental caries in different populations (children, adolescents, and adults) and when tested against different reference standards.\nSearch methods\nCochrane Oral Health's Information Specialist undertook a search of the following databases: MEDLINE Ovid (1946 to 26 April 2019); Embase Ovid (1980 to 26 April 2019); US National Institutes of Health Ongoing Trials Register (ClinicalTrials.gov, to 26 April 2019); and the World Health Organization International Clinical Trials Registry Platform (to 26 April 2019). We studied reference lists as well as published systematic review articles.\nSelection criteria\nWe included diagnostic accuracy studies that compared electrical conductance devices with a reference standard of histology or an enhanced visual examination. This included prospective studies that evaluated the diagnostic accuracy of single index tests and studies that directly compared two or more index tests. We included studies using previously extracted teeth or those that recruited participants with teeth believed to be sound or with early lesions limited to enamel.\nStudies that explicitly recruited participants with more advanced lesions that were obviously into dentine or frankly cavitated were excluded.\nData collection and analysis\nTwo review authors extracted data independently using a piloted study data extraction form based on the Quality Assessment of Diagnostic Accuracy Studies 2 (QUADAS-2). Sensitivity and specificity with 95% confidence intervals (CIs) were reported for each study. This information was displayed as coupled forest plots, and plotted as summary receiver operating characteristic (SROC) plots, displaying the sensitivity-specificity points for each study. Due to variability in thresholds we estimated diagnostic accuracy using hierarchical summary receiver operating characteristic (HSROC) methods.\nMain results\nWe included seven studies reporting a total of 719 tooth sites or surfaces, with an overall prevalence of the target condition of 73% (528 tooth sites or surfaces). The included studies evaluated two index tests: the electronic caries monitor (ECM) (four studies, 475 tooth surfaces) and CarieScan Pro (three studies, 244 tooth surfaces). Six studies used histology as the reference standard, one used an enhanced visual examination. No study was considered to be at low risk of bias across all four domains or low concern for applicability or both. All studies were at high (five studies) or unclear (two studies) risk of bias for the patient selection domain. We judged two studies to be at unclear risk of bias for the index test domain, and one study to be at high risk of bias for the reference standard and flow and timing domains. We judged three studies to be at low concern for applicability for patient selection, and all seven studies to be of low concern for reference standard and flow and timing domains.\nStudies were synthesised using a hierarchical method for meta-analysis. There was variability in the results of the individual studies, with sensitivities which ranged from 0.55 to 0.98 and specificities from 0 to 1.00. These extreme values of specificity may be explained by a low number of healthy tooth surfaces in the included samples. The diagnostic odds ratio (DOR) was 15.65 (95% CI 1.43 to 171.15), and indicative of the variability in the included studies. Through meta-regression we observed no meaningful difference in accuracy according to device type or dentition. Due to the small number of studies we were unable to formally investigate other potential sources of heterogeneity.\nWe judged the certainty of the evidence as very low, and downgraded for risk of bias due to limitations in the design and conduct of the included studies, imprecision arising from the relatively small number of surfaces studied, and inconsistency due to the variability of results.\nAuthors' conclusions\nThe design and conduct of studies to determine the diagnostic accuracy of methods to detect and diagnose caries in situ is particularly challenging. The evidence base to support the detection and diagnosis of caries with electrical conductance devices is sparse. Newer electrical conductance devices show promise and further research at the enamel caries threshold using a robust study design to minimise bias is warranted. In terms of applicability, any future studies should be carried out in a clinical setting to provide a realistic assessment within the oral cavity where plaque, staining, and restorations can be problematic.\n\nGiven the provided abstract, please respond to the following query: \"What is the aim of this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The aim of this Cochrane Review was to find out how accurate electrical conductance devices (non-invasive devices that send an electrical current to the surface of the tooth) are for detecting and diagnosing early tooth decay as part of the dental 'check-up' for children and adults who visit their general dentist. Researchers in Cochrane included seven studies published between 1997 and 2018 to answer this question." } ]
query-laysum
6720
[ { "role": "user", "content": "Abstract: Background\nSevere coronavirus disease 2019 (COVID-19) can cause thrombotic events that lead to severe complications or death. Antiplatelet agents, such as acetylsalicylic acid, have been shown to effectively reduce thrombotic events in other diseases: they could influence the course of COVID-19 in general.\nObjectives\nTo assess the efficacy and safety of antiplatelets given with standard care compared to no treatment or standard care (with/without placebo) for adults with COVID-19.\nSearch methods\nWe searched the Cochrane COVID-19 Study Register (which comprises MEDLINE (PubMed), Embase, ClinicalTrials.gov, WHO ICTRP, medRxiv, CENTRAL), Web of Science, WHO COVID-19 Global literature on coronavirus disease and the Epistemonikos COVID-19 L*OVE Platform to identify completed and ongoing studies without language restrictions to December 2022.\nSelection criteria\nWe followed standard Cochrane methodology. We included randomised controlled trials (RCTs) evaluating antiplatelet agents for the treatment of COVID-19 in adults with COVID-19, irrespective of disease severity, gender or ethnicity.\nData collection and analysis\nWe followed standard Cochrane methodology.\nTo assess bias in included studies, we used the Cochrane risk of bias tool (RoB 2) for RCTs. We rated the certainty of evidence using the GRADE approach for the outcomes.\nMain results\nAntiplatelets plus standard care versus standard care (with/without placebo)\nAdults with a confirmed diagnosis of moderate to severe COVID-19\nWe included four studies (17,541 participants) that recruited hospitalised people with a confirmed diagnosis of moderate to severe COVID-19. A total of 8964 participants were analysed in the antiplatelet arm (either with cyclooxygenase inhibitors or P2Y12 inhibitors) and 8577 participants in the control arm. Most people were older than 50 years and had comorbidities such as hypertension, lung disease or diabetes. The studies were conducted in high- to lower middle-income countries prior to wide-scale vaccination programmes.\nAntiplatelets compared to standard care:\n– probably result in little to no difference in 28-day mortality (risk ratio (RR) 0.95, 95% confidence interval (CI) 0.85 to 1.05; 3 studies, 17,249 participants; moderate-certainty evidence). In absolute terms, this means that for every 177 deaths per 1000 people not receiving antiplatelets, there were 168 deaths per 1000 people who did receive the intervention (95% CI 151 to 186 per 1000 people);\n– probably result in little to no difference in worsening (new need for invasive mechanical ventilation or death up to day 28) (RR 0.95, 95% CI 0.90 to 1.01; 2 studies, 15,266 participants; moderate-certainty evidence);\n– probably result in little to no difference in improvement (participants discharged alive up to day 28) (RR 1.00, 95% CI 0.96 to 1.04; 2 studies, 15,454 participants; moderate-certainty evidence);\n– probably result in a slight reduction of thrombotic events at longest follow-up (RR 0.90, 95% CI 0.80 to 1.02; 4 studies, 17,518 participants; moderate-certainty evidence);\n– may result in a slight increase in serious adverse events at longest follow-up (Peto odds ratio (OR) 1.57, 95% CI 0.48 to 5.14; 1 study, 1815 participants; low-certainty evidence), but non-serious adverse events during study treatment were not reported;\n– probably increase the occurrence of major bleeding events at longest follow-up (Peto OR 1.68, 95% CI 1.29 to 2.19; 4 studies, 17,527 participants; moderate-certainty evidence).\nAdults with a confirmed diagnosis of asymptomatic SARS-CoV-2 infection or mild COVID-19\nWe included two RCTs allocating participants, of whom 4209 had confirmed mild COVID-19 and were not hospitalised. A total of 2109 participants were analysed in the antiplatelet arm (treated with acetylsalicylic acid) and 2100 participants in the control arm. No study included people with asymptomatic SARS-CoV-2 infection.\nAntiplatelets compared to standard care:\n– may result in little to no difference in all-cause mortality at day 45 (Peto OR 1.00, 95% CI 0.45 to 2.22; 2 studies, 4209 participants; low-certainty evidence);\n– may slightly decrease the incidence of new thrombotic events up to day 45 (Peto OR 0.37, 95% CI 0.09 to 1.46; 2 studies, 4209 participants; low-certainty evidence);\n– may make little or no difference to the incidence of serious adverse events up to day 45 (Peto OR 1.00, 95% CI 0.60 to 1.64; 1 study, 3881 participants; low-certainty evidence), but non-serious adverse events were not reported.\nThe evidence is very uncertain about the effect of antiplatelets on the following outcomes (compared to standard care plus placebo):\n– admission to hospital or death up to day 45 (Peto OR 0.79, 95% CI 0.57 to 1.10; 2 studies, 4209 participants; very low-certainty evidence);\n– major bleeding events up to longest follow-up (no event occurred in 328 participants; very low-certainty evidence).\nQuality of life and adverse events during study treatment were not reported.\nAuthors' conclusions\nIn people with confirmed or suspected COVID-19 and moderate to severe disease, we found moderate-certainty evidence that antiplatelets probably result in little to no difference in 28-day mortality, clinical worsening or improvement, but probably result in a slight reduction in thrombotic events. They probably increase the occurrence of major bleeding events. Low-certainty evidence suggests that antiplatelets may result in a slight increase in serious adverse events.\nIn people with confirmed COVID-19 and mild symptoms, we found low-certainty evidence that antiplatelets may result in little to no difference in 45-day mortality and serious adverse events, and may slightly reduce thrombotic events. The effects on the combined outcome admission to hospital or death up to day 45 and major bleeding events are very uncertain. Quality of life was not reported.\nIncluded studies were conducted in high- to lower middle-income settings using antiplatelets prior to vaccination roll-outs.\nWe identified a lack of evidence concerning quality of life assessments, adverse events and people with asymptomatic infection. The 14 ongoing and three completed, unpublished RCTs that we identified in trial registries address similar settings and research questions as in the current body of evidence. We expect to incorporate the findings of these studies in future versions of this review.\n\nGiven the provided abstract, please respond to the following query: \"How up-to-date is this evidence?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Our evidence is up-to-date to 22 December 2022." } ]
query-laysum
6721
[ { "role": "user", "content": "Abstract: Background\nWomen with a suspected large-for-dates fetus or a fetus with suspected macrosomia (birthweight greater than 4000 g) are at risk of operative birth or caesarean section. The baby is also at increased risk of shoulder dystocia and trauma, in particular fractures and brachial plexus injury. Induction of labour may reduce these risks by decreasing the birthweight, but may also lead to longer labours and an increased risk of caesarean section.\nObjectives\nTo assess the effects of a policy of labour induction at or shortly before term (37 to 40 weeks) for suspected fetal macrosomia on the way of giving birth and maternal or perinatal morbidity.\nSearch methods\nWe searched the Cochrane Pregnancy and Childbirth Group's Trials Register (31 January 2016), contacted trial authors and searched reference lists of retrieved studies.\nSelection criteria\nRandomised trials of induction of labour for suspected fetal macrosomia.\nData collection and analysis\nReview authors independently assessed trials for inclusion and risk of bias, extracted data and checked them for accuracy. We contacted study authors for additional information. For key outcomes the quality of the evidence was assessed using the GRADE approach.\nMain results\nWe included four trials, involving 1190 women. It was not possible to blind women and staff to the intervention, but for other 'Risk of bias' domains these studies were assessed as being at low or unclear risk of bias.\nCompared to expectant management, there was no clear effect of induction of labour for suspected macrosomia on the risk of caesarean section (risk ratio (RR) 0.91, 95% confidence interval (CI) 0.76 to 1.09; 1190 women; four trials, moderate-quality evidence) or instrumental delivery (RR 0.86, 95% CI 0.65 to 1.13; 1190 women; four trials, low-quality evidence). Shoulder dystocia (RR 0.60, 95% CI 0.37 to 0.98; 1190 women; four trials, moderate-quality evidence), and fracture (any) (RR 0.20, 95% CI 0.05 to 0.79; 1190 women; four studies, high-quality evidence) were reduced in the induction of labour group. There were no clear differences between groups for brachial plexus injury (two events were reported in the control group in one trial, low-quality evidence). There was no strong evidence of any difference between groups for measures of neonatal asphyxia; low five-minute infant Apgar scores (less than seven) or low arterial cord blood pH (RR 1.51, 95% CI 0.25 to 9.02; 858 infants; two trials, low-quality evidence; and, RR 1.01, 95% CI 0.46 to 2.22; 818 infants; one trial, moderate-quality evidence, respectively).\nMean birthweight was lower in the induction group, but there was considerable heterogeneity between studies for this outcome (mean difference (MD) -178.03 g, 95% CI -315.26 to -40.81; 1190 infants; four studies; I2 = 89%).\nFor outcomes assessed using GRADE, we based our downgrading decisions on high risk of bias from lack of blinding and imprecision of effect estimates.\nAuthors' conclusions\nInduction of labour for suspected fetal macrosomia has not been shown to alter the risk of brachial plexus injury, but the power of the included studies to show a difference for such a rare event is limited. Also antenatal estimates of fetal weight are often inaccurate so many women may be worried unnecessarily, and many inductions may not be needed. Nevertheless, induction of labour for suspected fetal macrosomia results in a lower mean birthweight, and fewer birth fractures and shoulder dystocia. The observation of increased use of phototherapy in the largest trial, should also be kept in mind.\nFindings from trials included in the review suggest that to prevent one fracture it would be necessary to induce labour in 60 women. Since induction of labour does not appear to alter the rate of caesarean delivery or instrumental delivery, it is likely to be popular with many women. In settings where obstetricians can be reasonably confident about their scan assessment of fetal weight, the advantages and disadvantages of induction at or near term for fetuses suspected of being macrosomic should be discussed with parents.\nAlthough some parents and doctors may feel the evidence already justifies induction, others may justifiably disagree. Further trials of induction shortly before term for suspected fetal macrosomia are needed. Such trials should concentrate on refining the optimum gestation of induction, and improving the accuracy of the diagnosis of macrosomia.\n\nGiven the provided abstract, please respond to the following query: \"What does this mean?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The number of births where the baby's shoulder became stuck (shoulder dystocia) or a bone was fractured (usually the clavicle, which heals well without consequences) were reduced in the induction of labour group. The evidence was assessed as moderate quality for shoulder dystocia and high quality for fracture. No clear differences between groups were reported for damage to the network of nerves that send signals from the spine to the shoulder, arm and hand (brachial plexus injury) of the baby (low-quality evidence due to very few events occurring) or signs of not enough oxygen during birth. A policy of labour induction reduced the average birthweight of babies by 178 g. The trials did not show any differences in the number of women who had caesarean sections or instrumental births. We conclude that there appear to be benefits, but there may also be some disadvantages of induction of labour shortly before term. The option of having an induction should be discussed with parents when their baby is suspected to be extra large.\nAlthough some parents and doctors may feel the existing evidence is sufficient to justify inducing labour, others may disagree. Further high-quality studies are needed in order to find out what is the best time to induce labour towards the end of pregnancy, and how to improve the accuracy in diagnosing macrosomia.\n" } ]
query-laysum
6722
[ { "role": "user", "content": "Abstract: Background\nCervical cancer ranks as the fourth leading cause of death from cancer in women. Historically, women with metastatic or recurrent cervical cancer have had limited treatment options. New anti-angiogenesis therapies, such as vascular endothelial growth factor (VEGF) targeting agents, offer an alternative strategy to conventional chemotherapy; they act by inhibiting the growth of new blood vessels, thereby restricting tumour growth by blocking the blood supply.\nObjectives\nTo assess the benefits and harms of VEGF targeting agents in the management of persistent, recurrent, or metastatic cervical cancer.\nSearch methods\nWe performed searches of the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, Embase, online registers of clinical trials, and abstracts of scientific meetings up until 27 May 2020.\nSelection criteria\nWe examined randomised controlled trials (RCTs) that evaluated the use of VEGF targeting agents alone or in combination with conventional chemotherapy or other VEGF targeting agents.\nData collection and analysis\nThree review authors independently screened the results of search strategies, extracted data, assessed risk of bias, and analysed data according to the standard methods expected by Cochrane. The certainty of evidence was assessed via the GRADE approach.\nMain results\nA total of 1634 records were identified. From these, we identified four studies with a total of 808 participants for inclusion. We also identified two studies that were awaiting classification and nine ongoing studies.\nBevacizumab plus chemotherapy versus chemotherapy\nTreatment with bevacizumab plus chemotherapy may result in lower risk of death compared to chemotherapy alone (hazard ratio (HR) 0.77, 95% confidence interval (CI) 0.62 to 0.95; 1 study, 452 participants; low-certainty evidence). However, there are probably more specific adverse events when compared to chemotherapy alone, including gastrointestinal perforations or fistulae (risk ratio (RR) 18.00, 95% CI 2.42 to 133.67; 1 study, 440 participants; moderate-certainty evidence); serious thromboembolic events (RR 4.5, 95% CI 1.55 to 13.08; 1 study, 440 participants; moderate-certainty evidence); and hypertension (RR 13.75, 95% CI 5.07 to 37.29; 1 study, 440 participants; moderate-certainty evidence). There may also be a higher incidence of serious haemorrhage (RR 5.00, 95% CI 1.11 to 22.56; 1 study, 440 participants; low-certainty evidence). In addition, the incidence of serious adverse events is probably higher (RR 1.44, 95% CI 1.16 to 1.79; 1 study, 439 participants; moderate-certainty evidence). The incremental cost-effectiveness ratio was USD 295,164 per quality-adjusted life-year (1 study, 452 participants; low-certainty evidence).\nCediranib plus chemotherapy versus chemotherapy\nTreatment with cediranib plus chemotherapy may or may not result in similar risk of death when compared to chemotherapy alone (HR 0.94, 95% CI 0.53 to 1.65; 1 study, 69 participants; low-certainty evidence). We found very uncertain results for the incidences of specific adverse events, including gastrointestinal perforations or fistulae (RR 3.27, 95% CI 0.14 to 77.57; 1 study, 67 participants; very low-certainty evidence); serious haemorrhage (RR 5.45, 95% CI 0.27 to 109.49; 1 study, 67 participants; very low-certainty evidence); serious thromboembolic events (RR 3.41, 95% CI 0.14 to 80.59; 1 study, 60 participants; very low-certainty evidence); and serious hypertension (RR 0.36, 95% CI 0.02 to 8.62; 1 study, 67 participants; very low-certainty evidence). In addition, there may or may not be a similar incidence of serious adverse events compared to chemotherapy alone (RR 1.15, 95% CI 0.75 to 1.78; 1 study, 67 participants; low-certainty evidence).\nApatinib plus chemotherapy or chemotherapy/brachytherapy versus chemotherapy or chemotherapy/brachytherapy\nTreatment with apatinib plus chemotherapy or chemotherapy/brachytherapy may or may not result in similar risk of death compared to chemotherapy alone or chemotherapy/brachytherapy alone (HR 0.90, 95% CI 0.51 to 1.60; 1 study, 52 participants; low-certainty evidence). However, hypertension events may occur at a higher incidence as compared to chemotherapy alone or chemotherapy/brachytherapy alone (RR 5.14, 95% CI 1.28 to 20.73; 1 study, 52 participants; low-certainty evidence).\nPazopanib plus lapatinib versus lapatinib\nTreatment with pazopanib plus lapatinib may result in higher risk of death compared to lapatinib alone (HR 2.71, 95% CI 1.16 to 6.31; 1 study, 117 participants; low-certainty evidence). We found very uncertain results for the incidences of specific adverse events, including gastrointestinal perforations or fistulae (RR 2.00, 95% CI 0.19 to 21.59; 1 study, 152 participants; very low-certainty evidence); haemorrhage (RR 2.00, 95% CI 0.72 to 5.58; 1 study, 152 participants; very low-certainty evidence); and thromboembolic events (RR 3.00, 95% CI 0.12 to 72.50; 1 study, 152 participants; very low-certainty evidence). In addition, the incidence of hypertension events is probably higher (RR 12.00, 95% CI 2.94 to 49.01; 1 study, 152 participants; moderate-certainty evidence). There may or may not be a similar incidence of serious adverse events as compared to lapatinib alone (RR 1.45, 95% CI 0.94 to 2.26; 1 study, 152 participants; low-certainty evidence).\nPazopanib versus lapatinib\nTreatment with pazopanib may or may not result in similar risk of death as compared to lapatinib (HR 0.96, 95% CI 0.67 to 1.38; 1 study, 152 participants; low-certainty evidence). We found very uncertain results for the incidences of specific adverse events, including gastrointestinal perforations or fistulae (RR 1.03, 95% CI 0.07 to 16.12; 1 study, 150 participants; very low-certainty evidence); haemorrhage (RR 1.03, 95% CI 0.31 to 3.40; 1 study, 150 participants; very low-certainty evidence); and thromboembolic events (RR 3.08, 95% CI 0.13 to 74.42; 1 study, 150 participants; very low-certainty evidence). In addition, the incidence of hypertension events is probably higher (RR 11.81, 95% CI 2.89 to 48.33; 1 study, 150 participants; moderate-certainty evidence). The risk of serious adverse events may or may not be similar as compared to lapatinib (RR 1.31, 95% CI 0.83 to 2.07; 1 study, 150 participants; low-certainty evidence).\nAuthors' conclusions\nWe found low-certainty evidence in favour of the use of bevacizumab plus chemotherapy. However, bevacizumab probably increases specific adverse events (gastrointestinal perforations or fistulae, thromboembolic events, hypertension) and serious adverse events. We found low-certainty evidence that does not support the use of cediranib plus chemotherapy, apatinib plus chemotherapy, apatinib plus chemotherapy/brachytherapy, or pazopanib monotherapy. We found low-certainty evidence suggesting that pazopanib plus lapatinib worsens outcomes. The VEGF inhibitors apatinib and pazopanib may increase the probability of hypertension events.\n\nGiven the provided abstract, please respond to the following query: \"What are the main results of the review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found four randomised controlled trials (RCTs) that met our inclusion criteria and enrolled 808 women.\nWe found one study, which included 452 women, that assessed the use of bevacizumab plus chemotherapy versus chemotherapy alone. The inclusion of bevacizumab may improve overall survival and probably increases the incidences of specific and serious adverse events.\nA second study analysed 69 women treated with cediranib plus chemotherapy versus chemotherapy alone. Cediranib may make little to no difference in survival, and it is uncertain whether it increases the incidences of specific or serious adverse events.\nAnother study with 59 women reported data on the use of apatinib plus chemotherapy or chemoradiotherapy (concurrent chemotherapy and radiotherapy) versus chemotherapy or chemoradiotherapy alone. Apatinib may make little to no difference in survival, but it exhibited promising efficacy for progression-free survival.\nWe found one study with 228 women that compared pazopanib plus lapatinib versus only lapatinib, or pazopanib versus lapatinib: pazopanib plus lapatinib may reduce survival and probably increases the incidence of hypertension; pazopanib alone may make little to no difference in survival and probably increases the incidence of hypertension.\nOverall the quality (certainty) of the evidence was low, as each comparison included only one study and most studies were small." } ]
query-laysum
6723
[ { "role": "user", "content": "Abstract: Background\nPatients undergoing haemodialysis (HD) through a central venous catheter (CVC) are exposed to several risks, being a catheter-related infection (CRI) and a CVC lumen thrombosis among the most serious. Standard of care regarding CVCs includes their sealing with heparin lock solutions to prevent catheter lumen thrombosis. Other lock solutions to prevent CRI, such as antimicrobial lock solutions, have proven useful with antibiotics solutions, but not as yet for non–antibiotic antimicrobial solutions. Furthermore, it is uncertain if these solutions have a negative effect on thrombosis incidence.\nObjectives\nTo assess the efficacy and safety of antimicrobial (antibiotic, non-antibiotic, or both) catheter lock solutions for preventing CRI in participants undergoing HD with a CVC.\nSearch methods\nWe searched the Cochrane Kidney and Transplant Specialised Register up to 18 December 2017 through contact with the Information Specialist using search terms relevant to this review. Studies in the Register are identified through searches of CENTRAL, MEDLINE, and EMBASE, conference proceedings, the International Clinical Trials Register (ICTRP) Search Portal, and ClinicalTrials.gov.\nSelection criteria\nWe included all randomised or quasi-randomised control trials (RCTs) comparing antimicrobial (antibiotic and non-antibiotic) lock solutions to standard lock solutions, in participants using a CVC for HD, without language restriction.\nData collection and analysis\nTwo authors independently assessed studies for eligibility, and two additional authors assessed for risk of bias and extracted data. We expressed results as rate ratios (RR) per 1000 catheter-days or 1000 dialysis sessions with 95% confidence intervals (CI). Statistical analyses were performed using the random-effects model.\nMain results\nThirty-nine studies, enrolling 4216 participants, were included in this review, however only 30 studies, involving 3392 participants, contained enough data to be meta-analysed. Risk of bias was low or unclear for most domains in the majority of the included studies.\nStudies compared antimicrobial lock solutions (antibiotic and non-antibiotic) to standard sealing solutions (usually heparin) of the CVC for HD. Fifteen studies used antibiotic lock solutions, 21 used non-antibiotic antimicrobial lock solutions, and 4 used both (antibiotic and non-antibiotic) lock solutions. Studies reported the incidence of CRI, catheter thrombosis, or both.\nAntimicrobial lock solutions probably reduces CRI per 1000 catheter-days (27 studies: RR 0.38, 95% CI 0.27 to 0.53; I2 = 54%; low certainty evidence), however antimicrobial lock solutions probably makes little or no difference to the risk of thrombosis per 1000 catheter days (14 studies: RR 0.79, 95% CI 0.52 to 1.22; I2 = 83%; very low certainty evidence). Subgroup analysis of antibiotic and the combination of both lock solutions showed that both probably reduced CRI per 1000 catheter-days (13 studies: RR 0.30, 95% CI: 0.22 to 0.42; I2 = 47%) and risk of thrombosis per 1000 catheter-days (4 studies: RR 0.26, 95% CI: 0.14 to 0.49; I2 = 0%), respectively. Non-antibiotic antimicrobial lock solutions probably reduced CRI per 1000 catheter-days for tunnelled CVC (9 studies: RR 0.60, 95% CI 0.40 to 0.91) but probably made little or no difference with non-tunnelled CVC (4 studies: RR 0.93, 95% CI 0.48 to 1.81). Subgroup analyses showed that antibiotic (5 studies: RR 0.76, 95% CI 0.42 to 1.38), non-antibiotic (8 studies: RR 0.85, 95% CI 0.44 to 1.66), and the combination of both lock solutions (3 studies: RR 0.63, 95% CI 0.22 to 1.81) made little or no difference to thrombosis per 1000 catheter-days compared to control lock solutions.\nAuthors' conclusions\nAntibiotic antimicrobial and combined (antibiotic-non antibiotic) lock solutions decreased the incidence of CRI compared to control lock solutions, whereas non-antibiotic lock solutions reduce CRI only for tunnelled CVC. The effect on thrombosis incidence is uncertain for all antimicrobial lock solutions. Our confidence in the evidence is low and very low; therefore, better-designed studies are needed to confirm the efficacy and safety of antimicrobial lock solutions.\n\nGiven the provided abstract, please respond to the following query: \"Conclusion\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Some antimicrobial (antibiotic and the combination of antibiotic-non antibiotic) lock solutions decrease the incidence of CRIs compared to heparin. Their effect on CVC permeability remains unclear. The quality of the studies is low and very low, respectively; therefore, more studies are needed to confirm the benefits and harms of antimicrobial lock solutions." } ]
query-laysum
6724
[ { "role": "user", "content": "Abstract: Background\nFibromyalgia is characterized by chronic widespread pain that leads to reduced physical function. Exercise training is commonly recommended as a treatment for management of symptoms. We examined the literature on resistance training for individuals with fibromyalgia. Resistance training is exercise performed against a progressive resistance with the intention of improving muscle strength, muscle endurance, muscle power, or a combination of these.\nObjectives\nTo evaluate the benefits and harms of resistance exercise training in adults with fibromyalgia. We compared resistance training versus control and versus other types of exercise training.\nSearch methods\nWe searched nine electronic databases (The Cochrane Library, MEDLINE, EMBASE, CINAHL, PEDro, Dissertation Abstracts, Current Controlled Trials, World Health Organization (WHO) International Clinical Trials Registry Platform, AMED) and other sources for published full-text articles. The date of the last search was 5 March 2013. Two review authors independently screened 1856 citations, 766 abstracts and 156 full-text articles. We included five studies that met our inclusion criteria.\nSelection criteria\nSelection criteria included: a) randomized clinical trial, b) diagnosis of fibromyalgia based on published criteria, c) adult sample, d) full-text publication, and e) inclusion of between-group data comparing resistance training versus a control or other physical activity intervention.\nData collection and analysis\nPairs of review authors independently assessed risk of bias and extracted intervention and outcome data. We resolved disagreements between the two review authors and questions regarding interpretation of study methods by discussion within the pairs or when necessary the issue was taken to the full team of 11 members. We extracted 21 outcomes of which seven were designated as major outcomes: multidimensional function, self reported physical function, pain, tenderness, muscle strength, attrition rates, and adverse effects. We evaluated benefits and harms of the interventions using standardized mean differences (SMD) or mean differences (MD) or risk ratios or Peto odds ratios and 95% confidence intervals (CI). Where two or more studies provided data for an outcome, we carried out a meta-analysis.\nMain results\nThe literature search yielded 1865 citations with five studies meeting the selection criteria. One of the studies that had three arms contributed data for two comparisons. In the included studies, there were 219 women participants with fibromyalgia, 95 of whom were assigned to resistance training programs. Three randomized trials compared 16 to 21 weeks of moderate- to high-intensity resistance training versus a control group. Two studies compared eight weeks of progressive resistance training (intensity as tolerated) using free weights or body weight resistance exercise versus aerobic training (ie, progressive treadmill walking, indoor and outdoor walking), and one study compared 12 weeks of low-intensity resistance training using hand weights (1 to 3 lbs (0.45 to 1.36 kg)) and elastic tubing versus flexibility exercise (static stretches to major muscle groups).\nStatistically significant differences (MD; 95% CI) favoring the resistance training interventions over control group(s) were found in multidimensional function (Fibromyalgia Impact Questionnaire (FIQ) total decreased 16.75 units on a 100-point scale; 95% CI -23.31 to -10.19), self reported physical function (-6.29 units on a 100-point scale; 95% CI -10.45 to -2.13), pain (-3.3 cm on a 10-cm scale; 95% CI -6.35 to -0.26), tenderness (-1.84 out of 18 tender points; 95% CI -2.6 to -1.08), and muscle strength (27.32 kg force on bilateral concentric leg extension; 95% CI 18.28 to 36.36).\nDifferences between the resistance training group(s) and the aerobic training groups were not statistically significant for multidimensional function (5.48 on a 100-point scale; 95% CI -0.92 to 11.88), self reported physical function (-1.48 units on a 100-point scale; 95% CI -6.69 to 3.74) or tenderness (SMD -0.13; 95% CI -0.55 to 0.30). There was a statistically significant reduction in pain (0.99 cm on a 10-cm scale; 95% CI 0.31 to 1.67) favoring the aerobic groups.\nStatistically significant differences were found between a resistance training group and a flexibility group favoring the resistance training group for multidimensional function (-6.49 FIQ units on a 100-point scale; 95% CI -12.57 to -0.41) and pain (-0.88 cm on a 10-cm scale; 95% CI -1.57 to -0.19), but not for tenderness (-0.46 out of 18 tender points; 95% CI -1.56 to 0.64) or strength (4.77 foot pounds torque on concentric knee extension; 95% CI -2.40 to 11.94). This evidence was classified low quality due to the low number of studies and risk of bias assessment. There were no statistically significant differences in attrition rates between the interventions. In general, adverse effects were poorly recorded, but no serious adverse effects were reported. Assessment of risk of bias was hampered by poor written descriptions (eg, allocation concealment, blinding of outcome assessors). The lack of a priori protocols and lack of care provider blinding were also identified as methodologic concerns.\nAuthors' conclusions\nThe evidence (rated as low quality) suggested that moderate- and moderate- to high-intensity resistance training improves multidimensional function, pain, tenderness, and muscle strength in women with fibromyalgia. The evidence (rated as low quality) also suggested that eight weeks of aerobic exercise was superior to moderate-intensity resistance training for improving pain in women with fibromyalgia. There was low-quality evidence that 12 weeks of low-intensity resistance training was superior to flexibility exercise training in women with fibromyalgia for improvements in pain and multidimensional function. There was low-quality evidence that women with fibromyalgia can safely perform moderate- to high-resistance training.\n\nGiven the provided abstract, please respond to the following query: \"Research question\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "What are the effects of resistance exercise training for people with fibromyalgia on wellness, symptoms, fitness, and adverse effects?" } ]
query-laysum
6725
[ { "role": "user", "content": "Abstract: Background\nLow-back pain (LBP) is responsible for considerable personal suffering due to pain and reduced function, as well as the societal burden due to costs of health care and lost work productivity. For the vast majority of people with LBP, no specific anatomical cause can be reliably identified. For these people with non-specific LBP there are numerous treatment options, few of which have been shown to be effective in reducing pain and disability. The muscle energy technique (MET) is a treatment technique used predominantly by osteopaths, physiotherapists and chiropractors which involves alternating periods of resisted muscle contractions and assisted stretching. To date it is unclear whether MET is effective in reducing pain and improving function in people with LBP.\nObjectives\nTo examine the effectiveness of MET in the treatment of people with non-specific LBP compared with control interventions, with particular emphasis on subjective pain and disability outcomes.\nSearch methods\nCENTRAL, MEDLINE, EMBASE, five other databases and two trials registers were searched from inception to May and June 2014 together with reference checking and citation searching of relevant systematic reviews.\nSelection criteria\nRandomised controlled trials assessing the effect of MET on pain or disability in patients with non-specific LBP were included.\nData collection and analysis\nTwo authors independently assessed the risk of bias and extracted the data. Meta-analysis was performed where clinical homogeneity was sufficient. The quality of the evidence for each comparison was assessed with the GRADE approach.\nMain results\nThere were 12 randomised controlled trials with 14 comparisons included in the review, with a total sample of 500 participants across all comparisons. Included studies were typically very small (n = 20 to 72), all except one were assessed as being at high risk of bias, and all reported short-term outcomes. For the purposes of pooling, studies were divided into seven clinically homogenous comparisons according to the patient population (acute or chronic LBP) and the nature of the control intervention. Most of the comparisons (five out of seven) included only one study, one comparison had two studies, and one comparison included seven studies.\nThe meta-analyses provided low-quality evidence that MET provided no additional benefit when added to other therapies on the outcomes of chronic pain and disability in the short-term (weighted mean difference (WMD) for pain 0.00, 95% CI -2.97 to 2.98 on a 100-point scale; standardised mean difference (SMD) for disability -0.18, 95% CI -0.43 to 0.08, 7 studies, 232 participants). There was low-quality evidence that MET produced no clinically relevant differences in pain compared to sham MET (mean difference (MD) 14.20, 95% CI -10.14 to 38.54, 1 study, 20 participants). For the comparison of MET to other conservative therapies for acute non-specific LBP, there was very low-quality evidence of no clinically relevant difference for the outcomes of pain (MD -10.72, 95% CI -32.57 to 11.13, 2 studies, 88 participants) and functional status (MD 0.87, 95% CI -6.31 to 8.05, 1 study, 60 participants). For the comparison of MET to other conservative therapies for chronic non-specific LBP, there was low-quality evidence of no clinically relevant difference for the outcomes of pain (MD -9.70, 95% CI -20.20 to 0.80, 1 study, 30 participants) and functional status (MD -4.10, 95% CI -9.53 to 1.33, 1 study, 30 participants). There was low-quality evidence of no clinically relevant difference for the addition of MET to other interventions for acute non-specific LBP for the outcome of pain (MD -3, 95% CI -11.37 to 5.37, 1 study, 40 participants) and low-quality evidence of an effect in favour of MET for functional status (MD -17.6, 95% CI -27.05 to -8.15, 1 study, 40 participants). For chronic non-specific LBP, there was low-quality evidence of an effect in favour of MET for the addition of MET to other interventions for the outcomes of pain (MD -34.1, 95% CI -38.43 to -29.77, 1 study, 30 participants) and functional status (MD -22, 95% CI -27.41 to -16.59, 1 study, 30 participants). Lastly, there was low-quality evidence of no difference for the addition of MET to another manual intervention compared to the same intervention with other conservative therapies for the outcomes of pain (MD 5.20, 95% CI -3.03 to 13.43, 1 study, 20 participants) and functional status (MD 6.0, 95% CI -0.49 to 12.49, 1 study, 20 participants).\nNo study reported on our other primary outcome of general well-being. Seven studies reported that no adverse events were observed, whereas the other five studies did not report any information on adverse events.\nAuthors' conclusions\nThe quality of research related to testing the effectiveness of MET is poor. Studies are generally small and at high risk of bias due to methodological deficiencies. Studies conducted to date generally provide low-quality evidence that MET is not effective for patients with LBP. There is not sufficient evidence to reliably determine whether MET is likely to be effective in practice. Large, methodologically-sound studies are necessary to investigate this question.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The review authors could not find adequate evidence to make any definitive judgements about the safety or effectiveness of MET. Studies were generally too small and had a high risk of bias, producing unreliable answers about this therapy.\nThere is a need for larger, high-quality studies to determine the effectiveness and safety of MET.\nAt present there is no convincing evidence that MET is effective as a stand-alone therapy or improves the effectiveness as an accompaniment to other therapies." } ]
query-laysum
6726
[ { "role": "user", "content": "Abstract: Background\nRecent progress in understanding the genetic basis of breast cancer and widely publicized reports of celebrities undergoing risk-reducing mastectomy (RRM) have increased interest in RRM as a method of preventing breast cancer. This is an update of a Cochrane Review first published in 2004 and previously updated in 2006 and 2010.\nObjectives\n(i) To determine whether risk-reducing mastectomy reduces death rates from any cause in women who have never had breast cancer and in women who have a history of breast cancer in one breast, and (ii) to examine the effect of risk-reducing mastectomy on other endpoints, including breast cancer incidence, breast cancer mortality, disease-free survival, physical morbidity, and psychosocial outcomes.\nSearch methods\nFor this Review update, we searched Cochrane Breast Cancer's Specialized Register, MEDLINE, Embase and the WHO International Clinical Trials Registry Platform (ICTRP) on 9 July 2016. We included studies in English.\nSelection criteria\nParticipants included women at risk for breast cancer in at least one breast. Interventions included all types of mastectomy performed for the purpose of preventing breast cancer.\nData collection and analysis\nAt least two review authors independently abstracted data from each report. We summarized data descriptively; quantitative meta-analysis was not feasible due to heterogeneity of study designs and insufficient reporting. We analyzed data separately for bilateral risk-reducing mastectomy (BRRM) and contralateral risk-reducing mastectomy (CRRM). Four review authors assessed the methodological quality to determine whether or not the methods used sufficiently minimized selection bias, performance bias, detection bias, and attrition bias.\nMain results\nAll 61 included studies were observational studies with some methodological limitations; randomized trials were absent. The studies presented data on 15,077 women with a wide range of risk factors for breast cancer, who underwent RRM.\nTwenty-one BRRM studies looking at the incidence of breast cancer or disease-specific mortality, or both, reported reductions after BRRM, particularly for those women with BRCA1/2 mutations. Twenty-six CRRM studies consistently reported reductions in incidence of contralateral breast cancer but were inconsistent about improvements in disease-specific survival. Seven studies attempted to control for multiple differences between intervention groups and showed no overall survival advantage for CRRM. Another study showed significantly improved survival following CRRM, but after adjusting for bilateral risk-reducing salpingo-oophorectomy (BRRSO), the CRRM effect on all-cause mortality was no longer significant.\nTwenty studies assessed psychosocial measures; most reported high levels of satisfaction with the decision to have RRM but greater variation in satisfaction with cosmetic results. Worry over breast cancer was significantly reduced after BRRM when compared both to baseline worry levels and to the groups who opted for surveillance rather than BRRM, but there was diminished satisfaction with body image and sexual feelings.\nSeventeen case series reporting on adverse events from RRM with or without reconstruction reported rates of unanticipated reoperations from 4% in those without reconstruction to 64% in participants with reconstruction.\nIn women who have had cancer in one breast, removing the other breast may reduce the incidence of cancer in that other breast, but there is insufficient evidence that this improves survival because of the continuing risk of recurrence or metastases from the original cancer. Additionally, thought should be given to other options to reduce breast cancer risk, such as BRRSO and chemoprevention, when considering RRM.\nAuthors' conclusions\nWhile published observational studies demonstrated that BRRM was effective in reducing both the incidence of, and death from, breast cancer, more rigorous prospective studies are suggested. BRRM should be considered only among those at high risk of disease, for example, BRCA1/2 carriers. CRRM was shown to reduce the incidence of contralateral breast cancer, but there is insufficient evidence that CRRM improves survival, and studies that control for multiple confounding variables are recommended. It is possible that selection bias in terms of healthier, younger women being recommended for or choosing CRRM produces better overall survival numbers for CRRM. Given the number of women who may be over-treated with BRRM/CRRM, it is critical that women and clinicians understand the true risk for each individual woman before considering surgery. Additionally, thought should be given to other options to reduce breast cancer risk, such as BRRSO and chemoprevention when considering RRM.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Recent progress in understanding the genetic basis of breast cancer and widely publicized reports of celebrities undergoing RRM have increased interest in it as a method of preventing breast cancer." } ]
query-laysum
6727
[ { "role": "user", "content": "Abstract: Background\nTraditionally, cavitated carious lesions and those extending into dentine have been treated by 'complete' removal of carious tissue, i.e. non-selective removal and conventional restoration (CR). Alternative strategies for managing cavitated or dentine carious lesions remove less or none of the carious tissue and include selective carious tissue removal (or selective excavation (SE)), stepwise carious tissue removal (SW), sealing carious lesions using sealant materials, sealing using preformed metal crowns (Hall Technique, HT), and non-restorative cavity control (NRCC).\nObjectives\nTo determine the comparative effectiveness of interventions (CR, SE, SW, sealing of carious lesions using sealant materials or preformed metal crowns (HT), or NRCC) to treat carious lesions conventionally considered to require restorations (cavitated or micro-cavitated lesions, or occlusal lesions that are clinically non-cavitated but clinically/radiographically extend into dentine) in primary or permanent teeth with vital (sensitive) pulps.\nSearch methods\nAn information specialist searched four bibliographic databases to 21 July 2020 and used additional search methods to identify published, unpublished and ongoing studies.\nSelection criteria\nWe included randomised clinical trials comparing different levels of carious tissue removal, as listed above, against each other, placebo, or no treatment. Participants had permanent or primary teeth (or both), and vital pulps (i.e. no irreversible pulpitis/pulp necrosis), and carious lesions conventionally considered to need a restoration (i.e. cavitated lesions, or non- or micro-cavitated lesions radiographically extending into dentine). The primary outcome was failure, a composite measure of pulp exposure, endodontic therapy, tooth extraction, and restorative complications (including resealing of sealed lesions).\nData collection and analysis\nPairs of review authors independently screened search results, extracted data, and assessed the risk of bias in the studies and the overall certainty of the evidence using GRADE criteria. We measured treatment effects through analysing dichotomous outcomes (presence/absence of complications) and expressing them as odds ratios (OR) with 95% confidence intervals (CI). For failure in the subgroup of deep lesions, we used network meta-analysis to assess and rank the relative effectiveness of different interventions.\nMain results\nWe included 27 studies with 3350 participants and 4195 teeth/lesions, which were conducted in 11 countries and published between 1977 and 2020. Twenty-four studies used a parallel-group design and three were split-mouth. Two studies included adults only, 20 included children/adolescents only and five included both. Ten studies evaluated permanent teeth, 16 evaluated primary teeth and one evaluated both. Three studies treated non-cavitated lesions; 12 treated cavitated, deep lesions, and 12 treated cavitated but not deep lesions or lesions of varying depth.\nSeventeen studies compared conventional treatment (CR) with a less invasive treatment: SE (8), SW (4), two HT (2), sealing with sealant materials (4) and NRCC (1). Other comparisons were: SE versus HT (2); SE versus SW (4); SE versus sealing  with sealant materials (2); sealant materials versus no sealing (2).\nFollow-up times varied from no follow-up (pulp exposure during treatment) to 120 months, the most common being 12 to 24 months.\nAll studies were at overall high risk of bias.\nEffect of interventions\nSealing using sealants versus other interventions for non-cavitated or cavitated but not deep lesions\nThere was insufficient evidence of a difference between sealing with sealants and CR (OR 5.00, 95% CI 0.51 to 49.27; 1 study, 41 teeth, permanent teeth, cavitated), sealing versus SE (OR 3.11, 95% CI 0.11 to 85.52; 2 studies, 82 primary teeth, cavitated) or sealing versus no treatment (OR 0.05, 95% CI 0.00 to 2.71; 2 studies, 103 permanent teeth, non-cavitated), but we assessed all as very low-certainty evidence.\nHT, CR, SE, NRCC for cavitated, but not deep lesions in primary teeth\nThe odds of failure may be higher for CR than HT (OR 8.35, 95% CI 3.73 to 18.68; 2 studies, 249 teeth; low-certainty evidence) and lower for HT than NRCC (OR 0.19, 95% CI 0.05 to 0.74; 1 study, 84 teeth, very low-certainty evidence). There was insufficient evidence of a difference between SE versus HT (OR 8.94, 95% CI 0.57 to 139.67; 2 studies, 586 teeth) or CR versus NRCC (OR 1.16, 95% CI 0.50 to 2.71; 1 study, 102 teeth), both very low-certainty evidence.\nCR, SE, SW for deep lesions\nThe odds of failure were higher for CR than SW in permanent teeth (OR 2.06, 95% CI 1.34 to 3.17; 3 studies, 398 teeth; moderate-certainty evidence), but not primary teeth (OR 2.43, 95% CI 0.65 to 9.12; 1 study, 63 teeth; very low-certainty evidence).\nThe odds of failure may be higher for CR than SE in permanent teeth (OR 11.32, 95% CI 1.97 to 65.02; 2 studies, 179 teeth) and primary teeth (OR 4.43, 95% CI 1.04 to 18.77; 4 studies, 265 teeth), both very low-certainty evidence. Notably, two studies compared CR versus SE in cavitated, but not deep lesions, with insufficient evidence of a difference in outcome (OR 0.62, 95% CI 0.21 to 1.88; 204 teeth; very low-certainty evidence).\nThe odds of failure were higher for SW than SE in permanent teeth (OR 2.25, 95% CI 1.33 to 3.82; 3 studies, 371 teeth; moderate-certainty evidence), but not primary teeth (OR 2.05, 95% CI 0.49 to 8.62; 2 studies, 126 teeth; very low-certainty evidence).\nFor deep lesions, a network meta-analysis showed the probability of failure to be greatest for CR compared with SE, SW and HT.\nAuthors' conclusions\nCompared with CR, there were lower numbers of failures with HT and SE in the primary dentition, and with SE and SW in the permanent dentition. Most studies showed high risk of bias and limited precision of estimates due to small sample size and typically limited numbers of failures, resulting in assessments of low or very low certainty of evidence for most comparisons.\n\nGiven the provided abstract, please respond to the following query: \"What did we want to find out?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We wanted to know the best way for dentists to manage decay that has extended into dentine or cavitated." } ]
query-laysum
6729
[ { "role": "user", "content": "Abstract: Background\nStroke is the third leading cause of disability worldwide. Physical activity is important for secondary stroke prevention and for promoting functional recovery. However, people with stroke are more inactive than healthy age-matched controls. Therefore, interventions to increase activity after stroke are vital to reduce stroke-related disability.\nObjectives\nTo summarise the available evidence regarding the effectiveness of commercially available, wearable activity monitors and smartphone applications for increasing physical activity levels in people with stroke.\nSearch methods\nWe searched the Cochrane Stroke Group Trials Register, CENTRAL, MEDLINE, Embase, CINAHL, SPORTDiscus, and the following clinical trial registers: WHO International Clinical Trials Registry Platform, Clinical Trials, EU Clinical Trial Register, ISRCTN Registry, Australian and New Zealand Clinical Trial Registry, and Stroke Trials Registry to 3 March 2018. We also searched reference lists, Web of Science forward tracking, and Google Scholar, and contacted trial authors to obtain further data if required. We did not restrict the search on language or publication status.\nSelection criteria\nWe included all randomised controlled trials (RCTs) and randomised cross-over trials that included use of activity monitors versus no intervention, another type of intervention, or other activity monitor. Participants were aged 18 years or older with a diagnosis of stroke, in hospital or living in the community. Primary outcome measures were steps per day and time in moderate-to-vigorous intensity activity. Secondary outcomes were sedentary time, time spent in light intensity physical activity, walking duration, fatigue, mood, quality of life, community participation and adverse events. We excluded upper limb monitors that only measured upper limb activity.\nData collection and analysis\nWe followed standard Cochrane methodology to analyse and interpret the data. At least two authors independently screened titles and abstracts for inclusion. We resolved disagreements by consulting a third review author. We extracted the following data from included studies into a standardised template: type of study, participant population, study setting, intervention and co-interventions, time-frame, and outcomes. We graded levels of bias as high, low, or unclear, and assessed the quality of evidence for each outcome using the GRADE approach.\nMain results\nWe retrieved 28,098 references, from which we identified 29 potential articles. Four RCTs (in 11 reports) met the inclusion criteria.The sample sizes ranged from 27 to 135 (total 245 participants). Time poststroke varied from less than one week (n = 1), to one to three months (n = 2), or a median of 51 months (n = 1). Stroke severity ranged from a median of one to six on the National Institutes of Health Stroke Scale (NIHSS). Three studies were conducted in inpatient rehabilitation, and one was in a university laboratory. All studies compared use of activity monitor plus another intervention (e.g. a walking retraining programme or an inpatient rehabilitation programme) versus the other intervention alone. Three studies reported on the primary outcome of daily step counts.\nThere was no clear effect for the use of activity monitors in conjunction with other interventions on step count in a community setting (mean difference (MD) -1930 steps, 95% confidence interval (CI) -4410 to 550; 1 RCT, 27 participants; very low-quality evidence), or in an inpatient rehabilitation setting (MD 1400 steps, 95% CI -40 to 2840; 2 RCTs, 83 participants; very low-quality evidence). No studies reported the primary outcome moderate-to-vigorous physical activity, but one did report time spent in moderate and vigorous intensity activity separately: this study reported that an activity monitor in addition to usual inpatient rehabilitation increased the time spent on moderate intensity physical activity by 4.4 minutes per day (95% CI 0.28 to 8.52; 1 RCT, 48 participants; low-quality evidence) compared with usual rehabilitation alone, but there was no clear effect for the use of an activity monitor plus usual rehabilitation for increasing time spent in vigorous intensity physical activity compared to usual rehabilitation (MD 2.6 minutes per day, 95% CI -0.8 to 6; 1 RCT, 48 participants; low-quality evidence). The overall risk of bias was low, apart from high-risk for blinding of participants and study personnel. None of the included studies reported any information relating to adverse effects.\nAuthors' conclusions\nOnly four small RCTs with 274 participants (three in inpatient rehabilitation and one in the community) have examined the efficacy of activity monitors for increasing physical activity after stroke. Although these studies showed activity monitors could be incorporated into practice, there is currently not enough evidence to support the use of activity monitors to increase physical activity after stroke.\n\nGiven the provided abstract, please respond to the following query: \"Search date\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched for studies up to 3 March 2018." } ]
query-laysum
6730
[ { "role": "user", "content": "Abstract: Background\nSacrococcygeal pilonidal sinus disease is a common debilitating condition that predominantly affects young adults, with a profound impact on their activities of daily living. The condition is treated surgically, and in some cases the wound in the natal cleft is left open to heal by itself. Many dressings and topical agents are available to aid healing of these wounds.\nObjectives\nTo assess the effects of dressings and topical agents for the management of open wounds following surgical treatment for sacrococcygeal pilonidal sinus in any care setting.\nSearch methods\nIn March 2021, we searched the Cochrane Wounds Specialised Register, CENTRAL, MEDLINE, Embase and EBSCO CINAHL Plus. We also searched clinical trials registries for ongoing and unpublished studies, and we scanned reference lists of included studies, reviews, meta-analyses and health technology reports to identify additional studies. There were no restrictions with respect to language, date of publication or study setting.\nSelection criteria\nWe included parallel-group randomised controlled trials (RCTs) only. We included studies with participants who had undergone any type of sacrococcygeal pilonidal sinus disease surgery and were left with an open wound.\nData collection and analysis\nWe used the standard methodological procedures expected by Cochrane. We used GRADE to assess the certainty of the evidence for each outcome.\nMain results\nWe included 11 RCTs comprising 932 participants. Two studies compared topical negative pressure wound therapy (TNPWT) with conventional open wound healing, two studies compared platelet-rich plasma with sterile absorbent gauze, and the other seven studies compared various dressings and topical agents. All studies were at high risk of bias in at least one domain, whilst one study was judged to be at low risk of bias in all but one domain. All studies were conducted in secondary care. Mean participant ages were between 20 and 30 years, and nearly 80% of participants were male. No studies provided data on quality of life, cost-effectiveness, pain at first dressing change or proportion of wounds healed at 6 or 12 months, and very few adverse effects were recorded in any study.\nIt is unclear whether TNPWT reduces time to wound healing compared with conventional open wound healing (comparison 1), as the certainty of evidence is very low. The two studies provided conflicting results, with one study showing benefit (mean difference (MD) −24.01 days, 95% confidence interval (CI) −35.65 to −12.37; 19 participants), whilst the other reported no difference. It is also unclear whether TNPWT has any effect on the proportion of wounds healed by 30 days (risk ratio (RR) 3.60, 95% CI 0.49 to 26.54; 19 participants, 1 study; very low-certainty evidence). Limited data were available for our secondary outcomes time to return to normal daily activities and recurrence rate; we do not know whether TNPWT has any effect on these outcomes.\nLietofix cream may increase the proportion of wounds that heal by 30 days compared with an iodine dressing (comparison 4; RR 8.06, 95% CI 1.05 to 61.68; 205 participants, 1 study; low-certainty evidence). The study did not provide data on time to wound healing.\nWe do not know whether hydrogel dressings reduce time to wound healing compared with wound cleaning with 10% povidone iodine (comparison 5; MD −24.54 days, 95% CI −47.72 to −1.36; 31 participants, 1 study; very low-certainty evidence). The study did not provide data on the proportion of wounds healed. It is unclear whether hydrogel dressings have any effect on adverse effects as the certainty of the evidence is very low.\nPlatelet-rich plasma may reduce time to wound healing compared with sterile absorbent gauze (comparison 6; MD −19.63 days, 95% CI −34.69 to −4.57; 210 participants, 2 studies; low-certainty evidence). No studies provided data on the proportion of wounds healed. Platelet-rich plasma may reduce time to return to normal daily activities (MD −15.49, 95% CI −28.95 to −2.02; 210 participants, 2 studies; low-certainty evidence).\nZinc oxide mesh may make little or no difference to time to wound healing compared with placebo (comparison 2; median 54 days in the zinc oxide mesh group versus 62 days in the placebo mesh group; low-certainty evidence). We do not know whether zinc oxide mesh has an effect on the proportion of wounds healed by 30 days as the certainty of the evidence is very low (RR 2.35, 95% CI 0.49 to 11.23).\nIt is unclear whether gentamicin-impregnated collagen sponge reduces time to wound healing compared with no dressing (comparison 7; MD −1.40 days, 95% CI −5.05 to 2.25; 50 participants, 1 study; very low-certainty evidence). The study did not provide data on the proportion of wounds healed.\nDialkylcarbamoyl chloride (DACC)-coated dressings may make little or no difference to time to wound healing compared with alginate dressings (comparison 8; median 69 (95% CI 62 to 72) days in the DACC group versus 71 (95% CI 69 to 85) days in the alginate group; 1 study, 246 participants; low-certainty evidence).\nOne study compared a polyurethane foam hydrophilic dressing with an alginate dressing (comparison 3) whilst another study compared a hydrocolloid dressing with an iodine dressing (comparison 9). It is unclear whether either intervention has any effect on time to wound healing as the certainty of evidence is very low.\nAuthors' conclusions\nAt present, the evidence that any of the dressings or topical agents contained in this review have a benefit on time to wound healing, the proportion of wounds that heal at a specific time point or on any of the secondary outcomes of our review ranges from low certainty to very low certainty. There is low-certainty evidence on the benefit on wound healing of platelet-rich plasma from two studies and of Lietofix cream and hydrogel dressings from single studies. Further studies are required to investigate these interventions further.\n\n\nGiven the provided abstract, please respond to the following query: \"What did we do?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We included participants of any age and either sex who had been treated in any care setting. We searched for studies where:\n- participants had been treated for pilonidal sinus disease of the buttocks and were left with an open wound;\n- different dressings and topical agents were compared to see how effective they were for helping wounds to heal." } ]
query-laysum
6731
[ { "role": "user", "content": "Abstract: Background\nPermissive hypoxaemia describes a concept in which a lower level of arterial oxygenation (PaO2) than usual is accepted to avoid the detrimental effects of high fractional inspired oxygen and invasive mechanical ventilation. Currently however, no specific threshold is known that defines permissive hypoxaemia, and its use in adults remains formally untested. The importance of this systematic review is thus to determine whether any substantial evidence is available to support the notion that permissive hypoxaemia may improve clinical outcomes in mechanically ventilated critically ill patients.\nObjectives\nWe assessed whether permissive hypoxaemia (accepting a lower PaO2 than is current practice) in mechanically ventilated critically ill patients affects patient morbidity and mortality. We planned to conduct subgroup and sensitivity analyses and to examine the role of bias to determine the level of evidence provided.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL) 2013, Issue 11, part of The Cochrane Library; MEDLINE (1954 to November 2013); EMBASE (1980 to November 2013); CINAHL (1982 to November 2013) and ISI Web of Science (1946 to November 2013). We combined the sensitive search strategies described in the Cochrane Handbook for Systematic Reviews of Interventions to search for randomized controlled trials (RCTs) in MEDLINE and EMBASE. For ongoing trials, we also searched the following databases: MetaRegister of ControlledTrials and the National Research Register. We applied no language restrictions.\nSelection criteria\nRCTs and quasi-RCTs that compared outcomes for mechanically ventilated critically ill participants, in which the intervention group was targeted to be hypoxaemic relative to the control group, and the control group was normoxaemic or was mildly hypoxaemic, were eligible for inclusion in this review. Exact values defining 'conventional' and 'permissive hypoxaemia' groupings were purposely not specified, and the manner in which these oxygenation goals were achieved also was not specified. We did state however that the intervention group required a target oxygenation level lower than that of the control group, and that the control group target levels should be in the range of normoxaemia or mild hypoxaemia (not hyperoxaemia).\nData collection and analysis\nWe used standard methodological procedures expected by The Cochrane Collaboration. Using the results of the above searches, two review authors (EG-K and KM) independently screened all titles and abstracts for eligibility and duplication. No discrepancies were encountered, nor was it necessary for review authors to contact the first author of any trial to ask for additional information.\nMain results\nOur search strategy yielded a total of 2419 results. After exclusion of duplications, 1651 candidate studies were identified. Screening of titles and abstracts revealed that no studies met our inclusion criteria.\nAuthors' conclusions\nThis comprehensive review failed to identify any relevant studies evaluating permissive hypoxaemia versus normoxaemia in mechanically ventilated critically ill participants. Therefore we are unable to support or refute the hypothesis that this treatment strategy is of benefit to patients.\nGiven the substantial amount of provocative evidence derived from related clinical contexts (resuscitation, myocardial infarction, stroke), we believe that this review highlights an important unanswered question within critical care.  In the presence of two competing harms (hypoxia and hyperoxia), it will be important to carefully evaluate the safety and feasibility of permissive hypoxaemia before proceeding to efficacy and effectiveness trials.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "A common feature of people who become severely unwell and require admission to the ICU/CCU is lack of oxygen in the blood. Regardless of the initial reason that caused them to become unwell, people on the ICU/CCU suffer from the effects of low oxygen levels; however the treatments that we can currently offer are frequently ineffective and may even be harmful. High levels of oxygen are toxic, and the ventilators used to deliver oxygen may cause physical damage to the lungs. Conversely, lower levels of oxygen in the blood than are considered normal are not necessarily harmful and may be seen in people who subsequently fully recover, or in healthy people at altitude. We therefore wanted to ascertain whether any research had been done to examine whether allowing low blood oxygen levels, as opposed to normal blood oxygen levels, in ventilated severely ill people on the ICU/CCU altered their morbidity and mortality." } ]
query-laysum
6732
[ { "role": "user", "content": "Abstract: Background\nThe projected rise in the incidence of type 2 diabetes mellitus (T2DM) could develop into a substantial health problem worldwide. Whether diet, physical activity or both can prevent or delay T2DM and its associated complications in at-risk people is unknown.\nObjectives\nTo assess the effects of diet, physical activity or both on the prevention or delay of T2DM and its associated complications in people at increased risk of developing T2DM.\nSearch methods\nThis is an update of the Cochrane Review published in 2008. We searched the CENTRAL, MEDLINE, Embase, ClinicalTrials.gov, ICTRP Search Portal and reference lists of systematic reviews, articles and health technology assessment reports. The date of the last search of all databases was January 2017. We continuously used a MEDLINE email alert service to identify newly published studies using the same search strategy as described for MEDLINE up to September 2017.\nSelection criteria\nWe included randomised controlled trials (RCTs) with a duration of two years or more.\nData collection and analysis\nWe used standard Cochrane methodology for data collection and analysis. We assessed the overall quality of the evidence using GRADE.\nMain results\nWe included 12 RCTs randomising 5238 people. One trial contributed 41% of all participants. The duration of the interventions varied from two to six years. We judged none of the included trials at low risk of bias for all 'Risk of bias' domains.\nEleven trials compared diet plus physical activity with standard or no treatment. Nine RCTs included participants with impaired glucose tolerance (IGT), one RCT included participants with IGT, impaired fasting blood glucose (IFG) or both, and one RCT included people with fasting glucose levels between 5.3 to 6.9 mmol/L. A total of 12 deaths occurred in 2049 participants in the diet plus physical activity groups compared with 10 in 2050 participants in the comparator groups (RR 1.12, 95% CI 0.50 to 2.50; 95% prediction interval 0.44 to 2.88; 4099 participants, 10 trials; very low-quality evidence). The definition of T2DM incidence varied among the included trials. Altogether 315 of 2122 diet plus physical activity participants (14.8%) developed T2DM compared with 614 of 2389 comparator participants (25.7%) (RR 0.57, 95% CI 0.50 to 0.64; 95% prediction interval 0.50 to 0.65; 4511 participants, 11 trials; moderate-quality evidence). Two trials reported serious adverse events. In one trial no adverse events occurred. In the other trial one of 51 diet plus physical activity participants compared with none of 51 comparator participants experienced a serious adverse event (low-quality evidence). Cardiovascular mortality was rarely reported (four of 1626 diet plus physical activity participants and four of 1637 comparator participants (the RR ranged between 0.94 and 3.16; 3263 participants, 7 trials; very low-quality evidence). Only one trial reported that no non-fatal myocardial infarction or non-fatal stroke had occurred (low-quality evidence). Two trials reported that none of the participants had experienced hypoglycaemia. One trial investigated health-related quality of life in 2144 participants and noted that a minimal important difference between intervention groups was not reached (very low-quality evidence). Three trials evaluated costs of the interventions in 2755 participants. The largest trial of these reported an analysis of costs from the health system perspective and society perspective reflecting USD 31,500 and USD 51,600 per quality-adjusted life year (QALY) with diet plus physical activity, respectively (low-quality evidence). There were no data on blindness or end-stage renal disease.\nOne trial compared a diet-only intervention with a physical-activity intervention or standard treatment. The participants had IGT. Three of 130 participants in the diet group compared with none of the 141 participants in the physical activity group died (very low-quality evidence). None of the participants died because of cardiovascular disease (very low-quality evidence). Altogether 57 of 130 diet participants (43.8%) compared with 58 of 141 physical activity participants (41.1%) group developed T2DM (very low-quality evidence). No adverse events were recorded (very low-quality evidence). There were no data on non-fatal myocardial infarction, non-fatal stroke, blindness, end-stage renal disease, health-related quality of life or socioeconomic effects.\nTwo trials compared physical activity with standard treatment in 397 participants. One trial included participants with IGT, the other trial included participants with IGT, IFG or both. One trial reported that none of the 141 physical activity participants compared with three of 133 control participants died. The other trial reported that three of 84 physical activity participants and one of 39 control participants died (very low-quality evidence). In one trial T2DM developed in 58 of 141 physical activity participants (41.1%) compared with 90 of 133 control participants (67.7%). In the other trial 10 of 84 physical activity participants (11.9%) compared with seven of 39 control participants (18%) developed T2DM (very low-quality evidence). Serious adverse events were rarely reported (one trial noted no events, one trial described events in three of 66 physical activity participants compared with one of 39 control participants - very low-quality evidence). Only one trial reported on cardiovascular mortality (none of 274 participants died - very low-quality evidence). Non-fatal myocardial infarction or stroke were rarely observed in the one trial randomising 123 participants (very low-quality evidence). One trial reported that none of the participants in the trial experienced hypoglycaemia. One trial investigating health-related quality of life in 123 participants showed no substantial differences between intervention groups (very low-quality evidence). There were no data on blindness or socioeconomic effects.\nAuthors' conclusions\nThere is no firm evidence that diet alone or physical activity alone compared to standard treatment influences the risk of T2DM and especially its associated complications in people at increased risk of developing T2DM. However, diet plus physical activity reduces or delays the incidence of T2DM in people with IGT. Data are lacking for the effect of diet plus physical activity for people with intermediate hyperglycaemia defined by other glycaemic variables. Most RCTs did not investigate patient-important outcomes.\n\nGiven the provided abstract, please respond to the following query: \"Quality of the evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "All included trials had deficiencies in the way that they were conducted or how key items were reported. For diet plus physical activity compared with standard treatment, we found rather good evidence that the development of new type 2 diabetes was reduced or delayed. For the other comparisons the number of participants was small, resulting in a high risk of random errors (play of chance)." } ]
query-laysum
6733
[ { "role": "user", "content": "Abstract: Background\nCell transplantation offers a potential therapeutic approach to the repair and regeneration of damaged vascular and cardiac tissue after acute myocardial infarction (AMI). This has resulted in multiple randomised controlled trials (RCTs) across the world.\nObjectives\nTo determine the safety and efficacy of autologous adult bone marrow stem cells as a treatment for acute myocardial infarction (AMI), focusing on clinical outcomes.\nSearch methods\nThis Cochrane review is an update of a previous version (published in 2012). We searched the Cochrane Central Register of Controlled Trials (CENTRAL 2015, Issue 2), MEDLINE (1950 to March 2015), EMBASE (1974 to March 2015), CINAHL (1982 to March 2015) and the Transfusion Evidence Library (1980 to March 2015). In addition, we searched several international and ongoing trial databases in March 2015 and handsearched relevant conference proceedings to January 2011.\nSelection criteria\nRCTs comparing autologous bone marrow-derived cells with no cells in patients diagnosed with AMI were eligible.\nData collection and analysis\nTwo review authors independently screened all references, assessed the risk of bias of the included trials and extracted data. We conducted meta-analyses using random-effects models throughout. We analysed outcomes at short-term (less than 12 months) and long-term (12 months or more) follow-up. Dichotomous outcomes are reported as risk ratio (RR) and continuous outcomes are reported as mean difference (MD) or standardised MD (SMD). We performed sensitivity analyses to evaluate the results in the context of the risk of selection, performance and attrition bias. Exploratory subgroup analysis investigated the effects of baseline cardiac function (left ventricular ejection fraction, LVEF) and cell dose, type and timing of administration, as well as the use of heparin in the final cell solution.\nMain results\nForty-one RCTs with a total of 2732 participants (1564 cell therapy, 1168 controls) were eligible for inclusion. Cell treatment was not associated with any changes in the risk of all-cause mortality (34/538 versus 32/458; RR 0.93, 95% CI 0.58 to 1.50; 996 participants; 14 studies; moderate quality evidence), cardiovascular mortality (23/277 versus 18/250; RR 1.04, 95% CI 0.54 to 1.99; 527 participants; nine studies; moderate quality evidence) or a composite measure of mortality, reinfarction and re-hospitalisation for heart failure (24/262 versus 33/235; RR 0.63, 95% CI 0.36 to 1.10; 497 participants; six studies; moderate quality evidence) at long-term follow-up. Statistical heterogeneity was low (I2 = 0% to 12%). Serious periprocedural adverse events were rare and were generally unlikely to be related to cell therapy. Additionally, cell therapy had no effect on morbidity, quality of life/performance or LVEF measured by magnetic resonance imaging. Meta-analyses of LVEF measured by echocardiography, single photon emission computed tomography and left ventricular angiography showed evidence of differences in mean LVEF between treatment groups although the mean differences ranged between 2% and 5%, which are accepted not to be clinically relevant. Results were robust to the risk of selection, performance and attrition bias from individual studies.\nAuthors' conclusions\nThe results of this review suggest that there is insufficient evidence for a beneficial effect of cell therapy for AMI patients. However, most of the evidence comes from small trials that showed no difference in clinically relevant outcomes. Further adequately powered trials are needed and until then the efficacy of this intervention remains unproven.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "In this updated systematic review we analysed data from a total of 41 trials with over 2700 patients. Evaluation of the currently available evidence indicates that this treatment may not lead to improvement when compared to standard treatment, as measured by the frequency of deaths, heart attacks and/or heart failure requiring re-hospitalisation following treatment, as well as tests of heart function, in the short and long term." } ]
query-laysum
6734
[ { "role": "user", "content": "Abstract: Background\nMonoclonal antibodies (mAbs) are laboratory-produced molecules derived from the B cells of an infected host. They are being investigated as a potential therapy for coronavirus disease 2019 (COVID-19).\nObjectives\nTo assess the effectiveness and safety of SARS-CoV-2-neutralising mAbs for treating patients with COVID-19, compared to an active comparator, placebo, or no intervention. To maintain the currency of the evidence, we will use a living systematic review approach.\nA secondary objective is to track newly developed SARS-CoV-2-targeting mAbs from first tests in humans onwards.\nSearch methods\nWe searched MEDLINE, Embase, the Cochrane COVID-19 Study Register, and three other databases on 17 June 2021. We also checked references, searched citations, and contacted study authors to identify additional studies. Between submission and publication, we conducted a shortened randomised controlled trial (RCT)-only search on 30 July 2021.\nSelection criteria\nWe included studies that evaluated SARS-CoV-2-neutralising mAbs, alone or combined, compared to an active comparator, placebo, or no intervention, to treat people with COVID-19. We excluded studies on prophylactic use of SARS-CoV-2-neutralising mAbs.\nData collection and analysis\nTwo authors independently assessed search results, extracted data, and assessed risk of bias using the Cochrane risk of bias tool (RoB2). Prioritised outcomes were all-cause mortality by days 30 and 60, clinical progression, quality of life, admission to hospital, adverse events (AEs), and serious adverse events (SAEs). We rated the certainty of evidence using GRADE.\nMain results\nWe identified six RCTs that provided results from 17,495 participants with planned completion dates between July 2021 and December 2031. Target sample sizes varied from 1020 to 10,000 participants. Average age was 42 to 53 years across four studies of non-hospitalised participants, and 61 years in two studies of hospitalised participants.\nNon-hospitalised individuals with COVID-19\nFour studies evaluated single agents bamlanivimab (N = 465), sotrovimab (N = 868), regdanvimab (N = 307), and combinations of bamlanivimab/etesevimab (N = 1035), and casirivimab/imdevimab (N = 799). We did not identify data for mortality at 60 days or quality of life. Our certainty of the evidence is low for all outcomes due to too few events (very serious imprecision).\nBamlanivimab compared to placebo\nNo deaths occurred in the study by day 29. There were nine people admitted to hospital by day 29 out of 156 in the placebo group compared with one out of 101 in the group treated with 0.7 g bamlanivimab (risk ratio (RR) 0.17, 95% confidence interval (CI) 0.02 to 1.33), 2 from 107 in the group treated with 2.8 g (RR 0.32, 95% CI 0.07 to 1.47) and 2 from 101 in the group treated with 7.0 g (RR 0.34, 95% CI 0.08 to 1.56). Treatment with 0.7 g, 2.8 g and 7.0 g bamlanivimab may have similar rates of AEs as placebo (RR 0.99, 95% CI 0.66 to 1.50; RR 0.90, 95% CI 0.59 to 1.38; RR 0.81, 95% CI 0.52 to 1.27). The effect on SAEs is uncertain. Clinical progression/improvement of symptoms or development of severe symptoms were not reported.\nBamlanivimab/etesevimab compared to placebo\nThere were 10 deaths in the placebo group and none in bamlanivimab/etesevimab group by day 30 (RR 0.05, 95% CI 0.00 to 0.81). Bamlanivimab/etesevimab may decrease hospital admission by day 29 (RR 0.30, 95% CI 0.16 to 0.59), may result in a slight increase in any grade AEs (RR 1.15, 95% CI 0.83 to 1.59) and may increase SAEs (RR 1.40, 95% CI 0.45 to 4.37). Clinical progression/improvement of symptoms or development of severe symptoms were not reported.\nCasirivimab/imdevimab compared to placebo\nCasirivimab/imdevimab may reduce hospital admissions or death (2.4 g: RR 0.43, 95% CI 0.08 to 2.19; 8.0 g: RR 0.21, 95% CI 0.02 to 1.79). We are uncertain of the effect on grades 3-4 AEs (2.4 g: RR 0.76, 95% CI 0.17 to 3.37; 8.0 g: RR 0.50, 95% CI 0.09 to 2.73) and SAEs (2.4 g: RR 0.68, 95% CI 0.19 to 2.37; 8.0 g: RR 0.34, 95% CI 0.07 to 1.65). Mortality by day 30 and clinical progression/improvement of symptoms or development of severe symptoms were not reported.\nSotrovimab compared to placebo\nWe are uncertain whether sotrovimab has an effect on mortality (RR 0.33, 95% CI 0.01 to 8.18) and invasive mechanical ventilation (IMV) requirement or death (RR 0.14, 95% CI 0.01 to 2.76). Treatment with sotrovimab may reduce the number of participants with oxygen requirement (RR 0.11, 95 % CI 0.02 to 0.45), hospital admission or death by day 30 (RR 0.14, 95% CI 0.04 to 0.48), grades 3-4 AEs (RR 0.26, 95% CI 0.12 to 0.60), SAEs (RR 0.27, 95% CI 0.12 to 0.63) and may have little or no effect on any grade AEs (RR 0.87, 95% CI 0.66 to 1.16).\nRegdanvimab compared to placebo\nTreatment with either dose (40 or 80 mg/kg) compared with placebo may decrease hospital admissions or death (RR 0.45, 95% CI 0.14 to 1.42; RR 0.56, 95% CI 0.19 to 1.60, 206 participants), but may increase grades 3-4 AEs (RR 2.62, 95% CI 0.52 to 13.12; RR 2.00, 95% CI 0.37 to 10.70). 80 mg/kg may reduce any grade AEs (RR 0.79, 95% CI 0.52 to 1.22) but 40 mg/kg may have little to no effect (RR 0.96, 95% CI 0.64 to 1.43). There were too few events to allow meaningful judgment for the outcomes mortality by 30 days, IMV requirement, and SAEs.\nHospitalised individuals with COVID-19\nTwo studies evaluating bamlanivimab as a single agent (N = 314) and casirivimab/imdevimab as a combination therapy (N = 9785) were included.\nBamlanivimab compared to placebo\nWe are uncertain whether bamlanivimab has an effect on mortality by day 30 (RR 1.39, 95% CI 0.40 to 4.83) and SAEs by day 28 (RR 0.93, 95% CI 0.27 to 3.14). Bamlanivimab may have little to no effect on time to hospital discharge (HR 0.97, 95% CI 0.78 to 1.20) and mortality by day 90 (HR 1.09, 95% CI 0.49 to 2.43). The effect of bamlanivimab on the development of severe symptoms at day 5 (RR 1.17, 95% CI 0.75 to 1.85) is uncertain. Bamlanivimab may increase grades 3-4 AEs at day 28 (RR 1.27, 95% CI 0.81 to 1.98). We assessed the evidence as low certainty for all outcomes due to serious imprecision, and very low certainty for severe symptoms because of additional concerns about indirectness.\nCasirivimab/imdevimab with usual care compared to usual care alone\nTreatment with casirivimab/imdevimab compared to usual care probably has little or no effect on mortality by day 30 (RR 0.94, 95% CI 0.87 to 1.02), IMV requirement or death (RR 0.96, 95% CI 0.90 to 1.04), nor alive at hospital discharge by day 30 (RR 1.01, 95% CI 0.98 to 1.04). We assessed the evidence as moderate certainty due to study limitations (lack of blinding). AEs and SAEs were not reported.\nAuthors' conclusions\nThe evidence for each comparison is based on single studies. None of these measured quality of life. Our certainty in the evidence for all non-hospitalised individuals is low, and for hospitalised individuals is very low to moderate. We consider the current evidence insufficient to draw meaningful conclusions regarding treatment with SARS-CoV-2-neutralising mAbs.\nFurther studies and long-term data from the existing studies are needed to confirm or refute these initial findings, and to understand how the emergence of SARS-CoV-2 variants may impact the effectiveness of SARS-CoV-2-neutralising mAbs. Publication of the 36 ongoing studies may resolve uncertainties about the effectiveness and safety of SARS-CoV-2-neutralising mAbs for the treatment of COVID-19 and possible subgroup differences.\n\nGiven the provided abstract, please respond to the following query: \"What did we find?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found six active studies including a total of 17,495 people. Four studies investigated non-hospitalised people with no symptoms or mild COVID-19. Two studies investigated hospitalised people with moderate to severe COVID-19. Studies took place across the world. Three studies were funded by pharmaceutical companies. The monoclonal antibodies they studied were bamlanivimab, etesevimab, casirivimab and imdevimab, sotrovimab, regdanvimab. We did not identify data for mortality at 60 days and quality of life." } ]
query-laysum
6735
[ { "role": "user", "content": "Abstract: Background\nOutcome after acute spontaneous (non-traumatic) intracerebral haemorrhage (ICH) is influenced by haematoma volume. ICH expansion occurs in about 20% of people with acute ICH. Early haemostatic therapy might improve outcome by limiting ICH expansion. This is an update of a Cochrane Review first published in 2006, and last updated in 2018.\nObjectives\nTo examine 1. the effects of individual classes of haemostatic therapies, compared with placebo or open control, in adults with acute spontaneous ICH, and 2. the effects of each class of haemostatic therapy according to the use and type of antithrombotic drug before ICH onset.\nSearch methods\nWe searched the Cochrane Stroke Trials Register, CENTRAL (2022, Issue 8), MEDLINE Ovid, and Embase Ovid on 12 September 2022. To identify further published, ongoing, and unpublished randomised controlled trials (RCTs), we scanned bibliographies of relevant articles and searched international registers of RCTs in September 2022.\nSelection criteria\nWe included RCTs of any haemostatic intervention (i.e. procoagulant treatments such as clotting factor concentrates, antifibrinolytic drugs, platelet transfusion, or agents to reverse the action of antithrombotic drugs) for acute spontaneous ICH, compared with placebo, open control, or an active comparator.\nData collection and analysis\nWe used standard Cochrane methods. Our primary outcome was death/dependence (modified Rankin Scale (mRS) 4 to 6) by day 90. Secondary outcomes were ICH expansion on brain imaging after 24 hours, all serious adverse events, thromboembolic adverse events, death from any cause, quality of life, mood, cognitive function, Barthel Index score, and death or dependence measured on the Extended Glasgow Outcome Scale by day 90.\nMain results\nWe included 20 RCTs involving 4652 participants: nine RCTs of recombinant activated factor VII (rFVIIa) versus placebo/open control (1549 participants), eight RCTs of antifibrinolytic drugs versus placebo/open control (2866 participants), one RCT of platelet transfusion versus open control (190 participants), and two RCTs of prothrombin complex concentrates (PCC) versus fresh frozen plasma (FFP) (47 participants). Four (20%) RCTs were at low risk of bias in all criteria.\nFor rFVIIa versus placebo/open control for spontaneous ICH with or without surgery there was little to no difference in death/dependence by day 90 (risk ratio (RR) 0.88, 95% confidence interval (CI) 0.74 to 1.05; 7 RCTs, 1454 participants; low-certainty evidence). We found little to no difference in ICH expansion between groups (RR 0.81, 95% CI 0.56 to 1.16; 4 RCTs, 220 participants; low-certainty evidence). There was little to no difference in all serious adverse events and death from any cause between groups (all serious adverse events: RR 0.81, 95% CI 0.30 to 2.22; 2 RCTs, 87 participants; very low-certainty evidence; death from any cause: RR 0.78, 95% CI 0.56 to 1.08; 8 RCTs, 1544 participants; moderate-certainty evidence).\nFor antifibrinolytic drugs versus placebo/open control for spontaneous ICH, there was no difference in death/dependence by day 90 (RR 1.00, 95% CI 0.93 to 1.07; 5 RCTs, 2683 participants; high-certainty evidence). We found a slight reduction in ICH expansion with antifibrinolytic drugs for spontaneous ICH compared to placebo/open control (RR 0.86, 95% CI 0.76 to 0.96; 8 RCTs, 2866 participants; high-certainty evidence). There was little to no difference in all serious adverse events and death from any cause between groups (all serious adverse events: RR 1.02, 95% CI 0.75 to 1.39; 4 RCTs, 2599 participants; high-certainty evidence; death from any cause: RR 1.02, 95% CI 0.89 to 1.18; 8 RCTs, 2866 participants; high-certainty evidence). There was little to no difference in quality of life, mood, or cognitive function (quality of life: mean difference (MD) 0, 95% CI -0.03 to 0.03; 2 RCTs, 2349 participants; mood: MD 0.30, 95% CI -1.98 to 2.57; 2 RCTs, 2349 participants; cognitive function: MD -0.37, 95% CI -1.40 to 0.66; 1 RCTs, 2325 participants; all high-certainty evidence).\nPlatelet transfusion likely increases death/dependence by day 90 compared to open control for antiplatelet-associated ICH (RR 1.29, 95% CI 1.04 to 1.61; 1 RCT, 190 participants; moderate-certainty evidence). We found little to no difference in ICH expansion between groups (RR 1.32, 95% CI 0.91 to 1.92; 1 RCT, 153 participants; moderate-certainty evidence). There was little to no difference in all serious adverse events and death from any cause between groups (all serious adverse events: RR 1.46, 95% CI 0.98 to 2.16; 1 RCT, 190 participants; death from any cause: RR 1.42, 95% CI 0.88 to 2.28; 1 RCT, 190 participants; both moderate-certainty evidence).\nFor PCC versus FFP for anticoagulant-associated ICH, the evidence was very uncertain about the effect on death/dependence by day 90, ICH expansion, all serious adverse events, and death from any cause between groups (death/dependence by day 90: RR 1.21, 95% CI 0.76 to 1.90; 1 RCT, 37 participants; ICH expansion: RR 0.54, 95% CI 0.23 to 1.22; 1 RCT, 36 participants; all serious adverse events: RR 0.27, 95% CI 0.02 to 3.74; 1 RCT, 5 participants; death from any cause: RR 0.49, 95% CI 0.16 to 1.56; 2 RCTs, 42 participants; all very low-certainty evidence).\nAuthors' conclusions\nIn this updated Cochrane Review including 20 RCTs involving 4652 participants, rFVIIa likely results in little to no difference in reducing death or dependence after spontaneous ICH with or without surgery; antifibrinolytic drugs result in little to no difference in reducing death or dependence after spontaneous ICH, but result in a slight reduction in ICH expansion within 24 hours; platelet transfusion likely increases death or dependence after antiplatelet-associated ICH; and the evidence is very uncertain about the effect of PCC compared to FFP on death or dependence after anticoagulant-associated ICH. Thirteen RCTs are ongoing and are likely to increase the certainty of the estimates of treatment effect.\n\nGiven the provided abstract, please respond to the following query: \"What is an intracerebral haemorrhage?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "More than one-tenth of all strokes are caused by intracerebral haemorrhage. The bigger the haemorrhage, the more likely it is to be fatal. Roughly one-fifth of intracerebral haemorrhages enlarge significantly, most during the first three hours after the bleed started." } ]
query-laysum
6736
[ { "role": "user", "content": "Abstract: Background\nInterleukin-1 (IL-1) blocking agents have been used for treating severe coronavirus disease 2019 (COVID-19), on the premise that their immunomodulatory effect might be beneficial in people with COVID-19.\nObjectives\nTo assess the effects of IL-1 blocking agents compared with standard care alone or with placebo on effectiveness and safety outcomes in people with COVID-19.\nWe will update this assessment regularly.\nSearch methods\nWe searched the Cochrane COVID-19 Study Register and the COVID-19 L-OVE Platform (search date 5 November 2021). These sources are maintained through regular searches of MEDLINE, Embase, CENTRAL, trial registers and other sources. We also checked the World Health Organization International Clinical Trials Registry Platform, regulatory agency websites, Retraction Watch (search date 3 November 2021).\nSelection criteria\nWe included randomised controlled trials (RCTs) evaluating IL-1 blocking agents compared with standard care alone or with placebo for people with COVID-19, regardless of disease severity.\nData collection and analysis\nWe followed Cochrane methodology. The protocol was amended to reduce the number of outcomes considered. Two researchers independently screened and extracted data and assessed the risk of bias with the Cochrane Risk of Bias 2 tool. We rated the certainty of evidence using the GRADE approach for the critical outcomes of clinical improvement (Day 28; ≥ D60); WHO Clinical Progression Score of level 7 or above (i.e. the proportion of participants with mechanical ventilation +/- additional organ support OR death) (D28; ≥ D60); all-cause mortality (D28; ≥ D60); incidence of any adverse events; and incidence of serious adverse events.\nMain results\nWe identified four RCTs of anakinra (three published in peer-reviewed journals, one reported as a preprint) and two RCTs of canakinumab (published in peer-reviewed journals). All trials were multicentre (2 to 133 centres). Two trials stopped early (one due to futility and one as the trigger for inferiority was met). The median/mean age range varied from 58 to 68 years; the proportion of men varied from 58% to 77%. All participants were hospitalised; 67% to 100% were on oxygen at baseline but not intubated; between 0% and 33% were intubated at baseline. We identified a further 16 registered trials with no results available, of which 15 assessed anakinra (four completed, four terminated, five ongoing, three not recruiting) and one (completed) trial assessed canakinumab.\nEffectiveness of anakinra for people with COVID-19\nAnakinra probably results in little or no increase in clinical improvement at D28 (risk ratio (RR) 1.08, 95% confidence interval (CI) 0.97 to 1.20; 3 RCTs, 837 participants; absolute effect: 59 more per 1000 (from 22 fewer to 147 more); moderate-certainty evidence.\nThe evidence is uncertain about an effect of anakinra on 1) the proportion of participants with a WHO Clinical Progression Score of level 7 or above at D28 (RR 0.67, 95% CI 0.36 to 1.22; 2 RCTs, 722 participants; absolute effect: 55 fewer per 1000 (from 107 fewer to 37 more); low-certainty evidence) and ≥ D60 (RR 0.54, 95% CI 0.30 to 0.96; 1 RCT, 606 participants; absolute effect: 47 fewer per 1000 (from 72 fewer to 4 fewer) low-certainty evidence); and 2) all-cause mortality at D28 (RR 0.69, 95% CI 0.34 to 1.39; 2 RCTs, 722 participants; absolute effect: 32 fewer per 1000 (from 68 fewer to 40 more); low-certainty evidence).\nThe evidence is very uncertain about an effect of anakinra on 1) the proportion of participants with clinical improvement at ≥ D60 (RR 0.93, 95% CI 0.78 to 1.12; 1 RCT, 115 participants; absolute effect: 59 fewer per 1000 (from 186 fewer to 102 more); very low-certainty evidence); and 2) all-cause mortality at ≥ D60 (RR 1.03, 95% CI 0.68 to 1.56; 4 RCTs, 1633 participants; absolute effect: 8 more per 1000 (from 84 fewer to 147 more); very low-certainty evidence).\nSafety of anakinra for people with COVID-19\nAnakinra probably results in little or no increase in adverse events (RR 1.02, 95% CI 0.94 to 1.11; 2 RCTs, 722 participants; absolute effect: 14 more per 1000 (from 43 fewer to 78 more); moderate-certainty evidence).\nThe evidence is uncertain regarding an effect of anakinra on serious adverse events (RR 0.95, 95% CI 0.58 to 1.56; 2 RCTs, 722 participants; absolute effect: 12 fewer per 1000 (from 104 fewer to 138 more); low-certainty evidence).\nEffectiveness of canakinumab for people with COVID-19\nCanakinumab probably results in little or no increase in clinical improvement at D28 (RR 1.05, 95% CI 0.96 to 1.14; 2 RCTs, 499 participants; absolute effect: 42 more per 1000 (from 33 fewer to 116 more); moderate-certainty evidence).\nThe evidence of an effect of canakinumab is uncertain on 1) the proportion of participants with a WHO Clinical Progression Score of level 7 or above at D28 (RR 0.72, 95% CI 0.44 to 1.20; 2 RCTs, 499 participants; absolute effect: 35 fewer per 1000 (from 69 fewer to 25 more); low-certainty evidence); and 2) all-cause mortality at D28 (RR:0.75; 95% CI 0.39 to 1.42); 2 RCTs, 499 participants; absolute effect: 20 fewer per 1000 (from 48 fewer to 33 more); low-certainty evidence).\nThe evidence is very uncertain about an effect of canakinumab on all-cause mortality at ≥ D60 (RR 0.55, 95% CI 0.16 to 1.91; 1 RCT, 45 participants; absolute effect: 112 fewer per 1000 (from 210 fewer to 227 more); very low-certainty evidence).\nSafety of canakinumab for people with COVID-19\nCanakinumab probably results in little or no increase in adverse events (RR 1.02; 95% CI 0.86 to 1.21; 1 RCT, 454 participants; absolute effect: 11 more per 1000 (from 74 fewer to 111 more); moderate-certainty evidence).\nThe evidence of an effect of canakinumab on serious adverse events is uncertain (RR 0.80, 95% CI 0.57 to 1.13; 2 RCTs, 499 participants; absolute effect: 44 fewer per 1000 (from 94 fewer to 28 more); low-certainty evidence).\nAuthors' conclusions\nOverall, we did not find evidence for an important beneficial effect of IL-1 blocking agents. The evidence is uncertain or very uncertain for several outcomes. Sixteen trials of anakinra and canakinumab with no results are currently registered, of which four are completed, and four terminated. The findings of this review are updated on the COVID-NMA platform (covid-nma.com).\n\nGiven the provided abstract, please respond to the following query: \"What is interleukin-1 and what is its role in COVID-19?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Interleukin-1 (IL-1) is a type of protein called a cytokine, which helps to regulate the body’s immune system. In particular, IL-1 triggers inflammation to help fight infection. In COVID-19, as the immune system fights the virus, the lungs and airways become inflamed, causing breathing difficulties. However, in some people, the immune system can over-react (called a ‘cytokine storm’) and produce dangerously high levels of inflammation and tissue damage. This can lead to severe breathing difficulties, organ failure and death." } ]
query-laysum
6737
[ { "role": "user", "content": "Abstract: Background\nHepatitis B virus (HBV) infection, a liver disease caused by hepatitis B virus, may lead to serious complications such as cirrhosis and hepatocellular carcinoma. People with HBV infection may have co-infections including HIV and other hepatitis viruses (hepatitis C or D), and co-infection may increase the risk of all-cause mortality. Chronic HBV infection increases morbidity and psychological stress and is an economic burden on people with chronic hepatitis B and their families. Radix Sophorae flavescentis, an herbal medicine, is administered most often in combination with other drugs or herbs. It is believed that it decreases discomfort and prevents replication of the virus in people with chronic hepatitis B. However, the benefits and harms of Radix Sophorae flavescentis for patient-centred outcomes are not known, and its wide usage has never been established with rigorous review methodology.\nObjectives\nTo assess the benefits and harms of Radix Sophorae flavescentis versus placebo or no intervention in people with chronic hepatitis B.\nSearch methods\nWe searched the Cochrane Hepato-Biliary Group Controlled Trials Register, CENTRAL, MEDLINE Ovid, Embase Ovid, LILACS, Science Citation Index Expanded, Conference Proceedings Citation Index - Science, China National Knowledge Infrastructure (CNKI), Chongqing VIP (CQVIP), Wanfang Data, and SinoMed. We also searched the World Health Organization International Clinical Trials Registry Platform (www.who.int/ictrp), ClinicalTrials.gov (www.clinicaltrials.gov/), and the Chinese Clinical Trial Registry for ongoing or unpublished trials. We conducted the last search in December 2018.\nSelection criteria\nWe included randomised clinical trials, irrespective of publication status, language, or blinding, comparing Radix Sophorae flavescentis versus no intervention or placebo in people with chronic hepatitis B. We excluded polyherbal blends containing Radix Sophorae flavescentis. We allowed co-interventions when the co-interventions were administered equally to all intervention groups.\nData collection and analysis\nWe used standard methodological procedures expected by The Cochrane Collaboration. Review authors in pairs retrieved data from individual published reports and after correspondence with investigators. Our primary outcomes were all-cause mortality, serious adverse events, and health-related quality of life. Our secondary outcomes were hepatitis B-related mortality, hepatitis B-related morbidity, and adverse events considered 'not to be serious'. We presented meta-analysed results as risk ratios (RRs) with 95% confidence intervals (CIs). We assessed risk of bias using domains with pre-defined definitions. We conducted Trial Sequential Analyses to control the risk of random errors. We used GRADE methodology to evaluate our certainty in the evidence (i.e. \"the extent of our confidence that the estimates of the effect are correct or are adequate to support a particular decision or recommendation\").\nMain results\nWe included 35 randomised clinical trials with 3556 participants. One trial compared Radix Sophorae flavescentis with placebo; the remaining 34 trials compared effects of Radix Sophorae flavescentis in addition to a co-intervention versus the same co-intervention. The included trials assessed heterogenous forms and ways of administering Radix Sophorae flavescentis (e.g. oral capsules, oral tablets, intravenous infusion, intramuscular injection, acupoint (a specifically chosen site of acupuncture) injection) with treatment duration of 1 to 24 months. Two of the trials included children up to 14 years old. Participants in two trials had cirrhosis in addition to chronic hepatitis B. All trials were assessed at high risk of bias, and certainty of the evidence for all outcomes was very low.\nOnly one of the 35 trials assessed mortality; no deaths occurred. Ten trials assessed serious adverse events; no serious adverse events occurred. None of the trials reported health-related quality of life, hepatitis B-related mortality, or morbidity. Adverse events considered 'not to be serious' was an outcome in 19 trials; nine of these trials had zero events in both groups. Radix Sophorae flavescentis versus placebo or no intervention showed no difference in effects on adverse events considered 'not to be serious' (RR 1.10, 95% CI 0.76 to 1.59; I² = 49%; 10 trials, 1050 participants). Radix Sophorae flavescentis showed a reduction in the proportion of participants with detectable HBV-DNA (RR 0.61, 95% CI 0.55 to 0.68; I² = 56%; 29 trials, 2914 participants) and in the proportion of participants with detectable HBeAg (hepatitis B e-antigen) (RR 0.71, 95% CI 0.66 to 0.76; I² = 19%; 20 trials, 2129 participants).\nSeven of the 35 randomised clinical trials received academic funding from government or hospital. Four trials received no funding. The remaining 24 trials provided no information on funding.\nAdditionally, 432 trials lacked the methodological information needed to ensure inclusion of these trials in our review.\nAuthors' conclusions\nThe included trials lacked data on health-related quality of life, hepatitis B-related mortality, and hepatitis B-related morbidity. The effects of Radix Sophorae flavescentis on all-cause mortality and on the proportion of participants with serious adverse events and adverse events considered 'not to be serious' remain unclear. We advise caution in interpreting results showing that Radix Sophorae flavescentis reduced the proportion of people with detectable HBV-DNA and detectable HBeAg because the trials reporting on these outcomes are at high risk of bias and both outcomes are non-validated surrogate outcomes. We were unable to obtain information on the design and conduct of a large number of trials; therefore, we were deterred from including them in our review. Undisclosed funding may influence trial results and may lead to poor trial design. Given the wide usage of Radix Sophorae flavescentis, we need large, unbiased, high-quality placebo-controlled randomised trials in which patient-centred outcomes are assessed.\n\nGiven the provided abstract, please respond to the following query: \"Search date\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "December 2018 was the date of the last search." } ]
query-laysum
6738
[ { "role": "user", "content": "Abstract: Background\nLaparoscopic adrenalectomy is an accepted treatment worldwide for adrenal gland disease in adults. The transperitoneal approach is more common. The retroperitoneal approach may be preferred, to avoid entering the peritoneum, but no clear advantage has been demonstrated so far.\nObjectives\nTo assess the effects of laparoscopic transperitoneal adrenalectomy (LTPA) versus laparoscopic retroperitoneal adrenalectomy (LRPA) for adrenal tumours in adults.\nSearch methods\nWe searched CENTRAL, MEDLINE, Embase, ICTRP Search Portal, and ClinicalTrials.gov to 3 April 2018. We applied no language restrictions.\nSelection criteria\nTwo review authors independently scanned the abstract, title, or both sections of every record retrieved to identify randomised controlled trials (RCTs) on laparoscopic adrenalectomy for preoperatively assessed adrenal tumours. Participants were affected by corticoid and medullary, benign and malignant, functional and silent tumours or masses of the adrenal gland, which were assessed by both laboratory and imaging studies.\nData collection and analysis\nTwo review authors independently extracted data, assessed trials for risk of bias, and evaluated overall study quality using GRADE criteria. We calculated the risk ratio (RR) for dichotomous outcomes, or the mean difference (MD) for continuous variables, and corresponding 95% confidence interval (CI). We primarily used a random-effects model for pooling data.\nMain results\nWe examined 1069 publications, scrutinized 42 full-text publications or records, and included five RCTs. Altogether, 244 participants entered the five trials; 127 participants were randomised to retroperitoneal adrenalectomy and 117 participants to transperitoneal adrenalectomy. Two trials had a follow-up of nine months, and three trials a follow-up of 31 to 70 months. Most participants were women, and the average age was around 40 years. Three trials reported all-cause mortality; in two trials, there were no deaths, and in one trial with six years of follow-up, four participants died in the LRPA group and one participant in the LTPA group (164 participants; low-certainty evidence). The trials did not report all-cause morbidity. Therefore, we analysed early and late morbidity, and included specific adverse events under these outcome measures. The results were inconclusive between LRPA and LTPA for early morbidity (usually reported within 30 to 60 days after surgery; RR 0.56, 95% CI 0.27 to 1.16; P = 0.12; 5 trials, 244 participants; very low-certainty evidence). Nine out of 127 participants (7.1%) in the LRPA group, compared with 16 out of 117 participants (13.7%) in the LTPA group experienced an adverse event. Participants in the LRPA group may have a lower risk of developing late morbidity (reported as latest available follow-up; RR 0.12, 95% CI 0.01 to 0.92; P = 0.04; 3 trials, 146 participants; very low-quality evidence). None of the 78 participants in the LRPA group, compared with 7 of the 68 participants (10.3%) in the LTPA group experienced an adverse event.\nNone of the trials reported health-related quality of life. The results were inconclusive for socioeconomic effects, assessed as time to return to normal activities and length of hospital stay, between the intervention and comparator groups (very low-certainty evidence). Participants who had LRPA may have had an earlier start on oral fluid or food intake (MD -8.6 hr, 95% CI -13.5 to -3.7; P = 0.0006; 2 trials, 89 participants), and ambulation (MD -5.4 hr, 95% CI -6.8 to -4.0 hr; P < 0.0001; 2 trials, 89 participants) than those in the LTPA groups. Postoperative and operative parameters (duration of surgery, operative blood loss, conversion to open surgery) showed inconclusive results between the intervention and comparator groups.\nAuthors' conclusions\nThe body of evidence on laparoscopic retroperitoneal adrenalectomy compared with laparoscopic transperitoneal adrenalectomy is limited. Late morbidity might be reduced following laparoscopic retroperitoneal adrenalectomy, but we are uncertain about this effect because of very low-quality evidence. The effects on other key outcomes, such as all-cause mortality, early morbidity, socioeconomic effects, and operative and postoperative parameters are uncertain. LRPA might show a shorter time to oral fluid or food intake and time to ambulation, but we are uncertain whether this finding can be replicated. New long-term RCTs investigating additional data, such as health-related quality of life, surgeons' level of experience, treatment volume of surgical centres, and details on techniques used are needed.\n\nGiven the provided abstract, please respond to the following query: \"Certainty of the evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We are uncertain which adrenalectomy technique is best, mainly because of the small number of studies, small number of participants, and some systematic errors in the majority of our analysed studies. New studies should especially investigate health-related quality of life. Surgeons' level of experience and treatment volume of surgical centres might also influence results." } ]
query-laysum
6739
[ { "role": "user", "content": "Abstract: Background\nCollagenous colitis is a cause of chronic diarrhea. This updated review was performed to identify therapies for collagenous colitis that have been assessed in randomized controlled trials (RCTs).\nObjectives\nThe primary objective was to assess the benefits and harms of treatments for collagenous colitis.\nSearch methods\nWe searched CENTRAL, the Cochrane IBD Group Specialized Register, MEDLINE and EMBASE from inception to 7 November 2016.\nSelection criteria\nWe included RCTs comparing a therapy with placebo or active comparator for the treatment of active or quiescent collagenous colitis.\nData collection and analysis\nData were independently extracted by two authors. The primary outcome was clinical response or maintenance of response as defined by the included studies. Secondary outcome measures included histological response, quality of life and the occurrence of adverse events. Risk ratios (RR) and 95% confidence intervals (CI) were calculated for dichotomous outcomes. The Cochrane risk of bias tool was used to assess bias. The overall quality of the evidence was assessed using the GRADE criteria.\nMain results\nTwelve RCTs (476 participants) were included. These studies assessed bismuth subsalicylate, Boswellia serrata extract, mesalamine, cholestyramine, probiotics, prednisolone and budesonide therapy. Four studies were low risk of bias. One study assessing mesalamine and cholestyramine was judged to be high risk of bias due to no blinding. The other studies had an unclear risk of bias for random sequence generation (five studies) allocation concealment (six studies), blinding (one study), incomplete outcome data (one study) and selective outcome reporting (one study). Clinical response occurred in 100% (4/4) of patients who received bismuth subsalicylate (nine 262 mg tablets daily for 8 weeks) compared to 0% (0/5) of patients who received placebo (1 study; 9 participants; RR 10.80, 95% CI 0.75 to 155.93; GRADE = very low). Clinical response occurred in 44% (7/16) of patients who received Boswellia serrata extract (three 400 mg/day capsules for 8 weeks) compared to 27% (4/15) of patients who received placebo (1 study; 31 participants; RR 1.64, 95% CI 0.60 to 4.49; GRADE = low). Clinical response occurred in 80% (24/30) of budesonide patients compared to 44% (11/25) of mesalamine patients (1 study; 55 participants; RR 1.82, 95% CI 1.13 to 2.93; GRADE = low). Histological response was observed in 87% (26/30) of budesonide patients compared to 44% (11/25) of mesalamine patients (1 study, 55 participants; RR 1.97, 95% CI 1.24 to 3.13; GRADE = low). There was no difference between the two treatments with respect to adverse events (RR 0.69, 95% CI 0.43 to 1.10; GRADE = low), withdrawals due to adverse events (RR 0.09, 95% CI 0.01 to 1.65; GRADE = low) and serious adverse events (RR 0.12, 95% CI 0.01 to 2.21; GRADE = low). Clinical response occurred in 44% (11/25) of mesalamine patients (3 g/day) compared to 59% (22/37) of placebo patients (1 study; 62 participants; RR 0.74, 95% CI 0.44 to 1.24; GRADE = low). Histological response was observed in 44% (11/25) and 51% (19/37) of patients receiving mesalamine and placebo, respectively (1 study; 62 participants; RR 0.86, 95% CI 0.50 to 1.47; GRADE = low). There was no difference between the two treatments with respect to adverse events (RR 1.26, 95% CI 0.84 to 1.88; GRADE = low), withdrawals due to adverse events (RR 5.92, 95% CI 0.70 to 49.90; GRADE = low) and serious adverse events (RR 4.44, 95% CI 0.49 to 40.29; GRADE = low). Clinical response occurred in 63% (5/8) of prednisolone (50 mg/day for 2 weeks) patients compared to 0% (0/3) of placebo patients (1 study, 11 participants; RR 4.89, 95% CI 0.35 to 68.83; GRADE = very low). Clinical response occurred in 29% (6/21) of patients who received probiotics (2 capsules containing 0.5 x 1010 CFU each of L. acidophilus LA-5 and B. animalis subsp. lactis strain BB-12 twice daily for 12 weeks) compared to 13% (1/8) of placebo patients (1 study, 29 participants, RR 2.29, 95% CI 0.32 to 16.13; GRADE = very low). Clinical response occurred in 73% (8/11) of patients who received mesalamine (800 mg three times daily) compared to 100% (12/12) of patients who received mesalamine + cholestyramine (4 g daily) (1 study, 23 participants; RR 0.74, 95% CI 0.50 to 1.08; GRADE = very low). Clinical response occurred in 81% (38/47) of patients who received budesonide (9 mg daily in a tapering schedule for 6 to 8 weeks) compared to 17% (8/47) of placebo patients (3 studies; 94 participants; RR 4.56, 95% CI 2.43 to 8.55; GRADE = low). Histological response was higher in budesonide participants (72%, 34/47) compared to placebo (17%, 8/47) (RR 4.15, 95% CI 2.25 to 7.66; GRADE = low). Clinical response was maintained in 68% (57/84) of budesonide patients compared to 20% (18/88) of placebo patients (3 studies, 172 participants, RR 3.30 95% CI 2.13 to 5.09; GRADE = low). Histological response was maintained in 48% (19/40) of budesonide patients compared to 15% (6/40) of placebo patients (2 studies; 80 participants; RR 3.17, 95% CI 1.44 to 6.95; GRADE = very low). No difference was found between budesonide and placebo for adverse events (5 studies; 290 participants; RR 1.18, o95% CI 0.92 to 1.51; GRADE = low), withdrawals due to adverse events (5 studies, 290 participants; RR 0.97, 95% CI 0.43 to 2.17; GRADE = very low) or serious adverse events (4 studies, 175 participants; RR 1.11, 95% CI 0.15 to 8.01; GRADE = very low). Adverse effects reported in the budesonide studies include nausea, vomiting, neck pain, abdominal pain, excessive sweating and headache. Adverse effects reported in the mesalamine studies included nausea and skin rash. Adverse effects in the prednisolone study included abdominal pain, headache, sleep disturbance, mood change and weight gain.\nAuthors' conclusions\nLow quality evidence suggests that budesonide may be effective for inducing and maintaining clinical and histological response in patients with collagenous colitis. We are uncertain about the benefits and harms of therapy with bismuth subsalicylate, Boswellia serrata extract, mesalamine with or without cholestramine, prednisolone and probiotics. These agents and other therapies require further study.\n\nGiven the provided abstract, please respond to the following query: \"What is collagenous colitis?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Collagenous colitis is a type of microscopic colitis, a condition characterized by chronic watery non-bloody diarrhea. People with collagenous colitis have a normal appearing bowel when assessed by an endoscope (a camera used to look at the bowel); but have microscopic inflammation of the bowel when assessed by a biopsy (a tissue sample taken during endoscopy). The cause of this disorder is unknown." } ]
query-laysum
6740
[ { "role": "user", "content": "Abstract: Background\nIn vitro fertilisation (IVF) is a treatment for unexplained subfertility but is invasive, expensive, and associated with risks.\nObjectives\nTo evaluate the effectiveness and safety of IVF versus expectant management, unstimulated intrauterine insemination (IUI), and IUI with ovarian stimulation using gonadotropins, clomiphene citrate (CC), or letrozole in improving pregnancy outcomes.\nSearch methods\nWe searched following databases from inception to November 2021, with no language restriction: Cochrane Gynaecology and Fertility Register, CENTRAL, MEDLINE, Embase, PsycINFO, CINAHL. We searched reference lists of articles and conference abstracts.\nSelection criteria\nRandomised controlled trials (RCTs) comparing effectiveness of IVF for unexplained subfertility with expectant management, unstimulated IUI, and stimulated IUI.\nData collection and analysis\nWe followed standard Cochrane methods.\nMain results\nIVF versus expectant management (two RCTs)\nWe are uncertain whether IVF improves live birth rate (LBR) and clinical pregnancy rate (CPR) compared to expectant management (odds ratio (OR) 22.0, 95% confidence interval (CI) 2.56 to 189.37; 1 RCT; 51 women; very low-quality evidence; OR 3.24, 95% CI 1.07 to 9.8; 2 RCTs; 86 women; I2 = 80%; very low-quality evidence). Adverse effects were not reported. Assuming 4% LBR and 12% CPR with expectant management, these would be 8.8% to 9% and 13% to 58% with IVF.\nIVF versus unstimulated IUI (two RCTs)\nIVF may improve LBR compared to unstimulated IUI (OR 2.47, 95% CI 1.19 to 5.12; 2 RCTs; 156 women; I2 = 60%; low-quality evidence). We are uncertain whether there is a difference between IVF and IUI for multiple pregnancy rate (MPR) (OR 1.03, 95% CI 0.04 to 27.29; 1 RCT; 43 women; very low-quality evidence) and miscarriage rate (OR 1.72, 95% CI 0.14 to 21.25; 1 RCT; 43 women; very low-quality evidence). No study reported ovarian hyperstimulation syndrome (OHSS). Assuming 16% LBR, 3% MPR, and 6% miscarriage rate with unstimulated IUI, these outcomes would be 18.5% to 49%, 0.1% to 46%, and 0.9% to 58% with IVF.\nIVF versus IUI + ovarian stimulation with gonadotropins (6 RCTs), CC (1 RCT), or letrozole (no RCTs)\nStratified analysis was based on pretreatment status.\nTreatment-naive women\nThere may be little or no difference in LBR between IVF and IUI + gonadotropins (1 IVF to 2 to 3 IUI cycles: OR 1.19, 95% CI 0.87 to 1.61; 3 RCTs; 731 women; I2 = 0%; low-quality evidence; 1 IVF to 1 IUI cycle: OR 1.63, 95% CI 0.91 to 2.92; 2 RCTs; 221 women; I2 = 54%; low-quality evidence); or between IVF and IUI + CC (OR 2.51, 95% CI 0.96 to 6.55; 1 RCT; 103 women; low-quality evidence). Assuming 42% LBR with IUI + gonadotropins (1 IVF to 2 to 3 IUI cycles) and 26% LBR with IUI + gonadotropins (1 IVF to 1 IUI cycle), LBR would be 39% to 54% and 24% to 51% with IVF. Assuming 15% LBR with IUI + CC, LBR would be 15% to 54% with IVF.\nThere may be little or no difference in CPR between IVF and IUI + gonadotropins (1 IVF to 2 to 3 IUI cycles: OR 1.17, 95% CI 0.85 to 1.59; 3 RCTs; 731 women; I2 = 0%; low-quality evidence; 1 IVF to 1 IUI cycle: OR 4.59, 95% CI 1.86 to 11.35; 1 RCT; 103 women; low-quality evidence); or between IVF and IUI + CC (OR 3.58, 95% CI 1.51 to 8.49; 1 RCT; 103 women; low-quality evidence). Assuming 48% CPR with IUI + gonadotropins (1 IVF to 2 to 3 IUI cycles) and 17% with IUI + gonadotropins (1 IVF to 1 IUI cycle), CPR would be 44% to 60% and 28% to 70% with IVF. Assuming 21% CPR with IUI + CC, CPR would be 29% to 69% with IVF.\nThere may be little or no difference in multiple pregnancy rate (MPR) between IVF and IUI + gonadotropins (1 IVF to 2 to 3 IUI cycles: OR 0.82, 95% CI 0.38 to 1.77; 3 RCTs; 731 women; I2 = 0%; low-quality evidence; 1 IVF to 1 IUI cycle: OR 0.76, 95% CI 0.36 to 1.58; 2 RCTs; 221 women; I2 = 0%; low-quality evidence); or between IVF and IUI + CC (OR 0.64, 95% CI 0.17 to 2.41; 1 RCT; 102 women; low-quality evidence).\nWe are uncertain if there is a difference in OHSS between IVF and IUI + gonadotropins with 1 IVF to 2 to 3 IUI cycles (OR 6.86, 95% CI 0.35 to 134.59; 1 RCT; 207 women; very low-quality evidence); and there may be little or no difference in OHSS with 1 IVF to 1 IUI cycle (OR 1.22, 95% CI 0.36 to 4.16; 2 RCTs; 221 women; I2 = 0%; low-quality evidence). There may be little or no difference between IVF and IUI + CC (OR 1.53, 95% CI 0.24 to 9.57; 1 RCT; 102 women; low-quality evidence).\nWe are uncertain if there is a difference in miscarriage rate between IVF and IUI + gonadotropins with 1 IVF to 2 to 3 IUI cycles (OR 0.31, 95% CI 0.03 to 3.04; 1 RCT; 207 women; very low-quality evidence); and there may be little or no difference with 1 IVF to 1 IUI cycle (OR 1.16, 95% CI 0.44 to 3.02; 1 RCT; 103 women; low-quality evidence). There may be little or no difference between IVF and IUI + CC (OR 1.48, 95% CI 0.54 to 4.05; 1 RCT; 102 women; low-quality evidence).\nIn women pretreated with IUI + CC\nIVF may improve LBR compared with IUI + gonadotropins (OR 3.90, 95% CI 2.32 to 6.57; 1 RCT; 280 women; low-quality evidence). Assuming 22% LBR with IUI + gonadotropins, LBR would be 39% to 65% with IVF.\nIVF may improve CPR compared with IUI + gonadotropins (OR 14.13, 95% CI 7.57 to 26.38; 1 RCT; 280 women; low-quality evidence). Assuming 30% CPR with IUI + gonadotropins, CPR would be 76% to 92% with IVF.\nAuthors' conclusions\nIVF may improve LBR over unstimulated IUI. Data should be interpreted with caution as overall evidence quality was low.\n\nGiven the provided abstract, please respond to the following query: \"What are the limitations of the evidence?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We have low confidence in the evidence because there were relatively few studies, with low numbers of participants." } ]
query-laysum
6741
[ { "role": "user", "content": "Abstract: Background\nEndometrial cancer is one of the most common gynaecological cancers in developed countries. Treatment of advanced endometrial cancer usually involves radiotherapy, chemotherapy, endocrine therapy or a combination of these. However, survival outcomes are poor in advanced or metastatic disease. Better systemic treatment options are needed to improve survival and safety outcomes for these women. The PI3K/AKT/mTOR pathway is a frequently altered signalling pathway in endometrial cancer. Single-arm studies have reported some encouraging results of the PI3K/AKT/mTOR inhibition in advanced or recurrent endometrial cancer.\nObjectives\nTo assess the efficacy and safety of PI3K/AKT/mTOR inhibitor-containing regimens in women with locally-advanced, metastatic or recurrent endometrial cancer.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials, MEDLINE and Embase to 16 January 2019; and the World Health Organization's International Clinical Trials Registry Platform (WHO ICTRP) and ClinicalTrials.gov in July 2018. We also reviewed reference lists from included studies and endometrial cancer guidelines.\nSelection criteria\nWe included randomised controlled trials (RCTs) comparing a regimen with a PI3K/AKT/mTOR inhibitor (either alone or in combination with other treatments, such as chemotherapy or hormonal therapy) versus a comparator regimen without a PI3K/AKT/mTOR inhibitor. There were no restrictions on which comparator(s) were included.\nData collection and analysis\nWe extracted data independently, and assessed risks of bias and the certainty of the evidence. The primary outcome measures were progression-free survival and toxicity (grade 3/4 where available). We derived hazard ratios (HRs) for time-to-event outcomes and risk ratios (RRs) for dichotomous outcomes. Secondary outcomes included overall survival, objective tumour response rate, quality of life and treatment-related death. We used GRADEproGDT to assess the certainty of the evidence for the most important outcomes (by first-line and second/third-line therapy for progression-free survival and overall survival).\nMain results\nWe included two RCTs involving 361 women. One study assessed the effects of the mTOR inhibitor temsirolimus, in combination with carboplatin/paclitaxel versus carboplatin/paclitaxel and bevacizumab in treatment-naïve women with advanced or recurrent endometrial cancer. The second study compared the mTOR inhibitor ridaforolimus alone versus progestin or investigator choice of chemotherapy in women who had received prior treatment for metastatic or recurrent endometrial cancer. We identified five ongoing studies on the effects of PI3K and AKT inhibitors, metformin and dual mTOR inhibitors.\nFor first-line therapy, an mTOR inhibitor-containing regimen may worsen progression-free survival (HR 1.43, 95% CI 1.06 to 1.93; 1 study, 231 participants; low-certainty evidence), while for second/third-line therapy, an mTOR inhibitor probably improves progression-free survival compared to chemotherapy or endocrine therapy (HR 0.53, 95% CI 0.31 to 0.91; 1 study, 95 participants; moderate-certainty evidence). Data on toxicity were available from both studies: administering an mTOR inhibitor regimen may increase the risk of grade 3/4 mucositis (RR 10.42, 95% CI 1.34 to 80.74; 2 studies, 357 participants; low-certainty evidence), but may result in little to no difference in risk of anaemia or interstitial pneumonitis (low-certainty evidence for both toxicities). Overall, event rates were low. For first-line therapy, an mTOR inhibitor-containing regimen may result in little to no difference in overall survival compared to chemotherapy (HR 1.32, 95% CI 0.98 to 1.781 study, 231 participants; low-certainty evidence). The finding was similar for second/third-line therapy (HR 1.06, 95% CI 0.70 to 1.61; 1 study, 130 participants; low-certainty evidence). Administering mTOR inhibitor-containing regimens may result in little to no difference in tumour response compared to chemotherapy or hormonal therapy in first-line or second/third-line therapy (first line: RR 0.93, 95% CI 0.75 to 1.17; 1 study, 231 participants; second/third line: RR 0.22, 95% CI 0.01 to 4.40; 1 study, 61 participants; low-certainty evidence).\nNeither study collected or reported quality-of-life data.\nAuthors' conclusions\nTwo RCTs have been reported to date, with low certainty of evidence. In a recurrent disease setting, mTOR inhibitors may result in improved progression-free survival, but we found no clear benefit in overall survival or tumour response rate. We await the publication of at least five ongoing studies investigating the role of PI3K/AKT/mTOR inhibitors in advanced or recurrent endometrial cancer before any conclusions can be drawn on their use.\n\nGiven the provided abstract, please respond to the following query: \"How up-to-date is this review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We searched for studies that had been published up to January 2019." } ]
query-laysum
6742
[ { "role": "user", "content": "Abstract: Background\nThere are many accepted airway clearance techniques (ACTs) for managing the respiratory health of people with cystic fibrosis (CF); none of which demonstrate superiority. Other Cochrane Reviews have reported short-term effects related to mucus transport, but no evidence supporting long-term benefits. Exercise is an alternative ACT thought to produce shearing forces within the lung parenchyma, which enhances mucociliary clearance and the removal of viscous secretions.\nRecent evidence suggests that some people with CF are using exercise as a substitute for traditional ACTs, yet there is no agreed recommendation for this. Additionally, one of the top 10 research questions identified by people with CF is whether exercise can replace other ACTs.\nSystematically reviewing the evidence for exercise as a safe and effective ACT will help people with CF decide whether to incorporate this strategy into their treatment plans and potentially reduce their treatment burden. The timing of this review is especially pertinent given the shifting landscape of CF management with the advent of highly-effective small molecule therapies, which are changing the way people with CF are cared for.\nObjectives\nTo compare the effect of exercise to other ACTs for improving respiratory function and other clinical outcomes in people with CF and to assess the potential adverse effects associated with this ACT.\nSearch methods\nOn 28 February 2022, we searched the Cochrane Cystic Fibrosis Trials Register, compiled from electronic database searches and handsearching of journals and conference abstract books. We also searched the reference lists of relevant articles and reviews.\nWe searched online clinical trial registries on 15 February 2022.\nWe emailed authors of studies awaiting classification or potentially eligible abstracts for additional information on 1 February 2021.\nSelection criteria\nWe selected randomised controlled studies (RCTs) and quasi-RCTs comparing exercise to another ACT in people with CF for at least two treatment sessions.\nData collection and analysis\nTwo review authors independently extracted data and assessed risk of bias for the included studies. They assessed the certainty of the evidence using GRADE. Review authors contacted investigators for further relevant information regarding their publications.\nMain results\nWe included four RCTs. The 86 participants had a wide range of disease severity (forced expiratory volume in one second (FEV1) ranged from 54% to 95%) and were 7 to 41 years old. Two RCTs were cross-over and two were parallel in design. Participants in one RCT were hospitalised with an acute respiratory exacerbation, whilst the participants in three RCTs were clinically stable. All four RCTs compared exercise either alone or in combination with another ACT, but these were too diverse to allow us to combine results. The certainty of the evidence was very low; we downgraded it due to low participant numbers and high or unclear risks of bias across all domains.\nExercise versus active cycle of breathing technique (ACBT)\nOne cross-over trial (18 participants) compared exercise alone to ACBT. There was no change from baseline in our primary outcome FEV1, although it increased in the exercise group before returning to baseline after 30 minutes; we are unsure if exercise affected FEV1 as the evidence is very low-certainty. Similar results were seen for other measures of lung function. No adverse events occurred during the exercise sessions (very low-certainty evidence). We are unsure if ACBT was perceived to be more effective or was the preferred ACT (very low-certainty evidence). 24-hour sputum volume was less in the exercise group than with ACBT (secondary outcome). Exercise capacity, quality of life, adherence, hospitalisations and need for additional antibiotics were not reported.\nExercise plus postural drainage and percussion (PD&P) versus PD&P only\nTwo trials (55 participants) compared exercise and PD&P to PD&P alone. At two weeks, one trial narratively reported a greater increase in FEV1 % predicted with PD&P alone. At six months, the other trial reported a greater increase with exercise combined with PD&P, but did not provide data for the PD&P group. We are uncertain whether exercise with PD&P improves FEV1 as the certainty of evidence is very low. Other measures of lung function did not show clear evidence of effect. One trial reported no difference in exercise capacity (maximal work rate) after two weeks. No adverse events were reported (1 trial, 17 participants; very low-certainty evidence). Adherence was high, with all PD&P sessions and 96% of exercise sessions completed (1 trial, 17 participants; very low-certainty evidence). There was no difference between groups in 24-hour sputum volume or in the mean duration of hospitalisation, although the six-month trial reported fewer hospitalisations due to exacerbations in the exercise and PD&P group. Quality of life, ACT preference and need for antibiotics were not reported.\nExercise versus underwater positive expiratory pressure (uPEP)\nOne trial (13 participants) compared exercise to uPEP (also known as bubble PEP). No adverse events were recorded in either group (very low-certainty evidence). Trial investigators reported that participants perceived exercise as more fatiguing but also more enjoyable than bubble PEP (very low-certainty evidence). There were no differences found in the total weight of sputum collected during treatment sessions. The trial did not report the primary outcomes (FEV1, quality of life, exercise capacity) or the secondary outcomes (other measures of lung function, adherence, need for antibiotics or hospitalisations).\nAuthors' conclusions\nAs one of the top 10 research questions identified by clinicians and people with CF, it is important to systematically review the literature regarding whether or not exercise is an acceptable and effective ACT, and whether it can replace traditional methods. We identified an insufficient number of trials to conclude whether or not exercise is a suitable alternative ACT, and the diverse design of included trials did not allow for meta-analysis of results. The evidence is very low-certainty, so we are uncertain about the effectiveness of exercise as an ACT. Longer studies examining outcomes that are important to people with CF are required to answer this question.\n\nGiven the provided abstract, please respond to the following query: \"Search date\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The evidence is current to 15 February 2022." } ]
query-laysum
6743
[ { "role": "user", "content": "Abstract: Background\nDrug- and alcohol-related impairment in the workplace has been linked to an increased risk of injury for workers. Randomly testing populations of workers for these substances has become a practice in many jurisdictions, with the intention of reducing the risk of workplace incidents and accidents. Despite the proliferation of random drug and alcohol testing (RDAT), there is currently a lack of consensus about whether it is effective at preventing workplace injury, or improving other non-injury accident outcomes in the work place.\nObjectives\nTo assess the effectiveness of workplace RDAT to prevent injuries and improve non-injury accident outcomes (unplanned events that result in damage or loss of property) in workers compared with no workplace RDAT.\nSearch methods\nWe conducted a systematic literature search to identify eligible published and unpublished studies. The date of the last search was 1 November 2020. We searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, Embase, two other databases, Google Scholar, and three trials registers. We also screened the reference lists of relevant publications known to us.\nSelection criteria\nStudy designs that were eligible for inclusion in our review included randomised controlled trials (RCTs), cluster-randomised trials (CRTs), interrupted time-series (ITS) studies, and controlled before-after (CBA) studies. Studies needed to evaluate the effectiveness of RDAT in preventing workplace injury or improving other non-injury workplace outcomes. We also considered unpublished data from clinical trial registries. We included employees working in all safety-sensitive occupations, except for commercial drivers, who are the subject of another Cochrane Review.\nData collection and analysis\nIndependently, two review authors used a data collection form to extract relevant characteristics from the included study. They then analysed a line graph included in the study of the prevalence rate of alcohol violations per year. Independently, the review authors completed a GRADE assessment, as a means of rating the quality of the evidence.\nMain results\nAlthough our searching originally identified 4198 unique hits, only one study was eligible for inclusion in this review. This was an ITS study that measured the effect of random alcohol testing (RAT) on the test positivity rate of employees of major airlines in the USA from 1995 to 2002. The study included data from 511,745 random alcohol tests, and reported no information about testing for other substances. The rate of positive results was the only outcome of interest reported by the study.\nThe average rate of positive results found by RAT increased from 0.07% to 0.11% when the minimum percentage of workers who underwent RAT annually was reduced from 25% to 10%. Our analyses found this change to be a statistically significant increase (estimated change in level, where the level reflects the average percentage points of positive tests = 0.040, 95% confidence interval 0.005 to 0.075; P = 0.031). Our GRADE assessment, for the observed effect of lower minimum testing percentages associating with a higher rate of positive test results, found the quality of the evidence to be 'very low' across the five GRADE domains. The one included study did not address the following outcomes of interest: fatal injuries; non-fatal injuries; non-injury accidents; absenteeism; and adverse effects associated with RDAT.\nAuthors' conclusions\nIn the aviation industry in the USA, the only setting for which the eligible study reported data, there was a statistically significant increase in the rate of positive RAT results following a reduction in the percentage of workers tested, which we deem to be clinically relevant. This result suggests an inverse relationship between the proportion of positive test results and the rate of testing, which is consistent with a deterrent effect for testing. No data were reported on adverse effects related to RDAT.\nWe could not draw definitive conclusions regarding the effectiveness of RDAT for employees in safety-sensitive occupations (not including commercial driving), or with safety-sensitive job functions. We identified only one eligible study that reflected one industry in one country, was of non-randomised design, and tested only for alcohol, not for drugs or other substances. Our GRADE assessment resulted in a 'very low' rating for the quality of the evidence on the only outcome reported. The paucity of eligible research was a major limitation in our review, and additional studies evaluating the effect of RDAT on safety outcomes are needed.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "From 1995 to 1997, random tests for alcohol included 25% of the relevant airlines' workforce each year. During this period, the average percentage of employees who tested positive for alcohol was 0.07%. From 1998 to 2002, the proportion of the workforce tested each year dropped to 10%. During this period, the average percentage of employees who tested positive for alcohol increased to 0.11%.\nThis means that when the airlines randomly tested a larger percentage of employees per year, a smaller proportion of them tested positive for alcohol. This relationship between the frequency of alcohol testing and the proportion of positive tests is what we would expect to see if testing has a deterrent effect, though one study alone cannot prove that there is a deterrent effect.\nThis study did not provide any information for other areas of interest to us, specifically:\n- fatal injuries;\n- non-fatal injuries (in which people are physically injured, but do not die);\n- 'non-injury accidents'; that is, accidents in which people are not injured, but property, processes, materials, and/or the environment are damaged;\n- absenteeism; and\n- unwanted, or adverse, events associated with RDAT, including impacts on privacy, confidentiality and employee perceptions." } ]
query-laysum
6744
[ { "role": "user", "content": "Abstract: Background\nThe World Health Organization (WHO) recommends undertaking 150 minutes of moderate-intensity physical activity per week, but most people do not. Workplaces present opportunities to influence behaviour and encourage physical activity, as well as other aspects of a healthy lifestyle. A pedometer is an inexpensive device that encourages physical activity by providing feedback on daily steps, although pedometers are now being largely replaced by more sophisticated devices such as accelerometers and Smartphone apps. For this reason, this is the final update of this review.\nObjectives\nTo assess the effectiveness of pedometer interventions in the workplace for increasing physical activity and improving long-term health outcomes.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials, MEDLINE, Embase, the Cumulative Index to Nursing and Allied Health Literature (CINAHL), Occupational Safety and Health (OSH) UPDATE, Web of Science, ClinicalTrials.gov, and the WHO International Clinical Trials Registry Platform from the earliest record to December 2016. We also consulted the reference lists of included studies and contacted study authors to identify additional records. We updated this search in May 2019, but these results have not yet been incorporated. One more study, previously identified as an ongoing study, was placed in 'Studies awaiting classification'.\nSelection criteria\nWe included randomised controlled trials (RCTs) of workplace interventions with a pedometer component for employed adults, compared to no or minimal interventions, or to alternative physical activity interventions. We excluded athletes and interventions using accelerometers. The primary outcome was physical activity. Studies were excluded if physical activity was not measured.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. When studies presented more than one physical activity measure, we used a pre-specified list of preferred measures to select one measure and up to three time points for analysis. When possible, follow-up measures were taken after completion of the intervention to identify lasting effects once the intervention had ceased. Given the diversity of measures found, we used ratios of means (RoMs) as standardised effect measures for physical activity.\nMain results\nWe included 14 studies, recruiting a total of 4762 participants. These studies were conducted in various high-income countries and in diverse workplaces (from offices to physical workplaces). Participants included both healthy populations and those at risk of chronic disease (e.g. through inactivity or overweight), with a mean age of 41 years. All studies used multi-component health promotion interventions. Eleven studies used minimal intervention controls, and four used alternative physical activity interventions. Intervention duration ranged from one week to two years, and follow-up after completion of the intervention ranged from three to ten months.\nMost studies and outcomes were rated at overall unclear or high risk of bias, and only one study was rated at low risk of bias. The most frequent concerns were absence of blinding and high rates of attrition.\nWhen pedometer interventions are compared to minimal interventions at follow-up points at least one month after completion of the intervention, pedometers may have no effect on physical activity (6 studies; very low-certainty evidence; no meta-analysis due to very high heterogeneity), but the effect is very uncertain. Pedometers may have effects on sedentary behaviour and on quality of life (mental health component), but these effects were very uncertain (1 study; very low-certainty evidence).\nPedometer interventions may slightly reduce anthropometry (body mass index (BMI) -0.64, 95% confidence interval (CI) -1.45 to 0.18; 3 studies; low-certainty evidence). Pedometer interventions probably had little to no effect on blood pressure (systolic: -0.08 mmHg, 95% CI -3.26 to 3.11; 2 studies; moderate-certainty evidence) and may have reduced adverse effects (such as injuries; from 24 to 10 per 100 people in populations experiencing relatively frequent events; odds ratio (OR) 0.50, 95% CI 0.30 to 0.84; low-certainty evidence). No studies compared biochemical measures or disease risk scores at follow-up after completion of the intervention versus a minimal intervention.\nComparison of pedometer interventions to alternative physical activity interventions at follow-up points at least one month after completion of the intervention revealed that pedometers may have an effect on physical activity, but the effect is very uncertain (1 study; very low-certainty evidence). Sedentary behaviour, anthropometry (BMI or waist circumference), blood pressure (systolic or diastolic), biochemistry (low-density lipoprotein (LDL) cholesterol, total cholesterol, or triglycerides), disease risk scores, quality of life (mental or physical health components), and adverse effects at follow-up after completion of the intervention were not compared to an alternative physical activity intervention.\nSome positive effects were observed immediately at completion of the intervention periods, but these effects were not consistent, and overall certainty of evidence was insufficient to assess the effectiveness of workplace pedometer interventions.\nAuthors' conclusions\nExercise interventions can have positive effects on employee physical activity and health, although current evidence is insufficient to suggest that a pedometer-based intervention would be more effective than other options. It is important to note that over the past decade, technological advancement in accelerometers as commercial products, often freely available in Smartphones, has in many ways rendered the use of pedometers outdated. Future studies aiming to test the impact of either pedometers or accelerometers would likely find any control arm highly contaminated. Decision-makers considering allocating resources to large-scale programmes of this kind should be cautious about the expected benefits of incorporating a pedometer and should note that these effects may not be sustained over the longer term.\nFuture studies should be designed to identify the effective components of multi-component interventions, although pedometers may not be given the highest priority (especially considering the increased availability of accelerometers). Approaches to increase the sustainability of intervention effects and behaviours over a longer term should be considered, as should more consistent measures of physical activity and health outcomes.\n\nGiven the provided abstract, please respond to the following query: \"What are the results of our review?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "" } ]
query-laysum
6745
[ { "role": "user", "content": "Abstract: Background\nAlcohol use disorder (AUD) is one of the most widespread psychiatric disorders leading to detrimental consequences to people with this disorder and others. Worldwide, the prevalence of heavy episodic drinking (30-day prevalence of at least one occasion of 60 g of pure alcohol intake among current drinkers) is estimated at 20% and the prevalence of AUD at 5% of the adult general population, with highest prevalence in Europe and North America. Therapeutic approaches, including pharmacotherapy, play an important role in treating people with AUD.\nThis is an update of a Cochrane Review first published in 2018.\nObjectives\nTo evaluate the benefits and harms of baclofen on achieving and maintaining abstinence or reducing alcohol consumption in people with AUD compared to placebo, no treatment or any other pharmacological relapse prevention treatment.\nSearch methods\nWe used standard, extensive Cochrane search methods. The latest search was 22 November 2021.\nSelection criteria\nRandomised controlled trials (RCTs) of at least four weeks' treatment duration and 12 weeks' overall study duration comparing baclofen for AUD treatment with placebo, no treatment or other treatments.\nData collection and analysis\nWe used standard Cochrane methods. Our primary outcomes were 1. relapse, 2. frequency of use, 3. amount of use, 4. adverse events, 5. dropouts from treatment and 6. dropouts from treatment due to adverse events. Our secondary outcomes were 7. craving, 8. anxiety, 9. depression and 10. frequency of most relevant adverse events.\nMain results\nWe included 17 RCTs (1818 participants) with a diagnosis of alcohol dependence according to the Diagnostic and Statistical Manual of Mental Disorders, 4th edition or International Classification of Diseases 10th edition criteria. Mean age was 46.5 years and 70% were men. Ten studies compared baclofen to placebo or another medication; seven compared two baclofen doses to placebo or another medication. Globally, 15 studies compared baclofen to placebo, two baclofen to acamprosate and two baclofen to naltrexone. In 16 studies, participants received psychosocial treatments.\nWe judged most studies at low risk of selection, performance, detection (subjective outcome), attrition and reporting bias.\nTen studies detoxified participants before treatment; in seven studies, participants were still drinking at the beginning of treatment. Treatment duration was 12 weeks for 15 RCTs and longer in two studies. Baclofen daily dose was 30 mg to 300 mg: 10 RCTs used low doses (30 mg or less); eight RCTs medium doses (above 30 and 100 mg or less) and four RCTs high doses (above 100 mg).\nCompared to placebo, moderate-certainty evidence found that baclofen probably decreases the risk to relapse (risk ratio (RR) 0.87, 95% confidence interval (CI) 0.77 to 0.99; 12 studies, 1057 participants). This result was confirmed among detoxified participants but not among other subgroups of participants.\nHigh-certainty evidence found that baclofen increases the percentage of days abstinent (mean difference (MD) 9.07, 95% CI 3.30 to 14.85; 16 studies, 1273 participants). This result was confirmed among all subgroups of participants except non-detoxified or those who received medium doses.\nThere was no difference between baclofen and placebo in the other primary outcomes: heavy drinking days (standardised mean difference (SMD) −0.18, 95% CI −0.48 to 0.11; 13 studies, 840 participants; moderate-certainty evidence); number of drinks per drinking days (MD −0.45, 95% CI −1.20 to 0.30; 9 studies, 392 participants; moderate-certainty evidence); number of participants with at least one adverse event (RR 1.05, 95% CI 0.99 to 1.11; 10 studies, 738 participants; high-certainty evidence); dropouts (RR 0.88, 95% CI 0.74 to 1.03; 17 studies, 1563 participants; high-certainty evidence); dropouts due to adverse events (RR 1.39, 95% CI 0.89 to 2.18; 16 studies, 1499 participants; high-certainty evidence). These results were confirmed by subgroup analyses except than for the dropouts that resulted lower among participants who received high doses of baclofen and studies longer than 12 weeks.\nCompared to placebo, there was no difference in craving (SMD −0.16, 95% CI −0.37 to 0.04; 17 studies, 1275 participants), anxiety (MD −0.01, 95% CI −0.14 to 0.11; 15 studies, 1123 participants) and depression (SMD 0.07, 95% CI −0.12 to 0.27; 11 studies, 1029 participants).\nConcerning the specific adverse events, baclofen increases fatigue, dizziness, somnolence/sedation, dry mouth, paraesthesia and muscle spasms/rigidity. There was no difference in the other adverse events.\nCompared to acamprosate, one study (60 participants) found no differences in any outcomes but the evidence was very uncertain: relapse (RR 1.25, 95% CI 0.71 to 2.20; very low-certainty evidence); number of participants with at least one adverse event (RR 0.63, 95% CI 0.23 to 1.69; very low-certainty evidence); dropouts (RR 0.56, 95% CI 0.21 to 1.46; very low-certainty evidence); dropouts due to adverse events (RR 0.33, 95% CI 0.01 to 7.87; very low-certainty evidence) and craving (MD 5.80, 95% CI −11.84 to 23.44); and all the adverse events evaluated.\nCompared to naltrexone, baclofen may increase the risk of relapse (RR 2.50, 95% CI 1.12 to 5.56; 1 study, 60 participants; very low-certainty evidence) and decrease the number of participants with at least one adverse event (RR 0.35, 95% CI 0.15 to 0.80; 2 studies, 80 participants; very low-certainty evidence) but the evidence is very uncertain. One study (60 participants) found no difference between baclofen and naltrexone in the dropouts at the end of treatment (RR 1.00, 95% CI 0.32 to 3.10; very low-certainty evidence), craving (MD 2.08, 95% CI −3.71 to 7.87), and all the adverse events evaluated.\nAuthors' conclusions\nBaclofen likely reduces the risk of relapse to any drinking and increases the percentage of abstinent days, mainly among detoxified participants. It does not increase the number of participants with at least one adverse event, those who dropout for any reason or due to adverse events. It probably does not reduce number of heavy drinking days and the number of drinks per drinking days. Current evidence suggests that baclofen may help people with AUD in maintaining abstinence. The results of comparisons of baclofen with acamprosate and naltrexone were mainly based on only one study.\n\nGiven the provided abstract, please respond to the following query: \"Summary of results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found 17 studies with 1818 participants with AUD. The duration of the interventions ranged from three months to one year. Five studies were in the USA; two each in Australia, France, India and Italy; and one each in Germany, Israel and the Netherlands.\nDoses of baclofen were 30 mg a day to 300 mg a day, and, in some cases, the doses were increased during the treatment. Seventeen studies compared baclofen to placebo (dummy medication), two compared baclofen to acamprosate (medication used to treat AUD), and two compared baclofen to naltrexone (medication used to treat AUD).\nCompared with placebo, baclofen probably helps people with AUD at reducing the risk of relapse and increasing the rate of abstinent days. These effects may be more evident among detoxified than non-detoxified people with AUD. Baclofen probably makes no difference in the rate of heavy drinking days; drinks per drinking days; craving, anxiety and depression severity; number of participants who dropout from treatment for any reason; those who dropped out due to adverse events (side effects); or the number of participants with at least one adverse event. Baclofen may increase adverse events such as fatigue, vertigo (a feeling that everything is spinning), sleepiness, dry mouth, numbness and muscle spasm but we found no differences between baclofen and placebo for other adverse events." } ]
query-laysum
6746
[ { "role": "user", "content": "Abstract: Background\nInfants born preterm (before 37 weeks' gestation) have poorer outcomes than infants at term, particularly if born before 32 weeks. Early cord clamping has been standard practice over many years, and enables quick transfer of the infant to neonatal care. Delayed clamping allows blood flow between the placenta, umbilical cord and baby to continue, and may aid transition. Keeping baby at the mother's side enables neonatal care with the cord intact and this, along with delayed clamping, may improve outcomes. Umbilical cord milking (UCM) is proposed for increasing placental transfusion when immediate care for the preterm baby is needed. This Cochrane Review is a further update of a review first published in 2004 and updated in 2012.\nObjectives\nTo assess the effects on infants born at less than 37 weeks' gestation, and their mothers of: 1) delayed cord clamping (DCC) compared with early cord clamping (ECC) both with immediate neonatal care after cord clamping; 2) DCC with immediate neonatal care with cord intact compared with ECC with immediate neonatal care after cord clamping; 3) DCC with immediate neonatal care after cord clamping compared with UCM; 4) UCM compared with ECC with immediate neonatal care after cord clamping.\nSearch methods\nWe searched the Cochrane Pregnancy and Childbirth Group Trials Register, ClinicalTrials.gov, the WHO International Clinical Trials Registry Platform (ICTRP) (10 November 2017), and reference lists of retrieved studies. We updated the search in November 2018 and added nine new trial reports to the awaiting classification section to be assessed at the next update.\nSelection criteria\nRandomised controlled trials (RCTs) comparing delayed with early clamping of the umbilical cord (with immediate neonatal care after cord clamping or with cord intact) and UCM for births before 37 weeks' gestation. Quasi-RCTs were excluded.\nData collection and analysis\nTwo review authors independently assessed trials for inclusion and risk of bias, extracted data and checked them for accuracy. Random-effects are used in all meta-analyses. Review authors assessed the certainty of the evidence using the GRADE approach.\nMain results\nThis update includes forty-eight studies, involving 5721 babies and their mothers, with data available from 40 studies involving 4884 babies and their mothers. Babies were between 24 and 36+6 weeks' gestation at birth and multiple births were included. The data are mostly from high-income countries. Delayed clamping ranged between 30 to 180 seconds, with most studies delaying for 30 to 60 seconds. Early clamping was less than 30 seconds and often immediate. UCM was mostly before cord clamping but some were milked after cord clamping. We undertook subgroup analysis by gestation and type of intervention, and sensitivity analyses by low risk of selection and attrition bias.\nAll studies were high risk for performance bias and many were unclear for other aspects of risk of bias. Certainty of the evidence using GRADE was mostly low, mainly due to imprecision and unclear risk of bias.\nDelayed cord clamping (DCC) versus early cord clamping (ECC) both with immediate neonatal care after cord clamping (25 studies, 3100 babies and their mothers)\nDCC probably reduces the number of babies who die before discharge compared with ECC (average risk ratio (aRR) 0.73, 95% confidence interval (CI) 0.54 to 0.98, 20 studies, 2680 babies (moderate certainty)).\nNo studies reported on 'Death or neurodevelopmental impairment' in the early years'.\nDCC may make little or no difference to the number of babies with severe intraventricular haemorrhage (IVH grades 3 and 4) (aRR 0.94, 95% CI 0.63 to 1.39, 10 studies, 2058 babies, low certainty) but slightly reduces the number of babies with any grade IVH (aRR 0.83, 95% CI 0.70 to 0.99, 15 studies, 2333 babies, high certainty).\nDCC has little or no effect on chronic lung disease (CLD) (aRR 1.04, 95% CI 0.94 to 1.14, 6 studies, 1644 babies, high certainty).\nDue to insufficient data, we were unable to form conclusions regarding periventricular leukomalacia (PVL) (aRR 0.58, 95% CI 0.26 to 1.30, 4 studies, 1544 babies, low certainty) or maternal blood loss of 500 mL or greater (aRR 1.14, 95% CI 0.07 to 17.63, 2 studies, 180 women, very low certainty).\nWe identified no important heterogeneity in subgroup or sensitivity analyses.\nDelayed cord clamping (DCC) with immediate neonatal care with cord intact versus early cord clamping (ECC) (one study, 276 babies and their mothers)\nThere are insufficient data to be confident in our findings, but DCC with immediate neonatal care with cord intact may reduce the number of babies who die before discharge, although the data are also compatible with a slight increase in mortality, compared with ECC (aRR 0.47, 95% CI 0.20 to 1.11, 1 study, 270 babies, low certainty). DCC may also reduce the number of babies who die or have neurodevelopmental impairment in early years (aRR 0.61, 95% CI 0.39 to 0.96, 1 study, 218 babies, low certainty). There may be little or no difference in: severe IVH; all grades IVH; PVL; CLD; maternal blood loss ≥ 500 mL, assessed as low certainty mainly due to serious imprecision.\nDelayed cord clamping (DCC) with immediate neonatal care after cord clamping versus umbilical cord milking (UCM) (three studies, 322 babies and their mothers) and UCM versus early cord clamping (ECC) (11 studies, 1183 babies and their mothers)\nThere are insufficient data for reliable conclusions about the comparative effects of UCM compared with delayed or early clamping (mostly low or very low certainty).\nAuthors' conclusions\nDelayed, rather than early, cord clamping may reduce the risk of death before discharge for babies born preterm. There is insufficient evidence to show what duration of delay is best, one or several minutes, and therefore the optimum time to clamp the umbilical cord remains unclear. Whilst the current evidence supports not clamping the cord before 30 seconds at preterm births, future trials could compare different lengths of delay. Immediate neonatal care with the cord intact requires further study, and there are insufficient data on UCM.\nThe nine new reports awaiting further classification may alter the conclusions of the review once assessed.\n\nGiven the provided abstract, please respond to the following query: \"Why is it important?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Babies born before 37 weeks, or preterm, have poorer health outcomes than babies born at term, particularly if they are born before 32 weeks. Babies born preterm can experience problems with the functioning of many of their major organs including their lungs, gut and hearts. They have a greater risk of dying or having long-term problems such as cerebral palsy. After birth, the babies may need blood transfusions and drugs to strengthen their heart contractions (inotropes) and to raise their blood pressure. It is important to try to find ways of improving the health of these tiny babies.\nEarly clamping of the umbilical cord has been standard practice over many years. It allows the baby to be transferred quickly to care from a specialised team of doctors either at the side of the room or in another room. Yet, delayed clamping for half to three or more minutes allows continuing blood flow between the mother and her baby, and this may help the baby to adjust to breathing air. Squeezing blood along the umbilical cord towards the baby (milking the cord), can boost the baby's blood volume, and this may improve the baby's health. We wanted to see if there are any benefits or harms from either waiting to clamp or milking the cord." } ]
query-laysum
6747
[ { "role": "user", "content": "Abstract: Background\nTreatment guidelines for asthma recommend inhaled corticosteroids (ICS) as first-line therapy for children with persistent asthma. Although ICS treatment is generally considered safe in children, the potential systemic adverse effects related to regular use of these drugs have been and continue to be a matter of concern, especially the effects on linear growth.\nObjectives\nTo assess the impact of ICS on the linear growth of children with persistent asthma and to explore potential effect modifiers such as characteristics of available treatments (molecule, dose, length of exposure, inhalation device) and of treated children (age, disease severity, compliance with treatment).\nSearch methods\nWe searched the Cochrane Airways Group Specialised Register of trials (CAGR), which is derived from systematic searches of bibliographic databases including CENTRAL, MEDLINE, EMBASE, CINAHL, AMED and PsycINFO; we handsearched respiratory journals and meeting abstracts. We also conducted a search of ClinicalTrials.gov and manufacturers' clinical trial databases to look for potential relevant unpublished studies. The literature search was conducted in January 2014.\nSelection criteria\nParallel-group randomised controlled trials comparing daily use of ICS, delivered by any type of inhalation device for at least three months, versus placebo or non-steroidal drugs in children up to 18 years of age with persistent asthma.\nData collection and analysis\nTwo review authors independently performed study selection, data extraction and assessment of risk of bias in included studies. We conducted meta-analyses using the Cochrane statistical package RevMan 5.2 and Stata version 11.0. We used the random-effects model for meta-analyses. We used mean differences (MDs) and 95% CIs as the metrics for treatment effects. A negative value for MD indicates that ICS have suppressive effects on linear growth compared with controls. We performed a priori planned subgroup analyses to explore potential effect modifiers, such as ICS molecule, daily dose, inhalation device and age of the treated child.\nMain results\nWe included 25 trials involving 8471 (5128 ICS-treated and 3343 control) children with mild to moderate persistent asthma. Six molecules (beclomethasone dipropionate, budesonide, ciclesonide, flunisolide, fluticasone propionate and mometasone furoate) given at low or medium daily doses were used during a period of three months to four to six years. Most trials were blinded and over half of the trials had drop out rates of over 20%.\nCompared with placebo or non-steroidal drugs, ICS produced a statistically significant reduction in linear growth velocity (14 trials with 5717 participants, MD -0.48 cm/y, 95% CI -0.65 to -0.30, moderate quality evidence) and in the change from baseline in height (15 trials with 3275 participants; MD -0.61 cm/y, 95% CI -0.83 to -0.38, moderate quality evidence) during a one-year treatment period.\nSubgroup analysis showed a statistically significant group difference between six molecules in the mean reduction of linear growth velocity during one-year treatment (Chi² = 26.1, degrees of freedom (df) = 5, P value < 0.0001). The group difference persisted even when analysis was restricted to the trials using doses equivalent to 200 μg/d hydrofluoroalkane (HFA)-beclomethasone. Subgroup analyses did not show a statistically significant impact of daily dose (low vs medium), inhalation device or participant age on the magnitude of ICS-induced suppression of linear growth velocity during a one-year treatment period. However, head-to-head comparisons are needed to assess the effects of different drug molecules, dose, inhalation device or patient age. No statistically significant difference in linear growth velocity was found between participants treated with ICS and controls during the second year of treatment (five trials with 3174 participants; MD -0.19 cm/y, 95% CI -0.48 to 0.11, P value 0.22). Of two trials that reported linear growth velocity in the third year of treatment, one trial involving 667 participants showed similar growth velocity between the budesonide and placebo groups (5.34 cm/y vs 5.34 cm/y), and another trial involving 1974 participants showed lower growth velocity in the budesonide group compared with the placebo group (MD -0.33 cm/y, 95% CI -0.52 to -0.14, P value 0.0005). Among four trials reporting data on linear growth after treatment cessation, three did not describe statistically significant catch-up growth in the ICS group two to four months after treatment cessation. One trial showed accelerated linear growth velocity in the fluticasone group at 12 months after treatment cessation, but there remained a statistically significant difference of 0.7 cm in height between the fluticasone and placebo groups at the end of the three-year trial.\nOne trial with follow-up into adulthood showed that participants of prepubertal age treated with budesonide 400 μg/d for a mean duration of 4.3 years had a mean reduction of 1.20 cm (95% CI -1.90 to -0.50) in adult height compared with those treated with placebo.\nAuthors' conclusions\nRegular use of ICS at low or medium daily doses is associated with a mean reduction of 0.48 cm/y in linear growth velocity and a 0.61-cm change from baseline in height during a one-year treatment period in children with mild to moderate persistent asthma. The effect size of ICS on linear growth velocity appears to be associated more strongly with the ICS molecule than with the device or dose (low to medium dose range). ICS-induced growth suppression seems to be maximal during the first year of therapy and less pronounced in subsequent years of treatment. However, additional studies are needed to better characterise the molecule dependency of growth suppression, particularly with newer molecules (mometasone, ciclesonide), to specify the respective role of molecule, daily dose, inhalation device and patient age on the effect size of ICS, and to define the growth suppression effect of ICS treatment over a period of several years in children with persistent asthma.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We included in this review trials comparing daily use of corticosteroids, delivered by any type of inhalation device for at least three months, versus placebo or non-steroidal drugs in children up to 18 years of age with persistent asthma." } ]
query-laysum
6748
[ { "role": "user", "content": "Abstract: Background\nAlmost 358,000 women die each year in childbirth, mainly in low-income countries. More than half of all maternal deaths occur within 24 hours of giving birth; severe bleeding in the postpartum period is the single most important cause. Depending on the rate of blood loss and other factors, such as pre-existing anaemia, untreated postpartum haemorrhage (PPH) can lead to hypovolaemic shock, multi-organ dysfunction, and maternal death, within two to six hours.\nThis review investigated different methods for estimating blood loss. The most common method of measuring blood loss during the third stage of labour is visual estimation, during which the birth attendant makes a quantitative or semi-quantitative estimate of the amount of blood lost. In direct blood collection, all blood lost during the third stage of labour (except for the placenta and membranes) is contained in a disposable, funnelled, plastic collector bag, which is attached to a plastic sheet, and placed under the woman's buttocks. When the bleeding stops, there are two options: the bag can be weighed (also called gravimetric technique), or the bag can be calibrated, allowing for a direct measurement. A more precise measurement of blood loss is haemoglobin concentration (Hb) in venous blood sampling and spectrophotometry. With the dye dilution technique, a known quantity of dye is injected into the vein and its plasmatic concentration is monitored after the uterus stops bleeding. Using nuclear medicine, a radioactive tracer is injected, and its concentration is monitored after the uterus stops bleeding. Although hypothetically, these advanced methods could provide a better quantification of blood loss, they are difficult to perform and are not accessible in most settings.\nObjectives\nTo evaluate the effect of alternative methods to estimate blood loss during the third stage of labour, to help healthcare providers reduce the adverse consequences of postpartum haemorrhage after vaginal birth.\nSearch methods\nWe searched Cochrane Pregnancy and Childbirth's Trials Register (2 February 2018), ClinicalTrials.gov and the WHO International Clinical Trials Registry Platform (ICTRP; 21 March 2018), and reference lists of retrieved studies.\nSelection criteria\nAll randomised trials, including cluster-randomised trials, evaluating methods for estimating blood loss after vaginal birth.\nData collection and analysis\nTwo review authors independently assessed trials for inclusion and risk of bias, extracted data, and checked them for accuracy.\nMain results\nThe search retrieved 62 reports in total. Of these, we assessed 12 reports in full, corresponding to six trials. We included three trials and excluded one; two trials are ongoing.\nThe included trials were conducted in hospital settings. Two trials were conducted in India; the third trial was a large cluster-randomised trial, which took place in 13 European countries. Overall, we judged the included trials to be at a low risk of bias. One study evaluated the use of calibrated drapes versus visual estimation, another evaluated the use of calibrated drapes versus the gravimetric technique (weight of blood-soaked materials), therefore, we were unable to pool the data from the two studies. The third study did not measure any of the outcomes of interest, so did not contribute data to the analyses.\nDirect measurement using calibrated drapes versus visual estimation\nOne cluster-randomised controlled trial in 13 western European countries, with over 25,000 women, examined this comparison.\nThe trial did not report on postpartum anaemia (defined as Hb lower than 9 mg/dL), blood loss greater than 500 mL, or maternal infection.\nModerate-quality evidence suggests there is probably little or no difference between groups in: severe morbidity (coagulopathy, organ failure, intensive care unit admission; adjusted risk ratio (RR) 0.82, 95% confidence interval (CI) 0.48 to 1.39); the risk of blood transfusion (adjusted RR 0.82, 95% CI 0.46 to 1.46); the use of plasma expanders (adjusted RR 0.77, 95% CI 0.42 to 1.42); and the use of therapeutic uterotonics (adjusted RR 0.87, 95% CI 0.42 to 1.76).\nDirect measurement using calibrated drapes (Excellent BRASSS-V Drape™) versus gravimetric technique\nOne randomised controlled trial in India, with 900 women, examined this comparison.\nThe trial did not report on postpartum anaemia (defined as Hb lower than 9 mg/dL), severe morbidity, or maternal infection.\nHigh-quality evidence showed that using calibrated drapes improved the detection of blood loss greater than 500 mL when compared with the gravimetric technique (RR 1.86, 95% CI 1.11 to 3.11). Low-quality evidence suggests there may be little or no difference in the risk of blood transfusion between the two groups (RR 1.00, 95% CI 0.06 to 15.94), or in the use of plasma expanders, reported as intravenous fluids given for PPH treatment (RR 0.67; 95% CI 0.19 to 2.35). High-quality evidence showed little or no difference in the use of therapeutic uterotonics (RR 1.01, 95% CI 0.90 to 1.13), but the use of therapeutic uterotonics was extremely high in both arms of the study (57% and 56%).\nAuthors' conclusions\nOverall, the evidence in this review is insufficient to support the use of one method over another for blood loss estimation after vaginal birth. In general, the quality of evidence for our predefined outcomes ranged from low to high quality, with downgrading decisions due to imprecision. The included trials did not report on many of our primary and secondary outcomes.\nIn trials that evaluate methods for estimating blood loss during vaginal birth, we believe it is important to measure their impact on clinical maternal and neonatal outcomes, along with their diagnostic accuracy. This body of knowledge needs further, well designed, appropriately powered, randomised controlled trials that correlate blood loss with relevant clinical outcomes, such as those listed in this review.\n\nGiven the provided abstract, please respond to the following query: \"What does this mean?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "There was insufficient evidence to support the use of one method over another to estimate blood loss after vaginal birth. There is a need for high quality trials that measure important outcomes, such as those listed in this review." } ]
query-laysum
6749
[ { "role": "user", "content": "Abstract: Background\nThe efficacy of decongestants, antihistamines and nasal irrigation in children with clinically diagnosed acute sinusitis has not been systematically evaluated.\nObjectives\nTo determine the efficacy of decongestants, antihistamines or nasal irrigation in improving symptoms of acute sinusitis in children.\nSearch methods\nWe searched CENTRAL (2014, Issue 5), MEDLINE (1950 to June week 1, 2014) and EMBASE (1950 to June 2014).\nSelection criteria\nWe included randomized controlled trials (RCTs) and quasi-RCTs, which evaluated children younger than 18 years of age with acute sinusitis, defined as 10 to 30 days of rhinorrhea, congestion or daytime cough. We excluded trials of children with chronic sinusitis and allergic rhinitis.\nData collection and analysis\nTwo review authors independently assessed each study for inclusion.\nMain results\nOf the 662 studies identified through the electronic searches and handsearching, none met all the inclusion criteria.\nAuthors' conclusions\nThere is no evidence to determine whether the use of antihistamines, decongestants or nasal irrigation is efficacious in children with acute sinusitis. Further research is needed to determine whether these interventions are beneficial in the treatment of children with acute sinusitis.\n\nGiven the provided abstract, please respond to the following query: \"Use of statistics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Not applicable." } ]
query-laysum
6750
[ { "role": "user", "content": "Abstract: Background\nRadical cystectomy (RC) is the primary surgical treatment for muscle-invasive urothelial carcinoma of the bladder. This major operation is typically associated with an extended hospital stay, a prolonged recovery period and potentially major complications. Nutritional interventions are beneficial in some people with other types of cancer and may be of value in this setting too.\nObjectives\nTo assess the effects of perioperative nutrition in people undergoing radical cystectomy for the treatment of bladder cancer.\nSearch methods\nWe performed a comprehensive search using multiple databases (Evidence Based Medicine Reviews, MEDLINE, Embase, AMED, CINAHL), trials registries, other sources of grey literature, and conference proceedings published up to 22 February 2019, with no restrictions on the language or status of publication.\nSelection criteria\nWe included parallel-group randomised controlled trials (RCTs) of adults undergoing RC for bladder cancer. The intervention was any perioperative nutrition support.\nData collection and analysis\nTwo review authors independently assessed studies for inclusion, extracted data, and assessed risk of bias and the quality of evidence using GRADE. Primary outcomes were postoperative complications at 90 days and length of hospital stay. The secondary outcome was mortality up to 90 days after surgery. When 90-day outcome data were not available, we reported 30-day data.\nMain results\nThe search identified eight trials including 500 participants. Six trials were conducted in the USA and two in Europe.\n1. Parenteral nutrition (PN) versus oral nutrition: based on one study with 157 participants, PN may increase postoperative complications within 30 days (risk ratio (RR) 1.40, 95% confidence interval (CI) 1.07 to 1.82; low-quality evidence). We downgraded the quality of evidence for serious study limitations (unclear risk of selection, performance and selective reporting bias) and serious imprecision. This corresponds to 198 more complications per 1000 participants (95% CI 35 more to 405 more). Length of hospital stay may be similar (mean difference (MD) 0.5 days higher, CI not reported; low-quality evidence).\n2. Immuno-enhancing nutrition versus standard nutrition: based on one study including 29 participants, immuno-enhancing nutrition may reduce 90-day postoperative complications (RR 0.31, 95% CI 0.08 to 1.23; low-quality evidence). These findings correspond to 322 fewer complications per 1000 participants (95% CI 429 fewer to 107 more). Length of hospital stay may be similar (MD 0.20 days, 95% CI 1.69 lower to 2.09 higher; low-quality evidence). We downgraded the quality of evidence of both outcomes for very serious imprecision.\n3. Preoperative oral nutritional support versus normal diet: based on one study including 28 participants, we are very uncertain if preoperative oral supplements reduces postoperative complications. We downgraded quality for serious study limitations (unclear risk of selection, performance, attrition and selective reporting bias) and very serious imprecision. The study did not report on length of hospital stay.\n4. Early postoperative feeding versus standard postoperative management: based on one study with 102 participants, early postoperative feeding may increase postoperative complications (very low-quality evidence) but we are very uncertain of this finding. We downgraded the quality of evidence for serious study limitations (unclear risk of selection and performance bias) and very serious imprecision. Length of hospital stay may be similar (MD 0.95 days less, CI not reported; low-quality evidence). We downgraded the quality of evidence for serious study limitations (unclear risk of selection and performance bias) and serious imprecision.\n5. Amino acid with dextrose versus dextrose: based on two studies with 104 participants, we are very uncertain whether amino acids reduce postoperative complications (very low-quality evidence). We are also very uncertain whether length of hospital stay is similar (very low-quality evidence). We downgraded the quality of evidence for both outcomes for serious study limitations (unclear and high risk of selection bias; unclear risk of performance, detection and selective reporting bias), serious indirectness related to the patient population and very serious imprecision.\n6. Branch chain amino acids versus dextrose only: based on one study including 19 participants, we are very uncertain whether complication rates are similar (very low-quality evidence). We downgraded the quality of evidence for serious study limitations (unclear risk of selection, performance, detection, attrition and selective reporting bias), serious indirectness related to the patient population and very serious imprecision. The study did not report on length of hospital stay.\n7. Perioperative oral nutritional supplements versus oral multivitamin and mineral supplement: based on one study with 61 participants, oral supplements compared to a multivitamin and mineral supplement may slightly decrease postoperative complications (low-quality evidence). These findings correspond to 135 fewer occurrences per 1000 participants (95% CI 256 fewer to 65 more). Length of hospital stay may be similar (low-quality evidence). We downgraded the quality of evidence of both outcomes for study limitations and imprecision.\nAuthors' conclusions\nBased on few, small and dated studies, with serious methodological limitations, we found limited evidence for a benefit of perioperative nutrition interventions. We rated the quality of evidence as low or very low, which underscores the urgent need for high-quality research studies to better inform nutritional support interventions for people undergoing surgery for bladder cancer.\n\nGiven the provided abstract, please respond to the following query: \"1. Feeding into a vein versus oral nutrition\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "based on one study that included 157 people, we found that feeding into a vein may increase complications after surgery. However, there may be little or no difference in length of hospital stay." } ]
query-laysum
6751
[ { "role": "user", "content": "Abstract: Background\nHereditary haemochromatosis is a genetic disorder related to proteins involved in iron transport, resulting in iron load and deposition of iron in various tissues of the body. This iron overload leads to complications including liver cirrhosis (and related complications such as liver failure and hepatocellular carcinoma), cardiac failure, cardiac arrhythmias, impotence, diabetes, arthritis, and skin pigmentation. Phlebotomy (venesection or 'blood letting') is the currently recommended treatment for hereditary haemochromatosis. The optimal treatment of hereditary haemochromatosis remains controversial.\nObjectives\nTo assess the comparative benefits and harms of different interventions in the treatment of hereditary haemochromatosis through a network meta-analysis and to generate rankings of the available treatments according to their safety and efficacy. However, we found only one comparison. Therefore, we did not perform the network meta-analysis and we assessed the comparative benefits and harms of different interventions using standard Cochrane methodology.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, Embase, Science Citation Index Expanded, World Health Organization International Clinical Trials Registry Platform, and randomised clinical trials registers to March 2016 to identify randomised clinical trials on treatments for hereditary haemochromatosis.\nSelection criteria\nWe included only randomised clinical trials (irrespective of language, blinding, or publication status) in participants with hereditary haemochromatosis. We excluded trials which included participants who had previously undergone liver transplantation. We considered any of the various interventions compared with each other or with inactive treatment.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. We calculated the odds ratio (OR) and rate ratio with 95% confidence intervals (CI) using both fixed-effect and random-effects models with RevMan 5 based on available-participant analysis. We assessed risk of bias according to Cochrane, controlled risk of random errors with Trial Sequential Analysis, and assessed the quality of the evidence using GRADE.\nMain results\nThree trials with 146 participants met the inclusion criteria of this review. Two parallel group trials with 100 participants provided information on one or more outcomes. The remaining trial was a cross-over trial, with no usable data for analysis. All the trials were at high risk of bias. Overall, all the evidence was of very low quality. All three trials compared erythrocytapheresis (removal of red cells only, instead of whole blood) versus phlebotomy. Two of the trials shared the same first author. The mean or median age in the three trials ranged from 42 to 55 years. None of the trials reported whether the included participants were symptomatic or asymptomatic or a mixture of both. Two trials were conducted in people who were haemochromatosis treatment-naive. The trial that provided most data for this review excluded people with malignancy, heart failure, and serious cardiac arrhythmias. We found no trials assessing iron-chelating agents.\nOnly one of the trials with 38 participants reported no short-term mortality and no serious adverse events at the end of the short-term follow-up (eight months). Two trials reported the proportion of people with adverse events: 10/49 (20.4%) in the erythrocytapheresis group versus 11/51 (21.6%) in the phlebotomy group. One of these two trials provided data on adverse event rates (42.1 events per 100 participants with erythrocytapheresis versus 52.6 events per 100 participants with phlebotomy). There was no evidence of differences in the proportion of people with adverse events and the number of adverse events (serious and non-serious) between the groups (proportion of people with adverse events: OR 0.93, 95% CI 0.36 to 2.43; participants = 100; trials = 2; number of adverse events: rate ratio 0.80, 95% CI 0.32 to 2.03; participants = 38; trial = 1). There was no difference between the groups regarding short-term health-related quality of life (mean difference (MD) 1.00, 95% CI -10.80 to 12.80; participants = 38; trials = 1). This outcome was measured using EQ-VAS (range: 0 to 100 where a higher score indicates better health-related quality of life). None of the trials reported mortality beyond one year, health-related quality of life beyond one year, liver transplantation, decompensated liver disease, cirrhosis, hepatocellular carcinoma, diabetes, or cardiovascular complications during the long-term follow-up.\nThe two trials that provided data for this review were funded by parties with no vested interest in the results; the source of funding of the third trial was not reported.\nAuthors' conclusions\nThere is currently insufficient evidence to determine whether erythrocytapheresis is beneficial or harmful compared with phlebotomy. Phlebotomy has less equipment requirements and remains the treatment of choice in people with hereditary haemochromatosis who require blood letting in some form. However, it should be noted that there is no evidence from randomised clinical trials that blood letting in any form is beneficial in people with hereditary haemochromatosis. Having said this, a trial including no treatment is unlikely to be conducted. Future trials should compare different frequencies of phlebotomy and erythrocytapheresis versus phlebotomy with and without different iron-chelating agents compared with each other, and with placebo. Such trials should include long-term follow-up of participants (e.g. using national record linkage databases) to determine whether treatments are beneficial or harmful in terms of clinical outcomes such as deaths, health-related quality of life, liver damage and its consequences, heart damage and its consequences, and other outcomes that are of importance to people with hereditary haemochromatosis.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Hereditary haemochromatosis is an inherited genetic disorder (derived from one's parents) resulting in excessive iron accumulation in the body. Some people develop liver damage leading to liver failure, heart damage leading to heart failure, impotence (inability for a man to have an erection or orgasm), diabetes, arthritis (joint pain and swelling), and skin pigmentation (colouring) because of excessive iron accumulation. Several treatments are used to treat hereditary haemochromatosis but the best way is not clear. We searched for randomised clinical trials (well-design clinical studies where people are randomly put into one of two or more treatment groups) reported to March 2016. We included trials in which participants had not had a liver transplant. Apart from using standard Cochrane methods which allow comparison of only two treatments at a time (direct comparison), we planned to use an advanced method which allows comparison of the many different treatments which are individually compared in the trials (network meta-analysis). However, because there was only one comparison, we used standard Cochrane methodology." } ]
query-laysum
6752
[ { "role": "user", "content": "Abstract: Background\nAttention deficit hyperactivity disorder (ADHD) is a prevalent neurobiological condition, characterised by behavioral and cognitive symptoms such as inattention, impulsivity and/or excessive activity. The syndrome is commonly accompanied by psychiatric comorbidities and is associated with educational and occupational underachievement.\nAlthough psychostimulant medications are the mainstay of treatment for ADHD, not all adults respond optimally to, or can tolerate, these medicines. Thus, alternative non-stimulant treatment approaches for ADHD have been explored. One of these alternatives is bupropion, an aminoketone antidepressant and non-competitive antagonism of nicotinic acetylcholine receptors. Bupropion is registered for the treatment of depression and smoking cessation, but is also used off-label to treat ADHD.\nObjectives\nTo assess the effects and safety of bupropion for the treatment of adults with ADHD.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE, Embase, and seven other databases in February 2017. We also searched three trials registers and three online theses portals. In addition, we checked references of included studies and contacted study authors to identify potentially relevant studies that were missed by our search.\nSelection criteria\nWe included all randomised controlled trials (RCTs) that evaluated the effects (including adverse effects) of bupropion compared to placebo in adults with ADHD.\nData collection and analysis\nTwo review authors (WV, GB) independently screened records and extracted data using a data extraction sheet that we tested in a pilot study. We extracted all relevant data on study characteristics and results. We assessed risks of bias using the Cochrane 'Risk of bias' tool, and assessed the overall quality of evidence using the GRADE approach. We used a fixed-effect model to pool the results across studies.\nMain results\nWe included six studies with a total of 438 participants. Five studies were conducted in the USA, and one in Iran. All studies evaluated a long-acting version of bupropion, with the dosage ranging from 150 mg up to 450 mg daily. Study intervention length varied from six to 10 weeks. Four studies explicitly excluded participants with psychiatric comorbidity and one study included only participants with opioid dependency. Four studies were funded by industry, but the impact of this on study results is unknown. Two studies were publicly funded and in one of these studies, the lead author was a consultant for several pharmaceutical companies and also received investigator-driven funding from two companies, however none of these companies manufacture bupropion. We judged none of the studies to be free of bias because for most risk of bias domains the study reports failed to provide sufficient details. Using the GRADE approach, we rated the overall quality of evidence as low. We downgraded the quality of the evidence because of serious risk of bias and serious imprecision due to small sample sizes.\nWe found low-quality evidence that bupropion decreased the severity of ADHD symptoms (standardised mean difference -0.50, 95% confidence interval (CI) -0.86 to -0.15, 3 studies, 129 participants), and increased the proportion of participants achieving clinical improvement (risk ratio (RR) 1.50, 95% CI 1.13 to 1.99, 4 studies, 315 participants), and reporting an improvement on the Clinical Global Impression - Improvement scale (RR 1.78, 95% CI 1.27 to 2.50, 5 studies, 337 participants). There was low-quality evidence that the proportion of participants who withdrew due to any adverse effect was similar in the bupropion and placebo groups (RR 1.20, 95% CI 0.35 to 4.10, 3 studies, 253 participants). The results were very similar when using a random-effects model and when we analysed only studies that excluded participants with a psychiatric comorbidity.\nAuthors' conclusions\nThe findings of this review, which compared bupropion to placebo for adult ADHD, indicate a possible benefit of bupropion. We found low-quality evidence that bupropion decreased the severity of ADHD symptoms and moderately increased the proportion of participants achieving a significant clinical improvement in ADHD symptoms. Furthermore, we found low-quality evidence that the tolerability of bupropion is similar to that of placebo. In the pharmacological treatment of adults with ADHD, extended- or sustained-release bupropion may be an alternative to stimulants. The low-quality evidence indicates uncertainty with respect to the pooled effect estimates. Further research is very likely to change these estimates. More research is needed to reach more definite conclusions as well as clarifying the optimal target population for this medicine. Treatment response remains to be reported in a DSM5-diagnosed population. There is also a lack of knowledge on long-term outcomes.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We included six randomised controlled trials (RCTs), that is, studies in which participants are randomly allocated to one of two or more treatment groups. Five of the studies took place in the USA, and the sixth in Iran. The studies included 438 people with ADHD. All evaluated a long-acting version of bupropion, that is to say, a version of the drug is absorbed slowly, and can therefore be taken just once a day. This simple dosage suits people with ADHD, as the illness may make it difficult for them to remember to take their medication.\nThe duration of the studies varied between six and 10 weeks. All participants were diagnosed with ADHD, and often had other mental health problems. In one study, all participants had ADHD and were addicted to opioids (a drug that relieves pain)." } ]
query-laysum
6753
[ { "role": "user", "content": "Abstract: Background\nSteroids have been used widely since the early 1970s for the treatment of adult-onset minimal change disease (MCD). Recently, newer agents have been used in adult MCD aiming to reduce the risk of adverse effects. The response rates to immunosuppressive agents in adult MCD are more variable than in children. The optimal agent, dose, and duration of treatment for the first episode of nephrotic syndrome, or for disease relapse(s) have not been determined. This is an update of a review first published in 2008.\nObjectives\nWe aimed to 1) evaluate the benefits and harms of different agents, including both immunosuppressive and non-immunosuppressive agents, in adults with MCD causing the nephrotic syndrome; and 2) evaluate the efficacy of interventions on 'time-to-remission' of nephrotic syndrome, in adults with MCD causing the nephrotic syndrome.\nSearch methods\nWe searched the Cochrane Kidney and Transplant Register of Studies up to 21 July 2021 through contact with the Information Specialist using search terms relevant to this review. Studies in the Register are identified through searches of CENTRAL, MEDLINE, and EMBASE, conference proceedings, the International Clinical Trials Register (ICTRP) Search Portal and ClinicalTrials.gov.\nSelection criteria\nRandomised controlled trials (RCTs) and quasi-RCTs of any intervention for MCD with nephrotic syndrome in adults over 18 years were included. Studies comparing different types, routes, frequencies, and duration of immunosuppressive agents and non-immunosuppressive agents were assessed.\nData collection and analysis\nTwo authors independently assessed study quality and extracted data. Statistical analyses were performed using the random-effects model and results were expressed as a risk ratio (RR) for dichotomous outcomes, or mean difference (MD) for continuous data with 95% confidence intervals (CI). Confidence in the evidence was assessed using the Grading of Recommendations Assessment, Development and Evaluation (GRADE) approach.\nMain results\nFifteen RCTs (769 randomised participants) were identified; four studies evaluated different prednisolone regimens, eight studies evaluated the calcineurin inhibitors (CNIs) (tacrolimus or cyclosporin), two studies evaluated enteric-coated mycophenolate sodium (EC-MPS) and one study evaluated levamisole. In all but two studies of non-corticosteroid agents, reduced-dose prednisolone was given with the treatment agent and the comparator was high-dose prednisolone. In the risk of bias assessment, 11 and seven studies were at low risk of bias for sequence generation and allocation concealment, respectively. No studies were at low risk of performance bias and eight studies were at low risk of detection bias. Thirteen, 10 and six studies were at low risk of attrition bias, reporting bias and other bias, respectively.\nCompared with no specific treatment, it is uncertain whether prednisolone increases the number with complete remission (1 study, 28 participants: RR 1.44, 95% CI 0.95 to 2.19), complete or partial remission (1 study, 28 participants: RR 1.38, 95% CI 0.98 to 1.95), subsequent relapse (1 study, 28 participants: RR 0.75, 95% CI 0.48 to 1.17), or reduces the adverse effects because the certainty of the evidence is very low. Compared with oral prednisolone alone, it is uncertain whether intravenous methylprednisolone and prednisolone increase the number with complete remission (2 studies, 35 participants: RR 1.76, 95% CI 0.17 to 18.32; I² = 90%), relapse (two studies, 19 participants. RR 1.18, 95% CI 0.65 to 2.15; I² = 0%) or adverse events because the certainty of the evidence is very low.\nCompared with prednisolone alone, CNIs with reduced-dose prednisolone or without prednisolone probably make little or no difference to the number achieving complete remission (8 studies; 492 participants: RR 0.99, 95% CI 0.93 to 1.05; I² = 0%), complete or partial remission (4 studies, 269 participants: RR 1.01, 95% CI 0.96 to 1.05; I² = 0%), or relapse (7 studies; 422 participants: RR 0.73, 95% CI 0.51 to 1.03; I² = 0%) (moderate certainty evidence), may reduce the risk of obesity or Cushing's Syndrome (5 studies; 388 participants: RR 0.11, 95% CI 0.02 to 0.59; I² = 45%) and the risk of acne (4 studies; 270 participants: RR 0.15, 95% CI 0.03 to 0.67; I² = 0%) (low certainty evidence); and had uncertain effects on diabetes or hyperglycaemia, hypertension, and acute kidney injury (AKI) (low certainty evidence).\nCompared with prednisolone alone, EC-MPS with reduced-dose prednisolone probably make little or no difference to the number undergoing complete remission at 4 weeks (1 study, 114 participants: RR 1.12, 95% CI 0.84 to 1.50), and at 24 weeks probably make little or no difference to the number undergoing complete remission (2 studies, 134 participants: RR 1.12, 95% CI 0.84 to 1.38; I² = 0%) (moderate certainty evidence), complete or partial remission (2 studies 134 participants: RR 0.92, 95% CI 0.75 to 1.12; I² = 0%), relapse (2 studies, 83 participants: RR 0.50, 95% CI 0.07 to 3.74; I² = 56%) (low certainty evidence); or to the adverse events of new-onset glucose intolerance, death, or AKI (low certainty evidence).\nOne study (24 participants) compared levamisole and prednisolone with prednisolone in patients with relapsing disease. The authors identified no differences in mean relapse rate or adverse effects but no standard deviations were provided.\nAuthors' conclusions\nThis updated review has identified evidence for the efficacy and adverse effects of CNIs and EC-MPS with or without reduced-dose prednisolone compared with prednisolone alone for the induction of remission in adults with MCD and nephrotic syndrome with some reductions in steroid-associated adverse events. RCT data on the efficacy and adverse effects of rituximab in adults with MCD are awaited. Further, adequately powered RCTs are required to determine the relative efficacies of CNIs and EC-MPS and to evaluate these medications in patients with relapsing or steroid-resistant disease.\n\nGiven the provided abstract, please respond to the following query: \"Conclusions\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found that data included in this updated review indicate that calcineurin inhibitors and enteric-coated mycophenolate sodium may be effective in achieving remission in adults with nephrotic syndrome due to MCD with some reduction in the side effects of high dose prednisolone therapy. Studies of these therapies in relapsing disease and in MCD resistant to prednisolone therapy are required." } ]
query-laysum
6754
[ { "role": "user", "content": "Abstract: Background\nNeonates might be exposed to numerous painful procedures due to diagnostic reasons, therapeutic interventions, or surgical procedures.\nOptions for pain management include opioids, non-pharmacological interventions, and other drugs. Morphine, fentanyl, and remifentanil are the opioids most often used in neonates. However, negative impact of opioids on the structure and function of the developing brain has been reported.\nObjectives\nTo evaluate the benefits and harms of opioids in term or preterm neonates exposed to procedural pain, compared to placebo or no drug, non-pharmacological intervention, other analgesics or sedatives, other opioids, or the same opioid administered by a different route.\nSearch methods\nWe used standard, extensive Cochrane search methods. The latest search date was December 2021.\nSelection criteria\nWe included randomized controlled trials conducted in preterm and term infants of a postmenstrual age (PMA) up to 46 weeks and 0 days exposed to procedural pain where opioids were compared to 1) placebo or no drug; 2) non-pharmacological intervention; 3) other analgesics or sedatives; 4) other opioids; or 5) the same opioid administered by a different route.\nData collection and analysis\nWe used standard Cochrane methods. Our primary outcomes were pain assessed with validated methods and any harms. We used a fixed-effect model with risk ratio (RR) for dichotomous data and mean difference (MD) for continuous data, and their confidence intervals (CI). We used GRADE to assess the certainty of the evidence for each outcome.\nMain results\nWe included 13 independent studies (enrolling 823 newborn infants): seven studies compared opioids to no treatment or placebo (the main comparison in this review), two studies to oral sweet solution or non-pharmacological intervention, and five studies (of which two were part of the same study) to other analgesics and sedatives. All studies were performed in a hospital setting.\nOpioids compared to placebo or no drug\nCompared to placebo, opioids probably reduce pain score assessed with the Premature Infant Pain Profile (PIPP)/PIPP-Revised (PIPP-R) scale during the procedure (MD −2.58, 95% CI −3.12 to −2.03; 199 participants, 3 studies; moderate-certainty evidence); may reduce Neonatal Infant Pain Scale (NIPS) during the procedure (MD −1.97, 95% CI −2.46 to −1.48; 102 participants, 2 studies; low-certainty evidence); and may result in little to no difference in pain score assessed with the Douleur Aiguë du Nouveau-né (DAN) scale one to two hours after the procedure (MD −0.20, 95% CI −2.21 to 1.81; 42 participants, 1 study; low-certainty evidence). The evidence is very uncertain about the effect of opioids on pain score assessed with the PIPP/PIPP-R scale up to 30 minutes after the procedure (MD 0.14, 95% CI −0.17 to 0.45; 123 participants, 2 studies; very low-certainty evidence) or one to two hours after the procedure (MD −0.83, 95% CI −2.42 to 0.75; 54 participants, 2 studies; very low-certainty evidence). The evidence is very uncertain about the effect of opioids on episodes of bradycardia (RR 3.19, 95% CI 0.14 to 72.69; 172 participants, 3 studies; very low-certainty evidence). Opioids may result in an increase in episodes of apnea compared to placebo (RR 3.15, 95% CI 1.08 to 9.16; 199 participants, 3 studies; low-certainty evidence): with one study reporting a concerning increase in severe apnea (RR 7.44, 95% CI 0.42 to 132.95; 31 participants, 1 study; very low-certainty). The evidence is very uncertain about the effect of opioids on episodes of hypotension (RR not estimable, risk difference 0.00, 95% CI −0.06 to 0.06; 88 participants, 2 studies; very low-certainty evidence). No studies reported parent satisfaction with care provided in the neonatal intensive care unit (NICU).\nOpioids compared to non-pharmacological intervention\nThe evidence is very uncertain about the effect of opioids on pain score assessed with the Crying Requires oxygen Increased vital signs Expression Sleep (CRIES) scale during the procedure when compared to facilitated tucking (MD −4.62, 95% CI −6.38 to −2.86; 100 participants, 1 study; very low-certainty evidence) or sensorial stimulation (MD 0.32, 95% CI −1.13 to 1.77; 100 participants, 1 study; very low-certainty evidence). The other main outcomes were not reported.\nOpioids compared to other analgesics or sedatives\nThe evidence is very uncertain about the effect of opioids on pain score assessed with the PIPP/PIPP-R during the procedure (MD −0.29, 95% CI −1.58 to 1.01; 124 participants, 2 studies; very low-certainty evidence); up to 30 minutes after the procedure (MD −1.10, 95% CI −2.82 to 0.62; 12 participants, 1 study; very low-certainty evidence); and one to two hours after the procedure (MD −0.17, 95% CI −2.22 to 1.88; 12 participants, 1 study; very low-certainty evidence). No studies reported any harms. The evidence is very uncertain about the effect of opioids on episodes of apnea during (RR 3.27, 95% CI 0.85 to 12.58; 124 participants, 2 studies; very low-certainty evidence) and after the procedure (RR 2.71, 95% CI 0.11 to 64.96; 124 participants, 2 studies; very low-certainty evidence) and on hypotension (RR 1.34, 95% CI 0.32 to 5.59; 204 participants, 3 studies; very low-certainty evidence). The other main outcomes were not reported.\nWe identified no studies comparing different opioids (e.g. morphine versus fentanyl) or different routes for administration of the same opioid (e.g. morphine enterally versus morphine intravenously).\nAuthors' conclusions\nCompared to placebo, opioids probably reduce pain score assessed with PIPP/PIPP-R scale during the procedure; may reduce NIPS during the procedure; and may result in little to no difference in DAN one to two hours after the procedure. The evidence is very uncertain about the effect of opioids on pain assessed with other pain scores or at different time points. The evidence is very uncertain about the effect of opioids on episodes of bradycardia, hypotension or severe apnea. Opioids may result in an increase in episodes of apnea. No studies reported parent satisfaction with care provided in the NICU. The evidence is very uncertain about the effect of opioids on any outcome when compared to non-pharmacological interventions or to other analgesics. We identified no studies comparing opioids to other opioids or comparing different routes of administration of the same opioid.\n\nGiven the provided abstract, please respond to the following query: \"What did we want to find out?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We wanted to find out the effect of opioids in babies exposed to painful procedures, compared to:\n• no treatment or placebo;\n• non-drug treatments (such as sweet solutions);\n• other drugs;\n• different types of opioids;\n• or same opioid administered by a different route, for example by mouth compared to by injection." } ]
query-laysum
6755
[ { "role": "user", "content": "Abstract: Background\nAdult female Anopheles mosquitoes can transmit Plasmodium parasites that cause malaria. Some fish species eat mosquito larvae and pupae. In disease control policy documents, the World Health Organization (WHO) includes biological control of malaria vectors by stocking ponds, rivers, and water collections near where people live with larvivorous fish to reduce Plasmodium parasite transmission. In the past, the Global Fund has financed larvivorous fish programmes in some countries, and, with increasing efforts in eradication of malaria, policymakers may return to this option. Therefore, we assessed the evidence base for larvivorous fish programmes in malaria control.\nObjectives\nTo evaluate whether introducing larvivorous fish to anopheline larval habitats impacts Plasmodium parasite transmission. We also sought to summarize studies that evaluated whether introducing larvivorous fish influences the density and presence of Anopheles larvae and pupae in water sources.\nSearch methods\nWe searched the Cochrane Infectious Diseases Group Specialized Register; the Cochrane Central Register of Controlled Trials (CENTRAL), published in the Cochrane Library; MEDLINE (PubMed); Embase (Ovid); CABS Abstracts; LILACS; and the metaRegister of Controlled Trials (mRCT) up to 6 July 2017. We checked the reference lists of all studies identified by the search. We examined references listed in review articles and previously compiled bibliographies to look for eligible studies. Also we contacted researchers in the field and the authors of studies that met the inclusion criteria for additional information regarding potential studies for inclusion and ongoing studies. This is an update of a Cochrane Review published in 2013.\nSelection criteria\nRandomized controlled trials (RCTs) and non-RCTs, including controlled before-and-after studies, controlled time series, and controlled interrupted time series studies from malaria-endemic regions that introduced fish as a larvicide and reported on malaria in the community or the density of the adult anopheline population. In the absence of direct evidence of an effect on transmission, we performed a secondary analysis on studies that evaluated the effect of introducing larvivorous fish on the density or presence of immature anopheline mosquitoes (larvae and pupae forms) in water sources to determine whether this intervention has any potential that may justify further research in the control of malaria vectors.\nData collection and analysis\nTwo review authors independently screened each article by title and abstract, and examined potentially relevant studies for inclusion using an eligibility form. At least two review authors independently extracted data and assessed risk of bias of included studies. If relevant data were unclear or were not reported, we contacted the study authors for clarification. We presented data in tables, and we summarized studies that evaluated the effects of introducing fish on anopheline immature density or presence, or both. We used the GRADE approach to summarize the certainty of the evidence. We also examined whether the included studies reported any possible adverse impact of introducing larvivorous fish on non-target native species.\nMain results\nWe identified no studies that reported the effects of introducing larvivorous fish on the primary outcomes of this review: malaria infection in nearby communities, entomological inoculation rate, or on adult Anopheles density.\nFor the secondary analysis, we examined the effects of introducing larvivorous fish on the density and presence of anopheline larvae and pupae in community water sources, and found 15 small studies with a follow-up period between 22 days and five years. These studies were undertaken in Sri Lanka (two studies), India (three studies), Ethiopia (one study), Kenya (two studies), Sudan (one study), Grande Comore Island (one study), Korea (two studies), Indonesia (one study), and Tajikistan (two studies). These studies were conducted in a variety of settings, including localized water bodies (such as wells, domestic water containers, fishponds, and pools (seven studies); riverbed pools below dams (two studies)); rice field plots (five studies); and water canals (two studies). All included studies were at high risk of bias. The research was insufficient to determine whether larvivorous fish reduce the density of Anopheles larvae and pupae (12 studies, unpooled data, very low certainty evidence). Some studies with high stocking levels of fish seemed to arrest the increase in immature anopheline populations, or to reduce the number of immature anopheline mosquitoes, compared with controls. However, this finding was not consistent, and in studies that showed a decrease in immature anopheline populations, the effect was not always consistently sustained. In contrast, some studies reported larvivorous fish reduced the number of water sources withAnopheles larvae and pupae (five studies, unpooled data, low certainty evidence).\nNone of the included studies reported effects of larvivorous fish on local native fish populations or other species.\nAuthors' conclusions\nWe do not know whether introducing larvivorous fish reduces malaria transmission or the density of adult anopheline mosquito populations.\nIn research studies that examined the effects on immature anopheline stages of introducing fish to potential malaria vector larval habitats, high stocking levels of fish may reduce the density or presence of immature anopheline mosquitoes in the short term. We do not know whether this translates into impact on malaria transmission. Our interpretation of the current evidence is that countries should not invest in fish stocking as a stand alone or supplementary larval control measure in any malaria transmission areas outside the context of research using carefully controlled field studies or quasi-experimental designs. Such research should examine the effects on native fish and other non-target species.\n\nGiven the provided abstract, please respond to the following query: \"Key messages\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We do not know if introducing fish that eat mosquito larvae and pupae has an impact on the number of people with malaria or on the adult Anopheles mosquito population." } ]
query-laysum
6756
[ { "role": "user", "content": "Abstract: Background\nAcute diarrhoea is a leading cause of death for children under five years of age. Most deaths are caused by excessive fluid and electrolyte losses. Racecadotril is an anti-secretory drug that has been used for acute diarrhoea in children as an adjunct to oral rehydration therapy.\nObjectives\nTo assess the efficacy and safety of racecadotril for treating acute diarrhoea in children under five years of age.\nSearch methods\nWe searched the Cochrane Infectious Diseases Group Specialized Register; the Cochrane Central Register of Controlled Trials (CENTRAL, published in the Cochrane Library Issue 3, March 2019); MEDLINE; Embase; LILACS; ClinicalTrials.gov; and the World Health Organization International Clinical Trials Registry Platform (WHO ICTRP), up to 4 March 2019, for clinical trials regardless of publication language or status.\nSelection criteria\nRandomized controlled trials (RCTs) that compared racecadotril to placebo or no intervention in addition to standard care (oral rehydration therapy) in children under five with acute diarrhoea. The primary outcomes were failure of oral rehydration, duration of diarrhoea, and number of stools. The secondary outcomes were stool output, length of the hospital stay, and adverse events.\nData collection and analysis\nTwo review authors independently assessed trial eligibility, extracted the data and assessed risk of bias. We presented dichotomous data with risk ratios (RR) and continuous data with mean difference (MD) or standardized mean difference (SMD). Where appropriate, we combined trials with meta-analysis and used a random-effects model if there was significant heterogeneity (I² ≥ 50%). We assessed the certainty of the evidence using the GRADE approach.\nMain results\nSeven RCTs with a total of 1140 participants met the inclusion criteria. The trials were carried out on children aged three months to five years, in outpatient and inpatient facilities from France, Spain, Peru, India, Kenya, and Ecuador. The efficacy and safety of racecadotril were compared to placebo or no treatment. Racecadotril may reduce the risk of rehydration failure (RR 0.41, 95% CI 0.13 to 1.23; 2 RCTs, 192 participants; low-certainty evidence). Data on duration of diarrhoea, number of stools in the first 48 hours are insufficient to reach a conclusion; stool output in the first 48 hours appears to be lower in the two trials measuring this, although the data is not combinable. Length of hospital stay was similar in two studies measuring this, and overall there was no evidence that racecadotril increased overall rate of adverse events (RR 0.90, 95% CI 0.66 to 1.22; 5 RCTs, 688 participants; low-certainty evidence). Most adverse events in the racecadotril group were mild or moderate.\nAuthors' conclusions\nRacecadotril seems to be a safe drug but has little benefit in improving acute diarrhoea in children under five years of age. Current evidence does not support routine use of racecadotril in management of acute diarrhoea in children under five outside of the context of placebo controlled RCTs.\n\nGiven the provided abstract, please respond to the following query: \"Key messages\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Racecadotril may reduce the risk of rehydration failure. We are uncertain whether or not it influences number of bowel movements or duration of diarrhoea." } ]
query-laysum
6757
[ { "role": "user", "content": "Abstract: Background\nRenin inhibitors (RIs) reduce blood pressure more than placebo, with the magnitude of this effect thought to be similar to that for angiotensin converting enzyme (ACE) inhibitors. However, a drug's efficacy in lowering blood pressure cannot be considered as a definitive indicator of its effectiveness in reducing mortality and morbidity. The effectiveness and safety of RIs compared to ACE inhibitors in treating hypertension is unknown.\nObjectives\nTo evaluate the benefits and harms of renin inhibitors compared to ACE inhibitors in people with primary hypertension.\nSearch methods\nThe Cochrane Hypertension Group Information Specialist searched the following databases for randomized controlled trials up to August 2020: the Cochrane Hypertension Specialized Register, the Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE (from 1946), Embase (from 1974), the World Health Organization International Clinical Trials Registry Platform, and ClinicalTrials.gov. We also contacted authors of relevant papers about further published and unpublished work. The searches had no language restrictions.\nSelection criteria\nWe included randomized, active-controlled, double-blinded studies (RCTs) with at least four weeks follow-up in people with primary hypertension, which compared renin inhibitors with ACE inhibitors and reported morbidity, mortality, adverse events or blood pressure outcomes. We excluded people with proven secondary hypertension.\nData collection and analysis\nTwo review authors independently selected the included trials, evaluated the risks of bias and entered the data for analysis.\nMain results\nWe include 11 RCTs involving 13,627 participants, with a mean baseline age from 51.5 to 74.2 years. Follow-up duration ranged from four weeks to 36.6 months.\nThere was no difference between RIs and ACE inhibitors for the outcomes: all-cause mortality: risk ratio (RR) 1.05, 95% confidence interval (CI) 0.93 to 1.18; 5 RCTs, 5962 participants; low-certainty evidence; total myocardial infarction: RR 0.86, 95% CI 0.22 to 3.39; 2 RCTs, 957 participants; very low-certainty evidence; adverse events: RR 0.98, 95% CI 0.93 to 1.03; 10 RTCs, 6007 participants;  moderate-certainty evidence; serious adverse events: RR 1.21, 95% CI 0.89 to 1.64; 10 RTCs, 6007 participants; low-certainty evidence; and withdrawal due to adverse effects: RR 0.85, 95% CI 0.68 to 1.06; 10 RTCs, 6008 participants; low-certainty evidence. No data were available for total cardiovascular events, heart failure, stroke, end-stage renal disease or change in heart rate. Low-certainty evidence suggested that RIs reduced systolic blood pressure: mean difference (MD) −1.72, 95% CI −2.47 to −0.97; 9 RCTs, 5001 participants;  and diastolic blood pressure: MD −1.18, 95% CI −1.65 to −0.72; 9 RCTs, 5001 participants,  to a greater extent than ACE inhibitors, but we judged this to be more likely due to bias than a true effect.\nAuthors' conclusions\nFor the treatment of hypertension, we have low certainty that renin inhibitors (RI) and angiotensin converting enzyme (ACE) inhibitors do not differ for all-cause mortality and myocardial infarction. We have low to moderate certainty that they do not differ for adverse events. Small reductions in blood pressure with renin inhibitors compared to ACE inhibitors are of low certainty.  More independent, large, long-term trials are needed to compare RIs with ACE inhibitors, particularly assessing morbidity and mortality outcomes, but also on blood pressure-lowering effect.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We included randomized, active-controlled, double-blinded studies (RCTs) for this review. We include 11 RCTs involving 13,627 participants, with an average age from 51 to 74 years. Length of follow-up ranged from four weeks to 36 months." } ]
query-laysum
6758
[ { "role": "user", "content": "Abstract: Background\nPeople with inflammatory bowel disease (IBD) require intensive follow-up with frequent consultations after diagnosis. IBD telehealth management includes consulting by phone, instant messenger, video, text message, or web-based services. Telehealth can be beneficial for people with IBD, but may have its own set of challenges. It is important to systematically review the evidence on the types of remote or telehealth approaches that can be deployed in IBD. This is particularly relevant following the coronavirus disease 2019 (COVID-19) pandemic, which led to increased self- and remote-management.\nObjectives\nTo identify the communication technologies used to achieve remote healthcare for people with inflammatory bowel disease and to assess their effectiveness.\nSearch methods\nOn 13 January 2022, we searched CENTRAL, Embase, MEDLINE, three other databases, and three trials registries with no limitations on language, date, document type, or publication status.\nSelection criteria\nAll published, unpublished, and ongoing randomised controlled trials (RCTs) that evaluated telehealth interventions targeted at people with IBD versus any other type of intervention or no intervention.\nWe did not include studies based on digital patient information resources or education resources, unless they formed part of a wider package including an element of telehealth. We excluded studies where remote monitoring of blood or faecal tests was the only form of monitoring.\nData collection and analysis\nTwo review authors independently extracted data from the included studies and assessed their risk of bias. We analysed studies on adult and paediatric populations separately. We expressed the effects of dichotomous outcomes as risk ratios (RRs) and the effects of continuous outcomes as mean differences (MDs) or standardised mean differences (SMDs), each with their 95% confidence intervals (CIs). We assessed the certainty of the evidence using GRADE methodology.\nMain results\nWe included 19 RCTs with a total of 3489 randomised participants, aged eight to 95 years. Three studies examined only people with ulcerative colitis (UC), two studies examined only people with Crohn's disease (CD), and the remaining studies examined a mix of IBD patients. Studies considered a range of disease activity states. The length of the interventions ranged from six months to two years. The telehealth interventions were web-based and telephone-based.\nWeb-based monitoring versus usual care\nTwelve studies compared web-based disease monitoring to usual care.\nThree studies, all in adults, provided data on disease activity. Web-based disease monitoring (n = 254) is probably equivalent to usual care (n = 174) in reducing disease activity in people with IBD (SMD 0.09, 95% CI −0.11 to 0.29). The certainty of the evidence is moderate.\nFive studies on adults provided dichotomous data that we could use for a meta-analysis on flare-ups. Web-based disease monitoring (n = 207/496) is probably equivalent to usual care (n = 150/372) for the occurrence of flare-ups or relapses in adults with IBD (RR 1.09, 95% CI 0.93 to 1.27). The certainty of the evidence is moderate. One study provided continuous data. Web-based disease monitoring (n = 465) is probably equivalent to usual care (n = 444) for the occurrence of flare-ups or relapses in adults with CD (MD 0.00 events, 95% CI −0.06 to 0.06). The certainty of the evidence is moderate. One study provided dichotomous data on flare-ups in a paediatric population. Web-based disease monitoring (n = 28/84) may be equivalent to usual care (n = 29/86) for the occurrence of flare-ups or relapses in children with IBD (RR 0.99, 95% CI 0.65 to 1.51). The certainty of the evidence is low.\nFour studies, all in adults, provided data on quality of life. Web-based disease monitoring (n = 594) is probably equivalent to usual care (n = 505) for quality of life in adults with IBD (SMD 0.08, 95% CI −0.04 to 0.20). The certainty of the evidence is moderate.\nBased on continuous data from one study in adults, we found that web-based disease monitoring probably leads to slightly higher medication adherence compared to usual care (MD 0.24 points, 95% CI 0.01 to 0.47). The results are of moderate certainty. Based on continuous data from one paediatric study, we found no difference between web-based disease monitoring and usual care in terms of their effect on medication adherence (MD 0.00, 95% CI −0.63 to 0.63), although the evidence is very uncertain. When we meta-analysed dichotomous data from two studies on adults, we found no difference between web-based disease monitoring and usual care in terms of their effect on medication adherence (RR 0.87, 95% CI 0.62 to 1.21), although the evidence is very uncertain.\nWe were unable to draw any conclusions on the effects of web-based disease monitoring compared to usual care on healthcare access, participant engagement, attendance rate, interactions with healthcare professionals, and cost- or time-effectiveness. The certainty of the evidence is very low.\nAuthors' conclusions\nThe evidence in this review suggests that web-based disease monitoring is probably no different to standard care in adults when considering disease activity, occurrence of flare-ups or relapse, and quality of life. There may be no difference in these outcomes in children, but the evidence is limited. Web-based monitoring probably increases medication adherence slightly compared to usual care.\nWe are uncertain about the effects of web-based monitoring versus usual care on our other secondary outcomes, and about the effects of the other telehealth interventions included in our review, because the evidence is limited.\nFurther studies comparing web-based disease monitoring to standard care for the clinical outcomes reported in adults are unlikely to change our conclusions, unless they have longer follow-up or investigate under-reported outcomes or populations. Studies with a clearer definition of web-based monitoring would enhance applicability, enable practical dissemination and replication, and enable alignment with areas identified as important by stakeholders and people affected by IBD.\n\nGiven the provided abstract, please respond to the following query: \"What next?\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "No further studies comparing web-based care to usual care in adults are necessary, unless they last for longer periods of time or give more details that would help clinicians adopt them anywhere in the world. This includes details on the type and number of staff needed, resources, equipment, costs, accessibility, and data security. More studies on children may be useful, as well as studies that examine differences based on sex and social or financial status. In any case, future studies should concentrate on measuring the results that matter most to people with inflammatory bowel disease and their care providers." } ]
query-laysum
6759
[ { "role": "user", "content": "Abstract: Background\nElectromechanical- and robot-assisted gait-training devices are used in rehabilitation and might help to improve walking after stroke. This is an update of a Cochrane Review first published in 2007 and previously updated in 2017.\nObjectives\nPrimary\n• To determine whether electromechanical- and robot-assisted gait training versus normal care improves walking after stroke\nSecondary\n• To determine whether electromechanical- and robot-assisted gait training versus normal care after stroke improves walking velocity, walking capacity, acceptability, and death from all causes until the end of the intervention phase\nSearch methods\nWe searched the Cochrane Stroke Group Trials Register (last searched 6 January 2020); the Cochrane Central Register of Controlled Trials (CENTRAL; 2020 Issue 1), in the Cochrane Library; MEDLINE in Ovid (1950 to 6 January 2020); Embase (1980 to 6 January 2020); the Cumulative Index to Nursing and Allied Health Literature (CINAHL; 1982 to 20 November 2019); the Allied and Complementary Medicine Database (AMED; 1985 to 6 January 2020); Web of Science (1899 to 7 January 2020); SPORTDiscus (1949 to 6 January 2020); the Physiotherapy Evidence Database (PEDro; searched 7 January 2020); and the engineering databases COMPENDEX (1972 to 16 January 2020) and Inspec (1969 to 6 January 2020). We handsearched relevant conference proceedings, searched trials and research registers, checked reference lists, and contacted trial authors in an effort to identify further published, unpublished, and ongoing trials.\nSelection criteria\nWe included all randomised controlled trials and randomised controlled cross-over trials in people over the age of 18 years diagnosed with stroke of any severity, at any stage, in any setting, evaluating electromechanical- and robot-assisted gait training versus normal care.\nData collection and analysis\nTwo review authors independently selected trials for inclusion, assessed methodological quality and risk of bias, and extracted data. We assessed the quality of evidence using the GRADE approach. The primary outcome was the proportion of participants walking independently at follow-up.\nMain results\nWe included in this review update 62 trials involving 2440 participants. Electromechanical-assisted gait training in combination with physiotherapy increased the odds of participants becoming independent in walking (odds ratio (random effects) 2.01, 95% confidence interval (CI) 1.51 to 2.69; 38 studies, 1567 participants; P < 0.00001; I² = 0%; high-quality evidence) and increased mean walking velocity (mean difference (MD) 0.06 m/s, 95% CI 0.02 to 0.10; 42 studies, 1600 participants; P = 0.004; I² = 60%; low-quality evidence) but did not improve mean walking capacity (MD 10.9 metres walked in 6 minutes, 95% CI -5.7 to 27.4; 24 studies, 983 participants; P = 0.2; I² = 42%; moderate-quality evidence). Electromechanical-assisted gait training did not increase the risk of loss to the study during intervention nor the risk of death from all causes. Results must be interpreted with caution because (1) some trials investigated people who were independent in walking at the start of the study, (2) we found variation between trials with respect to devices used and duration and frequency of treatment, and (3) some trials included devices with functional electrical stimulation. Post hoc analysis showed that people who are non-ambulatory at the start of the intervention may benefit but ambulatory people may not benefit from this type of training. Post hoc analysis showed no differences between the types of devices used in studies regarding ability to walk but revealed differences between devices in terms of walking velocity and capacity.\nAuthors' conclusions\nPeople who receive electromechanical-assisted gait training in combination with physiotherapy after stroke are more likely to achieve independent walking than people who receive gait training without these devices. We concluded that eight patients need to be treated to prevent one dependency in walking. Specifically, people in the first three months after stroke and those who are not able to walk seem to benefit most from this type of intervention. The role of the type of device is still not clear. Further research should consist of large definitive pragmatic phase 3 trials undertaken to address specific questions about the most effective frequency and duration of electromechanical-assisted gait training, as well as how long any benefit may last. Future trials should consider time post stroke in their trial design.\n\nGiven the provided abstract, please respond to the following query: \"Why we did this Cochrane Review\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Gait-training devices allow people who can't walk to have intensive walking practice, without the need for a lot of physical support from their therapist as they practice walking.\nWe wanted to find out if gait-training devices could help people improve their walking after a stroke." } ]
query-laysum
6760
[ { "role": "user", "content": "Abstract: Background\nEarly diagnosis and treatment of lower respiratory tract infections are the mainstay of management of lung disease in cystic fibrosis. When sputum samples are unavailable, treatment relies mainly on cultures from oropharyngeal specimens; however, there are concerns regarding the sensitivity of these to identify lower respiratory organisms.\nBronchoscopy and related procedures (including bronchoalveolar lavage) though invasive, allow the collection of lower respiratory specimens from non-sputum producers. Cultures of bronchoscopic specimens provide a higher yield of organisms compared to those from oropharyngeal specimens. Regular use of bronchoscopy and related procedures may help in a more accurate diagnosis of lower respiratory tract infections and guide the selection of antimicrobials, which may lead to clinical benefits.\nThis is an update of a previous review.\nObjectives\nTo evaluate the use of bronchoscopy-guided antimicrobial therapy in the management of lung infection in adults and children with cystic fibrosis.\nSearch methods\nWe searched the Cochrane Cystic Fibrosis Trials Register, compiled from electronic database searches and handsearching of journals and conference abstract books. Date of latest search: 30 August 2018.\nWe also searched three registries of ongoing studies and the reference lists of relevant articles and reviews. Date of latest search: 10 April 2018.\nSelection criteria\nWe included randomized controlled studies including people of any age with cystic fibrosis, comparing outcomes following therapies guided by the results of bronchoscopy (and related procedures) with outcomes following therapies guided by the results of any other type of sampling (including cultures from sputum, throat swab and cough swab).\nData collection and analysis\nTwo review authors independently selected studies, assessed their risk of bias and extracted data. We contacted study investigators for further information. The quality of the evidence was assessed using the GRADE criteria.\nMain results\nThe search identified 11 studies, but we only included one study enrolling infants with cystic fibrosis under six months of age and diagnosed through newborn screening (170 enrolled); participants were followed until they were five years old (data from 157 children). The study compared outcomes following therapy directed by bronchoalveolar lavage for pulmonary exacerbations with standard treatment based on clinical features and oropharyngeal cultures.\nWe considered this study to have a low risk of bias; however, the statistical power to detect a significant difference in the prevalence of Pseudomonas aeruginosa was limited due to the prevalence (of Pseudomonas aeruginosa isolation in bronchoalveolar lavage samples at five years age) being much lower in both the groups compared to that which was expected and which was used for the power calculation. The sample size was adequate to detect a difference in high-resolution computed tomography scoring. The quality of evidence for the key parameters was graded as low except high-resolution computed tomography scoring and cost of care analysis, which were graded as moderate quality.\nAt five years of age, there was no clear benefit of bronchoalveolar lavage-directed therapy on lung function z scores or nutritional parameters. Evaluation of total and component high-resolution computed tomography scores showed no significant difference in evidence of structural lung disease in the two groups.\nIn addition, this study did not show any difference between the number of isolates of Pseudomonas aeruginosa per child per year diagnosed in the bronchoalveolar lavage-directed therapy group compared to the standard therapy group. The eradication rate following one or two courses of eradication treatment was comparable in the two groups, as were the number of pulmonary exacerbations. However, the number of hospitalizations was significantly higher in the bronchoalveolar lavage-directed therapy group, but the mean duration of hospitalizations was significantly less compared to the standard therapy group.\nMild adverse events were reported in a proportion of participants, but these were generally well-tolerated. The most common adverse event reported was transient worsening of cough after 29% of procedures. Significant clinical deterioration was documented during or within 24 hours of bronchoalveolar lavage in 4.8% of procedures.\nAuthors' conclusions\nThis review, limited to a single, well-designed randomized controlled study, shows no clear evidence to support the routine use of bronchoalveolar lavage for the diagnosis and management of pulmonary infection in pre-school children with cystic fibrosis compared to the standard practice of providing treatment based on results of oropharyngeal culture and clinical symptoms. No evidence was available for adult and adolescent populations.\n\nGiven the provided abstract, please respond to the following query: \"Search date\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The evidence is current to: 10 April 2018." } ]
query-laysum
6761
[ { "role": "user", "content": "Abstract: Background\nDiaphragm-triggered non-invasive respiratory support, commonly referred to as NIV-NAVA (non-invasive neurally adjusted ventilatory assist), uses the electrical activity of the crural diaphragm to trigger the start and end of a breath. It provides variable inspiratory pressure that is proportional to an infant's changing inspiratory effort. NIV-NAVA has the potential to provide effective, non-invasive, synchronised, multilevel support and may reduce the need for invasive ventilation; however, its effects on short- and long-term outcomes, especially in the preterm infant, are unclear.\nObjectives\nTo assess the effectiveness and safety of diaphragm-triggered non-invasive respiratory support in preterm infants (< 37 weeks' gestation) when compared to other non-invasive modes of respiratory support (nasal intermittent positive pressure ventilation (NIPPV); nasal continuous positive airway pressure (nCPAP); high-flow nasal cannulae (HFNC)), and to assess preterm infants with birth weight less than 1000 grams or less than 28 weeks' corrected gestation at the time of intervention as a sub-group.\nSearch methods\nWe used the standard search strategy of Cochrane Neonatal to search the Cochrane Central Register of Controlled Trials (CENTRAL 2019, Issue 5), MEDLINE via PubMed (1946 to 10 May 2019), Embase (1947 to 10 May 2019), and CINAHL (1982 to 10 May 2019). We also searched clinical trials databases, conference proceedings, and the reference lists of retrieved articles for randomised controlled trials (RCTs) and quasi-randomised trials.\nSelection criteria\nRandomised and quasi-randomised controlled trials that compared diaphragm-triggered non-invasive versus other non-invasive respiratory support in preterm infants.\nData collection and analysis\nTwo review authors independently selected trials, assessed trial quality and extracted data from included studies. We performed fixed-effect analyses and expressed treatment effects as mean difference (MD), risk ratio (RR), and risk difference (RD) with 95% confidence intervals (CIs). We used the generic inverse variance method to analyse specific outcomes for cross-over trials. We used the GRADE approach to assess the certainty of evidence.\nMain results\nThere were two small randomised controlled trials including a total of 23 infants eligible for inclusion in the review. Only one trial involving 16 infants included in the analysis reported on either of the primary outcomes of the review. This found no difference in failure of modality between NIV-NAVA and NIPPV (RR 0.33, 95% CI 0.02 to 7.14; RD −0.13, 95% CI -0.41 to 0.16; 1 study, 16 infants; heterogeneity not applicable).\nBoth trials reported on secondary outcomes of the review, specific for cross-over trials (total 22 infants; 1 excluded due to failure of initial modality). One study involving seven infants reported a significant reduction in maximum FiO₂ with NIV-NAVA compared to NIPPV (MD −4.29, 95% CI −5.47 to −3.11; heterogeneity not applicable). There was no difference in maximum electric activity of the diaphragm (Edi) signal between modalities (MD −1.75, 95% CI −3.75 to 0.26; I² = 0%) and a significant increase in respiratory rate with NIV-NAVA compared to NIPPV (MD 7.22, 95% CI 0.21 to 14.22; I² = 72%) on a meta-analysis of two studies involving a total of 22 infants. The included studies did not report on other outcomes of interest.\nAuthors' conclusions\nDue to limited data and very low certainty evidence, we were unable to determine if diaphragm-triggered non-invasive respiratory support is effective or safe in preventing respiratory failure in preterm infants. Large, adequately powered randomised controlled trials are needed to determine if diaphragm-triggered non-invasive respiratory support in preterm infants is effective or safe.\n\nGiven the provided abstract, please respond to the following query: \"Quality of evidence\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We were not able to make any meaningful conclusions in this review due to limited data and very low quality evidence. Large, high-quality studies are needed to determine whether diaphragm-triggered non-invasive respiratory support can prevent respiratory failure." } ]
query-laysum
6762
[ { "role": "user", "content": "Abstract: Background\nThe finding that exercise is inversely related to metabolic syndrome after transplantation is novel and suggests that exercise interventions might provide a means for reducing metabolic syndrome complications in liver transplantation recipients. The use of exercise for increasing the physical activity daily levels by more frequent, higher intensity, and longer duration of training sessions, or the sum of these components may be necessary to counteract the effects of the pretransplant reduced activity, metabolic disturbances, and post-transplant immunosuppression, as well as improve physical function and aerobic capacity following liver transplantation. Regular physical activity has a long-term positive impact on recovery following various surgical procedures including transplantation, giving people the opportunity to return to an active life with their families, in society, and in their professional life. Likewise, specific muscle strength training may attenuate the loss of strength after liver transplantation.\nObjectives\nTo evaluate the benefits and harms of exercise-based interventions in adults after liver transplantation compared to no exercise, sham interventions, or another type of exercise.\nSearch methods\nWe used standard, extensive Cochrane search methods. The latest search date was 2 September 2022.\nSelection criteria\nWe included randomised clinical trials in liver transplantation recipients comparing any type of exercise with no exercise, sham interventions, or another type of exercise.\nData collection and analysis\nWe used standard Cochrane methods. Our primary outcomes were 1. all-cause mortality; 2. serious adverse events; and 3. health-related quality of life. Our secondary outcomes were 4. a composite of cardiovascular mortality and cardiac disease; 5. aerobic capacity; 6. muscle strength; 7. morbidity; 8. non-serious adverse events; and 9. cardiovascular disease post-transplantation. We assessed risk of bias of the individual trials using RoB 1, described the interventions using the TIDieR checklist, and used GRADE to assess certainty of evidence.\nMain results\nWe included three randomised clinical trials. The trials randomised 241 adults with liver transplantation, of which 199 participants completed the trials. The trials were conducted in the USA, Spain, and Turkey. They compared exercise versus usual care. The duration of the interventions ranged from two to 10 months. One trial reported that 69% of participants who received the exercise intervention were adherent to the exercise prescription. A second trial reported a 94% adherence to the exercise programme, with participants attending 45/48 sessions. The remaining trial reported a 96.8% adherence to the exercise intervention during the hospitalisation period.\nTwo trials received funding; one from the National Center for Research Resources (US) and the other from Instituto de Salud Carlos III (Spain). The remaining trial did not receive funding.\nAll trials were at an overall high risk of bias, derived from high risk of selective reporting bias and attrition bias in two trials. The results on all-cause mortality showed a higher risk of death in the exercise group versus the control group, but these results are very uncertain (risk ratio (RR) 3.14, 95% confidence interval (CI) 0.74 to 13.37; 2 trials, 165 participants; I² = 0%; very low-certainty evidence). The trials did not report data on serious adverse events excluding mortality or non-serious adverse events. However, all trials reported that there were no adverse effects associated with exercise. We are very uncertain on whether exercise compared with usual care has a beneficial or harmful effect on health-related quality of life assessed using the 36-item Short Form Physical Functioning subscale at the end of the intervention (mean difference (MD) 10.56, 95% CI −0.12 to 21.24; 2 trials, 169 participants; I² = 71%; very low-certainty evidence). None of the trials reported data on composite of cardiovascular mortality and cardiovascular disease, and cardiovascular disease post-transplantation. We are very uncertain if there are differences in aerobic capacity in terms of VO2peak at the end of the intervention between groups (MD 0.80, 95% CI −0.80 to 2.39; 3 trials, 199 participants; I² = 0%; very low-certainty evidence). We are very uncertain if there are differences in muscle strength at end of the intervention between groups (MD 9.91, 95% CI −3.68 to 23.50; 3 trials, 199 participants; I² = 44%; very low-certainty evidence). One trial measured perceived fatigue using the Checklist Individual Strength (CIST). Participants in the exercise group showed a clinically important lower degree of fatigue perception than participants in the control group, with a mean reduction of 40 points in the CIST (95% CI 15.62 to 64.38; 1 trial, 30 participants).\nWe identified three ongoing studies.\nAuthors' conclusions\nBased on very low-certainty evidence in our systematic review, we are very uncertain of the role of exercise training (aerobic, resistance-based exercises, or both) in affecting mortality, health-related quality of life, and physical function (i.e. aerobic capacity and muscle strength) in liver transplant recipients. There were few data on the composite of cardiovascular mortality and cardiovascular disease, cardiovascular disease post-transplantation, and adverse event outcomes. We lack larger trials with blinded outcome assessment, designed according to the SPIRIT statement and reported according to the CONSORT statement.\n\nGiven the provided abstract, please respond to the following query: \"Background\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Levels of physical activity tend to decrease in people who receive a liver transplant. The benefits and harms of exercise interventions to protect against the development of heart and lung diseases, hypertension, type II diabetes, dementia, non-alcoholic fatty liver disease (conditions caused by a build-up of fat in the liver), cancer, or other life-threatening diseases which may develop rapidly have not yet been well studied." } ]
query-laysum
6763
[ { "role": "user", "content": "Abstract: Background\nRapid implementation of robotic transabdominal surgery has resulted in the need for re-evaluation of the most suitable form of anaesthesia. The overall objective of anaesthesia is to minimize perioperative risk and discomfort for patients both during and after surgery. Anaesthesia for patients undergoing robotic assisted surgery is different from anaesthesia for patients undergoing open or laparoscopic surgery; new anaesthetic concerns accompany robotic assisted surgery.\nObjectives\nTo assess outcomes related to the choice of total intravenous anaesthesia (TIVA) or inhalational anaesthesia for adults undergoing transabdominal robotic assisted laparoscopic gynaecological, urological or gastroenterological surgery.\nSearch methods\nWe searched the Cochrane Central Register of Controlled Trials (CENTRAL; 2016 Issue 5), Ovid MEDLINE (1946 to May 2016), Embase via OvidSP (1982 to May 2016), the Cumulative Index to Nursing and Allied Health Literature (CINAHL) via EBSCOhost (1982 to May 2016) and the Institute for Scientific Information (ISI) Web of Science (1956 to May 2016). We also searched the International Standard Randomized Controlled Trial Number (ISRCTN) Registry and Clinical trials gov for ongoing trials (May 2016).\nSelection criteria\nWe searched for randomized controlled trials (RCTs) including adults, aged 18 years and older, of both genders, treated with transabdominal robotic assisted laparoscopic gynaecological, urological or gastroenterological surgery and focusing on outcomes of TIVA or inhalational anaesthesia.\nData collection and analysis\nWe used standard methodological procedures of Cochrane. Study findings were not suitable for meta-analysis.\nMain results\nWe included three single-centre, two-arm RCTs involving 170 participants. We found one ongoing trial. All included participants were male and were undergoing radical robotic assisted laparoscopic radical prostatectomy (RALRP). The men were between 50 and 75 years of age and met criteria for American Society of Anesthesiologists physical classification scores (ASA) I, ll and III.\nWe found evidence showing no clinically meaningful differences in postoperative pain between the two types of anaesthetics (mean difference (MD) in visual analogue scale (VAS) scores at one to six hours was -2.20 (95% confidence interval (CI) -10.62 to 6.22; P = 0.61) in a sample of 62 participants from one study. Low-quality evidence suggests that propofol reduces postoperative nausea and vomiting (PONV) over the short term (one to six hours after surgery) after RALRP compared with inhalational anaesthesia (sevoflurane, desflurane) (MD -1.70, 95% CI -2.59 to -0.81; P = 0.0002).\nWe found low-quality evidence suggesting that propofol may prevent an increase in intraocular pressure (IOP) after pneumoperitoneum and steep Trendelenburg positioning compared with sevoflurane (MD -3.90, 95% CI -6.34 to -1.46; P = 0.002) with increased IOP from baseline to 30 minutes in steep Trendelenburg. However, it is unclear whether this surrogate outcome translates directly to clinical avoidance of ocular complications during surgery. No studies addressed the secondary outcomes of adverse effects, all-cause mortality, respiratory or circulatory complications, cognitive dysfunction, length of stay or costs. Overall the quality of evidence was low to very low, as all studies were small, single-centre trials providing unclear descriptions of methods.\nAuthors' conclusions\nIt is unclear which anaesthetic technique is superior - TIVA or inhalational - for transabdominal robotic assisted surgery in urology, gynaecology and gastroenterology, as existing evidence is scarce, is of low quality and has been generated from exclusively male patients undergoing robotic radical prostatectomy.\nAn ongoing trial, which includes participants of both genders with a focus on quality of recovery, might have an impact on future evidence related to this topic.\n\nGiven the provided abstract, please respond to the following query: \"Study funding sources\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Studies were funded by the institution, or the funding sources were not stated." } ]
query-laysum
6764
[ { "role": "user", "content": "Abstract: Background\nDifferentiating both typhoid (Salmonella Typhi) and paratyphoid (Salmonella Paratyphi A) infection from other causes of fever in endemic areas is a diagnostic challenge. Although commercial point-of-care rapid diagnostic tests (RDTs) for enteric fever are available as alternatives to the current reference standard test of blood or bone marrow culture, or to the widely used Widal Test, their diagnostic accuracy is unclear. If accurate, they could potentially replace blood culture as the World Health Organization (WHO)-recommended main diagnostic test for enteric fever.\nObjectives\nTo assess the diagnostic accuracy of commercially available rapid diagnostic tests (RDTs) and prototypes for detecting Salmonella Typhi or Paratyphi A infection in symptomatic persons living in endemic areas.\nSearch methods\nWe searched the Cochrane Infectious Diseases Group Specialized Register, MEDLINE, Embase, Science Citation Index, IndMED, African Index Medicus, LILACS, ClinicalTrials.gov, and the World Health Organization (WHO) International Clinical Trials Registry Platform (ICTRP) up to 4 March 2016. We manually searched WHO reports, and papers from international conferences on Salmonella infections. We also contacted test manufacturers to identify studies.\nSelection criteria\nWe included diagnostic accuracy studies of enteric fever RDTs in patients with fever or with symptoms suggestive of enteric fever living in endemic areas. We classified the reference standard used as either Grade 1 (result from a blood culture and a bone marrow culture) or Grade 2 (result from blood culture and blood polymerase chain reaction, or from blood culture alone).\nData collection and analysis\nTwo review authors independently extracted the test result data. We used a modified QUADAS-2 extraction form to assess methodological quality. We performed a meta-analysis when there were sufficient studies for the test and heterogeneity was reasonable.\nMain results\nThirty-seven studies met the inclusion criteria and included a total of 5080 participants (range 50 to 1732). Enteric fever prevalence rates in the study populations ranged from 1% to 75% (median prevalence 24%, interquartile range (IQR) 11% to 46%). The included studies evaluated 16 different RDTs, and 16 studies compared two or more different RDTs. Only three studies used the Grade 1 reference standard, and only 11 studies recruited unselected febrile patients. Most included studies were from Asia, with five studies from sub-Saharan Africa. All of the RDTs were designed to detect S.Typhi infection only.\nMost studies evaluated three RDTs and their variants: TUBEX in 14 studies; Typhidot (Typhidot, Typhidot-M, and TyphiRapid-Tr02) in 22 studies; and the Test-It Typhoid immunochromatographic lateral flow assay, and its earlier prototypes (dipstick, latex agglutination) developed by the Royal Tropical Institute, Amsterdam (KIT) in nine studies. Meta-analyses showed an average sensitivity of 78% (95% confidence interval (CI) 71% to 85%) and specificity of 87% (95% CI 82% to 91%) for TUBEX; and an average sensitivity of 69% (95% CI 59% to 78%) and specificity of 90% (95% CI 78% to 93%) for all Test-It Typhoid and prototype tests (KIT). Across all forms of the Typhidot test, the average sensitivity was 84% (95% CI 73% to 91%) and specificity was 79% (95% CI 70% to 87%). When we based the analysis on the 13 studies of the Typhidot test that either reported indeterminate test results or where the test format means there are no indeterminate results, the average sensitivity was 78% (95% CI 65% to 87%) and specificity was 77% (95% CI 66% to 86%). We did not identify any difference in either sensitivity or specificity between TUBEX, Typhidot, and Test-it Typhoid tests when based on comparison to the 13 Typhidot studies where indeterminate results are either reported or not applicable. If TUBEX and Test-it Typhoid are compared to all Typhidot studies, the sensitivity of Typhidot was higher than Test-it Typhoid (15% (95% CI 2% to 28%), but other comparisons did not show a difference at the 95% level of CIs.\nIn a hypothetical cohort of 1000 patients presenting with fever where 30% (300 patients) have enteric fever, on average Typhidot tests reporting indeterminate results or where tests do not produce indeterminate results will miss the diagnosis in 66 patients with enteric fever, TUBEX will miss 66, and Test-It Typhoid and prototype (KIT) tests will miss 93. In the 700 people without enteric fever, the number of people incorrectly diagnosed with enteric fever would be 161 with Typhidot tests, 91 with TUBEX, and 70 with Test-It Typhoid and prototype (KIT) tests. The CIs around these estimates were wide, with no difference in false positive results shown between tests.\nThe quality of the data for each study was evaluated using a standardized checklist called QUADAS-2. Overall, the certainty of the evidence in the studies that evaluated enteric fever RDTs was low.\nAuthors' conclusions\nIn 37 studies that evaluated the diagnostic accuracy of RDTs for enteric fever, few studies were at a low risk of bias. The three main RDT tests and variants had moderate diagnostic accuracy. There was no evidence of a difference between the average sensitivity and specificity of the three main RDT tests. More robust evaluations of alternative RDTs for enteric fever are needed.\n\nGiven the provided abstract, please respond to the following query: \"Study characteristics\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Cochrane researchers searched the available literature up to 4 March 2016 and included 37 studies. Most studies recruited participants from South Asia. Most participants were adults, with 22 studies including children. All of the RDTs evaluated detected Salmonella Typhi (typhoid fever) only." } ]
query-laysum
6765
[ { "role": "user", "content": "Abstract: Background\nAs part of liver transplantation, immunosuppression (suppressing the host immunity) is given to prevent graft rejections resulting from the immune response of the body against transplanted organ or tissues from a different person whose tissue antigens are not compatible with those of the recipient. The optimal maintenance immunosuppressive regimen after liver transplantation remains uncertain.\nObjectives\nTo assess the comparative benefits and harms of different maintenance immunosuppressive regimens in adults undergoing liver transplantation through a network meta-analysis and to generate rankings of the different immunosuppressive regimens according to their safety and efficacy.\nSearch methods\nWe searched CENTRAL, MEDLINE, Embase, Science Citation Index Expanded, World Health Organization International Clinical Trials Registry Platform, and trials registers until October 2016 to identify randomised clinical trials on immunosuppression for liver transplantation.\nSelection criteria\nWe included only randomised clinical trials (irrespective of language, blinding, or publication status) in adult participants undergoing liver transplantation (or liver retransplantation) for any reason. We excluded trials in which participants had undergone multivisceral transplantation or participants with established graft rejections. We considered any of the various maintenance immunosuppressive regimens compared with each other.\nData collection and analysis\nWe performed a network meta-analysis with OpenBUGS using Bayesian methods and calculated the odds ratio, rate ratio, and hazard ratio (HR) with 95% credible intervals (CrI) based on an available-case analysis, according to National Institute of Health and Care Excellence Decision Support Unit guidance.\nMain results\nWe included a total of 26 trials (3842 participants) in the review, and 23 trials (3693 participants) were included in one or more outcomes in the review. The vast majority of the participants underwent primary liver transplantation. All of the trials were at high risk of bias, and all of the evidence was of low or very low quality. In addition, because of sparse data involving trials at high risk of bias, it is not possible to entirely rely on the results of the network meta-analysis. The trials included mainly participants undergoing primary liver transplantation of varied aetiologies. The follow-up in the trials ranged from 3 to 144 months. The most common maintenance immunosuppression used as a control was tacrolimus. There was no evidence of difference in mortality (21 trials; 3492 participants) or graft loss (15 trials; 2961 participants) at maximal follow-up between the different maintenance immunosuppressive regimens based on the network meta-analysis. In the direct comparison, based on a single trial including 222 participants, tacrolimus plus sirolimus had increased mortality (HR 2.76, 95% CrI 1.30 to 6.69) and graft loss (HR 2.34, 95% CrI 1.28 to 4.61) at maximal follow-up compared with tacrolimus. There was no evidence of differences in the proportion of people with serious adverse events (1 trial; 719 participants), proportion of people with any adverse events (2 trials; 940 participants), renal impairment (8 trials; 2233 participants), chronic kidney disease (1 trial; 100 participants), graft rejections (any) (16 trials; 2726 participants), and graft rejections requiring treatment (5 trials; 1025 participants) between the different immunosuppressive regimens. The network meta-analysis showed that the number of adverse events was lower with cyclosporine A than with many other immunosuppressive regimens (12 trials; 1748 participants), and the risk of retransplantation (13 trials; 1994 participants) was higher with cyclosporine A than with tacrolimus (HR 3.08, 95% CrI 1.13 to 9.90). None of the trials reported number of serious adverse events, health-related quality of life, or costs.\nFunding: 14 trials were funded by pharmaceutical companies who would benefit from the results of the trial; two trials were funded by parties who had no vested interest in the results of the trial; and 10 trials did not report the source of funding.\nAuthors' conclusions\nBased on low-quality evidence from a single small trial from direct comparison, tacrolimus plus sirolimus increases mortality and graft loss at maximal follow-up compared with tacrolimus. Based on very low-quality evidence from network meta-analysis, we found no evidence of difference between different immunosuppressive regimens. We found very low-quality evidence from network meta-analysis and low-quality evidence from direct comparison that cyclosporine A causes more retransplantation compared with tacrolimus. Future randomised clinical trials should be adequately powered; performed in people who are generally seen in the clinic rather than in highly selected participants; employ blinding; avoid postrandomisation dropouts or planned cross-overs; and use clinically important outcomes such as mortality, graft loss, renal impairment, chronic kidney disease, and retransplantation. Such trials should use tacrolimus as one of the control groups. Moreover, such trials ought to be designed in such a way as to ensure low risk of bias and low risks of random errors.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Several medical drugs were compared in the trials. We found no evidence of difference in the risk of death or graft loss between the different immunosuppressive regimens based on the network meta-analysis. In the direct comparison, based on a single trial including 222 participants, the risk of death and graft loss was higher with tacrolimus plus sirolimus than with tacrolimus alone. There was no evidence of differences between the various immunosuppressive regimens in percentage of people who developed serious adverse events, percentage of people who developed any adverse events, risk of poor kidney function requiring dialysis or kidney transplantation (kidney dysfunction), prolonged kidney disease, graft rejections requiring treatment, and any graft rejections. The number of adverse events was lower with cyclosporine A than with many other immunosuppressive regimens. The risk of retransplantation was higher with cyclosporine A than with tacrolimus. None of the trials reported number of serious adverse events, health-related quality of life, or costs.\nThere is significant uncertainty as to the optimal maintenance immunosuppressive regimen after liver transplantation; further well-designed randomised clinical trials are required. Future trials should be performed in people who are generally seen in the clinic rather than in highly selected participants and report clinically important outcomes such as death, graft loss, kidney dysfunction, long-term kidney disease, and retransplantation. Such trials should use tacrolimus as one of the control groups. Moreover, such trials ought to be designed in such a way as to ensure low risk of bias and low risks of random errors." } ]
query-laysum
6766
[ { "role": "user", "content": "Abstract: Background\nThe Cystic Fibrosis Foundation recommends both short-term and long-acting insulin therapy when cystic fibrosis-related diabetes (CFRD) has been diagnosed. Diagnosis is based on: an elevated fasting blood glucose level greater than 6.94 mmol/L (125 mg/dL); or oral glucose tolerance tests greater than 11.11 mmol/L (200 mg/dL) at two hours; or symptomatic diabetes for random glucose levels greater than 11.11 mmol/L (200 mg/dL); or glycated hemoglobin levels of at least 6.5%. This is an update of a previously published review.\nObjectives\nTo establish the effectiveness of insulin and oral agents for managing diabetes in people with cystic fibrosis in relation to blood sugar levels, lung function and weight management.\nSearch methods\nWe searched the Cochrane Cystic Fibrosis and Genetic Disorders Group's Trials Register comprising references identified from comprehensive electronic database searches and handsearches of relevant journals and abstract books of conference proceedings. We also handsearched abstracts from pulmonary symposia and the North American Cystic Fibrosis Conferences.\nDate of most recent register search: 10 September 2020.\nWe searched online trials registries; date of most recent searches: 21 March 2020.\nSelection criteria\nRandomized controlled trials comparing all methods of pharmacological diabetes therapy in people with diagnosed CFRD.\nData collection and analysis\nTwo authors independently extracted data and assessed the risk of bias in the included studies. Authors also used GRADE to assess the quality of the evidence.\nMain results\nThe searches identified 29 trials (45 references). Four included trials provide results: one short-term single-center cross-over trial (seven adults) comparing insulin with oral repaglinide and no medication in adults with CFRD and normal fasting glucose; one long-term multicenter trial (61 adults with CFRD) comparing insulin with oral repaglinide and placebo; one long-term multicenter trial (67 adults) comparing insulin with oral repaglinide; and one 12-week single-center cross-over trial (20 adults) comparing the long-acting insulin glargine to short-term neutral protamine Hagedorn insulin. Two ongoing trials of newly approved incretin mimics have been noted for possible future inclusion.\nDowngrading of the quality of the evidence was mainly due to risks of bias across all domains, but particularly due to concerns surrounding allocation concealment and selective reporting. There were also some concerns due to imprecision from small sample sizes and low event rates. Finally, there may be some bias due to the amounts of insulin and repaglinide given not being comparable.\nData from one trial comparing insulin to placebo (39 participants) did not show any difference between groups for the primary outcomes of blood glucose levels (very low-quality evidence), lung function (low-quality evidence) or nutritional status (low-quality evidence). Similarly, no differences between groups were seen for the secondary outcomes of number of hypoglycemic episodes (low-quality evidence), secondary infection complications or quality of life (QoL). These results were mirrored in the narrative reports for the second trial in this comparison (seven participants).\nData from the one-year trial comparing repaglinide to placebo (38 participants), showed no differences between groups for the primary outcomes of blood glucose levels (very low-quality evidence), lung function (low-quality evidence) and nutritional status (low-quality evidence). Also, no differences were seen between groups for the secondary outcomes of number of hypoglycemic episodes (low-quality evidence), secondary infection complications or QoL. These findings were mirrored in the narrative reports for the second trial (n = 7) in this comparison.\nThree trials compared insulin to repaglinide (119 participants). Data from one trial (n = 67) showed no difference in blood glucose levels at either 12 months (high-quality evidence) or 24 months; narrative reports from one trial (45 participants) reported no difference between groups, but the second trial (7 participants) reported a beneficial effect of insulin over repaglinide. Two trials (112 participants) found no difference between insulin and repaglinide in lung function or nutritional status (moderate-quality evidence). Two trials (56 participants) reported no difference in the number of hypoglycemic episodes (low-quality evidence). One trial (45 participants) reported no difference between groups in secondary infections and cystic fibrosis QoL.\nThe single trial comparing glargine to neutral protamine Hagedorn insulin did not report directly on the review's primary outcomes, but did report no differences between groups in post-prandial glucose values and weight; neither group reported infectious complications. There was no difference in episodes of hypoglycemia (very low-quality evidence) and while there was no difference reported in QoL, all participants opted to continue treatment with glargine after the trial was completed.\nMortality was not reported by any trial in any comparison, but death was not given as a reason for withdrawal in any trial.\nAuthors' conclusions\nThis review has not found any conclusive evidence that any agent has a distinct advantage over another in controlling hyperglycemia or the clinical outcomes associated with CFRD. Given the treatment burden already experienced by people with cystic fibrosis, oral therapy may be a viable treatment option.\nWhile some cystic fibrosis centers use oral medications to help control diabetes, the Cystic Fibrosis Foundation (USA) clinical practice guidelines support the use of insulin therapy and this remains the most widely-used treatment method. Randomized controlled trials specifically related to controlling diabetes and its impact on the course of pulmonary disease process in cystic fibrosis continue to be a high priority. Specifically, investigators should evaluate adherence to different therapies and also whether there is benefit in using additional hypoglycemic agents as well as the newly approved incretin mimics. Agents that potentiate insulin action, especially agents with additional anti-inflammatory potential should also be further investigated as adjuvant therapy to insulin.\n\nGiven the provided abstract, please respond to the following query: \"Key results\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "We found results for our primary outcomes of blood glucose levels, lung function and nutritional status and our secondary outcomes of pulmonary exacerbations (flare up of disease), complications and quality of life. We could only analyse limited results and did not find evidence that showed that any one treatment was better than another. Our analysis did not show that either insulin or repaglinide made any difference to any of our outcomes compared to placebo or no treatment at any time point. Similarly, most of our results did not show whether insulin or repaglinide was better, but the evidence was a little stronger that there was no difference in lung function, nutritional status or blood glucose levels at 12 months. Similarly, the evidence was not strong enough to show if there is any difference between glargine and neutral protamine Hagedorn insulin for any of our outcome measures. None of the trials reported if there were any deaths.\nGiven the treatment burden already experienced by people with CF, tablets may be a welcome alternative to injecting insulin. Longer-term trials with larger numbers of people are still needed to see how controlling CFRD affects lung function. Investigators should also assess how many people keep to their treatment plan when comparing insulin to tablets. Research should look at using the drugs by themselves or together with insulin to enhance insulin action, especially those agents with additional anti-inflammatory potential." } ]
query-laysum
6767
[ { "role": "user", "content": "Abstract: Background\nOn the American continent, cutaneous and mucocutaneous leishmaniasis (CL and MCL) are diseases associated with infection by several species of Leishmania parasites. Pentavalent antimonials remain the first-choice treatment. There are alternative interventions, but reviewing their effectiveness and safety is important as availability is limited. This is an update of a Cochrane Review first published in 2009.\nObjectives\nTo assess the effects of interventions for all immuno-competent people who have American cutaneous and mucocutaneous leishmaniasis (ACML).\nSearch methods\nWe updated our database searches of the Cochrane Skin Group Specialised Register, CENTRAL, MEDLINE, Embase, LILACS and CINAHL to August 2019. We searched five trials registers.\nSelection criteria\nRandomised controlled trials (RCTs) assessing either single or combination treatments for ACML in immuno-competent people, diagnosed by clinical presentation and Leishmania infection confirmed by smear, culture, histology, or polymerase chain reaction on a biopsy specimen. The comparators were either no treatment, placebo only, or another active compound.\nData collection and analysis\nWe used standard methodological procedures expected by Cochrane. Our key outcomes were the percentage of participants 'cured' at least three months after the end of treatment, adverse effects, and recurrence. We used GRADE to assess evidence certainty for each outcome.\nMain results\nWe included 75 studies (37 were new), totalling 6533 randomised participants with ATL. The studies were mainly conducted in Central and South America at regional hospitals, local healthcare clinics, and research centres. More male participants were included (mean age: roughly 28.9 years (SD: 7.0)). The most common confirmed species were L. braziliensis, L. panamensis, and L. mexicana. The most assessed interventions and comparators were non-antimonial systemics (particularly oral miltefosine) and antimonials (particularly meglumine antimoniate (MA), which was also a common intervention), respectively.\nThree studies included moderate-to-severe cases of mucosal leishmaniasis but none included cases with diffuse cutaneous or disseminated CL, considered the severe cutaneous form. Lesions were mainly ulcerative and located in the extremities and limbs. The follow-up (FU) period ranged from 28 days to 7 years. All studies had high or unclear risk of bias in at least one domain (especially performance bias). None of the studies reported the degree of functional or aesthetic impairment, scarring, or quality of life.\nCompared to placebo, at one-year FU, intramuscular (IM) MA given for 20 days to treat L. braziliensis and L. panamensis infections in ACML may increase the likelihood of complete cure (risk ratio (RR) 4.23, 95% confidence interval (CI) 0.84 to 21.38; 2 RCTs, 157 participants; moderate-certainty evidence), but may also make little to no difference, since the 95% CI includes the possibility of both increased and reduced healing (cure rates), and IMMA probably increases severe adverse effects such as myalgias and arthralgias (RR 1.51, 95% CI 1.17 to 1.96; 1 RCT, 134 participants; moderate-certainty evidence). IMMA may make little to no difference to the recurrence risk, but the 95% CI includes the possibility of both increased and reduced risk (RR 1.79, 95% CI 0.17 to 19.26; 1 RCT, 127 participants; low-certainty evidence).\nCompared to placebo, at six-month FU, oral miltefosine given for 28 days to treat L. mexicana, L. panamensis and L. braziliensis infections in American cutaneous leishmaniasis (ACL) probably improves the likelihood of complete cure (RR 2.25, 95% CI 1.42 to 3.38), and probably increases nausea rates (RR 3.96, 95% CI 1.49 to 10.48) and vomiting (RR 6.92, 95% CI 2.68 to 17.86) (moderate-certainty evidence). Oral miltefosine may make little to no difference to the recurrence risk (RR 2.97, 95% CI 0.37 to 23.89; low-certainty evidence), but the 95% CI includes the possibility of both increased and reduced risk (all based on 1 RCT, 133 participants).\nCompared to IMMA, at 6 to 12 months FU, oral miltefosine given for 28 days to treat L. braziliensis, L. panamensis, L. guyanensis and L. amazonensis infections in ACML may make little to no difference to the likelihood of complete cure (RR 1.05, 95% CI 0.90 to 1.23; 7 RCTs, 676 participants; low-certainty evidence). Based on moderate-certainty evidence (3 RCTs, 464 participants), miltefosine probably increases nausea rates (RR 2.45, 95% CI 1.72 to 3.49) and vomiting (RR 4.76, 95% CI 1.82 to 12.46) compared to IMMA. Recurrence risk was not reported.\nFor the rest of the key comparisons, recurrence risk was not reported, and risk of adverse events could not be estimated.\nCompared to IMMA, at 6 to 12 months FU, oral azithromycin given for 20 to 28 days to treat L. braziliensis infections in ACML probably reduces the likelihood of complete cure (RR 0.51, 95% CI 0.34 to 0.76; 2 RCTs, 93 participants; moderate-certainty evidence).\nCompared to intravenous MA (IVMA) and placebo, at 12 month FU, adding topical imiquimod to IVMA, given for 20 days to treat L. braziliensis, L. guyanensis and L. peruviana infections in ACL probably makes little to no difference to the likelihood of complete cure (RR 1.30, 95% CI 0.95 to 1.80; 1 RCT, 80 participants; moderate-certainty evidence).\nCompared to MA, at 6 months FU, one session of local thermotherapy to treat L. panamensis and L. braziliensis infections in ACL reduces the likelihood of complete cure (RR 0.80, 95% CI 0.68 to 0.95; 1 RCT, 292 participants; high-certainty evidence).\nCompared to IMMA and placebo, at 26 weeks FU, adding oral pentoxifylline to IMMA to treat CL (species not stated) probably makes little to no difference to the likelihood of complete cure (RR 0.86, 95% CI 0.63 to 1.18; 1 RCT, 70 participants; moderate-certainty evidence).\nAuthors' conclusions\nEvidence certainty was mostly moderate or low, due to methodological shortcomings, which precluded conclusive results. Overall, both IMMA and oral miltefosine probably result in an increase in cure rates, and nausea and vomiting are probably more common with miltefosine than with IMMA.\nFuture trials should investigate interventions for mucosal leishmaniasis and evaluate recurrence rates of cutaneous leishmaniasis and its progression to mucosal disease.\n\nGiven the provided abstract, please respond to the following query: \"What this means\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "The main findings of this review suggest that:\n- MA and miltefosine probably increase chances of complete healing; and\n- vomiting or nausea are probably more common with miltefosine, and joint or muscle ache is probably more common with MA.\nThe evidence was mostly of moderate certainty, so the true results are likely close to what we found. Evidence was limited by the inclusion of very few people in some studies, and participants or investigators knowing which treatments they were receiving." } ]
query-laysum
6768
[ { "role": "user", "content": "Abstract: Background\nTyphoid and paratyphoid are febrile illnesses, due to a bacterial infection, which remain common in many low- and middle-income countries. The World Health Organization (WHO) currently recommends the fluoroquinolone antibiotics in areas with known resistance to the older first-line antibiotics.\nObjectives\nTo evaluate fluoroquinolone antibiotics for treating children and adults with enteric fever.\nSearch methods\nWe searched The Cochrane Infectious Disease Group Specialized Register (February 2011); Cochrane Central Register of Controlled Trials (CENTRAL), published in The Cochrane Library (2011, Issue 2); MEDLINE (1966 to February 2011); EMBASE (1974 to February 2011); and LILACS (1982 to February 2011). We also searched the metaRegister of Controlled Trials (mRCT) in February 2011.\nSelection criteria\nRandomized controlled trials examining fluoroquinolone antibiotics, in people with blood, stool or bone marrow culture-confirmed enteric fever.\nData collection and analysis\nTwo authors independently assessed the trial's methodological quality and extracted data. We calculated risk ratios (RR) for dichotomous data and mean difference for continuous data with 95% confidence intervals (CI).\nComparative effectiveness has been interpreted in the context of; length of treatment, dose, year of study, known levels of antibiotic resistance, or proxy measures of resistance such as the failure rate in the comparator arm.\nMain results\nTwenty-six studies, involving 3033 patients, are included in this review.\nFluoroquinolones versus older antibiotics (chloramphenicol, co-trimoxazole, amoxicillin and ampicillin)\nIn one study from Pakistan in 2003-04, high clinical failure rates were seen with both chloramphenicol and co-trimoxazole, although resistance was not confirmed microbiologically. A seven-day course of either ciprofloxacin or ofloxacin were found to be superior. Older studies of these comparisons failed to show a difference (six trials, 361 participants).\nIn small studies conducted almost two decades ago, the fluoroquinolones were demonstrated to have fewer clinical failures than ampicillin and amoxicillin (two trials, 90 participants, RR 0.11, 95% CI 0.02 to 0.57).\nFluoroquinolones versus current second-line options (ceftriaxone, cefalexin, and azithromycin)\nThe two studies comparing a seven day course of oral fluoroquinolones with three days of intravenous ceftriaxone were too small to detect important differences between antibiotics should they exist (two trials, 89 participants).\nIn Pakistan in 2003-04, no clinical or microbiological failures were seen with seven days of either ciprofloxacin, ofloxacin or cefixime (one trial, 139 participants). In Nepal in 2005, gatifloxacin reduced clinical failure and relapse compared to cefixime, despite a high prevalence of NaR in the study population (one trial, 158 participants, RR 0.04, 95% CI 0.01 to 0.31).\nCompared to a seven day course of azithromycin, a seven day course of ofloxacin had a higher rate of clinical failures in populations with both multi-drug resistance (MDR) and nalidixic acid resistance (NaR) enteric fever in Vietnam in 1998-2002 (two trials, 213 participants, RR 2.20, 95% CI 1.23 to 3.94). However, a more recent study from Vietnam in 2004-05, detected no difference between gatifloxacin and azithromycin with both drugs performing well (one trial, 287 participants).\nAuthors' conclusions\nGenerally, fluoroquinolones performed well in treating typhoid, and maybe superior to alternatives in some settings. However, we were unable to draw firm general conclusions on comparative contemporary effectiveness given that resistance changes over time, and many studies were small. Policy makers and clinicians need to consider local resistance patterns in choosing a fluoroquinolone or alternative.\nThere is some evidence that the newest fluoroquinolone, gatifloxacin, remains effective in some regions where resistance to older fluoroquinolones has developed. However, the different fluoroquinolones have not been compared directly in trials in these settings.\n\nGiven the provided abstract, please respond to the following query: \"What the research says\". Use language that can be easily comprehended by individuals without domain knowledge:\n" }, { "role": "assistant", "content": "Effect of using fluoroquinolones:\nGenerally, fluoroquinolones are effective in typhoid.\nPolicy makers and clinicians will need to consider local antibiotic resistance when considering treatment options for enteric fever.\nOne relatively new fluoroquinolone, gatifloxacin, seems to remain effective in some regions where resistance to older fluoroquinolones has developed." } ]