instruction
stringlengths
10
664
context
stringlengths
1
5.66k
response
stringlengths
1
3.34k
category
stringclasses
1 value
Mesenchymal stem cell implantation in osteoarthritic knees: is fibrin glue effective as a scaffold?
The cell-based tissue engineering approach that uses mesenchymal stem cells (MSCs) has addressed the issue of articular cartilage repair in osteoarthritic (OA) knees. However, to improve outcomes, an advanced surgical procedure with tissue-engineered scaffolds may be needed to treat patients with large cartilage lesions. To investigate the clinical and second-look arthroscopic outcomes of the implantation of MSCs loaded in fibrin glue as a scaffold in patients with OA knees and to compare these outcomes with those of MSC implantation without a scaffold. Cohort study; Level of evidence, 3. This study retrospectively evaluated 54 patients (56 knees) who were examined with second-look arthroscopy after MSC implantation for cartilage lesions in their OA knees. Patients were divided into 2 groups: 37 patients (39 knees) were treated with MSC implantation without a scaffold (group 1), and 17 patients (17 knees) underwent implantation of MSCs loaded in fibrin glue as a scaffold (group 2). Clinical outcomes were evaluated according to the International Knee Documentation Committee (IKDC) score and the Tegner activity scale, and cartilage repair was assessed with the International Cartilage Repair Society (ICRS) grade. Statistical analyses were performed to identify various prognostic factors associated with the clinical and second-look arthroscopic outcomes. At final follow-up (mean, 28.6 months; range, 24-34 months), the mean IKDC score and Tegner activity scale in each group significantly improved: group 1, from 38.1±7.7 to 62.0±11.7 (IKDC) and from 2.5±0.9 to 3.5±0.8 (Tegner); group 2, from 36.1±6.2 to 64.4±11.5 (IKDC) and from 2.2±0.8 to 3.8±0.8 (Tegner) (P<.001 for all). According to the overall ICRS cartilage repair grades, 9 of the 39 lesions (23%) in group 1 and 12 of the 17 lesions (58%) in group 2 achieved a grade of I or II. There was a significant difference in ICRS grades between the groups (P=.028). Overweight (body mass index≥27.5 kg/m2) and large lesion size (≥5.7 cm2) were significant predictors of poor clinical and arthroscopic outcomes in group 1 (P<.05 for both). There was a similar trend in group 2, but the differences were not significant, possibly owing to the smaller sample size.
Clinical and arthroscopic outcomes of MSC implantation were encouraging for OA knees in both groups, although there were no significant differences in outcome scores between groups. However, at second-look arthroscopy, there were better ICRS grades in group 2.
closed_qa
Does shoulder impingement syndrome affect the shoulder kinematics and associated muscle activity in archers?
Archery related injuries, such as shoulder impingement syndrome are caused by repeated motion of the shoulder. The aim of this study was to analyze differences in the shoulder kinematics and the associated muscle activity between archers with shoulder impingement and uninjured archery players. Thirty male archers, who were divided into an impingement group and an uninjured group, were included in this study. The angle of scapular elevation, shoulder joint abduction, horizontal extension, and elbow joint flexion as well as the electromyographic activity of the upper trapezius, lower trapezius, deltoid middle, deltoid posterior, biceps brachii, and triceps brachii muscles at the point of stabilization during shooting were measured. Variables differing between impingement and uninjured groups were identified, and a stepwise regression analysis was performed to identify a combination of variables that effectively impingement syndrome. The results indicated that the angle of scapular elevation was significantly greater than that uninjured group (P<0.05). The angle of horizontal extension in the impingement group was significantly smaller than that in the uninjured group (P<0.05). The angle of elbow flexion in the impingement group was significantly smaller than that in the uninjured group (P<0.05). The levels of upper trapezius and deltoid middle muscle activity were significantly higher in the impingement group, while the level of lower trapezius muscle activity was significantly lower (P<0.05) when compared to the uninjured group. The impingement group had a greater angle of scapular elevation, smaller angle of horizontal extension, smaller angle of elbow flexion, higher the levels of upper trapezius, lower the levels of lower trapezius, higher deltoid middle muscle activity and higher UT/LT ratio (all differences were significant). A logistic model for predicting impingement syndrome showed that UT/LT ratio was significantly related impingement syndrome (P<0.05).
The authors concluded that archers with shoulder impingement syndrome exhibit different kinematics and muscle activity compared to uninjured archers. Therefore, in order to prevent shoulder joint impingement during archery, training is necessary what can make lower trapezius muscle activity increased to decrease the UT/LT ratio.
closed_qa
Are speed cameras able to reduce traffic noise disturbances?
Disturbance by traffic noise can result in health problems in the long run. However, the subjective perception of noise plays an important role in their development. The aim of this study was to determine if speed cameras are able to reduce subjective traffic noise disturbance of residents of high-traffic roads in Luebeck? In August 2012 a speed camera has been installed in 2 high-traffic roads in Luebeck (IG). Residents living 1.5 km in front of the installed speed cameras and behind them received a postal questionnaire to evaluate their subjective noise perception before (t0), 8 weeks (t1) and 12 months (t2) after the installation of the speed camera. As controls (CG) we asked residents of another high-traffic road in Luebeck without speed cameras and residents of 2 roads with several consecutive speed cameras installed a few years ago. Furthermore, objective measures of the traffic noise level were conducted. Response rates declined from 35.9% (t0) to 27.2% (t2). The proportion of women in the CG (61.4-63.7%) was significantly higher than in the IG (53.7-58.1%, p<0.05), and responders were significantly younger (46.5±20.5-50±22.0 vs. 59.1±17.0-60.5±16.9 years, p<0.05). A reduction of the perceived noise disturbance of 0.2 point, measured on a scale from 0 (no disturbance) to 10 (heavy disturbance), could be observed in both IG and CG. Directly asked, 15.2% of the IG and 19.3% of the CG reported a traffic noise reduction at t2. The objective measure shows a mean reduction of 0.6 dB at t1.
The change of noise level of 0.6 dB, which could only be experienced by direct comparison, is in line with the subjective noise perception. As sole method to reduce traffic noise (and for health promotion) a speed camera is insufficient.
closed_qa
Does cesarean section before the scheduled date increase the risk of neonatal morbidity?
Previous studies led to the recommendation to schedule planned elective cesarean deliveries at or after 39 weeks of gestation and not before 38 weeks. The question is whether this practice is appropriate in face of possible risks to the newborn should the pregnancy have to be ended by cesarean section before the scheduled date. To compare the outcomes of newborn infants who were delivered on their scheduled day by elective cesarean section versus those who required delivery earlier. This single-center retrospective study was based on medical records covering a period of 18 months. We compared the neonatal outcomes of 272 infants delivered by elective cesarean section as scheduled (at 38.8 +/- 0.8 weeks gestation)and 44 infants who had to be delivered earlier than planned j(at 37.9 +/- 1.1 weeks). We found no morbidity directly related to delivery by cesarean section before the scheduled date. There were no significant differences in the need for resuscitation after delivery. Although more of the infants who were delivered early were admitted to intensive care and overall stayed longer in the hospital (5.8 +/- 7.3 vs. 3.9 +/- 0.8 days, P<0.02), their more severe respiratory illness and subsequent longer hospitalization was the result of their younger gestational age. Transient tachypnea of the newborn was associated with younger gestational age at delivery in both groups.
We suggest continuing with the current recommendation to postpone elective cesarean singleton deliveries beyond 38-39 weeks of gestation whenever possible.
closed_qa
Are physician pagers an outmoded technology?
Pagers are the most commonly used method of communications in American hospitals. However, its financial cost and efficiency is unknown. To evaluate the efficiency of conventional hospital pagers and to estimate the financial cost of time wasted by the use of these pagers. We conducted a survey among 100 clinicians, nurses and pharmacists in our community teaching hospital, estimating the time spent in the process of sending and responding to pages and the financial equivalent of this time, and evaluating the potential advantages of hospital-based wireless telephones compared with traditional pagers. A total of 70 clinicians completed the survey for a response rate of 70%. The average time spent per daytime shift in using the paging system was between 48 and 66 minutes for physicians, 120 minutes for nurses and 165 minutes for pharmacists. The financial cost of time lost for a single medical ward for one month was estimated to be $2,732-$17,250, depending on the case scenario.
Our study suggests that the traditional paging system is an inefficient means of communication between clinicians and hospital staff and that a switch to direct phone calls might be far more cost-effective. Similar considerations probably apply to most hospitals that still use traditional pagers.
closed_qa
Does a Nintendo Wii exercise program provide similar exercise demands as a traditional pulmonary rehabilitation program in adults with COPD?
The chronic obstructive pulmonary disease (COPD) population can experience lower activity and fitness levels than the non-COPD population. The Nintendo Wii may be an appropriate at-home training device for the COPD population, which could be used as a supplement for a pulmonary rehabilitation program. This study was a randomized, within-subject, cross-over study involving 10 adults with COPD previously enrolled in St Paul's Hospital's pulmonary rehabilitation program. This study attempted to determine if specific Wii activities resulted in similar energy expenditures to that of a more traditional pulmonary rehabilitation activity. Participants completed two 15-min exercise interventions in a single session, with a washout period of 30 min in-between. The interventions were an experimental Wii intervention and a traditional treadmill intervention. There was no significant difference in total energy expenditure between the two 15-min exercise interventions [mean difference 36.3 joules; 95% confidence interval (CI): 31.4, 104]. There was no significant difference in heart rate (mean difference -0.167 beats per minute; 95% CI: -4.83, 4.50), rating of perceived exertion (mean difference 0.100; 95% CI: -0.416, 0.616) and Borg dyspnea scale (mean difference 0.267; 95% CI: -0.004, 0.537) between the two 15-min exercise interventions. There was a significant difference in SpO2 between the two 15-min exercise interventions (Wii intervention mean difference 2.33% > treadmill intervention; 95% CI: 1.52, 3.15).
Gaming technology can provide an exercise program that has similar cardiovascular demands to traditional pulmonary rehabilitation programs for patients with COPD. Further research is necessary to address feasibility and long-term adherence.
closed_qa
Could light meal jeopardize laboratory coagulation tests?
Presently the necessity of fasting time for coagulation tests is not standardized. Our hypothesis is that this can harm patient safety. This study is aimed at evaluating whether a light meal (i.e. breakfast) can jeopardize laboratory coagulation tests. A blood sample was firstly collected from 17 fasting volunteers (12 h). Immediately after blood collection, the volunteers consumed a light meal. Then samples were collected at 1, 2 and 4 h after the meal. Coagulation tests included: activated partial thromboplastin time (APTT), prothrombin time (PT), fibrinogen (Fbg), antithrombin III (AT), protein C (PC) and protein S (PS). Differences between samples were assessed by Wilcoxon ranked-pairs test. The level of statistical significance was set at P<0.05. Mean % differences were determined and differences between and baseline and 1, 2 and 4h samples were compared with reference change value (RCV). A significantly higher % activity of AT was observed at 1 h and 4 h after meal vs. baseline specimen [113 (104-117) and 111 (107-120) vs. 109 (102-118), respectively; P = 0.029 and P = 0.016]. APTT at 2 h was found significantly lower than baseline samples [32.0 (29.9-34.8) vs. 34.1 (32.2-35.2), respectively; P = 0.041]. The results of both Fbg and PS tests were not influenced by a light meal. Furthermore, no coagulation tests had significant variation after comparison with RCV.
A light meal does not influence the laboratory coagulation tests we assessed, but we suggest that the laboratory quality managers standardize the fasting time for all blood tests at 12 hours, to completely metabolize the lipids intake.
closed_qa
Does hypericin boost the efficacy of high-power laser?
Lasers are widely used in treating symptomatic benign prostatic hyperplasia. In current practice, potassium titanyl phosphate (KTP) lasers are the most common type of laser systems used. The aim here was to evaluate the rapid effect of high-power laser systems after application of hypericin. Experimental animal study conducted in the Department of Urology, Gülhane Military Medical Academy, Ankara, Turkey, in 2012. Sixteen rats were randomized into four groups: 120 W KTP laser + hypericin; 120 W KTP laser alone; 80 W KTP laser + hypericin; and 80 W KTP laser alone. Hypericin was given intraperitoneally two hours prior to laser applications. The laser incisions were made through the quadriceps muscle of the rats. The depth and the width of the laser incisions were evaluated histologically and recorded. To standardize the effects of the laser, we used the ratio of depth to width. These new values showed us the depth of the laser application per unit width. The new values acquired were evaluated statistically. Mean depth/width values were 231.6, 173.6, 214.1 and 178.9 in groups 1, 2, 3 and 4, respectively. The most notable result was that higher degrees of tissue penetration were achieved in the groups with hypericin (P<0.05).
The encouraging results from our preliminary study demonstrated that hypericin may improve the effects of KTP laser applications.
closed_qa
Do value thresholds for oncology drugs differ from nononcology drugs?
In the past decade, many oncologic drugs have been approved that extend life and/or improve patients' quality of life. However, new cancer drugs are often associated with high price and increased medical spending. For example, in 2010, the average annual cost of care for breast cancer in the final stage of disease was reported to be $94,284, and the total estimated cost in the United States was $16.50 billion. To determine whether value threshold, as defined by the incremental cost-effectiveness ratio (ICER), differed between oncology and other therapeutic areas. The PubMed database was searched for articles published between January 2003 and December 2013 with calculated ICER for therapeutic drug entities in a specific therapeutic area. The search term used was "ICER" and "United States." From 275 results, only those articles that reported ICERs using quality-adjusted life-years (QALY) were included. In addition, only those articles that used a U.S. payer perspective were retained. Among those, nondrug therapy articles and review articles were excluded. The mean ICER and value threshold for oncologic drugs and nononcologic drugs were evaluated for the analysis. From 54 articles selected for analysis, 13 pertained to drugs in oncology therapeutics, and the remaining 41 articles addressed ICER for drugs in other therapeutic areas. The mean and median of ICERs calculated for cancer-specific drug intervention was $138,582/QALY and $55,500/QALY, respectively, compared with $49,913/QALY and $31,000/QALY, respectively, for noncancer drugs. Among the cancer drugs, 45.0% had ICERs below $50,000/QALY and 70.0% below $100,000/QALY. In comparison, 72.0% of noncancer drugs showed ICERs below $50,000/QALY, and 90.0% had ICERs below $100,000/QALY. When a specific threshold was mentioned, it was in the range of $100,000-$150,000 in cancer drugs, whereas drugs in other therapeutic areas used traditional threshold value within the range of $50,000-$100,000.
The average ICER reported for cancer drugs was more than 2-fold greater than the average ICER for noncancer drugs. In general, articles that addressed the relative value of oncologic pharmaceuticals used higher value thresholds and reported higher ICERs than articles evaluating noncancer drugs.
closed_qa
A case study in generic drug use: should there be risk adjustment in incentive payments for the use of generic medications?
Encouraging generic drug use has reduced health care costs for payers and consumers, but the availability of therapeutically interchangeable medications or generic medications of choice is not equal across disease states. The extent to which systems of care are able to substitute with generics is not well understood. To (a) define and measure the maximum generic rate (MGR) of currently prescribed drugs within an academic medical group in and (b) illustrate differences across drugs associated with selected underlying diseases. Prescription claims data were examined from an academic medical group in Chicago, Illinois. Based on pharmacologic and therapeutic criteria, drugs were classified into 2 categories-potentially substitutable and not potentially substitutable-based on whether the drugs are branded forms of the same chemical entities that are available as generics or are therapeutically interchangeable with other medications that have different chemical compositions but the same mechanisms of action and potential efficacy. A medication was considered potentially substitutable if it (a) did not have a narrow therapeutic index as defined by the FDA; (b) did not belong to 1 of 6 protected classes of drugs in the Medicare D provisions; (c) was substitutable with a generic medication containing the same chemical entity; or (d) was therapeutically interchangeable with a therapeutically equivalent medication. MGR was defined as the percentage of prescriptions that could potentially be prescribed in generic form. This rate was examined overall and across drugs known to be associated with illustrative diseases including hypertension, diabetes mellitus, and obstructive lung diseases. The MGR ranged from 100% for drugs used in hypertension to 26.7% for drugs used in obstructive lung diseases. The MGR was 83.6%.
Payers wishing to promote generic substitution should incorporate the potential for substitution of clinically appropriate generic medications as part of incentives for generic utilization to avoid unintended consequences of using a fixed target rate. A practical methodology for determining an MGR is offered.
closed_qa
Is there an association between the high-risk medication star ratings and member experience CMS star ratings measures?
Methods to achieve high star ratings for the High-Risk Medication (HRM) measure are thought to result in unintended consequences and to compromise several member experience measures that ultimately put at risk the plan sponsor's Medicare Part D Centers for Medicare Medicaid (CMS) star rating. To determine if HRM scores are associated with relevant member experience measure scores. This is a cross-sectional analysis utilizing CMS 2013 and 2014 plan star ratings reports (2011 and 2012 benefit year data) for Medicare Advantage prescription drug (MA-PD) plans and prescription drug plans (PDPs). Medicare contracts with complete data for all measures of interest in 2013 and 2014 star ratings reports were included (N = 443). Bivariate linear regressions were performed for each of 2 independent variables: (1) 2014 HRM score and (2) 2013 to 2014 change in HRM score. Dependent variables were the 2014 scores for "Getting Needed Prescription Drugs," "Complaints about Drug Plan," "Rating of Drug Plan," and "Members Choosing to Leave the Plan." The bivariate linear regressions demonstrated weak positive associations between the 2014 HRM score and each of the 2014 member experience measures that explained 0.5% to 4% (R2) of variance of these measures. The bivariate regressions for the 2013 to 2014 change in the HRM score and 2014 member experience measures of interest demonstrated associations accounting for 1% to 8% of variance (R2). The greatest associations were observed between each independent variable and the 2014 "Getting Needed Prescription Drugs" score with correlation coefficients of 0.21 and 0.29.
HRM star ratings and change in HRM star ratings are weakly correlated with member experience measures in concurrent measurement periods. Plan sponsors may be more aggressive in HRM utilization management, since it is unlikely to negatively impact CMS summary star ratings.
closed_qa
Does nonsurgical treatment improve longitudinal outcomes of lateral epicondylitis over no treatment?
Lateral epicondylitis is a painful tendinopathy for which several nonsurgical treatment strategies are used. Superiority of these nonsurgical treatments over nontreatment has not been definitively established.QUESTIONS/ We asked whether nonsurgical treatment of lateral epicondylitis compared with observation only or placebo provides (1) better overall improvement, (2) less need for escape interventions, (3) better outcome scores, and (4) improved grip strength at intermediate- to long-term followup. The English-language literature was searched using PubMed and the Cochrane Central Register of Controlled Trials. Randomized-controlled trials (RCTs) comparing any form of nonsurgical treatment with either observation only or placebo at followup of at least 6 months were included. Nonsurgical treatments included injections (corticosteroid, platelet-rich plasma, autologous blood, sodium hyaluronate, or glycosaminoglycan polysulfate), physiotherapy, shock wave therapy, laser, ultrasound, corticosteroid iontophoresis, topical glyceryl trinitrate, or oral naproxen. Methodologic quality was assessed with the Consolidated Standards of Reporting Trials (CONSORT) checklist, and 22 RCTs containing 2280 patients were included. Pooled analyses were performed to evaluate overall improvement; requirement for escape interventions (treatment of any kind, outside consultation, and surgery); outcome scores (Patient-Rated Tennis Elbow Evaluation [PRTEE]; DASH; Pain-Free Function Index [PFFI]; EuroQoL [EQ]-5D; and overall function); and maximum and pain-free grip strength. Sensitivity analyses were performed using only trials of excellent or good quality. Heterogeneity analyses were performed, and funnel plots were constructed to assess for publication bias. Nonsurgical treatment was not favored over nontreatment based on overall improvement (risk ratio [RR] = 1.05 [0.96-1.15]; p = 0.32), need for escape treatment (RR = 1.50 [0.84-2.70]; p = 0.17), PRTEE scores (mean difference [MD]= 1.47, [0.68-2.26]; p<0.001), DASH scores (MD = -2.69, [-15.80 to 10.42]; p = 0.69), PFFI scores (standardized mean difference [SMD]= 0.25, [-0.32 to 0.81]; p = 0.39), overall function using change-from-baseline data (SMD = 0.11, [-0.14 to 0.36]; p = 0.37) and final data (SMD = -0.16, [-0.79 to 0.47]; p = 0.61), EQ-5D scores (SMD = 0.08, [-0.52 to 0.67]; p = 0.80), maximum grip strength using change-from-baseline data (SMD = 0.12, [-0.11 to 0.35]; p = 0.31) and final data (SMD = 4.37, [-0.65 to 9.38]; p = 0.09), and pain-free grip strength using change-from-baseline data (SMD = -0.20, [-0.84 to 0.43]; p = 0.53) and final data (SMD = -0.03, [-0.61 to 0.54]; p = 0.91).
Pooled data from RCTs indicate a lack of intermediate- to long-term clinical benefit after nonsurgical treatment of lateral epicondylitis compared with observation only or placebo.
closed_qa
KIR genotype distribution among symptomatic patients with and without Helicobacter pylori infection: is there any role for the B haplotype?
Contact of peripheral blood lymphocytes with Helicobacter pylori was proved to induce non- major histocompatibility complex-restricted cytotoxicity and natural killer cells are thought to play an important role in the immunity against H. pylori. In this research, we investigated any possible association between killer immunoglobulin-like receptors (KIR) genotypes and H. pylori infection. KIR genotype was analysed in 101 Lebanese symptomatic patients (51 H. pylori positive and 50 H. pylori-negative) using the KIR Genotyping SSP kit. Among the H. pylori-positive patients, the AA, AB and BB genotypical frequencies were, respectively, 43.14%, 41.18% and 15.68% with an A:B ratio of 1.76:1. The AA, AB and BB genotypes frequencies for H. pylori-negative individuals were 18%, 62% and 20%, respectively, with an A:B ratio of 0.96:1. No significant difference between patients and controls was detected.
We noticed a reduced distribution of A haplotype among the 'H. pylori-negative' patients as compared with the "H. pylori-positive" group. This is the first study in the international literature that targets the correlation between KIR genotypes and H. pylori.
closed_qa
Does noninvasive ventilation delivery in the ward provide early effective ventilation?
Although noninvasive ventilation (NIV) is increasingly used in general wards, limited information exists about its ability to provide effective ventilation in this setting. We aim to evaluate NIV delivered in the ward by assessing (1) overall time of application and occurrence of adverse events and (2) differences between daytime and nighttime NIV application. We studied subjects with hypercapnic acute hypercapnic respiratory failure not fulfilling strict criteria for ICU admission, and excluded those who interrupted NIV prior to 48 h. Time spent on NIV, presence and extent of air leaks, and occurrence of desaturations were assessed for the overall study period, and compared between daytime and nighttime. We enrolled 42 subjects, 25 of whom received NIV for at least 48 h and were included in the data analysis. NIV was successful for 20 subjects, who did not reach criteria for ICU admission. Both PaCO2 and pH significantly improved on average after 2 h and at the end of the study period. NIV was applied for 64.5% of the overall study period and had absent or compensated air leaks for 62.3% of the overall 48-h period. NIV was applied for 55.8% of daytime and for 79.3% of nighttime (P<.01). Effective NIV application was significantly longer overnight (76.9%) than during daytime (53.2%) (P<.01).
In selected subjects with hypercapnic acute respiratory failure not fulfilling criteria for ICU admission, the application of NIV in the ward is feasible; in addition, NIV can be safely administered overnight.
closed_qa
Is neck circumference measurement an indicator for abdominal obesity?
Neck circumference (NC) measurement is one of the simple screening measurements which can be used as an index of upper body fat distribution to identify obesity. The aim of this study was to determine the relationship between neck circumferences and obesity. A total 411 volunteer adults participated in this study (174 men, 237 women). A questionnaire which consisted of anthropometric measurements and demographic features was used. Patients with NC ≥37 cm for men and ≥34 cm for women require evaluation of overweight status. The percentages of the men and women with BMI ≥ 25 kg/m(2) were 55.2% and 27.0% respectively and with high neck circumferences were 85.1% and 38.8%, respectively. The percentages of the men and women with high waist circumference were 31.6% and 79.3%, respectively. In both gender there were positive significant correlations between neck circumference, body weight (men, r=0.576; women, r=0.702; p=0.000), waist circumferences (men, r=0.593; women r=0.667; p=0.000), hip circumferences (men, r=0.568; women, r=0.617; p=0.000) and BMI (men, r=0.587; women, r=0.688; p=0.000).
This study indicates that NC was associated with body weight, BMI, waist and hip circumferences and waist/hip ratio for men and women. A significant association was found between NC and conventional overweight and obesity indexes. NC was associated with waist/hip ratio for men and women.
closed_qa
Management of LBP at primary care level in South Africa: up to standards?
Primary Health Care (PHC) is well suited for management of low back pain (LBP). Prevalence of (chronic) LBP is suspected to be high among visitors of the South African primary care centers, but currently no information exists on prevalence or guideline adherence. To establish if treatment received for LBP in public PHC in the Cape Town area compares with international evidence based guidelines. Cluster randomization determined the 8 community health centres where the study took place. A measurement tool was developed and validated for this population. Descriptive analysis and logistic regression analytical techniques were applied. 489 participants (mean age: 44.8) were included in this study. Lifetime prevalence was 73.2% and 26.3% suffered from chronic low back pain (CLBP) . Pain medication was the only form of treatment received by 90% of the sample. Interventions received seemed to be unrelated to type of LBP (acute, sub acute and chronic). Referral to physiotherapy, education and advice to stay active were rarely done. Participants expressed low satisfaction with treatment.
Current management of LBP at PHC level appears to be ineffective and not conform guidelines. Further South African research should focus on barriers as well as measures to be taken for implementation of LBP guidelines.
closed_qa
Does chronic rhinosinusitis increase the risk of lung cancer?
Chronic rhinosinusitis is one of the most common chronic inflammatory diseases of the upper airway. A previous study of chronic rhinosinusitis and the risks of lung cancer was based on a self-reported questionnaire concerning rhinosinusitis. Population-based cohort studies of the correlation between chronic rhinosinusitis and the adenocarcinoma subtype of lung cancer have been limited. In the present study, we used a population-based database to investigate the risks related to the adenocarcinoma subtype of lung cancer among patients with chronic rhinosinusitis. We identified 13 072 patients who were diagnosed with chronic rhinosinusitis in 1998-2010 as the exposure group. There were 52 288 randomly selected patients as the comparison cohort. We used data from the Taiwan National Health Insurance Research Database; the Taiwan National Health Insurance programme offers health-care services to 99% of the 23 million people residing in Taiwan. We compared the incidence of the adenocarcinoma subtype of lung cancer between the two cohorts. The cumulative incidence and hazard ratios of developing the adenocarcinoma subtype of lung cancer were calculated. This study included 13 072 participants with chronic rhinosinusitis and 52 288 participants with non-chronic rhinosinusitis. The risk of the adenocarcinoma subtype of lung cancer was higher in the chronic rhinosinusitis cohort than in the non-chronic rhinosinusitis cohort, with an adjusted hazard ratio of 3.52 after controlling for age and gender.
This large population-based cohort study demonstrated that patients in Taiwan with previous chronic rhinosinusitis are at greater risk of developing the adenocarcinoma subtype of lung cancer.
closed_qa
Is human T-lymphotropic virus type 1 infection associated with hearing loss?
Human T-lymphotropic virus type 1 (HTLV-1) infection is endemic in the northeast area of Iran. Although various neurological disturbances have been reported in HTLV-1 infection, possible audiovestibular involvement during this infection has not yet been studied. Case control study. Sixty-eight cases in three groups including 24 HTLV-1-infected patients with HTLV-1- associated myelopathy/tropical spastic paraparesis (HAM/TSP) (group 1), 23 HTLV-1-infected cases without clinical presentation (group 2), and 21 normal individuals (group 3) entered our study. A complete history of hearing-related disorders and a profile of audiologic tests, including pure-tone audiometry (PTA) with high frequencies, speech reception threshold (SRT), and auditory brainstem response (ABR) were taken. Subjective audiovestibular complaints of participants showed a significant difference among HAM/TSP patients and the two other groups regarding hearing loss and tinnitus, but not vertigo or aural fullness. Hearing evaluation by SRT and PTA in all frequencies showed a significant difference between HAM/TSP patients (group 1) and the controls (group 3). The difference was also significant between asymptomatic cases (group 2) and the controls only in PTA frequencies above 4 kHz. Auditory brainstem-evoked potential did not show any significant differences among the groups regarding latency of I, III, and V waves and interwave differences.
HTLV-1 infection, particularly in those with a clinical presentation, appears to accompany hearing loss. Based on the results of PTA and ABR tests, this study may suggest a cochlear source of hearing impairment rather than neural problems.
closed_qa
Is preeclampsia associated with fetal malformation?
To examine fetal malformations in mother-infant pairs with and without pregnancy-related hypertension. This was an observational, population-based study of women delivering a singleton at our hospital. Specific fetal malformations identified in women with gestational hypertension or preeclampsia were compared to those without pregnancy-related hypertension. Women with chronic hypertension, superimposed preeclampsia on chronic hypertension and pregestational diabetes were excluded. Between March 2002 and December 2012, a total of 151 997 women delivered, and 10 492 (7%) had preeclampsia, 4282 (3%) had gestational hypertension and 137,223 (90%) were referent normotensive controls. Women with preeclampsia were significantly more likely to deliver infants with malformations when compared to normotensive controls (2.5% versus 1.6%, p < 0.001), whereas women with gestational hypertension were not (1.9% versus 1.6%, p = 0.16). The overall risk for fetal malformation associated with preeclampsia remained significant following logistic regression for age, race, parity and maternal body-habitus (adjusted OR 1.5; 95% CI: 1.3-1.7). Only single-organ system malformations - microcephaly and hypospadias - remained associated with preeclampsia (p < 0.001), and fetal growth restriction was a co-factor for both.
Preeclampsia was associated with increased rates of fetal malformations when compared to normotensive women - specifically microcephaly and hypospadias. These associations appear predominantly as a consequence of impaired fetal growth.
closed_qa
Tranfusion risk: is "two-step" vaginal delivery a risk for postpartum hemorrhage?
In the active management strategy of third stage of labor, the optimal timing for clamping the umbilical cord after birth has been a subject of controversy. We want to evaluate if "two-step" delivery is a risk factor for postpartum hemorrhage (PPH), defined as need of transfusion, comparing to operative delivery, elective caesarean delivery and emergency caesarean delivery. This is a retrospective cohort study conducted in division of Perinatal Medicine, Policlinico Abano Terme. We evaluated the need of transfusion in all cases of PPH verified in all single deliveries between January 2011 and December 2012. The main outcome measure was blood loss and red blood cell transfusion. We found 17 cases of PPH (0.88%). The distribution of PPH in relation to mode of delivery was 0.71%, 2.46% and 1.98% respectively for two-step vaginal delivery (RR = 0.81 (0.56-1.22)), emergency cesarean section (RR = 2.88 (1.27-7.77)) and operative vaginal delivery (RR = 2.88 (0.59-5.66)). In labor induction there is a stronger relative risk association between PPH and as emergency cesarean delivery (p < 0.05) as operative vaginal delivery (p < 0.05).
"Two-step" delivery approach did not increase the risk of PPH with respect to operative delivery, elective caesarean section and emergency caesarean section.
closed_qa
Complex febrile crises: should we change the way we act?
Febrile seizures are one of the most frequent reasons why patients visit the healthcare specialist. Up until now, patients with complex febrile seizures (CFS) have been hospitalised, bearing in mind the higher percentages of epilepsy and acute complications that were classically reported. Today there are studies that back the idea of being less invasive in the management of these patients. AIMS. To describe the characteristics of patients hospitalised due to CFS and to propose a new protocol to be followed in dealing with such cases. The medical records of patients hospitalised because of CFS (January 2010-December 2013) were analysed retrospectively. Epidemiological and clinical data are presented, together with information from complementary tests and about development. CFS account for 4.2% of all neuropaediatric cases of admittance to hospital in (67 patients). Mean age at the time of the event: 25 months. A pathological family history existed in 47% of cases, and 31% had a previous personal history of febrile seizures. The CFS lasted less than five minutes in 54% of patients; there were also recurrences, most of them with a total of two crises and during the first day (CFS due to recurrence are the most frequent). None of the complementary tests that were carried out were of any use as a diagnostic aid during the acute phase. During their follow-up, five patients presented complications. Patients with a family history of febrile seizures presented a higher risk of epilepsy or recurrence (p = 0.02), with no significant differences as regards age, number of seizures, febrile interval, epileptic status or type of CFS.
The CFS are not associated with greater acute complications, and the complementary examinations do not allow high-risk patients to be distinguished at an early stage. Hospitalising them could be avoided in the absence of other clinical signs and symptoms, and thus be limited to selected cases.
closed_qa
Is Hepatic Resection for Large or Multifocal Intrahepatic Cholangiocarcinoma Justified?
The role of surgical resection for patients with large or multifocal intrahepatic cholangiocarcinoma (ICC) remains unclear. This study evaluated the long-term outcome of patients who underwent hepatic resection for large (≥7 cm) or multifocal (≥2) ICC. Between 1990 and 2013, 557 patients who underwent liver resection for ICC were identified from a multi-institutional database. Clinicopathologic characteristics, operative details, and long-term survival data were evaluated. Of the 557 patients, 215 (38.6 %) had a small, solitary ICC (group A) and 342 (61.4 %) had a large or multifocal ICC (group B). The patients in group B underwent an extended hepatectomy more frequently (16.9 vs. 30.4 %; P<0.001). At the final pathology exam, the patients in group B were more likely to show evidence of vascular invasion (22.5 vs. 38.5 %), direct invasion of contiguous organs (6.5 vs. 12.9 %), and nodal metastasis (13.3 vs. 21.0 %) (all P<0.05). Interestingly, the incidences of postoperative complications (39.3 vs. 46.8 %) and hospital mortality (1.1 vs. 3.7 %) were similar between the two groups (both P>0.05). The group A patients had better rates for 5-year overall survival (OS) (30.5 vs. 18.7 %; P<0.05) and disease-free survival (DFS) (22.6 vs. 8.2 %; P<0.05) than the group B patients. For the patients in group B, the factors associated with a worse OS included more than three tumor nodules [hazard ratio (HR), 1.56], nodal metastasis (HR, 1.47), and poor differentiation (HR, 1.48).
Liver resection can be performed safely for patients with large or multifocal ICC. The long-term outcome for these patients can be stratified on the basis of a prognostic score that includes tumor number, nodal metastasis, and poor differentiation.
closed_qa
Cervical lymph node metastasis in differentiated thyroid carcinoma: does it have an impact on disease-related morbid events?
The aim of the study was to evaluate the prognostic impact of cervical lymph node (CLN) metastasis in well-differentiated thyroid cancer (WDTC). A cohort of 164 patients who underwent thyroidectomy for WDTC in the last 12 years was studied. Patients were classified into those with CLN metastasis (group 1) and those without (group 2), with 82 patients in each group. Morbid events (recurrence and death) and prognostic risk factors were analyzed. The mean age of the whole group was 41.3±14.2 years, with 52.4% being at least 45 years old; 69.6% of the patients were female (P=0.02), with a female to male ratio of 2.3 : 1. Near-total thyroidectomy and radioactive iodine-131 therapy were performed in all 81.1% of patients in whom papillary cancer was found. Morbid events were found in 22/164 (13.4%) patients (17 recurrences and five deaths). All recurrences occurred in patients with the papillary subtype, whereas 4/5 deaths occurred in those with follicular thyroid cancer. Morbid events were found in 14/82 (17.1%) and 8/82 (9.6%) patients in group 1 and group 2, respectively (P=0.1). Comparison of morbid events between group 1 and group 2 showed statistically significant differences with respect to tumor size of 1 cm or more, multifocality, and positivity of the follow-up diagnostic tools at 6-12 months. Kaplan-Meier analysis revealed no significant difference between the two groups (P=0.4) over a mean follow-up period of 125.1±14.6 months. Multivariate Cox analysis showed that development of remote deposits, lymph node metastasis five or more nodes, extracapsular nodal invasion in at least three nodes, and positivity of the follow-up diagnostic tools at 6 months were the significant prognostic factors of morbid events.
Our study revealed the impact of CLN metastasis on the occurrence of morbid events in WDTC patients (especially in those with the papillary subtype) with node metastasis five or more, extracapsular nodal invasion in at least three nodes, and positivity of follow-up diagnostic tools at 6 months. Hence, these patients should be optimally managed and more closely monitored.
closed_qa
Can nasal decongestants improve eustachian tube function?
To evaluate the effect of nasal decongestants on eustachian tube (ET) opening. A prospective nonrandomized study. A tertiary referral center. Twenty-four patients (44 ears) with intact eardrums, 39 patients (43 ears) having a noninfected eardrum defect, and six patients with an upper airway infection. Nasal or intratympanal (in perforated ears) application of a nasal decongestant (xylometazoline 0.1%). Change of tube opening quality (yes or no; better or worse) measuring tube opening parameters (pressure, latency) using the Estève method and pressure equalization tests (swallowing at negative and positive external ear canal pressures). In most cases, nasal decongestion or intratympanal use of decongestants have no effect on ET opening. Improvement in tube opening is rather an exception and, in a minority of patients, a reduced ET function was evident.
Our acute studies revealed no improvement in eustachian ventilatory tube function with the administration of nasal decongestants.
closed_qa
HER2 immunohistochemical assessment with A0485 polyclonal antibody: is it time to refine the scoring criteria for the "2+" category?
An accurate determination of human epidermal growth factor receptor 2 (Her2) status in women with breast cancer is mandatory to identify patients who will benefit from trastuzumab-based therapy. Her2 immunohistochemical analysis (IHC) (performed with A0485 polyclonal antibody) on 943 invasive breast cancer cases was evaluated independently and blindly twice by 3 of us (V.A., I.P., and A.C.) according to DAKO scoring criteria. A total of 230 cases of invasive breast cancer scored 2+ at IHC and consequently evaluated by FISH were reviewed first independently, and then simultaneously by 3 of us (V.A., I.P., and A.C.) at a multiheaded microscope assessing the following parameters: overall signal intensity, granularity and continuity of membrane staining, and the presence of band-like membrane pattern in>25% of tumor cells. The frequencies of HER2 gene amplification for all the immunohistologic parameters (individually considered or in combination) were compared by Pearson χ analysis. Combinations of staining patterns did not give any statistically significant results, except when combining strong staining intensity and continuity of membrane signal. In fact, only 9 of the 86 cases with a weak-to-moderate staining intensity, which showed a fragmented membrane signal, resulted in being amplified by FISH, whereas 19 of the 51 cases presenting an overall strong IHC reaction and some extent of continuous membrane signal were FISH amplified (P=0.002).
Combined intensity and linearity of membrane signal, although limited, resulted in the best aid (P=0.0002) in making the final score decision in borderline IHC Her2 tests similar to what is envisaged in the Her2 scoring system for gastric cancer.
closed_qa
Concomitant MAZE procedure during cardiac surgical procedures: is there any survival advantage in conversion to sinus rhythm?
We retrospectively evaluated the conversion rate to SR and its correlation with long-term survival in 209 patients with chronic AF, who had a MAZE procedure during cardiac surgical procedures between the years 2006 and 2011 at our institution. The mean age was 67.2 ± 12.0 years and 52.2% were female (N. = 109). Perioperative mortality was 5.74% (N. = 12). In univariate analysis, significant risk factors for perioperative mortality were age (P = 0.0033), duration of perfusion time (P = 0.0093), elevated creatinine (≥ 1.6 mg/dL, P = .02), and cross clamp time (P = 0.016). In multivariate analysis age (HR 2.97) and duration of perfusion time (HR 1.48) were the only independent predictors of perioperative mortality. The overall one and five-year survival rates were 88% ± 2.2%, and 76% ± 3.3%, respectively. The one and five-year survival rates for patients who converted and were in sinus rhythm (SR) upon discharge (N. = 154) were 88% ± 2.6% and 80% ± 3.5%, respectively. While the one and five-year survival rates for patients who were still in AF upon discharge (N. = 55) were 94% ± 3% and 82% ± 6.6%, respectively, this survival difference was not statistically significant (P = 0.24). Significant risk factors for long-term mortality included DM (P = 0.023), preoperative MI (P = 0.043), preoperative renal insufficiency (creatinine, ≥ 1.6 mg/dL, P = 0.02) and asthma/COPD (P = 0.040). In multivariate analysis, age (HR 1.048) and preoperative MI (HR 1.948) were the only independent predictors of long-term mortality.
The surgical MAZE procedure has a high conversion rate, however, our data did not show improved survival in patients who converted to SR prior to discharge.
closed_qa
Are school children ready to donate blood?
Voluntary non-remunerated blood donors are considered the best among all different types of blood donors for improving the supply of safe blood. Though safe blood transfusion services have improved in Pakistan, but efforts are still required to optimize blood banks and improve recruitment of voluntary donors, such as senior school/college going students as a source of safe blood. This study looks into the awareness of senior school children concerning blood related issues including blood donation. This cross-sectional descriptive study enrolled 106 senior school students of private schools. Data were collected through self-administered questionnaire. Although 90% of the students considered blood as an important entity for saving lives still 56.8% had never thought of donating blood. Respondents had good knowledge regarding the possible spread of HIV/AIDS, and Hepatitis B and C through unsafe transfusions. Possible hindrances to donating blood included fear of needles, fear of acquiring disease, lack of knowledge regarding where to donate blood and lack of trust on blood banks. More than half of the students believed that blood should be bought from professional blood donors.
Senior school going children are not ready to donate blood. Lack of knowledge and prevailing misconceptions regarding blood transfusions need to be addressed and mechanisms to motivate and mobilize youth for becoming voluntary blood donors need to be established.
closed_qa
Do practice characteristics explain differences in morbidity estimates between electronic health record based general practice registration networks?
General practice based registration networks (GPRNs) provide information on population health derived from electronic health records (EHR). Morbidity estimates from different GPRNs reveal considerable, unexplained differences. Previous research showed that population characteristics could not explain this variation. In this study we investigate the influence of practice characteristics on the variation in incidence and prevalence figures between general practices and between GPRNs. We analyzed the influence of eight practice characteristics, such as type of practice, percentage female general practitioners, and employment of a practice nurse, on the variation in morbidity estimates of twelve diseases between six Dutch GPRNs. We used multilevel logistic regression analysis and expressed the variation between practices and GPRNs in median odds ratios (MOR). Furthermore, we analyzed the influence of type of EHR software package and province within one large national GPRN. Hardly any practice characteristic showed an effect on morbidity estimates. Adjusting for the practice characteristics did also not alter the variation between practices or between GPRNs, as MORs remained stable. The EHR software package 'Medicom' and the province 'Groningen' showed significant effects on the prevalence figures of several diseases, but this hardly diminished the variation between practices.
Practice characteristics do not explain the differences in morbidity estimates between GPRNs.
closed_qa
Do immunocompromised children benefit from having surgical lung biopsy performed?
Surgical lung biopsy is considered a gold standard for the evaluation of pulmonary disease in immunocompromised children. However, in the literature, its accuracy and the rate of complications vary. We aimed to evaluate the yield of surgical lung biopsies in the management of persistent pulmonary findings in immunocompromised children. We performed a retrospective review of clinical records of immunocompromised children who underwent surgical lung biopsies, and evaluated the impact that preoperative factors had on outcomes. Twenty-five patients underwent 27 surgical lung biopsies. The underlying immunodeficiency included allogeneic stem cell transplantation (n = 12), chemotherapy for solid tumors (n = 6), hematologic malignancy (n = 4), primary immunodeficiency (n = 4) and chronic steroid use (n = 1). Biopsies provided a specific histopathologic or microbiologic diagnosis in 10 cases (37%). No preoperative factor predicted a diagnostic biopsy. Five of the 27 biopsies were beneficial for the patients (18%). A major complication related to the procedure was reported for 1 biopsy (4%).
We conclude that surgical lung biopsy in pediatric immunocompromised patients appears to be safe, but has a relatively low diagnostic yield and an even lower yield with regards to the benefit it provides.
closed_qa
Is a specialist breathlessness service more effective and cost-effective for patients with advanced cancer and their carers than standard care?
Breathlessness is common in advanced cancer. The Breathlessness Intervention Service (BIS) is a multi-disciplinary complex intervention theoretically underpinned by a palliative care approach, utilising evidence-based non-pharmacological and pharmacological interventions to support patients with advanced disease. We sought to establish whether BIS was more effective, and cost-effective, for patients with advanced cancer and their carers than standard care. A single-centre Phase III fast-track single-blind mixed-method randomised controlled trial (RCT) of BIS versus standard care was conducted. Participants were randomised to one of two groups (randomly permuted blocks). A total of 67 patients referred to BIS were randomised (intervention arm n = 35; control arm n = 32 received BIS after a two-week wait); 54 completed to the key outcome measurement. The primary outcome measure was a 0 to 10 numerical rating scale for patient distress due to breathlessness at two-weeks. Secondary outcomes were evaluated using the Chronic Respiratory Questionnaire, Hospital Anxiety and Depression Scale, Client Services Receipt Inventory, EQ-5D and topic-guided interviews. BIS reduced patient distress due to breathlessness (primary outcome: -1.29; 95% CI -2.57 to -0.005; P = 0.049) significantly more than the control group; 94% of respondents reported a positive impact (51/53). BIS reduced fear and worry, and increased confidence in managing breathlessness. Patients and carers consistently identified specific and repeatable aspects of the BIS model and interventions that helped. How interventions were delivered was important. BIS legitimised breathlessness and increased knowledge whilst making patients and carers feel 'not alone'. BIS had a 66% likelihood of better outcomes in terms of reduced distress due to breathlessness at lower health/social care costs than standard care (81% with informal care costs included).
BIS appears to be more effective and cost-effective in advanced cancer than standard care.
closed_qa
Prospective Analysis of Payment per Hour in Head and Neck Reconstruction: Fiscally Feasible or Futile?
The authors assess the fiscal viability of complex head and neck reconstructive surgery by evaluating its financial reimbursement in the setting of resources used. The authors prospectively assessed provider reimbursement for consecutive patients undergoing head and neck reconstruction. Total care time was determined by adding 15 minutes to the operative time for each postoperative hospital day and each postoperative follow-up appointment within the 90-day global period. Physician reimbursement was divided by total care time hours to determine an hourly rate of reimbursement. A control group of patients undergoing carpal tunnel release was evaluated using the same methods described. A total of 50 patients met the inclusion criteria for study. The payer was Medicaid for nine patients (18 percent), Medicare for 19 patients (38 percent), and commercial for 22 patients (44 percent). The average provider revenue per case was $3241.01 ± $2500.65. For all patients, the mean operative time was 10.6 ± 3.87 hours and the mean number of postoperative hospital days was 15.1 ± 8.06. The mean reimbursement per total care time hour was $254 ± $199.87. Statistical analysis demonstrated difference in reimbursement per total care time hour when grouped by insurance type (p = 0.002) or flap type (p = 0.033). Of the 50 most recent patients to undergo carpal tunnel release, the average revenue per case was $785.27.
Total care time analysis demonstrates that physician reimbursement is not commensurate with resources used for complex head and neck reconstructive surgery.
closed_qa
Is inhaled colistin beneficial in ventilator associated pneumonia or nosocomial pneumonia caused by Acinetobacter baumannii?
In the present study, our objective was to evaluate and compare the clinical and microbiological results in patients receiving systemic and systemic plus inhaled colistin therapy due to nosocomial pneumonia (NP) or ventilator associated pneumonia (VAP) caused by Acinetobacter baumannii. A retrospective matched case-control study was performed at the ICUs at Izmir Katip Celebi University Ataturk Training and Research Hospital from January 2013 to December 2014. Eighty patients who received only systemic colistin were matched 43 patients who received systemic colistin combined with inhaled therapy. In 97.6 % of the patients colistin was co-administered with at least one additional antibiotic. The most frequently co-administered antibiotics were carbapenems (79.7 %). The patient groups did not differ significantly in terms of the non-colistin antibiotics used for treatment (p > 0.05). Acute renal injury was observed in 53.8 % and 48.8 % of the patients who received parenteral colistin or parenteral plus inhaler colistin, respectively (p = 0.603). There were no significant differences between the groups in terms of clinical success (p = 0.974), clinical failure (p = 0.291), or recurrence (p = 0.094). Only, a significantly higher partial clinical improvement rate was observed in the systemic colistin group (p = 0.009). No significant differences between the two groups in terms of eradication (p = 0.712), persistence (p = 0.470), or recurrence (p = 0.356) rates was observed. One-month mortality rate was similar in systemic (47.5 %) and systemic plus inhaled (53.5 %) treatment groups (p = 0.526).
Our results suggest that combination of inhaled colistin with intravenous colistin had no additional therapeutic benefit in terms of clinical or microbiological outcomes.
closed_qa
Active Staphylococcus aureus infection: Is it a contra-indication to the repair of complex hernias with synthetic mesh?
The management of chronic mesh infection is challenging and controversial. The use of synthetic material to repair the abdominal wall in the infected setting is not recommended, especially in the presence of active infection caused by Staphylococcus aureus. This is a prospective observational study designed to evaluate the outcomes in patients with active mesh infection caused by Staphylococcus aureus. Patients underwent simultaneous removal and replacement of polypropylene mesh. The treatment protocol included the complete removal of infected mesh, followed by the anatomical reconstruction, and reinforcement of the abdominal wall using a new onlay polypropylene mesh. Early and late wound complications, medical complications, and hernia recurrences were analyzed. From 2006 until 2014, 22 patients with a mean age of 57.2 years and mean BMI of 29,3 kg/m2 were studied. Sinuses were present in 21 patients. A recurrent ventral hernia was observed in 14 patients; two patients required a complex abdominal wall reconstruction due to enteric fistulas. Bowel resections or other potentially contaminated procedures were associated in 10 patients. Fourteen patients (63.6%) had an uneventful postoperative course; 5 (22.7%) patients had wound infections requiring debridement and three required partial (2) or total (1) mesh removal. Two patients died due to medical complications. Adverse results on long-term follow-up included one hernia recurrence after complete mesh removal and one persistent sinus after partial mesh removal requiring a reoperation to remove mesh remnants. All of the patients were considered free of infection after a mean follow-up of 44 months.
Synthetic mesh replacement in patients with active Staphylococcus aureus infection has an acceptable incidence of postoperative wound infection and prevents hernia recurrence. Large-pore polypropylene mesh is a suitable material to be used in the infected surgical field as an onlay graft.
closed_qa
Mobile revolution: a requiem for bleeps?
Effective communication is a vital part of good clinical care. Traditionally bleep systems have been used as the mainstay of communication. Mobile technology is increasingly seen as a quicker, easier and more reliable method of communication. Our objective was to assess the use of mobile devices within a typical National Health Service (NHS) hospital, discuss potential benefits and pitfalls, and develop suggestions for future improvements. A survey of 600 hospital doctors was conducted in a large NHS district general hospital between 1 May and 30 June 2015. The questionnaire explored the patterns of use, attitudes and impact of mobile communication, and identified potential risks and benefits of its wider adoption within the NHS. 92% of doctors use their personal mobile for hospital-related work. 95% share their personal number with colleagues, and 64% have it available through hospital switchboard. 77% use their personal mobile to discuss patient matters, and 48% are prevented from communicating effectively due to poor signal within the hospital. 90% are contacted when not at work on a weekly or daily basis regarding patients. 73% feel that traditional bleeps should be replaced with new mobile technologies.
Mobile phone usage is very common among doctors, and is the preferred method of communication within the hospital. Mobile technology has the potential to revolutionise communication and clinical care and should be embraced. The introduction of new technology will inevitably change existing hospital dynamics, and consequently may create a new set of challenges that will require further work to explore in the future.
closed_qa
Is ultrasonographic evaluation essential for diagnosis of retained products of conception after surgical abortion?
A total of 466 patients with a gestational age<10 weeks were followed up at the 7th day after MVA of unwanted pregnancies. The patients who had intense or moderate bleeding and other symptoms related to RPOC had a re-evacuation and all the patients were followed up until the next menstruation. Ultrasonographic evaluation was repeated weekly in asymptomatic patients with abnormal ultrasonographic findings (increased endometrial thickness, presence of hyperechogenic, mixed and hypoechogenic material) until a normal endometrial cavity was visualized. Out of the 466 patients, 15 (3.2%) had symptoms of RPOC at day 7 while the remaining 451 (96.8%) were asymptomatic and 20 (57.9%) had normal ultrasonographic findings. The 15 symptomatic patients (3.2%) had a repeated MVA. Nine of these 15 patients (60%) had endometrial echogenicity with mixed patterns, two (13.3%) had endometrial echogenicity with hyperechoic patterns, and the remaining four (26.7%) had normal endometrial echogenicity at day 7. Histopathologic examination of 12 of these 15 patients (80%) showed chorionic villi while no gestational tissue was noted in the remaining three of these patients. Endometrial thickness ≥ 10 mm on day 7 had 75% sensitivity and 100% specificity for diagnosis of RPOC in symptomatic women. None of the 451 asymptomatic patients developed any symptoms or needed any further intervention.
As abnormal ultrasonographic findings return to normal over time in asymptomatic patients, the diagnosis of RPOC should not be based on ultrasonographic findings.
closed_qa
Retinal detachments in southern New Zealand: do poorer patients have poorer outcomes?
To investigate associations between socioeconomic status, retinal detachment type and post-operative visual outcomes in southern New Zealand. A retrospective review of all cases of rhegmatogenous retinal detachments in Dunedin Hospital over two years was performed. Patient demographics and macula involvement at presentation were the primary outcome measures. The New Zealand Deprivation Index was used to group patients into low (30% least deprived), medium (middle 40%) and high (30% most deprived). Patients were excluded if they were not from New Zealand, or had traumatic detachments. During the study period, 95 retinal detachments in 94 patients were managed in Dunedin Hospital. Only 15% of retinal detachments occurred in the most deprived. More deprived patients had longer delays before assessment in hospital (mean of 29.8 days versus 10.1 days for the least deprived and 12.8 days for the medium category, overall p=0.025). There was no evidence of an association between deprivation and macula-off status (overall p=0.650) or visual acuity at one or three months (p=0.063 and p=0.328 respectively). Nor was there an association between referral pathway and macula-off status (p=0.242).
Retinal detachment in southern New Zealand may be less common amongst those with the most deprived socioeconomic status who also experience longer delays till first treatment; but there was no association between socioeconomic status and patients being macula-off at presentation, or having poorer visual outcomes. More targeted patient education towards our most deprived citizens may reduce delays in treatment, and result in better visual outcomes.
closed_qa
Predicted survival in patients with brain metastases from colorectal cancer: Is a current nomogram helpful?
To examine the clinical applicability of a new nomogram by comparing survival of patients with brain metastases from colorectal cancer treated with surgery and/or radiotherapy in the authors' institutions with nomogram-predicted median survival. Retrospective analysis of 64 patients treated with comparable approaches and during the same time period as the patients in the nomogram study. Points were assigned for age, performance status, number and site of brain metastases, as required for nomogram use. In 46 patients (72%), the observed survival was shorter than the predicted median. The median deviation was -1.4 months. The nomogram underestimated the survival of patients treated with radiosurgery/surgery by a median of 4.2 months, whereas it overestimated the survival of patients treated with whole-brain radiotherapy (WBRT) by a median of 2.1 months (p=0.0001). Nevertheless, all 5 patients with predicted median survival ≤3 months died within 3 months. Among 8 patients with predicted median survival>12 months, 6 (75%) survived for>12 months. Not all prognostic factors in the nomogram correlated with survival. In the multivariate Cox model, only performance status and number of brain metastases were significant, both with p=0.0001.
Despite differences in prognostic factors and survival of many individual patients, especially those with intermediate prognosis, the nomogram performed promising in poor- and good-prognosis patients. Evaluation of separate prediction tools for patients treated with WBRT and more aggressive local approaches appears warranted in order to minimize the influence of better local control of the brain metastases.
closed_qa
Urinalysis requests on the elderly residing in the Auckland community: tick box requesting?
Urinalysis for microscopy and culture is one of the most frequently requested tests for microbiology laboratories, particularly from elderly patients. This study sought to describe the clinical appropriateness of urinalysis from community-dwelling elderly patients and subsequent antibiotic prescription. Demographic, laboratory, and antibiotic prescription data were collected on all samples submitted from patients ≥ 70 years during August 2014 to Labtests Auckland. In addition, clinical data were collected by questionnaire from a subgroup of 200 patients. During August 2014, approximately 7% of the Auckland population aged ≥ 70 years had urinalysis submitted. Urine dipstick was not routinely performed before specimen submission, particularly from patients living at home rather than a long-term care facility, and nearly 50% of samples were not cultured due to absence of pyuria. Escherichia coli was isolated from 23% of female and 7% of male specimens. E. coli isolates from our cohort were less susceptible to all antibiotics tested against compared with all E. coli isolated from all urines in 2014. Clinical indications were absent in 40% of the subgroup of patients. Antibiotic prescription within 7 days of urinalysis was common (36%).
This study highlights the frequency of urinalysis testing among the elderly residing in the community. Clinical indications are often absent, and treatment of asymptomatic bacteriuria is likely to be contributing to excessive antibiotic prescription in this group of patients.
closed_qa
Does Congenital Heart Disease Affect Neurodevelopmental Outcomes in Children with Down Syndrome?
The impact that congenital heart disease (CHD) has on the neurodevelopment of children with Down syndrome (DS) is unknown and potentially has implications for targeted early intervention. This study assessed the relationship between CHD that required surgery in the first year of life and neurodevelopmental, behavioral and emotional functioning outcomes in children with DS. A retrospective chart review of 1092 children (0-18 years) with DS who visited a single institution from 8/08-8/13 was performed. Children who underwent at least one of nine neurodevelopmental (cognitive, language, developmental) or academic tests were included in the analysis (N = 178). Cohort was age-divided into infants/toddlers (0-2 years), preschoolers (3-5 years), and school age/adolescent (6-18 years). Test scores of children with DS who underwent cardiac surgery in the first year of life were compared to children with DS without CHD. T test, chi-square and Mann Whitney U tests were used where appropriate. Infants/toddlers with cardiac surgery had lower scores for receptive (P = .01), expressive (P = .021) and composite language (P < .001) compared to those with no CHD. Preschoolers with cardiac surgery had lower language scores and lower visual motor scores, although not statistically significant. In school age children with cardiac surgery there were no differences in IQ scores, language scores, or academic achievement scores compared to those without CHD. Also at school-age there was no difference in the incidence of ADHD, executive function or on internalizing and externalizing behavior scores.
Children with DS undergoing cardiac surgery during the first year demonstrated poorer neurodevelopmental outcomes as infants/toddler but had no difference at school age compared to children with DS without CHD. These results will guide early interventions to optimize neurodevelopmental outcomes in children with DS and will help with family counseling after CHD repair.
closed_qa
Is PET/CT Necessary in the Management of Early Breast Cancer?
Advanced imaging methods in early breast cancers are not recommended before surgery. In contrast to the accepted guidelines, some recent studies have shown some benefits with the use of PET/CT in early-stage breast cancer. In this study, we aimed to document the efficacy of PET/CT in detection of distant metastasis as well as other primary cancers. In this retrospective study, we reviewed the records of all women patients diagnosed with early breast cancer between March 2012 and December 2014. Besides demographics, we recorded the clinical TNM stage, histology of the tumor, and hormone receptor status. As PET/CT imaging is a routine procedure in our center for early breast cancer, tumor size, lymph node status, distant metastasis, and possible other primary malignancies detected by PET/CT were also recorded. Of the 419 women included in the study, 24.8% were clinically staged as stage I while the rest were stage II. Distant metastases were detected in 42 patients (10%). The yield of PET/CT in detecting metastasis was significant in stage II patients compared with stage I patients (12.4% vs 2.9%). In subgroup analysis of stage II patients, the performance of PET/CT in detecting metastasis was still evident in stage IIA patients (9.5%). In logistic regression analysis of the significant and near-significant factors (as detected by univariate analysis) effecting PET/CT detected distant metastasis, only nodal status (P = 0.053) was found to be significant.
We suggest the use of PET/CT in investigating metastasis in axilla positive and clinically stage II early breast cancer patients.
closed_qa
Can Monitoring Fetal Intestinal Inflammation Using Heart Rate Variability Analysis Signal Incipient Necrotizing Enterocolitis of the Neonate?
Necrotizing enterocolitis of the neonate is an acute inflammatory intestinal disease that can cause necrosis and sepsis. Chorioamnionitis is a risk factor of necrotizing enterocolitis. The gut represents the biggest vagus-innervated organ. Vagal activity can be measured via fetal heart rate variability. We hypothesized that fetal heart rate variability can detect fetuses with incipient gut inflammation. Prospective animal study. University research laboratory. Chronically instrumented near-term fetal sheep (n = 21). Animals were surgically instrumented with vascular catheters and electrocardiogram to allow manipulation and recording from nonanesthetized animals. In 14 fetal sheep, inflammation was induced with lipopolysaccharide (IV) to mimic chorioamnionitis. Fetal arterial blood samples were drawn at selected time points over 54 hours post lipopolysaccharide for blood gas and cytokines (interleukin-6 and tumor necrosis factor-α enzymelinked immunosorbent assay). Fetal heart rateV was quantified throughout the experiment. The time-matched fetal heart rate variability measures were correlated to the levels of interleukin-6 and tumor necrosis factor-α. Upon necropsy, ionized calcium binding adaptor molecule 1+ (Iba1+), CD11c+ (M1), CD206+ (M2 macrophages), and occludin (leakiness marker) immunofluorescence in the terminal ileum was quantified along with regional Iba1+ signal in the brain (microglia). Interleukin-6 peaked at 3 hours post lipopolysaccharide accompanied by mild cardiovascular signs of sepsis. At 54 hours, we identified an increase in Iba1+ and, specifically, M1 macrophages in the ileum accompanied by increased leakiness, with no change in Iba1 signal in the brain. Preceding this change on tissue level, at 24 hours, a subset of nine fetal heart rate variability measures correlated exclusively to the Iba+ markers of ileal, but not brain, inflammation. An additional fetal heart rate variability measure, mean of the differences of R-R intervals, correlated uniquely to M1 ileum macrophages increasing due to lipopolysaccharide.
We identified a unique subset of fetal heart rate variability measures reflecting 1.5 days ahead of time the levels of macrophage activation and increased leakiness in terminal ileum. We propose that such subset of fetal heart rate variability measures reflects brain-gut communication via the vagus nerve. Detecting such noninvasively obtainable organ-specific fetal heart rate variability signature of inflammation would alarm neonatologists about neonates at risk of developing necrotizing enterocolitis and sepsis. Clinical validation studies are required.
closed_qa
Refusal to participate in heart failure studies: do age and gender matter?
The objective of this retrospective study was to evaluate reasons heart failure patients decline study participation, to inform interventions to improve enrollment. Failure to enrol older heart failure patients (age>65) and women in studies may lead to sampling bias, threatening study validity. This study was a retrospective analysis of refusal data from four heart failure studies that enrolled 788 patients in four states. Chi-Square and a pooled t-test were computed to analyse refusal data (n = 300) obtained from heart failure patients who were invited to participate in one of the four studies but declined. Refusal reasons from 300 patients (66% men, mean age 65·33) included: not interested (n = 163), too busy (n = 64), travel burden (n = 50), too sick (n = 38), family problems (n = 14), too much commitment (n = 13) and privacy concerns (n = 4). Chi-Square analyses showed no differences in frequency of reasons (p>0·05) between men and women. Patients who refused were older, on average, than study participants.
Some reasons were patient-dependent; others were study-dependent. With 'not interested' as the most common reason, cited by over 50% of patients who declined, recruitment measures should be targeted at stimulating patients' interest. Additional efforts may be needed to recruit older participants. However, reasons for refusal were consistent regardless of gender.
closed_qa
Is it safe to perform completion lobectomy after diagnostic wedge resection using video-assisted thoracoscopic surgery?
The objective of this study was to assess the safety of video-assisted thoracoscopic surgery (VATS) completion lobectomy (CL) for non-small cell lung cancer (NSCLC) after diagnostic wedge resection by comparing with standard VATS lobectomy (SL). Data were retrieved from an institutional database of consecutive VATS lobectomies between January 1st 2007 and December 31st 2013. Patients were grouped into CL or SL. Patient characteristics, operative data, converted procedures, complications, and mortality was compared using Pearson Chi square, Fisher's exact test, or Mann-Whitney U test. In total 80 CL and 958 SLs were performed. There were no significant differences in median operating time, median chest drain duration or median length of stay. Median operative bleeding was 100 mL (IQR 50-238) in the CL group compared to 75 mL (IQR 25-200) in the SL group (p = 0.01). There were no differences between groups in major or minor complications. Median time from VATS wedge resection to CL was 33 days (IQR 27-41). Conversion rate was 1.3% in the CL group and 2.6% in the SL group (p = 0.72). 30-day mortality was 0 vs. 1.1 % for the CL group and the SL group accordingly (p>0.99).
This study comparing short-term surgical outcome and complications after surgical treatment of NSCLC indicates that VATS completion lobectomy after diagnostic wedge resection seems safe when looking at a relatively short time interval between the two procedures.
closed_qa
Can adverse maternal and perinatal outcomes be predicted when blood pressure becomes elevated?
For women with chronic or gestational hypertension in CHIPS (Control of Hypertension In Pregnancy Study, NCT01192412), we aimed to examine whether clinical predictors collected at randomization could predict adverse outcomes. This was a planned, secondary analysis of data from the 987 women in the CHIPS Trial. Logistic regression was used to examine the impact of 19 candidate predictors on the probability of adverse perinatal (pregnancy loss or high level neonatal care for>48 h, or birthweight<10th percentile) or maternal outcomes (severe hypertension, preeclampsia, or delivery at<34 or<37 weeks). A model containing all candidate predictors was used to start the stepwise regression process based on goodness of fit as measured by the Akaike information criterion. For face validity, these variables were forced into the model: treatment group ("less tight" or "tight" control), antihypertensive type at randomization, and blood pressure within 1 week before randomization. Continuous variables were represented continuously or dichotomized based on the smaller p-value in univariate analyses. An area-under-the-receiver-operating-curve (AUC ROC) of ≥0.70 was taken to reflect a potentially useful model. Point estimates for AUC ROC were<0.70 for all but severe hypertension (0.70, 95% CI 0.67-0.74) and delivery at<34 weeks (0.71, 95% CI 0.66-0.75). Therefore, no model warranted further assessment of performance.
CHIPS data suggest that when women with chronic hypertension develop an elevated blood pressure in pregnancy, or formerly normotensive women develop new gestational hypertension, maternal and current pregnancy clinical characteristics cannot predict adverse outcomes in the index pregnancy.
closed_qa
Does Preoperative Beta-blocker Use Influence Intraoperative Hemodynamic Profile and Post-operative Course of Liver Transplantation?
We compared the intraoperative hemodynamic profiles and outcomes of liver transplant (LT) patients receiving preoperative beta-blockers for portal hypertension with those of patients not receiving beta-blockers. Beat-to-beat hemodynamic data were recorded with the use of a lithium dilution cardiac output monitor, including heart rate (HR), cardiac output (CO), cardiac index (CI), systemic vascular resistance (SVR), SVR index, mean arterial pressure, stroke volume (SV), and SV index. Perioperative data included age, sex, etiology of cirrhosis, Model of End-Stage Liver Disease score, warm and cold ischemia times, and intraoperative blood and blood products replacement. Postoperative data included Acute Physiology and Chronic Health Evaluation II score, Intensive Care National Audit and Research Centre (ICNARC) score, length of intensive care unit stay, renal replacement treatment, and mortality. After exclusion, 215 patients not receiving and 93 patients receiving beta-blockers for portal hypertension were included. There was no significant difference between groups in demographic data. HR, CO, and CI were significantly lower in the beta-blocker group (P = .0001). Regarding outcomes, the ICNARC score was significantly higher in the beta-blocker group (P<.05).
Preoperative administration of beta-blockers attenuates the hyperdynamic circulation in LT patients; this is clearly seen through a reduction in intraoperative CO, CI, and HR. This decreased hyperdynamic state was not shown to relate to outcomes in any of the variables except for a higher ICNARC score reported in the group treated with beta-blockers.
closed_qa
Diastolic Dysfunction in Liver Cirrhosis: Prognostic Predictor in Liver Transplantation?
Patients with liver cirrhosis may develop cirrhotic cardiomyopathy (CC), characterized by blunted contractile responsiveness to stress, diastolic dysfunction (DD), and electrophysiological abnormalities. It may adversely affect the long-term prognosis of these patients. We conducted a retrospective analysis of patients undergoing liver transplantation (LT) for cirrhosis from January 2012 to June 2015. We analyzed demographic characteristics, the etiology of cirrhosis, Child-Pugh and Model for End-Stage Liver Disease (MELD) scores, the corrected QT (QTc) interval in the preoperative period, diastolic and systolic dysfunction, mortality and survival, and duration of mechanical ventilation and vasopressor support in the post-LT period. These variables were compared with diastolic dysfunction and prolongation of QTc, with the use of chi-square, Fisher, and Mann-Whitney U tests. The study included 106 patients, 80.2% male and overall average age 54.83 years. The median MELD score was 16, and Child-Pugh class C in 55.4%. Prolonged QTc interval before LT was present in 19% and DD in 35.8% of patients. QTc before LT or DD did not vary significantly with MELD or Child-Pugh score.
The patients in the pre-LT period presented with a significant incidence of DD, which can predispose them to adverse cardiac events. The presence of DD correlates with mortality after LT in patients with hepatic cirrhosis.
closed_qa
Are Surgeons Being Paid Fairly by Medicaid?
Both the Medicare (MCR) and Medicaid (MCD) programs turn 50 this year. Medicare has developed a national resource-based payment methodology for physicians' services, with broad input by specialty societies, and MCD payments are set by individual states by various means. We have conducted the first national comparison of payment methodology of MCD vs MCR for procedures commonly delivered by general surgeons. Using the most recent Centers for Medicare and Medicaid Services' Medicare data for frequency of allowed charges for general surgeons, we selected the most frequently billed procedures and gathered data from the 50 states for MCD and MCR payments. We determined the "Medicaid discount" (MCD payment minus MCR payment) expressed as dollars and percent, as well as dollars paid per relative value of work. We have discovered wide variations in MCD payments among states for the same procedures, demonstrating unexplained "discounts" of MCD payments in relationship to MCR. We found that MCD payments show wide variations across the states, with many states paying far less than MCR for common, essential procedures.
These findings call into question the fairness of MCD reimbursement for general surgery services in the United States. This discount to MCR could act as a disincentive for surgeons to care for some patients, based on the state of residence. These unexplained discounts could have considerable long-term effects for patients dependent on the MCD program. Our study should act as a stimulus for states to examine their payment methodologies to provide more uniform and fairer payments for surgical procedures.
closed_qa
Should we treat obesity in COPD?
Obesity is an established risk factor for poor health outcomes, but paradoxically in chronic obstructive pulmonary disease (COPD), it is associated with improved survival and lung function. A major evidence gap exisits to inform treatment recommendations for patients with COPD who are obese. We aimed to determine the effect of weight reduction involving a low-energy diet utilizing a partial meal replacement plan, coupled with resistance exercise training in obese COPD patients. In a proof of concept before-after clinical trial, obese (body mass index ≥30 kg/m(2) ) COPD patients received a 12 week weight reduction programme involving meal replacements, dietary counselling by a dietitian and resistance exercise training prescribed and supervised by a physiotherapist. Patients were reviewed face to face by the dietitian and physiotherapist every 2 weeks for counselling. Twenty-eight participants completed the intervention. Mean (standard deviation) body mass index was 36.3 kg/m(2) (4.6) at baseline and reduced by 2.4 kg/m(2) ((1.1) P < 0.0001) after the intervention. Importantly, skeletal muscle mass was maintained. Clinical outcomes improved with weight loss including exercise capacity, health status, dyspnea, strength and functional outcomes. There was also a significant reduction in the body mass index, obstruction, dyspnea and exercise score (BODE). Systemic inflammation measured by C-reactive protein however did not change.
In obese COPD patients, dietary energy restriction coupled with resistance exercise training results in clinically significant improvements in body mass index, exercise tolerance and health status, whilst preserving skeletal muscle mass. This novel study provides a framework for development of guidelines for the management of obese COPD patients and in guiding future research.
closed_qa
High-sensitivity cardiac troponin T: A biomarker for the early risk stratification of type-A acute aortic dissection?
High-sensitivity cardiac troponin is the most specific and sensitive biomarker of myocardial injury. However, no study has investigated whether the early concentration of high-sensitivity cardiac troponin is increased or is of value in predicting short-term prognosis in patients with type-A acute aortic dissection (AAD) in the emergency department. To measure the high-sensitivity cardiac troponin T (hs-TnT) concentration in patients with type-A AAD upon hospital admission, and to assess its value in predicting short-term prognosis. We enrolled consecutive patients with type-A AAD. Blood samples were collected on admission; hs-TnT concentrations were measured on the Elecsys 2010 system. High-sensitivity C-reactive protein (hs-CRP), D-dimer and other biochemical indicators were measured. Patients were divided into two groups according to hs-TnT concentration on admission (<or ≥0.014ng/mL). More than half (61.2%) of the 103 included patients had an hs-TnT concentration ≥0.014ng/mL. hs-TnT concentrations were significantly higher in those who died compared with survivors (0.292±0.516 vs. 0.069±0.154ng/mL; P=0.003). Multivariable Cox regression analysis suggested that hs-TnT is an independent factor for predicting in-hospital mortality risk (odds ratio: 2.202, 95% confidence interval: 1.111-4.367; P=0.024). Kaplan-Meier curves revealed a significant increase in hospital mortality in the hs-TnT(+) group compared with the hs-TnT(-) group (P=0.021). When hs-TnT was ≥0.042ng/mL, the sensitivity and specificity in predicting hospital short-term mortality were 70.8% and 76.4%, respectively.
Our study suggests that hs-TnT concentration could be used as an early biomarker for the risk stratification of patients with type-A AAD in the emergency department; the relationship between hs-TnT concentration and long-term prognosis needs further investigation.
closed_qa
Are parents' anxiety and depression related to child fussy eating?
To examine the association between parental anxiety and depression with child fussy eating-that is, consistent rejection of particular food items. This study was embedded in Generation R, a prospective cohort from fetal life onwards in the Netherlands. Population-based. 4746 4-year-old children and their parents. Parental internalising problems (ie, symptoms of anxiety and depression) were assessed with the Brief Symptoms Inventory during pregnancy and the preschool period (child age 3 years). The food fussiness scale of the Children's Eating Behaviour Questionnaire. Maternal anxiety during pregnancy and during the child's preschool period was related to higher food fussiness sum-scores in children. For instance, per point on the anxiety scale in pregnancy, children had on average a 1.02 higher sum-score (95% CI 0.59 to 1.46) on the food fussiness scale, after adjustment for confounders. Likewise, mothers' depressive symptoms at both time points were associated with fussy eating behaviour in their children (eg, in the antenatal period: per point on the depression scale, children had a 0.91 point higher sum-score on the food fussiness scale, 95% CI 0.49 to 1.33). We found largely similar associations between fathers' internalising problems and children's fussy eating. However, fathers' anxiety during the antenatal period was not related to child fussy eating.
Maternal and paternal internalising problems were prospectively associated with fussy eating in preschoolers. Healthcare practitioners should be aware that non-clinical symptoms of anxiety and depression in parents are risk factors for child fussy eating.
closed_qa
Is screening for urine infection in well infants with prolonged jaundice required?
The National Institute for Health and Care Excellence (NICE) neonatal jaundice guidance recommends a urine culture for investigation of babies with prolonged jaundice. However, the evidence cited for this guidance is limited. We aimed to review local data and the existing literature to identify evidence to either support or refute this guidance. We retrospectively reviewed 3 years of urine cultures from our outpatient prolonged jaundice clinic. We then conducted literature review with meta-analysis of studies presenting original data on urine tract infection (UTI) rates in jaundiced and prolonged jaundiced babies. From our local data, none of the 279 patients met our unit clinical criteria for UTI. Literature review revealed considerable differences worldwide in UTI rates in both jaundiced and prolonged jaundiced cases. Using pooled data from our literature review and our local population, the incidence of UTI in prolonged jaundiced babies is 0.21% (95% CI 0.0% to 0.73%) in the UK. This is significantly lower than the figure indicated from the data from elsewhere in the world, 8.21% (95% CI 4.36% to 13.0%).
The findings both from our local data and the current literature do not support the practice of routine screening for urine infection in well babies with prolonged jaundice. In view of the above, we no longer include urine culture in screening of well infants with prolonged jaundice. We hope that NICE will re-examine the evidence and recommend changes to their guidance on the role of routine screening for urine infection in babies with prolonged jaundice.
closed_qa
Does child temperament modify the overweight risk associated with parent feeding behaviors and child eating behaviors?
Child temperament is a measure of an individual's behavioral tendencies. The primary objective of this study was to examine whether child temperament modified the overweight risk associated with parent feeding behaviors and child eating behaviors. A sample of predominantly African American, Midwest families (N = 120) recruited from four metropolitan primary care clinics participated in this cross-sectional, mixed methods study. Parents reported on feeding practices, child eating behaviors, and child temperament. Difficult temperament was not statistically related to parent feeding practices or child eating behaviors (p>0.05). Tests of interaction indicated that the risk of child overweight differed by difficult temperament and easy temperament for two child eating behaviors (emotional eating and food fussiness, p<0.05). For example, the effect of food fussiness decreased the risk of overweight for difficult temperament children but increased overweight risk for easy temperament children. Further, the effect of emotional eating increased the risk of overweight for difficult temperament children but decreased overweight risk for easy temperament children.
Tailoring parent-level interventions to child temperament or promoting environments that trigger less reactive individual responses may be effective in lowering risk of child overweight.
closed_qa
Can Google Searches Predict the Popularity and Harm of Psychoactive Agents?
Predicting the popularity of and harm caused by psychoactive agents is a serious problem that would be difficult to do by a single simple method. However, because of the growing number of drugs it is very important to provide a simple and fast tool for predicting some characteristics of these substances. We were inspired by the Google Flu Trends study on the activity of the influenza virus, which showed that influenza virus activity worldwide can be monitored based on queries entered into the Google search engine. Our aim was to propose a fast method for ranking the most popular and most harmful drugs based on easily available data gathered from the Internet. We used the Google search engine to acquire data for the ranking lists. Subsequently, using the resulting list and the frequency of hits for the respective psychoactive drugs combined with the word "harm" or "harmful", we estimated quickly how much harm is associated with each drug. We ranked the most popular and harmful psychoactive drugs. As we conducted the research over a period of several months, we noted that the relative popularity indexes tended to change depending on when we obtained them. This suggests that the data may be useful in monitoring changes over time in the use of each of these psychoactive agents.
Our data correlate well with the results from a multicriteria decision analysis of drug harms in the United Kingdom. We showed that Google search data can be a valuable source of information to assess the popularity of and harm caused by psychoactive agents and may help in monitoring drug use trends.
closed_qa
Does serum uric acid act as a modulator of cerebrospinal fluid Alzheimer's disease biomarker related cognitive decline?
The association of serum uric acid, cerebrospinal fluid (CSF) biomarkers of Alzheimer's disease (AD) and longitudinal cognitive decline was evaluated using the AD Neuroimaging Initiative database. In 271 healthy subjects, 596 mild cognitive impairment patients and 197 AD patients, serum uric acid and CSF AD biomarkers were measured at baseline, and Mini-Mental State Examination and AD Assessment Scale - Cognitive Subscale (ADAS-cog) were assessed serially (mean duration, 2.9 years). The effect of uric acid on longitudinal cognitive decline was evaluated using linear mixed effect models for Mini-Mental State Examination and ADAS-cog scores in female and male subjects separately, with possible confounders controlled (model 1). To determine the effects of uric acid independent of CSF biomarker (Aβ1-42 or tau) and to test whether the detrimental effects of CSF biomarker differ according to uric acid, CSF biomarker and its interaction with uric acid were further included in model 1 (model 2). Higher levels of uric acid were associated with slower cognitive decline, particularly in the mild cognitive impairment and dementia subgroups, and more prominently in female subjects. Model 2 with CSF Aβ1-42 showed that higher levels of uric acid were associated with a slower cognitive decline and alleviated the detrimental effect of Aβ1-42 on cognitive decline. Model 2 with CSF tau showed that higher levels of uric acid alleviated the detrimental effect of tau on cognitive decline in female subjects but not in male subjects.
Higher levels of uric acid had protective effects on longitudinal cognitive decline independent of and interactively with CSF AD biomarkers.
closed_qa
Is Preventative Long-Segment Surgery for Multi-Level Spondylolysis Necessary?
For multi-level spondylolysis patients, surgeons commonly choose to fix all the segments with pars interarticularis defect even those without slippage and not responsible for clinical symptoms. In this study, we tried to study the necessity of the preventative long-segment surgery for the defected segment without slippage in treatment of multi-level spondylolysis patients from a biomechanical perspective. We established a bi-level spondylolysis model with pars defects at L4 and L5 segments, and simulated posterior lumbar interbody fusion (PLIF) and pedicle screw fixation at L5-S1 level. Then we compared the biomechanical changes at L4 segment before and after surgery in neutral, flexion, extension, lateral bending and axial rotation position. The stress on L4 pars interarticularis was very similar before and after surgery, and reached the highest in axial rotation. The L3-L4 intradiscal pressure was almost the same, while L4-L5 intradiscal pressure changed a little in lateral bending (increase from 1.993 to 2.160 MPa) and axial rotation (decrease from 1.639 to 1.307 MPa) after surgery. The PLIF surgery caused a little increase of range of motion at adjacent L4-L5 and L3-L4 levels, but the change is very tiny (1 degree).
The PLIF surgery will not cause significant biomechanical change at adjacent segment with pars defect in multi-level spondylolysis. On the contrary, excessive long-segment surgery will damage surrounding soft tissues which are important for maintaining the stability of spine. So a preventative long-segment surgery is not necessary for multi-level spondylolysis as long as there are no soft tissue degeneration signs at adjacent level.
closed_qa
Are all first-generation antipsychotics equally effective in treating schizophrenia?
Narrative, unsystematic reviews revealed no differences in efficacy between the various first-generation antipsychotics (FGAs) resulting in the psychopharmacological assumption of comparable efficacy between the different FGAs. We sought to determine if the assumption of comparable efficacy of all FGAs can be regarded as evidence-based using meta-analytic statistics. A systematic literature survey (Cochrane Schizophrenia Group trial register) was applied to identify all RCTs that compared oral haloperidol with another oral FGA in schizophrenia. Primary outcome was dichotomous treatment response. Secondary outcomes were symptom severity measured by rating scales, discontinuation rates, and specific adverse effects. Altogether, 79 RCTs with 4343 participants published between 1962 and 1999 were included. We found a significant between-group difference only between haloperidol and nemonapride, but not for the remaining 19 investigated FGAs. There were no significant differences for discontinuation rates.
As most of the single meta-analytic comparisons can be regarded as underpowered, the evidence for the assumption of comparable efficacy of all FGAs is inconclusive. We therefore cannot confirm or reject the statements of previous narrative, unsystematic reviews in this regard. Our findings were limited by the small sample size in the individual comparisons and the low methodological quality in many included studies.
closed_qa
Is self-rated health an independent prognostic factor of six-week mortality in older patients hospitalized for an acute condition?
To determine whether self-rated health is a prognostic factor of six-week mortality, independently of other known objective prognostic factors. The SAFMA study was a prospective cohort, which recruited patients from the University Hospital of Martinique Acute Care for Elders unit (French West Indies) from January to June 2012. Patients aged 75 or older and hospitalized for an acute condition were eligible. The outcome was time to death within the six-week follow-up. The main explanatory variable was self-rated health. Sociodemographic and clinical characteristics were considered as covariates. Cox's proportional hazards model was used. The mean age of the 223 patients included was 85.1 ± 5.5 years. Six-week mortality rate was 14.8 %; none were lost to follow-up. In total, 123 claimed "very good to good" health, and 100 "medium to very poor" health. Self-rated health was the only independent prognostic factor associated with 6-week mortality (hazard ratio 2.61; 95 % confidence interval 1.18-5.77; p = .02), when adjusting for known prognostic factors such as age, dimensions of the comprehensive geriatric assessment and comorbidity burden.
The association between self-rated health and short-term mortality could have implications for clinical practice, particularly in helping in the estimation of prognosis in acute care setting.
closed_qa
Targeted MRI-guided prostate biopsy: are two biopsy cores per MRI-lesion required?
This study evaluates the feasibility of performing less than two core biopsies per MRI-lesion when performing targeted MR-guided in-bore prostate biopsy. Retrospectively evaluated were 1545 biopsy cores of 774 intraprostatic lesions (two cores per lesion) in 290 patients (66 ± 7.8 years; median PSA 8.2 ng/ml) regarding prostate cancer (PCa) detection, Gleason score, and tumor infiltration of the first (FBC) compared to the second biopsy core (SBC). Biopsies were acquired under in-bore MR-guidance. For the biopsy cores, 491 were PCa positive, 239 of 774 (31 %) were FBC and 252 of 771 (33 %) were SBC (p = 0.4). Patient PCa detection rate based on the FBC vs. SBC were 46 % vs. 48 % (p = 0.6). For clinically significant PCa (Gleason score ≥4 + 3 = 7) the detection rate was 18 % for both, FBC and SBC (p = 0.9). Six hundred and eighty-seven SBC (89 %) showed no histologic difference. On the lesion level, 40 SBC detected PCa with negative FBC (7.5 %). Twenty SBC showed a Gleason upgrade from 3 + 3 = 6 to ≥3 + 4 = 7 (2.6 %) and 4 to ≥4 + 3 = 7 (0.5 %).
The benefit of a second targeted biopsy core per suspicious MRI-lesion is likely minor, especially regarding PCa detection rate and significant Gleason upgrading. Therefore, a further reduction of biopsy cores is reasonable when performing a targeted MR-guided in-bore prostate biopsy.
closed_qa
Serial structural MRI evaluation of arthroscopy rotator cuff repair: does Sugaya's classification correlate with the postoperative clinical outcomes?
Sugaya's classification is the most commonly used for postoperative evaluation of rotator cuff repairs. However, the correlation between this classification and clinical outcomes after supraspinatus tendon repair were not performed with serial MRI examinations in standardized time intervals. This prospective case series involved 54 patients undergoing repair of the supraspinatus tendon tear. Magnetic resonance imaging (MRI, 1.5 T) was used to determine the Sugaya's classifications at 3, 6, and 12 months, and these data were correlated with the visual analog scale for pain (VAS), Constant and University of California at Los Angeles (UCLA) assessments. Patients with types I, II, and ≥III of Sugaya's classification experienced pain of 1.27 ± 1.95, 1.00 ± 1.40, and 3.43 ± 3.36, respectively (p = 0.010), according to the VAS. The Constant and UCLA scales did not differ significantly. Type II predominated, though their percentage decreased over time (from 77.8 to 66.7 %), whereas type I became more frequent (from 1.9 to 20.4 %).
The pain was more intense in patients classified as types III, IV, or V of Sugaya's classification. The postoperative appearance of the supraspinatus tendon was not correlated with the Constant and UCLA scales. The occurrence of type II, the most prevalent, decreased over time, whereas the occurrence of type I increased; these differences were not significant. Level de evidence: prospective cohort evaluation-level III.
closed_qa
Can the optimal type of stent be predicted based on clinical risk factors?
The randomized BASKET-PROVE study showed no significant differences between sirolimus-eluting stents (SES), everolimus-eluting stents (EES), and bare-metal stents (BMS) with respect to the primary end point, rates of death from cardiac causes, or myocardial infarction (MI) at 2 years of follow-up, in patients requiring stenting of a large coronary artery. Clinical risk factors may affect clinical outcomes after percutaneous coronary interventions. We present a retrospective analysis of the BASKET-PROVE data addressing the question as to whether the optimal type of stent can be predicted based on a cumulative clinical risk score. A total of 2,314 patients (mean age 66 years) who underwent coronary angioplasty and implantation of ≥1 stents that were ≥3.0 mm in diameter were randomly assigned to receive SES, EES, or BMS. A cumulative clinical risk score was derived using a Cox model that included age, gender, cardiovascular risk factors (hypercholesterolemia, hypertension, family history of cardiovascular disease, diabetes, smoking), presence of ≥2 comorbidities (stroke, peripheral artery disease, chronic kidney disease, chronic rheumatic disease), a history of MI or coronary revascularization, and clinical presentation (stable angina, unstable angina, ST-segment elevation MI). An aggregate drug-eluting stent (DES) group (n = 1,549) comprising 775 patients receiving SES and 774 patients receiving EES was compared to 765 patients receiving BMS. Rates of death from cardiac causes or nonfatal MI at 2 years of follow-up were significantly increased in patients who were in the high tertile of risk stratification for the clinical risk score compared to those who were in the aggregate low-mid tertiles. In patients with a high clinical risk score, rates of death from cardiac causes or nonfatal MI were lower in patients receiving DES (2.4 per 100 person-years, 95% CI 1.6-3.6) compared with BMS (5.5 per 100 person-years, 95% CI 3.7-8.2, hazard ratio 0.45, 95% CI 0.26-0.80, P = .007). However, they were not significantly different between receivers of DES and BMS in patients in the low-mid risk tertiles.
This exploratory analysis suggests that, in patients who require stenting of a large coronary artery, use of a clinical risk score may identify those patients for whom DES use may confer a clinical advantage over BMS, beyond lower restenosis rates.
closed_qa
Is the Lecompte technique the last word on transposition of the great arteries repair for all patients?
To compare the Lecompte technique and the spiral anastomosis (complete anatomic correction) two decades after arterial switch operation (ASO). Nine patients after primary ASO with Lecompte and 6 selected patients after spiral anastomosis were evaluated 20.8 ± 2.1 years after ASO versus matched controls. Blood flow dynamics and flow profiles (e.g. vorticity, helicity) in the great arteries were quantified from time-resolved 3D magnetic resonance imaging (MRI) phase contrast flow measurements (4D flow MR) in addition to a comprehensive anatomical and functional cardiovascular MRI analysis. Compared with spiral reconstruction, patients with Lecompte showed more vortex formation, supranatural helical blood flow (relative helicity in aorta: 0.036 vs 0.089; P<0.01), a reduced indexed cross-sectional area of the left pulmonary artery (155 vs 85 mm²/m²; P<0.001) and more semilunar valve dysfunctions (n = 5 vs 1). There was no difference in elastic aortic wall properties, ventricular function, myocardial perfusion and myocardial fibrosis between the two groups. Cross-sectional area of the aortic sinus was larger in patients than in controls (669 vs 411 mm²/m²; P<0.01). In the spiral group, the pulmonary root was rotated after ASO more towards the normal left position (P<0.01).
In this study, selected patients with spiral anastomoses showed, two decades after ASO, better physiologically adapted blood flow dynamics, and attained a closer to normal anatomical position of their great arteries, as well as less valve dysfunction. Considering the limitations related to the small number of patients and the novel MRI imaging techniques, these data may provoke reconsidering the optimal surgical approaches to transposition of the great arteries repair.
closed_qa
Is volumetric 3-dimensional computed tomography useful to predict histological tumour invasiveness?
The purpose of this study was to use Hounsfield unit (HU) thresholds of computed tomography (CT) images to predict pathological lymph node metastasis and tumour invasiveness of cT1N0M0 lung adenocarcinoma on 3D evaluations. Preoperative CT images of 211 lesions of surgically resected cT1N0M0 lung adenocarcinoma were retrospectively examined. The tumour size was calculated in 1D, 2D and 3D views. Tumours with -300 HU and over were defined as 'solid tumours', and those between -800 and -301 HU were defined as 'ground glass opacity tumours'. Tumours with -800 HU and over were assumed to be the whole tumour entity. The proportion of 'solid tumour' within the whole tumour entity was also calculated as the 'solid tumour ratio'. These were compared with pathological information. Solid tumour size and ratio were positively correlated with microscopic invasion to pleura, vessels and lymphatics in all dimensional evaluations. Pathological lymph node metastases were also well predicted by solid tumour size and ratio in all dimensional evaluations. The P-values for the receiver operating characteristic (ROC) curves of 1D, 1D ×2, 2D and 3D evaluations were: solid tumour size P = 0.013, 0.014 and 0.032; and solid tumour ratio 0.016, 0.0032 and<0.0001. In comparisons of 1D, 2D and 3D evaluations, 'solid tumour size' of the area under the curve (AUC) of ROC to detect pathological lymph node metastases was not significant. However, strikingly, the 3D solid tumour ratio was found to be significantly more accurate for the prediction of pathological lymph node metastases than the 1D and 2D solid tumour ratios on ROC evaluation (AUC: 1D 0.736, 2D 0.803 and 3D 0.882; P-values for the AUC comparisons were P = 0.013 for 3D versus 1D and P = 0.022 for 3D versus 2D). The correlations of subtypes of adenocarcinoma and the 3D solid tumour ratio were also investigated. Subtypes of adenocarcinoma were well correlated with the 3D solid tumour ratio.
Preoperative 3D CT using threshold values of -800 and -300 HU was useful for predicting pathological lymph node metastases and tumour invasiveness of cT1N0M0 lung adenocarcinoma.
closed_qa
Aortic valve replacement with sutureless prosthesis: better than root enlargement to avoid patient-prosthesis mismatch?
Aortic valve replacement in patients with a small aortic annulus may result in patient-prosthesis mismatch (PPM). Aortic root enlargement (ARE) can reduce PPM, but leads to extended cardiac ischaemia times. Sutureless valves have the potential to prevent PPM while reducing cardiac ischaemia times. Between January 2007 and December 2011, a total of 128 patients with a small aortic annulus underwent surgery for aortic valve stenosis at our centre. Thirty-six (17% male, n = 6) patients received conventional valve replacement with ARE and 92 (16% male, n = 18) subjects received sutureless valve implantation (Sorin Perceval). We conducted a comparative, retrospective study with follow-up. The sutureless group showed a significantly higher age (79 years) than the ARE patients (62 years, P<0.001) and received significantly more concomitant cardiac procedures (33%, n = 30 vs 6%, n = 2, P = 0.001). The mean operation, cardiopulmonary bypass and cross-clamp times were significantly lower in sutureless patients (147 ± 42, 67 ± 26 and 35 ± 13 min, respectively) than in ARE patients (181 ± 41, 105 ± 29 and 70 ± 19 min, respectively, P<0.001). The mean postoperative effective orifice area (EOA) indexed to the body surface area was 0.91 ± 0.2 cm(2)/m(2) in ARE patients and 0.83 ± 0.14 cm(2)/m(2) in sutureless patients (P = 0.040). The rate of patients with severe PPM was 6% (n = 2) in ARE patients and 11% (n = 8%) in sutureless patients (not significant, n.s.). The 30-day mortality rates were 2% (n = 2) in sutureless patients and 6% (n = 2) in ARE patients (n.s.). The 1- and 5-year survival rates of the sutureless group were 92 and 54% years, respectively, whereas the 1- and 5-year survival rates of the ARE group were 76% (n.s.).
Although the sutureless valve patients received significantly more concomitant procedures, all operation-associated times were significantly shorter. Despite sutureless valve patients being older, the 30-day mortality and survival rates were comparable in the two groups. Since the indexed EOA was only slightly lower and the incidence of severe PPM was not significantly higher in the sutureless valve patients, we conclude that sutureless valve implantation is an alternative to conventional ARE to treat a small aortic annulus and avoid PPM, especially in geriatric patients who benefit from the quick implantation process.
closed_qa
Is cancer a good way to die?
Despite improved outcomes, cancer remains widely feared, often because of its association with a long and protracted death as opposed to the quick death that people associate with that other common cause of adult mortality: heart disease. Former editor-in-chief of the BMJ Richard Smith's view that 'cancer is the best way to die' therefore attracted much criticism. We examined middle-aged and older adults' agreement with this view and compared their attitudes towards dying from cancer versus heart disease in terms of which was a good death. This study was part of an online survey (February 2015) in a United Kingdom (UK) population sample of 50- to 70-year olds (n = 391), with sampling quotas for gender and education. Five characteristics of 'a good death' were selected from the end-of-life literature. Respondents were asked to rate the importance of each characteristic for their own death to ensure their relevance to a population sample and the likelihood of each for death from cancer and heart disease. We also asked whether they agreed with Smith's view. At least 95% of respondents considered the selected five characteristics important for their own death. Death from cancer was rated as more likely to provide control over what happens (p < 0.001), control over pain and other symptoms (p < 0.01), time to settle affairs (p < 0.001), and time to say goodbye to loved ones (p < 0.001) compared with death from heart disease, but there were no differences in expectation of living independently until death (p > 0.05). Almost half (40%) agreed that cancer is 'the best way to die', with no differences by age (p = 0.40), gender (p = 0.85), or education (p = 0.27).
Despite the media commotion, a surprisingly high proportion of middle-aged and older adults viewed cancer as 'the best way to die' and rated cancer death as better than heart disease. Given that one in two of us are likely to be diagnosed with cancer, conversations about a good death from cancer may in a small way mitigate fear of cancer. Future research could explore variations by type of cancer or heart disease and by previous experience of these illnesses in others.
closed_qa
Progressive radiolucent lines following the implantation of the cemented Rimfit acetabular component in total hip arthroplasty using the rimcutter technique: cause for concern?
The highly cross-linked polyethylene Exeter RimFit flanged cemented acetabular component was introduced in the United Kingdom in 2010. This study aimed to examine the rates of emergence of radiolucent lines observed when the Rimfit acetabular component was implanted at total hip arthroplasty (THA) using two different techniques: firstly, the 'rimcutter' technique in which the flange sits on a pre-prepared acetabular rim; and secondly, the 'trimmed flange' technique in which the flange is trimmed and the acetabular component is seated inside the rim of the acetabulum. The radiographs of 150 THAs (75 'rimcutter', 75 'trimmed flange') involving this component were evaluated to assess for radiolucencies at the cement/bone interface by three observers. Rimfit acetabular components implanted using the rimcutter technique had significantly higher rates of radiolucency than those introduced using the 'trimmed flange' technique one year post-operatively (one zone: 63/75 (84%) vs 17/75 (23%); two zones 42/75 (56%) vs 0/75 (0%); all three zones 17/75 (23%) vs 0/75 (0%):(all p<0.001).
On the basis of these findings, we have stopped using the 'rimcutter' technique when implanting the Rimfit acetabular component and have reverted to the 'trimmed flange' technique.
closed_qa
Do we really know our patient population in database research?
While use of large national clinical databases for orthopaedic trauma research has increased dramatically, there has been little study of the differences in populations contained therein. In this study we aimed to compare populations of patients with femoral shaft fractures across three commonly used national databases, specifically with regard to age and comorbidities. Patients were identified in the Nationwide Inpatient Sample (NIS), National Surgical Quality Improvement Program (NSQIP) and National Trauma Data Bank (NTDB). The distributions of age and Charleston comorbidity index (CCI) reflected a predominantly older population with more comorbidities in NSQIP (mean age 71.5; sd 15.6), mean CCI 4.9; sd 1.9) than in the NTDB (mean age 45.2; sd 21.4), mean CCI = 2.1; sd 2.0). Bimodal distributions in the NIS population showed a more mixed population (mean age 56.9; sd 24.9), mean CCI 3.2; sd 2.3). Differences in age and CCI were all statistically significant (p<0.001).
While these databases have been commonly used for orthopaedic trauma research, differences in the populations they represent are not always readily apparent. Care must be taken to understand fully these differences before performing or evaluating database research, as the outcomes they detail can only be analysed in context.
closed_qa
Outpatient Laparoscopic Appendectomy: Is It Time to End the Discussion?
Laparoscopic appendectomy is typically associated with inpatient hospitalization averaging between 1 and 2 days. In July 2010, a prospective protocol for outpatient laparoscopic appendectomy was adopted at our institution. Patients were dismissed from the post-anesthesia recovery room or day surgery if they met certain predefined criteria. Patients admitted to a hospital room as either full admission or observation status were considered failures of outpatient management. An IRB-approved, retrospective review of a prospective database was performed on all patients having laparoscopic appendectomy for uncomplicated appendicitis from July 2010 through December 2014. Study exclusions included age younger than 17 years, pregnancy, interval appendectomy, and gangrenous or perforated appendicitis. Patient demographics, success with outpatient management, morbidity, and readmissions were analyzed. Five hundred and sixty-three patients underwent laparoscopic appendectomy for uncomplicated appendicitis during this time frame. There were 281 men and 282 women, with a mean age of 35.5 years. Four hundred and eighty-four patients (86%) were managed as outpatients. Seventy-nine patients were admitted for pre-existing conditions (32 patients), postoperative morbidity (10 patients), physician discretion (6 patients), or lack of transportation or support at home (31 patients). Thirty-eight patients (6.7%) experienced postoperative morbidity. Seven patients (1.2%) were readmitted after outpatient management for transient fever, nausea/vomiting, migraine headache, urinary tract infection, partial small bowel obstruction, and deep venous thrombosis. There were no mortalities or reoperations. Including the readmissions, overall success with outpatient management was 85%.
Outpatient laparoscopic appendectomy can be performed with a high rate of success, low morbidity, and low readmission rate. This protocol has withstood the test of time. Widespread adoption has the potential for substantial health care savings.
closed_qa
Are physical activity, sedentary behaviors and sleep duration associated with body mass index-for-age and health-related quality of life among high school boys and girls?
Previous studies reported lower health-related quality of life (HRQOL) scores in overweight and obese adolescents compared to their normal weight counterparts; however, few studies investigated the association between obesity-related behaviors including physical activity and sedentary behaviors and HRQOL in adolescents. This study aimed at investigating the relationship between physical activity, sedentary behaviors, sleep duration and body mass index-for-age (BMI-for-age) and HRQOL among high school Tehranian students. A total of 465 high school students (48.8 % girls) were recruited from three different socio-economic zones in Tehran. The BMI-for-age was determined and physical activity and HRQOL were assessed using validated questionnaires including Quantification de l'Activite Physique en Altitude Chez les Enfants (QAPACE) and Pediatric Quality of Life Inventory (PedsQL) respectively. Over one third of students (38.5 %) were either overweight or obese. The means of all self- and parent-reported HRQOL scores were significantly lower in girls, compared to boys, except for the parent-reported social functioning subscale. Mean hours of daily sleeping were significantly higher in girls, compared to boys (8.16 ± 1.27 vs. 7.73 ± 1.22 respectively; p<0.05). Both girls and boys spent more time on sedentary activities than engaging in sport activities. During school and vacation periods, boys had significantly higher daily energy expenditure (DEE) compared to girls (p<0.05). Only DEE during school period had a significant inverse correlation with BMI-for-age in boys but not girls (r = -0.14, p<0.05). In addition, self-reported HRQOL scores were significantly associated with weekly hours adolescents spent on videogames/internet, listening to music and reading, watching TV, sports as well as DEE through sex-specific patterns. However according to parents' perspective only weekly hours spent on listening to music and readings and sport activities were significantly associated with their children HRQOL scores.
In summary, time spent on physical and sedentary activities were not associated with BMI-for-age, although both of these were associated with the HRQOL of high school students. The potential role of sedentary activities and physical activity should be considered in future interventions aimed at improving HRQOL in adolescents.
closed_qa
Correlates of unmet need for contraception in Bangladesh: does couples' concordance in household decision making matter?
A large body of literature has highlighted that women's household decision-making power is associated with better reproductive health outcomes, while most of the studies tend to measure such power from only women's point of view. Using both husband's and wife's matched responses to decision-making questions, this study examined the association between couples' concordant and discordant decision makings, and wife's unmet need for contraception in Bangladesh. This study used couple's data set (n=3336) from Bangladesh Demographic and Health Survey of 2007. Multivariate logistic regression was used to examine the likelihood of unmet need for contraception among married women of reproductive age. Study results suggested that couples who support the equalitarian power structure seemed to be more powerful in meeting the unmet demand for contraception. Logistic regression analysis revealed that compared to couple's concordant joint decision making, concordance in husband-only or other's involvement in decision making was associated with higher odds of unmet need for contraception. Wives exposed to family planning information discussed family planning more often with husbands, and those from richest households were less likely to have unmet need for contraception.
Couple's concordant joint decision making, reflecting the concept of equalitarian power structure, appeared to be a significant analytic category. Policy makers in the field of family planning may promote community-based outreach programs and communication campaigns for family planning focusing on egalitarian gender roles in the household.
closed_qa
Is strength training associated with mortality benefits?
The relationship between strength training (ST) behavior and mortality remains understudied in large, national samples, although smaller studies have observed that greater amounts of muscle strength are associated with lower risks of death. We aimed to understand the association between meeting ST guidelines and future mortality in an older US adult population. Data were analyzed from the 1997-2001 National Health Interview Survey (NHIS) linked to death certificate data in the National Death Index. The main independent variable was guideline-concordant ST (i.e. twice each week) and dependent variable was all-cause mortality. Covariates identified in the literature and included in our analysis were demographics, past medical history, and other health behaviors (including other physical activity). Given our aim to understand outcomes in older adults, analyses were limited to adults age 65years and older. Multivariate analysis was conducted using multiple logistic regression analysis. During the study period, 9.6% of NHIS adults age 65 and older (N=30,162) reported doing guideline-concordant ST and 31.6% died. Older adults who reported guideline-concordant ST had 46% lower odds of all-cause mortality than those who did not (adjusted odds ratio: 0.64; 95% CI: 0.57, 0.70; p<0.001). The association between ST and death remained after adjustment for past medical history and health behaviors.
Although a minority of older US adults met ST recommendations, guideline-concordant ST is significantly associated with decreased overall mortality. All-cause mortality may be significantly reduced through the identification of and engagement in guideline-concordant ST interventions by older adults.
closed_qa
Weight gain in Turner Syndrome: association to puberty induction?
Girls with Turner Syndrome (TS) treated or not treated with growth hormone (GH) are prone to overweight. Therefore, we hypothesize that puberty induction in TS is associated with weight gain. We analyzed weight changes (BMI-SDS) between onset of GH treatment and near adult height (NAH) in 887 girls with TS enrolled in KIGS (Pfizer International Growth Database). Puberty was induced with estrogens in 646 (72·8%) girls with TS. Weight status did not change significantly between GH treatment start and 1 year later (mean difference -0·02 BMI-SDS), but increased significantly (P<0·001) until NAH (+0·40 BMI-SDS). The BMI-SDS increased +0·21 until start of puberty (P<0·001). Girls with spontaneous and induced puberty showed similar BMI-SDS changes. Puberty induction at ≥12 years was associated with a significant (P<0·001) less increase of BMI-SDS (+0·7 BMI-SDS) between baseline and NAH compared to puberty induction at<12 year (+1·0 BMI-SDS). In multiple linear regression analyses changes of BMI-SDS between baseline and NAH were negatively associated with baseline BMI-SDS (P<0·001), GH doses (P = 0·015), and age at puberty induction (P<0·001), positively with years on GH treatment (P = 0·004), while duration and dose of estrogens, its route of administration (transdermal/oral), changes of height-SDS, thyroxin and oxandrolone treatment, and karyotype did not correlate significantly to changes of BMI-SDS in this time period.
Puberty does not seem to play a major role in weight gain in girls with TS since the majority of the increases in BMI-SDS occurred before puberty. However, late puberty induction seems to decrease the risk of weight gain.
closed_qa
Does the postcoital test predict pregnancy in WHO II anovulatory women?
To assess the capacity of the postcoital test (PCT) to predict pregnancy in WHO II anovulatory women who are ovulatory on clomiphene citrate (CC). In these women, an abnormal PCT result could be associated with lower pregnancy chances, but this has never been proven or refuted. Prospective cohort study was performed between December 2009 and September 2012 for all women who started ovulation induction with CC in one university clinic and two teaching hospitals in the Netherlands. A PCT was performed in one of the first three ovulatory cycles. Ovulation induction with CC was continued for at least six cycles. The PCT was judged to be positive if at least one progressive motile spermatozo was seen in one of five high power fields at 400× magnification. The primary outcome was time to ongoing pregnancy, within six ovulatory cycles. In 152 women the PCT was performed. 135 women had a reliable, well-timed PCT. The ongoing pregnancy rate was 44/107 (41%) for a positive and 10/28 (36%) for a negative PCT. The hazard rate for ongoing pregnancy was 1.3 (95% CI 0.64-2.5) for a positive versus a negative PCT. Thirty five of 77 (46%) women with clear mucus had an ongoing pregnancy versus 12 of 45 (27%) women in whom the mucus was not clear (HR 2.0; 95% CI 1.02-3.84, p=0.04).
The present study suggests that the outcome of the postcoital test in women with WHO-II anovulation that undergo ovulation induction with CC does not have a large effect on ongoing pregnancy chances over time.
closed_qa
Laparoscopic access overview: Is there a safest entry method?
Laparoscopy is a minimally invasive technique to access the abdominal cavity, for diagnostic or therapeutic applications. Optimizing the access technique is an important step for laparoscopic procedures. The aim of this study is to assess the outcomes of different laparoscopic access techniques and to identify the safest one. Laparoscopic access questionnaire was forwarded via e-mail to the 60 centers who are partners in working group for laparoscopic and robotic surgery of the Italian Urological Society (SIU) and their American and European reference centers. The response rate was 68.33%. The total number of procedures considered was 65.636. 61.5% of surgeons use Veress needle to create pneumoperitoneum. Blind trocar technique is the most commonly used, but has the greatest number of complications. Optical trocar technique seems to be the safest, but it's the less commonly used. The 28,2% of surgeons adopt open Hasson's technique. Total intra-operative complications rate was 3.3%. Open conversion rate was 0.33%, transfusion rate was 1.13%, and total post-operative complication rate was 2.53%.
Laparoscopic access is a safe technique with low complication rate. Most of complications can be managed conservatively or laparoscopically. The choice of access technique can affect the rate and type of complications and should be planned according to surgeon experience, safety of each technique and patient characteristics. All access types have perioperative complications. According with our study, optical trocar technique seems to be the safest.
closed_qa
Does the intestinal microbial community of Korean Crohn's disease patients differ from that of western patients?
Intestinal microbiota play an important role in maintaining the homeostasis of the host immune system. To analyze the alteration of the intestinal microbial community structure in Korean Crohn's disease (CD) patients, we performed a comparative metagenomic analysis between healthy people and CD patients using fecal samples and mucosal tissues of ileocecal valve. 16S rRNA genes from fecal samples or mucosal tissues of 35 CD patients and 15 healthy controls (HC) were amplified using a universal primer set and sequenced with GS FLX Titanium. The microbial composition and diversity of each sample were analyzed with the mothur pipeline, and the association between microbial community and clinical characteristics of the patients were investigated. The contribution of bacterial groups to the intestinal microbial composition differed between CD and HC, especially in fecal samples. Global structure and individual bacterial abundance of intestinal microbial community were different between feces and ileocecal tissues in HC. In CD patients with active stage, relative abundances of Gammaproteobacteria and Fusobacteria were higher in both fecal and mucosal tissue samples. Moreover, the intestinal microbial community structure was altered by anti-tumor necrosis factor (anti-TNF) treatment.
Our 16S rRNA sequence data demonstrate intestinal dysbiosis at the community level in Korean CD patients, which is similar to alterations of the intestinal microbial community seen in the western counterparts. Clinical disease activity and anti-TNF treatment might affect the intestinal microbial community structure in CD patients.
closed_qa
Brachiobasilic fistulae: an upper limb autologous option for everyone?
Autologous arteriovenous fistulae (AVF) are the vascular access option of choice. However, Brachiobasilic fistula (BBF) are complex procedures with variable outcomes reported in the literature. Our aim was to evaluate outcomes and morbidity associated with BBF in our population. Retrospective analysis of prospectively collected data was undertaken for all 51 BBF created in our institution between January 2010 and March 2013. These were compared to an age and sex-matched group of brachiocephalic fistula (BCF) created over a similar time period. Demographic, operative and outcome data was collected for all patients. The primary end points were primary functional patency, primary-assisted patency and secondary patency at 3, 6 and 12 months. Continuous data was compared using Student's t-test and categorical data using chi-squared test (p < 0.05 is significant). Mean patient age was 57 years (range: 21-82). About 45% were male. About 73% of patients were already on hemodialysis (HD) at the time of BBF creation. BBF was the secondary or tertiary access procedure in 60.8% of patients (n = 31). BBF were associated with significant operative burden with 82% of patients requiring a general anaesthetic and median hospital admission of 2 days (range: 1-4). About 75% had a single-stage procedure. About 23.5% (n = 12) required operative revision (superficialization or anastomotic revision for stenosis). Patients with BBF spent an additional 3.45 days/year in hospital as a result of access related complications compared to 0.56 days/year in the BCF group. Primary-assisted patency at 3, 6 and 12 months was better in the BCF group than BBF group (86%, 72% and 48% vs. 71%, 59%, 33%, respectively; p < 0.01) Primary functional and secondary patency rates at 3, 6 and 12 months were 74%, 58% and 32% and 78%, 63% and 41%, respectively, with an average of 0.33 procedures per AVF to maintain patency (10 fistuloplasties, 4 IR thrombectomy and 3 surgical thrombectomy). About 32% (n = 17) of BBF were never used for HD (seven patients had a patent AVF but did not require HD >18 months following creation; nine failed to adequately mature and one patient refused second stage superficialization).
BBF are associated with a significant initial operative burden and higher complication and poorer patency rates than BCF. About one-third of BBF created in this series were never used. Given the morbidity associated with this procedure, careful consideration should be given to creation of BBF in pre-dialysis patients particularly those in whom the rate of decline in renal function is slow.
closed_qa
Does Changing Examiner Stations During UK Postgraduate Surgery Objective Structured Clinical Examinations Influence Examination Reliability and Candidates' Scores?
Objective structured clinical examinations (OSCE) are widely used for summative assessment in surgery. Despite standardizing these as much as possible, variation, including examiner scoring, can occur which may affect reliability. In study of a high-stakes UK postgraduate surgical OSCE, we investigated whether examiners changing stations once during a long examining day affected marking, reliability, and overall candidates' scores compared with examiners who examined the same scenario all day. An observational study of 18,262 examiner-candidate interactions from the UK Membership of the Royal College of Surgeons examination was carried at 3 Surgical Colleges across the United Kingdom. Scores between examiners were compared using analysis of variance. Examination reliability was assessed with Cronbach's alpha, and the comparative distribution of total candidates' scores for each day was evaluated using t-tests of unit-weighted z scores. A significant difference was found in absolute scores differences awarded in the morning and afternoon sessions between examiners who changed stations at lunchtime and those who did not (p<0.001). No significant differences were found for the main effects of either broad content area (p = 0.290) or station content area (p = 0.450). The reliability of each day was not affected by examiner switching (p = 0.280). Overall, no difference was found in z-score distribution of total candidate scores and categories of examiner switching.
This large study has found that although the range of marks awarded varied when examiners change OSCE stations, examination reliability and the likely candidate outcome were not affected. These results may have implications for examination design and examiner experience in surgical OSCEs and beyond.
closed_qa
Is Zimbabwe ready to transition from anonymous unlinked sero-surveillance to using prevention of mother to child transmission of HIV (PMTCT) program data for HIV surveillance?
Prevention of mother-to-child transmission of HIV (PMTCT) programs collect socio-demographic and HIV testing information similar to that collected by unlinked anonymous testing sero-surveillance (UAT) in antenatal settings. Zimbabwe evaluated the utility of PMTCT data in replacing UAT. A UAT dataset was created by capturing socio-demographic, testing practices from the woman's booking-card and testing remnant blood at a laboratory from 1 June to 30 September 2012. PMTCT data were collected retrospectively from ANC registers. UAT and PMTCT data were linked by bar-code labels that were temporarily affixed to the ANC register. A questionnaire was used to obtain facility-level data at 53 sites. Pooled HIV prevalence was 15.8 % (95 % CI 15.3-16.4) among 17,349 women sampled by UAT, and 16.3 % (95 % CI 15.8 %-16.9 %) among 17,150 women in PMTCT datasets for 53 sites. Pooled national percent-positive agreement (PPA) was 91.2 %, and percent-negative agreement (PNA) was 98.7 % for 16,782 women with matched UAT and PMTCT data. Based on UAT methods, overall median prevalence was 12.9 % (Range 4.0 %-19.4 %) among acceptors and refusers of HIV test in PMTCT compared to 12.5 % ((Range 3.4 %-19.5 %) among acceptors in ANC registers. There were variations in prevalence by site.
Although, there is no statistical difference between pooled HIV prevalence in UAT compared to PMTCT program, the overall PPA of 91.2 % and PNA of 98.7 % fall below World Health Organisation (WHO) benchmarks of 97.6 % and 99.6 % respectively. Zimbabwe will need to strengthen quality assurance (QA) of rapid HIV testing and data collection practices. Sites with good performance should be prioritised for transitioning.
closed_qa
Does a distal perfusion cannula reduce ischaemic complications of extracorporeal membrane oxygenation?
Extracorporeal membrane oxygenation (ECMO) provides support to patients with severe but reversible cardiac or pulmonary failure. Vascular complications of ECMO are well recognized. We performed a retrospective review of 70 patients (mean age 48 years; 15-85) who received peripheral veno-arterial ECMO from 2004 to 2010 in a single centre. For statistical analysis, chi-squared test and multivariate binary logistic regression analysis were used to assess for association between response variables (i.e. limb ischaemia, ECMO site bleeding and deep vein thrombosis (DVT)) and possible predictive variables. There were 14 (20%) cases of acute limb ischaemia with no statistically significant relationship between acute limb ischaemia and independent variables. Thirty-three patients received distal limb cannulas (47%). There was no statistically significant association between limb ischaemia and presence of distal limb cannula (P = 0.8). Multivariate binary logistic regression analysis identified insertion by cutdown as a predictor of lower probability of insertion site bleeding (n = 12, odds ratio 0.24, P = 0.04). Seven cases of DVT were identified; multivariate binary logistic regression analysis identified insertion by cutdown (odds ratio 0.08, P = 0.03) and days of ECMO less than five (odds ratio 0.08, P = 0.04) as predictive factors for reduced rates of DVT.
Ischaemic complications of ECMO are common and occur despite the presence of a distal limb-perfusing cannula; however in our study the distal limb cannula was a limb-salvaging intervention in six patients. Prolonged time on ECMO is a risk factor for DVT, and a high index of suspicion must be maintained. Percutaneous insertion was associated with higher rates of bleeding and DVT.
closed_qa
Can a 'rural day' make a difference to GP shortage across rural Germany?
The 'Verbundweiterbildung(plus) Baden-Württemberg' (Verbundweiterbildung(plus)) - a structured general practice (GP) postgraduate training program - developed a 'rural day' as an intervention to ensure rural health exposure for GP trainees. This project report presents the outcome of the intervention: intention to work in a rural area. At the time of the study, 500 GP trainees were enrolled in Verbundweiterbildung(plus). Invitations to complete an internet-based questionnaire were distributed by email to the 274 GP trainees who had previously been asked to participate in one of eight rural days. The questionnaire contained 4 generic and 10 specific items for participants in the rural days and 4 items specific to non-participants. Mixed-methods analyses were performed using the Statistical Package for the Social Sciences. Of the 274 GP trainees invited to complete the questionnaire, 38 of 80 rural day participants and 72 of 194 non-participants completed the questionnaire (overall response rate 40%). Participants and non-participants did not differ significantly regarding age, sex and origin. Nearly half of the participants claimed the rural day changed their overall attitudes towards rural areas positively. However, there was no significant difference in the intention to work in a rural area for participants before and after the rural day experience.
The rural day increased positive attitudes towards rural areas in general but had no influence on intention to work in rural practice. More awareness and responsibility regarding workforce shortages in rural primary health care among political stakeholders, trainers and trainees is needed. Duration of the intervention should be extended and possibly complemented by offering rotations in rural practices to increase the effect. Enabling factors and barriers regarding working in rural areas are already known by political stakeholders, trainers and trainees and need to be more adequately addressed during GP postgraduate training.
closed_qa
Does the Epi-No(®) birth trainer prevent vaginal birth-related pelvic floor trauma?
Vaginal childbirth may result in levator ani injury secondary to overdistension during the second stage of labour. Other injuries include perineal and anal sphincter tears. Antepartum use of a birth trainer may prevent such injuries by altering the biomechanical properties of the pelvic floor. This study evaluates the effects of Epi-No(®) use on intrapartum pelvic floor trauma. Multicentre prospective randomised controlled trial. Two tertiary obstetric units in Australia. Nulliparous women carrying an uncomplicated singleton term pregnancy. Participants were assessed clinically and with 4D translabial ultrasound in the late third trimester, and again at 3-6 months postpartum. Women randomised to the intervention group were asked to use the Epi-No(®) device from 37 weeks of gestation until delivery. Levator ani, anal sphincter, and perineal trauma diagnosed clinically and/or with translabial ultrasound imaging. Of 660 women randomised, 504 (76.4%) returned for assessment at a mean of 5 months postpartum. There was no significant difference in the incidence of levator avulsion [12 versus 15%; relative risk (RR) 0.82, 95% confidence interval (95% CI) 0.51-1.32; absolute risk reduction (ARR) 0.03, 95% CI -0.04 to 0.09; P = 0.39], irreversible hiatal overdistension (13 versus 15%; RR 0.86, 95% CI 0.52-1.42; ARR 0.02, 95% CI -0.05 to 0.09; P = 0.51), clinical anal sphincter trauma (7 versus 6%; RR 1.12, 95% CI 0.49-2.60; ARR -0.01, 95% CI -0.05 to 0.06; P = 0.77), and perineal tears (51 versus 53%; RR 0.96, 95% CI 0.78-1.17; ARR 0.02, 95% CI -0.08 to 0.13; P = 0.65). A marginally higher rate of significant defects of the external anal sphincter on ultrasound was observed in the intervention group (21 versus 14%; RR 1.44, 95% CI 0.97-2.20; ARR -0.06, 95% CI -0.13 to 0.05; P = 0.07).
Antenatal use of the Epi-No(®) device is unlikely to be clinically beneficial in the prevention of intrapartum levator ani damage, or anal sphincter and perineal trauma.
closed_qa
Combining limb-sparing surgery with radiation therapy in high-grade soft tissue sarcoma of extremities - Is it effective?
Limb-sparing surgery in combination with radiation therapy is a well-established treatment for high-grade soft tissue sarcomas of the extremities. But selection of cases and optimal sequence of irradiation and surgery still remain controversial. 769 patients with a high-grade soft tissue sarcoma of the extremities, who underwent a limb-sparing surgery, were retrospectively reviewed. Group 1 (N = 89) was treated with neo-adjuvant radiation therapy, group 2 (N = 315) with adjuvant irradiation and group 3 (N = 365) with surgery alone. After a mean follow up of 45 months 95 local recurrences occurred resulting in a local recurrence-free survival of 83.2% after 5 years and 75.9% after 10 years. Contaminated surgical margins (Odds ratio: 2.42) and previous inadequate surgeries (Odds ratio: 1.89) were identified as risk factors for failed local control. Neo-adjuvant radiation therapy provides the best local recurrence-free rate for 5 years (90.0%), whereas after 10 years (78.3%) adjuvant irradiation showed better local control. The metastatic-free rate was independent from achieved surgical margins (p = 0.179). Group 1 showed the highest rate of revision surgery (9.0%), followed by group 3 (5.5%) and group 2 (4.4%) (p = 0.085). However, the rate of irradiation-correlated side effects was higher in group 2 (15.2%) than in group 1 (11.2%) (p = 0.221).
Surgery has to be effective for successful local control and remains the mainstay of the treatment in combination with neo-adjuvant as well as adjuvant irradiation. In really wide or even radical resections the benefit of radiation therapy can be discussed as the irradiation induced side effects are not negligible.
closed_qa
Is delayed surgery related to worse outcomes in native left-sided endocarditis?
Timing of surgery in the management of infective endocarditis is controversial, and there is still no definite conclusion on how early the surgery should be performed. This study focuses on the outcomes of surgery during the active period of infective endocarditis in consideration of the duration after diagnosis. One hundred and thirty-four patients with active native valve infective endocarditis who underwent surgery from January 2006 to December 2013 were reviewed retrospectively. They were divided in 2 groups based on timing of surgery: early group (first week after diagnosis, n = 37) and delayed group (2 to 6 weeks after diagnosis, n = 97). Compared to the delayed group, the early group had significantly more patients in New York Heart Association class IV (81% vs. 43.3%), more mechanically ventilated (54.1% vs. 18.6%), more on inotropic support (62.2% vs. 38.1%), and hence a worse EuroSCORE II (14.8% vs. 8.8%). Operative mortality was comparable (5.4% vs. 10.3%) and 7-year survival was similar (77.4% vs. 74.6%). On multivariable regression analysis, delayed surgery did not impact on short- and long-term outcomes. Preoperative cardiac arrest and infection with Haemophilus, Actinobacillus, Cardiobacterium, Eikenella, or Kingella were risk factors for higher operative mortality. Predictors of poor 7-year survival were diabetes mellitus and acute renal failure.
Delayed surgery is not associated with worse outcomes. Both early and delayed approaches are safe and provide acceptable results. Timing of surgery should be tailored to each patient's clinical status, not based on duration of endocarditis alone.
closed_qa
Tomographic evaluation of Hill-Sachs lesions: is there a correlation between different methods of measurement?
Several methods are currently available to evaluate and quantify the glenoid or humeral bone loss; however, none is universally accepted, particularly in the case of Hill-Sachs (HS) lesions. To establish whether there is correlation among different methods of measuring HS lesions, and to investigate the correlation between glenoid bone loss and the various HS lesion measurements and to assess the inter-observer reliability of such measurements. We assessed computed tomography (CT) or arthro-CT scans taken from individuals with recurrent anterior glenohumeral dislocation. The scans were independently assessed by two examiners. The parameters assessed were as follows: HS lesion width and depth on the axial and coronal planes, articular arc loss on the axial plane, and percentage of glenoid bone loss on the sagittal plane. Scans from 50 shoulders were assessed. The percentage of articular arc loss and HS lesion width on the axial plane were the only measurements that exhibited strong correlation (r = 0.83; P < 0.001). The values of the correlation coefficient corresponding to HS lesion depth on the coronal plane were the lowest. Most of the measurements exhibited moderate correlation. The inter-examiner reliability was good relative to all measurements except for HS lesion width and depth on the coronal plane, for which it was moderate.
The measurements of articular arc loss and HS lesion width on the axial plane exhibited strong correlation. The inter-examiner reliability relative to articular arc loss, HS lesion width and depth on the axial plane, and glenoid bone loss was good.
closed_qa
Mast Cell Concentrations in Peripheral and Central Giant Cell Granulomas: Is there any Angiogenetic Role?
In the maxillofacial region, giant cell granulomas occur in 2 clinical forms, central and peripheral. Despite histopathological similarity between these 2 forms totally different clinical behaviors have been reported. The present study was undertaken to compare mast cell and vascular concentrations in these pathologic lesions. In this cross-sectional descriptive study, 20 pathological samples of central giant cell granuloma (CGCG) and 20 samples of peripheral giant cell granuloma (PGCG) were selected and examined through toluidine blue staining for mast cell assessment and immunohistochemical staining by VEGEF antibody for comparing the number of mast cells. T-test, chi-squared test and backward multivariate linear regression were used for statistical analysis using SPSS 20. Statistical significance was set at P<0.05. This study showed significantly greater VEGF expression and mast cell concentrations in CGCG compared to PGCG cases. Also there was a significant correlation between VEGF expression and the concentration of mast cells. No relation was found between age, sex and site of the lesion and concentration of mast cells or VEGF expression.
It is feasible that higher concentrations of mast cells in CGCG versus PGCG samples might lead to more aggressive clinical behavior via vascular proliferation and angiogenesis. However, other biologic mechanisms should be considered in this situation.
closed_qa
Blood Samples of Peripheral Venous Catheter or The Usual Way: Do Infusion Fluid Alters the Biochemical Test Results?
Most blood tests require venous blood samples. Puncturing the vein also causes pain, infection, or damage to the blood, and lymph flow, or long-term healing. This study aimed to determine and compare the biochemical laboratory value of the blood samples that were provided through: peripheral vein infusion (PVI) receiving continuous intravenous fluid; and the usual method of blood sampling. This is an interventional, quasi-experimental, and controlled study. The selected study sample included 60 patients, who were hospitalized during 2014, in the Internal Medicine, part of Martyrs of Persian Gulf, teaching hospital at Bushehr. Three blood samples were taken from each patient that were provided through PVI line (5 ml blood collected at beginning of IVC and then another 5 cc), and another case was prepared by common blood sampling (control). All the samples were analyzed in terms of sodium, potassium, urea and creatinine using SPSS Ver.19 software, by paired t-test and Pearson's correlation coefficients. There was a statistically significant difference between the amount of sodium and potassium in the first blood samples taken from the intravenous infusion line and vein puncture .However, no significant differences were found among the biochemical amount in the second blood samples taken from the intravenous infusion line and vein puncture.
We can use blood samples taken from peripheral intravenous infusion lines after 5cc discarding from the first part of the sample for measuring the value of sodium, potassium, urea and creatinine.
closed_qa
Should ultrasound-guided fine needle aspiration be considered a first-line technique in assessing a thyroid nodule?
Thyroid nodule is a common finding in the general population. Fine-needle aspiration remains a crucial step in the assessment of any thyroid nodules. The aim of this study is to compare the accuracy of needle aspiration with and without ultrasound guidance (USFNA, FNA respectively). Retrospective chart review of 150 consecutive patients who had thyroid surgery. Fine needle aspiration biopsy, histopathology results alongside demographics and nodule sizes were recorded. There was no statistical difference in the nodule size of both groups. USFNA showed a significantly higher specificity and positive predictive values compared to FNA.
When readily available, USFNA should be considered in the investigations of a thyroid nodule.
closed_qa
Impact of a Comparative Study on the Management of Scoliosis in Duchenne Muscular Dystrophy: Are Corticosteroids Decreasing the Rate of Scoliosis Surgery in the United States?
A cross-sectional analysis. The aim of this study was to determine whether the surgical treatment for scoliosis due to Duchenne muscular dystrophy (DMD) has decreased over a recent 11-year period, specifically, after the wide acceptance of glucocorticoid treatment for DMD. DMD can result in a flaccid neuromuscular scoliosis that has been traditionally treated surgically. In 2004, a comparative study demonstrated that glucocorticoid treatment decreased the progression of scoliosis in DMD. We used the Nationwide Inpatient Sample from 2001 to 2012 to identify patients with DMD undergoing spinal fusion. Demographic information (age, hospital size, location, geographic status) was collected. We examined the distribution of patient and hospital characteristics among cohorts undergoing spinal fusion from 2001 to 2004 (period 1; before publication of the comparative study), 2005 to 2008 (period 2; immediately following publication of the comparative study), and 2009 to 2012 (period 3; moderate duration following publication of the comparative study). We identified 1874 males undergoing spinal fusion. During this period, the overall rate of DMD surgeries declined by 48%-from 1.87 surgeries in 2001 to 0.97 surgeries in 2012 per million US males per year. This decline was significantly pronounced following the publication of the comparative study [periods 2 and 3; For period 2 vs. period 1: incidence rate ratio (IRR) = 0.71, 95% confidence interval (95% CI) = 0.56-0.91, P = 0.01; For period 3 vs. period 1: IRR = 0.77, 95% CI = 0.61-0.97, P = 0.03].
Our study demonstrates a significant decrease in the rate of scoliosis surgery for DMD from 2001 to 2012. It appears that the decline in surgical treatment could be related to the publication and landmark study demonstrating decreased progression of scoliosis with glucocorticoid treatment.
closed_qa
CD81 and CD48 show different expression on blood eosinophils in systemic sclerosis: new markers for disease and pulmonary inflammation?
In systemic sclerosis (SSc)-related interstitial lung disease (ILD), elevated eosinophil counts in bronchoalveolar lavage are associated with a worse outcome. We hypothesized that eosinophils may be activated in the peripheral circulation, thereby increasing their recruitment to affected tissues and contributing to inflammation and fibrosis. The aim of this study was to characterize the blood eosinophils in SSc patients. Expression levels of surface markers CD11b, CD44, CD48, CD54, CD69, CD81, and HLA-DR on CD16(low)CD9(high)-expressing eosinophils were measured by flow cytometry in whole blood from SSc patients (n = 32) and controls (n = 11). Expression levels of CD54, CD69, and HLA-DR were undetectable in all groups. CD44 and CD11b expression levels were similar between groups. CD81 expression was lower in patients compared to controls independent of disease duration (p = 0.001). CD48 expression was increased in patients with a short disease duration (<2 years) compared to both controls (p = 0.042) and patients with longer disease duration (≥ 2 years; p = 0.027). In patients with short disease duration, increased CD48 expression was associated with alveolar inflammation as measured by an increased concentration of alveolar nitric oxide (r = 0.76, p = 0.003).
Blood eosinophils change phenotype during disease evolution in SSc, and CD48 expression may be used as a biomarker for pulmonary inflammation.
closed_qa
Is Tamsulosin Effective after Shock Wave Lithotripsy for Pediatric Renal Stones?
We assessed the effect of tamsulosin as an adjunctive therapy after shock wave lithotripsy for pediatric single renal pelvic stones. A total of 120 children with a unilateral single renal pelvic stone were included in a prospective randomized, controlled study. All children were randomized to 2 equal groups. Group 1 received tamsulosin (0.01 mg/kg once daily) as adjunctive therapy after shock wave lithotripsy in addition to paracetamol while group 2 received paracetamol only. Stone clearance was defined as no renal stone fragments or fragments less than 3 mm and no pelvicalyceal system dilatation. Our study included 69 boys and 51 girls with a median age of 3.5 years and a median stone size of 1.2 cm. There was no statistically significant difference between groups 1 and 2 in stone or patient criteria. Of the children 99 (82.5%) achieved stone clearance after the first session, including 50 in group 1 and 49 in group 2. All children in each group were cleared of stones after the second session. The overall complication rate was 14.2%. There was no statistically significant difference between single session stone clearance rates (p = 0.81) and complications rates (p = 0.432) in either group. On multivariate analysis using logistic regression smaller stone size (p = 0.016) and radiopaque stones (p = 0.019) were the only predictors of stone clearance at a single shock wave lithotripsy session. Tamsulosin therapy did not affect stone clearance (p = 0.649).
Tamsulosin does not seem to improve renal stone clearance. Smaller and radiopaque renal stones have more chance of clearance after shock wave lithotripsy for pediatric single renal pelvic stones.
closed_qa
Is Blast Injury a Modern Phenomenon?
Given the recent interest in blast injury spurred by returning soldiers from overseas conflicts, we sought to research the early historical descriptions of blast injuries and their treatments. Consideration was given to specific descriptions of survivors of closed head injury and their treatment. A review of the medical and nonmedical literature was undertaken, with particular emphasis on pre-1800 descriptions of volcanic eruptions and mining accidents. Compilations of accounts of the Etna eruptions dating from 126 BC were translated into English, and early mining texts from the 1600s and 1700s were reviewed. Accumulations of flammable gases were recorded in many medieval sources and this knowledge of toxic gas which could lead to blast injury was known in the mining community by 1316. No direct attribution of injuries to blast forces was present in the historical record examined before the 1300s, although mining accounts in the 1600s detail deaths due to blast. No specific descriptions of survivors of a closed head injury were found in the mining and volcanic eruption literature.
Descriptions and warnings of blast forces were commonly written about in the medieval and Renaissance mining communities. Personal narratives as early as 1316 recognize the traumatic effects of blast injury. No mining or volcanic blast descriptions before 1800 detailed severe closed head injury survivors, suggesting greater mortality than morbidity from blast injury in the premodern era. This review also uncovered that there was no historical treatment or remedy recommended to survivors of blast injury. Blast explosions resulting in injury or death were frequently described, although in simplistic terminology.
closed_qa
Is single incision midurethral sling effective in patients with low maximal urethral closure pressure?
To ascertain whether low preoperational maximal urethral closure pressure (MUCP) affects the outcomes of single incision sling (SIS) procedures and changes MUCP values postsurgery. There were 112 (MUCP ≥ 40 cmH2O, n = 88; MUCP<40 cmH2O, n = 24) consecutive women with urodynamic stress incontinence who had undergone SIS (MiniArc) procedures included in this study. The threshold of 40 cmH2O was used since it has been shown to be a significant risk factor for failed incontinence surgery. Clinical outcomes were assessed by the cough stress test, the 1-hour pad test, the Incontinence Impact Questionnaire-Short Form, the Urogenital Distress Inventory six-item questionnaire, the Sexual Questionnaire-SF, and postoperative changes in the urodynamic parameters. A comparison of the 1-year follow-up data is presented. Three months postsurgery, a significant decrease was observed in the 1-hour pad test, from 20.6 g preoperatively to 0.73 g postoperatively (p<0.001). The objective cure rate was 82.1% without any significant differences between the two groups (p = 0.202). At 3 months and 1 year after surgery, significantly decreasing Urogenital Distress Inventory six-item questionnaire and Incontinence Impact Questionnaire-Short Form, and increasing Sexual Questionnaire-SF scores were observed in both groups, without any significant differences between the two groups. No statistically significant difference in the subjective cured rate was noted between the two groups at the 3-month and 18.4 month follow-ups. The postoperative MUCP was significantly decreased in the MUCP ≥ 40 group (p<0.05) while significantly increased in the MUCP<40 group (p = 0.006).
These results suggest that SIS is a safe and highly effective treatment for urodynamic stress incontinence even in women with low MUCP at a mean follow-up of 18.4 months. Evaluation of the outcomes with more subjects after a longer follow-up period is necessary.
closed_qa
Are the interarytenoid muscles supplied by branches of both the recurrent and superior laryngeal nerves?
It has been generally accepted that the branches of the internal branch of the superior laryngeal nerve to the interarytenoid muscle are exclusively sensory. However, some experimental studies have suggested that these branches may contain motor axons, and therefore that the interarytenoid muscle is supplied by both the superior and recurrent laryngeal nerves. The aim of this work was to determine whether motor axons to the interarytenoid muscles are present in both laryngeal nerves. Basic research. Twelve human internal branches of the superior laryngeal nerve were dissected, and its branches to the interarytenoid muscle were removed and processed for choline-acetyltransferase immunohistochemistry, a method not used previously in studying the nerve fiber composition of the laryngeal nerves. The internal branch of the superior laryngeal nerve divided into two to five branches to the interarytenoid muscle. All branches contained motor axons, with the proportion of motor axons varying from 6% to 31%.
The present study confirms that the internal branch of the superior laryngeal nerve provides a motor innervation to the interarytenoid muscles.
closed_qa
Does intensified case finding increase tuberculosis case notification among children in resource-poor settings?
Tuberculosis (TB) is a major cause of morbidity and mortality in developing countries. Passive case detection in national TB programmes is associated with low case notification, especially in children. This study was undertaken to improve detection of childhood TB in resource-poor settings through intensified case-finding strategies. A community-based intervention was carried out in six states in Nigeria. The creation of TB awareness was undertaken, and work aids, guidelines, and diagnostic charts were produced, distributed, and used. Various cadres of health workers and ad hoc project staff were trained. Child contacts with TB patients were screened in their homes, and children presenting at various hospital units were screened for TB. Baseline and intervention data were collected for evaluation populations and control populations. Detection of childhood TB increased in the evaluation population during the intervention, with a mean quarterly increase of 4.0% [new smear positive (NSP), although the increasing trend was not statistically significant (χ(2)=1.8; p<.179)]. Additionally, there was a mean quarterly increase of 3% for all forms of TB, although the trend was not statistically significant (χ(2)=1.48; p<.224). Conversely, there was a decrease in case notification in the control population, with a mean decline of 3% (all forms). Compared to the baseline, there was an increase of 31% (all forms) and 22% (NSP) in the evaluation population.
Intensified case finding combined with capacity building, provision of work aids/guidelines, and TB health education can improve childhood-TB notification.
closed_qa
Do manual therapy techniques have a positive effect on quality of life in people with tension-type headache?
Controversy exists regarding the effectiveness of manual therapy for the relief of tension-type headache (TTH). However most studies have addressed the impact of therapy on the frequency and intensity of pain. No studies have evaluated the potentially significant effect on the patient's quality of life.AIM: To assess the quality of life of patients suffering from TTH treated for 4 weeks with different manual therapy techniques. Factorial, randomized, single-blinded, controlled clinical trial. Specialized center for the treatment of headache. Seventy-six (62 women) patients aged between 18 and 65 years (age: 39.9±10.9) with either episodic or chronic TTH. Patients were divided into four groups: suboccipital inhibitory pressure; suboccipital spinal manipulation; a combination of the two treatments; control. Quality of life was assessed using the SF-12 questionnaire (considering both the overall score and the different dimensions) at the beginning and end of treatment, and after a one month follow-up. Compared to baseline, the suboccipital inhibition treatment group showed a significant improvement in their overall quality of life at the one month follow-up and also showed specific improvement in the dimensions related to moderate physical activities, and in their emotional role. All the treatment groups, but not the control group, showed improvements in their physical role, bodily pain, and social functioning at the one month follow-up. Post treatment and at the one month follow-up, the combined treatment group showed improved vitality and the two treatment groups that involved manipulation showed improved mental health.
All three treatments were effective at changing different dimensions of quality of life, but the combined treatment showed the most change. The results support the effectiveness of treatments applied to the suboccipital region for patients with TTH.
closed_qa
Detection of locally radio-recurrent prostate cancer at multiparametric MRI: Can dynamic contrast-enhanced imaging be omitted?
The goal of this study was to assess the added value of dynamic contrast-enhanced (DCE) imaging in detecting locally radio-recurrent prostate cancer using multiparametric magnetic resonance imaging (mpMRI) at 3Tesla (T). We retrospectively analyzed 45 patients with rising prostate-specific antigen level after prostate radiotherapy who underwent mpMRI [T2-weighted (T2w), diffusion-weighted (Dw) and DCE imaging] at 3T before prostate biopsy. Four readers assigned a 5-level Likert score of cancer likelihood in 8 prostate sectors (6 sextants, 2 seminal vesicles) on T2w+Dw and T2w+Dw+DCE images. Biopsy results were used as the standard of reference. T2w+Dw and T2w+Dw+DCE imaging had similar areas under the receiver operating characteristic curves on per-sector (0.87-0.89 vs. 0.87-0.89; P=0.19-0.78) and per-lobe (0.82-0.94 vs. 0.80-0.91; P=0.21-0.84) analysis. Using a Likert score≥2/5 for diagnosis threshold, T2w+Dw+DCE imaging showed non-significantly higher sensitivities on per-sector (0.56-0.72 vs. 0.52-0.73, P=0.34-0.69) and per-lobe (0.80-0.90 vs. 0.73-0.88; P=0.63-0.99) analysis. It also showed non-significantly lower specificities on per-sector (0.74-0.89 vs. 0.82-0.89; P=0.09-0.99) and per-lobe (0.48-0.81 vs. 0.61-0.84; P=0.10-0.99) analysis. Weighted kappa values were respectively 0.57-0.70 and 0.55-0.66 for T2w+Dw and T2w+Dw+DCE imaging at the sector level, and 0.66-0.83 and 0.58-0.85 at the lobe level.
The use of DCE MR imaging tends to increase sensitivity and decrease specificity for all readers, but the differences are not significant.
closed_qa
Level of use and safety of botanical products for itching vulvar dermatoses. Are patch tests useful?
Topical remedies based on botanical ingredients are popular. To assess: (i) the usage of botanical substances in subjects affected with itching and chronic vulvar complaints; (ii) the incidence of side-effects associated with their use and the frequency of contact allergy; (iii) the diagnostic usefulness of patch testing. Sixty-six patients were provided with a questionnaire to assess the prevalence and type of topical botanical preparations used and the occurrence of adverse reactions. Patients were patch tested with (i) the Italian baseline series, (ii) a topical medicament series, and (iii) a botanical series. Forty-two patients (63.6%) reported the use of natural topical products on the vulva. Seven (16.7%) noted adverse reactions; 27 showed positive reactions with the baseline series; 14 (21.2%) had at least one relevant reaction, mainly to allergens in topical products and cosmetics; and 2 (3%) showed positive reactions to the botanical series. Of the 7 patients complaining of adverse effects of botanical products, 3 (42.8%) showed relevant sensitization.
The use of natural topical products is widespread among women affected with itching vulvar diseases. Contact dermatitis is a possible adverse effect. Botanical series are of questionable usefulness, owing to the wide variety of botanical ingredients.
closed_qa
Are Cure Rates for Breast Cancer Improved by Local and Regional Anesthesia?
Recent preclinical basic science studies suggest that patient tumor immunity is altered by general anesthesia (GA), potentially worsening cancer outcomes. A single retrospective review concluded that breast cancer patients receiving paravertebral block and GA had better cancer outcomes compared with patients receiving GA alone. This study has not been validated. We hypothesized that local or regional anesthesia (LRA) would be associated with better cancer outcomes compared with GA. We retrospectively reviewed a prospectively collected database to identify all stage 0-III breast cancer patients undergoing surgery in a single center during a 9-year period ending January 1, 2010. Patients were divided into 2 groups: those who received only LRA and those who received GA. Overall survival (OS), disease-free survival (DFS), and local regional recurrence (LRR) were calculated using the Kaplan-Meier method with log-rank comparison before and after propensity score matching. Median age of the 1107 patients who met study criteria was 64 years (range, 24-97 years). Median and longest follow-up were 5.5 and 12.5 years, respectively. General anesthesia was used for 461 patients (42%), and 646 (58%) received LRA. The point estimates of cumulative OS, DFS, and LRR "free" rates at 5 years for the GA and LRA groups were 85.5% and 87.1%, 94.2% and 96.1%, and 96.3% and 95.8%, respectively. Cox regression showed no significant differences between the 2 groups (GA and LRA) for the 3 outcomes: OS (hazard ratio [HR], 0.81; 95% confidence interval [CI], 0.59-1.10; P = 0.17), DFS (HR, 0.91; 95% CI, 0.55-1.76; P = 0.87), and LRR (HR, 1.73; 95% CI, 0.83-3.63; P = 0.15).
Breast cancer OS, DFS, and LRR were not affected by type of anesthesia in our institution. This result differs from that of the only prior published clinical report on this topic and does not provide clinical corroboration of the basic science studies that suggest oncologic benefits to LRA.
closed_qa
Do physicians correctly calculate thromboembolic risk scores?
Clinical risk scores, CHADS2 and CHA2 DS2 -VASc scores, are the established tools for assessing stroke risk in patients with atrial fibrillation (AF).AIM: The aim of this study is to assess concordance between manual and computer-based calculation of CHADS2 and CHA2 DS2 -VASc scores, as well as to analyse the patient categories using CHADS2 and the potential improvement on stroke risk stratification with CHA2 DS2 -VASc score. We linked data from Atrial Fibrillation Spanish registry FANTASIIA. Between June 2013 and March 2014, 1318 consecutive outpatients were recruited. We explore the concordance between manual scoring and computer-based calculation. We compare the distribution of embolic risk of patients using both CHADS2 and CHA2 DS2 -VASc scores The mean age was 73.8 ± 9.4 years, and 758 (57.5%) were male. For CHADS2 score, concordance between manual scoring and computer-based calculation was 92.5%, whereas for CHA2 DS2 -VASc score was 96.4%. In CHADS2 score, 6.37% of patients with AF changed indication on antithrombotic therapy (3.49% of patients with no treatment changed to need antithrombotic treatment and 2.88% of patients otherwise). Using CHA2 DS2 -VASc score, only 0.45% of patients with AF needed to change in the recommendation of antithrombotic therapy.
We have found a strong concordance between manual and computer-based score calculation of both CHADS2 and CHA2 DS2 -VASc risk scores with minimal changes in anticoagulation recommendations. The use of CHA2 DS2 -VASc score significantly improves classification of AF patients at low and intermediate risk of stroke into higher grade of thromboembolic score. Moreover, CHA2 DS2 -VASc score could identify 'truly low risk' patients compared with CHADS2 score.
closed_qa
Is baseline aerobic fitness associated with illness and attrition rate in military training?
Respiratory illnesses are a leading cause of morbidity and medical discharge in the military. This study aimed to investigate the effects of baseline aerobic fitness on haematological, salivary and mood variables, and simultaneously, in a novel approach, to identify factors precipitating illness and attrition rate in recruits during military training. Thirty-five healthy male recruits from an Army Training Regiment undertaking 12 weeks of training were prospectively investigated. Their 2.4 km run time (RT) was used as a surrogate of baseline aerobic fitness. Saliva and venous blood samples were analysed for secretory IgA, full blood counts and cell cytokine production (interleukin (IL) 6 and IL-8), respectively. Each recruit completed questionnaires on mood profile, and gastrointestinal and upper respiratory tract symptoms (URTS). Significant salivary and haematological perturbations were observed and coincided with increased duration of URTS/week and mood disturbance over this military training period. From Start to End: leucocyte count decreased by 28% (p<0.001); neutrophil percentage (%) decreased by 13% (p<0.01); lymphocyte % increased by 17% (p<0.05); the neutrophil:lymphocyte ratio decreased by 22% (p<0.01); eosinophil% increased by 71% (p<0.01). From Start to Mid to End: monocyte% increased by 68% at Mid (p<0.01) but only by 30% at End (p<0.01); IL-6 increased by 39% at Mid (p<0.01) and a further 61% by End. The 2.4 km RT was significantly associated with URTS duration (p<0.01). In addition, a 1-min increase in 2.4 km RT increased a recruit's risk 9.8-fold of developing URTS lasting, on average, 3.36 days/week. In recruits ranked with high-URTS duration their RT was 48 s slower (p<0.01) than those with low-URTS, and their attrition rate reached 45%.
The least fit recruits may have found training more physically demanding as reflected in the higher URTS duration, which may have led to a high attrition rate from the Army. It is worth considering that baseline aerobic fitness might be an important factor in illness development and attrition rate in recruits during this type of military training.
closed_qa
Should the BK polyomavirus cytopathic effect be best classified as atypical or benign in urine cytology specimens?
According to The Paris System for Reporting Urinary Cytology (TPS), the category of atypical urothelial cells (AUC) should not be applied to specimens in which cellular changes can be entirely attributed to the polyoma (BK) virus cytopathic effect (CPE). Until recently, cases with BK CPE at The Johns Hopkins Hospital were categorized as atypical urothelial cells of uncertain significance (AUC-US), which is equivalent to the TPS AUC category. This study was performed to determine how significantly the rate of AUC-US specimens would decrease if specimens with only BK CPE were classified as benign. Two reviewers and 1 adjudicator re-evaluated urinary tract specimens to determine whether sufficient cytological atypia justified an AUC-US diagnosis independent of the presence of BK CPE. For patients with surgical follow-up, the rate of high-grade urothelial carcinoma (HGUC) on tissue biopsy was tracked over a 5-year period. The reclassification rate of AUC-US cases with BK CPE as benign was 62.6%. The rate of subsequent HGUC was 6.0% for cases reclassified as benign and 10.0% for cases still classified as AUC-US. These rates were not significantly elevated in comparison with control cohorts among all-comers. However, for patients without a history of HGUC, the rate of HGUC on follow-up was significantly elevated in comparison with the rate for a benign control cohort and was similar to the rate for the AUC-US control cohort.
Reclassification as benign would have decreased the rate of AUC-US from 24.8% to 20.7% during the study year. However, the high rate of subsequent HGUC among nonsurveillance patients suggests that the reclassification of specimens with BK CPE in these patients may be inappropriate. Cancer Cytopathol 2016;124:436-42. © 2016 American Cancer Society.
closed_qa