context
stringlengths
151
5.1k
answer
stringlengths
3
2.87k
question
stringlengths
1
25.7k
The behavioural and socio-cultural processes underlying the association between socio-economic position (SEP) and body mass index (BMI) remain unclear. Occupational physical activity (OPA) is one plausible explanatory variable that has not been previously considered. 1) To examine the association between OPA and BMI, and 2) to examine whether OPA mediates the SEP-BMI association, in a Canadian population-based sample. This cross-sectional study was based on secondary analysis of the 2008 Canadian Community Health Survey data, focusing on adults (age 25-64) working at a job or business (men, n = 1,036; women, n = 936). BMI was based on measured height and weight and we derived a novel indicator of OPA from the National Occupational Classification Career Handbook. Our analytic technique was ordinary least squares regression, adjusting for a range of socio-demographic, health and behavioural covariates. OPA was marginally associated with BMI in women, such that women with medium levels of OPA tended to be lighter than women with low levels of OPA, in adjusted models. No associations between OPA and BMI were detected for males. Baron and Kenny's (1986) three conditions for testing mediation were not satisfied, and thus we were unable to proceed with testing OPA as a mediator.
Notwithstanding the small effects observed in women, overall the associations between OPA and BMI were neither clear nor strong, which could reflect conceptual and/or methodological reasons. Future research on this topic might incorporate other plausible explanatory variables (e.g., job-related psychosocial stress) and adopt a prospective design.
Occupational physical activity and body mass index (BMI) among Canadian adults: does physical activity at work help to explain the socio-economic patterning of body weight?
In the past 20 years, several factors were detected in the human seminal plasma and proposed as markers for spermatogenesis. Human chorionic gonadotropin (hCG) and its beta-subunit were therefore investigated, and their seminal levels were found to be higher than those detected in the serum and to correlate with sperm parameters. We designed a retrospective study to determine the suitability of hCG free beta-subunit concentration in the seminal plasma of fertile and infertile male patients as marker of spermatogenesis. A total of 79 infertile male patients were divided into four groups by their semen analysis results (group 1 [n=8]: azoospermia; group 2 [n=21]: severe oligozoospermia; group 3 [n=40]: oligoasthenospermia (OAS); group 4 [n=10]: asthenospermia) and 10 healthy volunteers of proven fertility were evaluated. The hCG free beta-subunit levels in the seminal plasma were found to be significantly higher (P<0.0001) in the control group in respect to those assayed in the infertile patients and showed a correlation with sperm count (r=0.5) and total motile sperm density (r=0.5). Twenty-five patients were on treatment with oral Mesterolone (100mg daily) plus Tamoxifen (20mg daily) for 3-6 months. Apart from a significant improvement (P<0.05) in sperm morphology, no significant changes in sperm count and motility were observed after the treatment in all the patients. In the seminal plasma of 10 patients who showed a significant increase in sperm count, hCG free beta-subunit levels were found to be significantly higher compared to those detected in the remaining patients (P<0.01). In all patients, these levels remained unchanged after the treatment.
The evidence regarding the positive correlation between hCG free beta-subunit levels in the seminal plasma and sperm concentration is consistent with the previous results regarding hCG levels. A previous study demonstrated that testosterone levels in seminal plasma correlated with sperm concentrations; from the same evidence regarding hCG we hypothesize that seminal plasma testosterone and hCG levels are correlated. Thus, hCG may play a paracrine role in the intratesticular regulation of testosterone secretion.
Human chorionic gonadotropin free beta-subunit in the human seminal plasma: a new marker for spermatogenesis?
Internal medicine fellowship programs have an incentive to select fellows who will ultimately publish. Whether an applicant's publication record predicts long term publishing remains unknown. Using records of fellowship bound internal medicine residents, we analyzed whether publications at time of fellowship application predict publications more than 3 years (2 years into fellowship) and up to 7 years after fellowship match. We calculate the sensitivity, specificity, positive and negative predictive values and likelihood ratios for every cutoff number of application publications, and plot a receiver operator characteristic curve of this test. Of 307 fellowship bound residents, 126 (41%) published at least one article 3 to 7 years after matching, and 181 (59%) of residents do not publish in this time period. The area under the receiver operator characteristic curve is 0.59. No cutoff value for application publications possessed adequate test characteristics.
The number of publications an applicant has at time of fellowship application is a poor predictor of who publishes in the long term. These findings do not validate the practice of using application publications as a tool for selecting fellows.
Can a resident's publication record predict fellowship publications?
To estimate the reliability of the Bone Conduction-HeadBand (BC-HB) test for predicting the postoperative functional outcome of a round-window (RW) vibroplasty. Within-subject comparison of the functional results of the BC-HB test, which is routinely used for the preoperative evaluation of a bone-conduction transducer, with an active middle ear implant (AMEI) placed onto the round window. Tertiary referral university hospital center. Seven patients with similar anatomic (absent stapes superstructure) and functional (moderate, mixed hearing loss) sequelae from open tympanoplasty technique. All subjects underwent preoperative audiologic assessment with the BC-HB. Subsequently, all subjects underwent surgical placement of an AMEI onto the round window. Pure tone and speech audiometry in quiet and noise were assessed. Additionally, evaluation of specific satisfactory targets was performed using the Client Oriented Scale of Improvement. Pure tone and speech audiometry in quiet established that both devices had very similar performance and provided remarkable improvement compared with the unaided condition. However, high-frequency gain and speech audiometry in noise demonstrated better performance with RW-AMEI.
In patients presenting with mixed hearing loss as a sequela from middle ear surgery, the preoperative BC-HB test may be helpful in predicting the final functional outcome and patient satisfaction with RW-AMEI.
Is the Bone-Conduction HeadBand test useful for predicting the functional outcome of a round window active middle ear implant?
The pulmonary embolism severity index (PESI) and the recently derived simplified PESI prognostic model have been developed to estimate the risk of 30-day mortality in patients with acute PE. We sought to assess if the PESI and simplified PESI prognostic models can accurately identify adverse events and to determine the rates of events in patients treated as outpatients. A retrospective cohort study of patients with acute pulmonary embolism (PE) presenting at the Ottawa Hospital (Canada) was conducted between 1 January 2007 and 31 December 2008. Two hundred and forty three patients were included. A total of 118 (48.6%) and 81 (33.3%) were classified as low risk patients using the original and simplified PESI prognostic models respectively. None of the low risk patients died within the 3months of follow-up. One hundred and fifteen (47.3%) patients were safely treated as outpatients with no deaths or bleeding episodes and only 1 recurrent event within the first 14days or after 30days of follow-up. Thirty four (29.6%) of these outpatients were classified as high risk patients according to the original PESI and 54 (47.0%) to the simplified PESI prognostic model.
Both PESI strategies accurately identify patients with acute PE who are at low risk and high risk for short-term adverse events. However, 30 to 47% of patients with acute PE and a high risk PESI score were safely managed as outpatients. Future research should be directed at developing tools that predict which patients would benefit from inpatient management.
Does the Pulmonary Embolism Severity Index accurately identify low risk patients eligible for outpatient treatment?
To describe the clinical course and steroid responsiveness of a patient with subacute proximal symmetric weakness, very high serum creatine kinase activity, and myopathic pattern with fibrillations in the electromyogram, whose muscle biopsy showed necrotizing myopathy, with practically no inflammation. Case report. Academic research. Diagnosis of muscular dystrophy was suggested; nevertheless, steroid treatment was initiated, and the patient recovered and gained normal strength. However, after a few years he stopped treatment, and all symptoms recurred. He developed severe proximal weakness of all limbs. Another biopsy showed similar findings, with no inflammation; still, he responded favorably to steroids and immunosuppressive medications. Currently on a low dose of prednisone and methotrexate, he has no neurological deficit.
The absence of inflammation in muscle biopsy may lead to misdiagnosis of muscular dystrophy; however, if the clinical impression is that of inflammatory myopathy, an immunomodulatory treatment should be initiated. During the past century, there has been much controversy about the diagnosis of polymyositis (PM). The debate is still ongoing. We present hereby a patient with typical course and clinical features of PM who underwent two muscle biopsies, several years apart, which showed necrotizing myopathy, practically without inflammation, leading to misdiagnosis of muscular dystrophy. This report brings up the dispute regarding the role of muscle biopsy in the diagnosis of PM.
Steroid-responsive myopathy: immune-mediated necrotizing myopathy or polymyositis without inflammation?
Childhood obesity and its consequences are a growing threat to national economies and health services. The aim of this study was to determine associations between waist-to-height ratio (WHtR) as a measure of central obesity, and health-related quality of life (HRQoL) and absenteeism of primary school children in the state of Baden-Württemberg, Germany. Cross-sectional data from 1888 first and second grade children (7.1±0.6 years) participating in the baseline measurements of the Baden-Württemberg Study were analyzed. Parents completed questionnaires including a rating of their children's HRQoL using KINDLR and EQ5D-Y VAS. Days of absence because of illness, and number of visits to a physician during the last year of school/kindergarten were asked, as well as the number of days parents took off work to care for their sick child. Anthropometric measurements were taken by trained staff. The Mann-Whitney-U test was used for statistical analysis of differences between WHtR groups. Logistic regression models were used to identify factors associated with sick days. A total of 158 (8.4%) children were centrally obese (WHtR ≥0.5). These children had significantly more sick days (9.05 vs. 6.84, p<0.001) and visits to a physician (3.58 vs. 2.91, p<0.05), but not days of parental absence than other children. According to regression analysis, sick days were also associated with age, migration status, physical activity pattern, maternal health awareness and family education level. Parent-rated HRQoL was significantly lower in centrally obese children for the EQ5D-Y VAS (88.1 vs. 91.6, p<0.001), and the KINDLR subscales 'school' (79.9 vs. 82.5, p<0.05) and 'friends' (75.4 vs. 78.3, p<0.05), but not for the total score.
Cross-sectional results show higher rates of absence, more visits to a physician and lower HRQoL in children with central obesity. Each missed day at school implies a hazard to academic achievement and each additional visit to a physician is related to higher health care costs. Thus, the negative impact of central obesity is already measurable in primary school children, which emphasizes the urgent need for early delivery of health promotion and targeted prevention.
Is central obesity associated with poorer health and health-related quality of life in primary school children?
To compare tacrolimus concentrations in blood samples using the MEIA II and MS methods at the lower and higher ends of the clinically relevant range. MEIA II was compared to our tandem MS/MS procedure by regression and difference plot analysis. Data from recently published CAP surveys comparing MEIA II with MS procedures are also analyzed. Comparison of MEIA II with our tandem MS/MS procedure by regression analysis yielded r values of 0.612 and 0.829 for tacrolimus concentrations below and above 9.0 ng/mL respectively, as determined by MEIA II. Below 9 ng/mL between day imprecision of tacrolimus controls gave CVs of 12.16 for MEIA II and 7.82 for tandem MS/MS. Addition of known amounts of tacrolimus to EDTA whole blood gave percent target values of 148 and 130 at concentrations of 5.0 and 17.5 ng/mL respectively as determined by MEIA II. Difference plots demonstrated variability in the mean bias for MEIA II of 43% and 36% at the lower and higher concentration ranges respectively. Analysis of the CAP surveys suggested that the relative positive bias and inter-laboratory variability with MEIA II was more pronounced when the MEIA II median value was below 9.0 ng/mL.
Our results along with the CAP survey data suggest that tacrolimus concentrations below 9.0 ng/mL measured by MEIA II are questionable and should be interpreted with caution.
IMx tacrolimus II assay: is it reliable at low blood concentrations?
To investigate whether radical nephrectomy (RN) and nephron-sparing surgery (NSS) for T1 renal cell carcinoma influence renal function, oncological outcome or survival rate. A retrospective study was performed, including 290 nephrectomies for tumours of a diameter of less than 7 cm; 174 radical nephrectomies were compared to 116 nephron-sparing surgeries. Preoperative and pathological data were compared between the two groups. The glomerular filtration rate was estimated using the abbreviated Modification of Diet and Renal Disease (MDRD4) study equation. The evolution of renal function was analysed from 6 months to 4 years after surgery, and the oncological outcomes were evaluated by means of cancer and non-cancer survival curves. The results showed a major impairment in renal function in the RN group compared to those who underwent NSS (25 vs 7 ml/min/1.73 m², 6 months after surgery), a difference that was maintained over time. Moreover, patients undergoing RN had a greater chance of developing renal failure. Overall, the survival curves showed a higher mortality rate for the RN group (p = 0.034), although the cancer-specific mortality rate did not show any statistically significant differences (p = 0.079).
For stage T1 renal cortical tumours, NSS should, whenever possible, be regarded as the primary therapeutic option, given that it obtains similar oncological outcomes to RN and preserves renal function, which seems to translate into a lower overall mortality rate.
Can partial nephrectomy preserve renal function and modify survival in comparison with radical nephrectomy?
Anatomical segmentectomy is again under evaluation for the cure of T1a N0 non-small cell lung cancer and carcinoid tumors. Whether anatomical segmentectomy does permit or not, an adequate resection of nodal stations for staging or cure is still pending. A case-matched study was ruled on patients with peripheral cT1a N0 M0 tumors that underwent anatomical segmentectomy or lobectomy. Dissection of lymph node stations 4, 5, 6, and 7 was identical in anatomical segmentectomy and lobectomy; stations 10, 11, 12, and 13 were also dissected carefully during anatomical segmentectomy. We individually matched 46 (69% men) anatomical segmentectomy with 46 (71% men) lobectomy for age, anatomical segment, and size of the tumor. The median (interquartile range) size of the resected lesions was 1.7 cm (1.35-1.95 cm) in anatomical segmentectomy and 1.6 cm (1.3-1.9 cm) (p = 0.96) in lobectomy. The anatomical segmentectomy and lobectomy resection margins were free of cancer. The median number (interquartile range) of total dissected lymph nodes was 12 (8-5-14) in anatomical segmentectomy compared with 13 (12-14.5) in lobectomy (p = 0.68), with a number of N1 nodes being 6 (4-7.5) and 7 (4.5-9.5) (p = 0.43), respectively, and N2 nodes 5.5 (4-7.7) and 5 (4-6.5) (p = 0.88). Only 1 patient of 46 (2%) anatomical segmentectomy was N1, whereas in lobectomy, 4% had N1 (2 patients). Freedom from recurrence at 36 months was 100% for anatomical segmentectomy and 93.5% for lobectomy (p = 0.33).
Anatomical segmentectomy for cT1a tumors compared with lobectomy procures an adequate number of N1 and N2 nodes for pathological examination. Cancer-specific survival was equivalent at 36 months.
Does anatomical segmentectomy allow an adequate lymph node staging for cT1a non-small cell lung cancer?
To evaluate whether prior pharmacy bargaining process strategies and pharmacy dependence on third parties affect the bargaining power of pharmacies in price negotiations with third parties. One-time survey. Random sample of 900 independent and small chain pharmacies in nine states: Colorado, Connecticut, Georgia, Kentucky, Minnesota, Oklahoma, Oregon, Pennsylvania, and Wisconsin. Two hundred sixteen of the returned surveys contained sufficient responses for this analysis. Survey data on pharmacy bargaining power and prior pharmacy bargaining strategies, pharmacy dependence, and market characteristics were analyzed using multiple regression in a previously developed and modified provider/third party bargaining model. Pharmacy bargaining power. Pharmacy bargaining power varied across our sample. Pharmacy bargaining power was positively related to whether a pharmacy previously bargained with the third parties, negatively related to prior requests for contract changes, and negatively related to the pharmacy's dependence on third parties in total.
Pharmacy bargaining power is related to the bargaining strategies employed by pharmacies during the previous year and the dependence of pharmacies on third party payers in total. With the prevalence of "take-it-or-leave-it" contracts from third parties, prior pharmacy bargaining behavior may affect the initial terms of the contracts that pharmacies are offered.
Third party bargaining and contract terms: a link over time?
To evaluate expressive vocabulary growth in hearing-impaired preschool children wearing hearing aids. Prospective analysis of the outcomes of children included in the 1994 German 'Goettinger Hoer-Sprachregister' (GHR) series, using a repeated-measures paradigm in six- to nine-month intervals (t1-t3). Twenty-seven children (aged 2.0-4.4 years) with bilateral sensorineural hearing loss (with averages at frequencies of 0.5, 1, 2 and 4 kHz of>20 to>90 dB in the better ear) from the 1994 GHR series. The children were diagnosed at a mean age of 31.4 months (standard deviation (SD) 10.6 months) and fitted with a binaural hearing aid at a mean age of 32.3 months (SD 10.5 months). Nonverbal intelligence was average (five missing data entries). Standardized, age-appropriate picture naming tests (the 'Sprachentwicklungstest für 2-jährige Kinder', the Kaufman Assessment Battery for Children subtest vocabulary, and the 'Aktiver Wortschatztest für drei- bis sechsjährige Kinder') were carried out at three time points and results compared with data from children with normal hearing. The test raw scores were converted to T scores (mean = 50; SD = 10). On average, the children scored far below the normative population at t1 (mean = 28.9; SD = 11.3) and slowly improved as they got older (at t3, mean = 34.1; SD = 16.1; p = 0.010). Children with mild or moderate hearing loss improved most notably (mean difference t1-t3; p = 0.001), except for one child of deaf parents. Two of the five mildly hearing-impaired children and two of the eleven moderately hearing-impaired children caught up with their normal hearing peers with regards to expressive vocabulary. Such expressive vocabulary achievements were not seen in any children with>70 dB hearing loss or in six of the eleven children (55 per cent) with a 40-70 dB hearing loss, despite receiving adequate personal amplification.
Testing expressive vocabulary size is a useful clinical tool in assessing linguistic lexical outcome.
Assessment of expressive vocabulary outcomes in hearing-impaired children with hearing aids: do bilaterally hearing-impaired children catch up?
To evaluate the effectiveness of ipsilateral lobectomy to treat unilateral, nontoxic, benign nodular goiter and to define predictive factors for recurrence. Patients undergoing thyroid lobectomy for unilateral, nontoxic, benign nodular goiter between 2002 and 2007 were included. Patients were excluded if coincidental thyroid cancer was detected at histopathologic examination and completion thyroidectomy was performed. Potential predictors of recurrence including age; sex; family history; preoperative volume of the thyroid gland; preoperative number, size, and ultrasonography characteristics of the nodules; duration of postoperative follow-up; postoperative use of thyroxine; and histopathologic diagnoses were recorded at baseline. Follow-up visits were scheduled every 3 months during the first year and every 6 months thereafter. Recurrent disease was defined as a hypoechogenic or hyperechogenic nodule larger than or equal to 3 mm detected in the remaining contralateral lobe during ultrasonography. Patients with a thyrotropin value greater than 5 mIU/L received thyroxine. Fine-needle aspiration biopsy was performed for nodules greater than 10 mm or for nodules with characteristics suggestive of malignancy. Reoperation was indicated if a nodule was greater than 3 cm in diameter, posed a risk of malignancy, or caused compression signs or symptoms. A total of 104 patients were included. Histopathologic diagnoses at initial operation were adenoma in 45 patients, colloidal nodular goiter in 45 patients, and chronic lymphocytic thyroiditis in 14 patients. Average duration of follow-up was 39.75 +/- 21.75 months (range, 5-87 months). Recurrence was seen in 63 patients (60.6%). Histopathologic characteristics of the lobectomy material (P<.001), preoperative volume of the thyroid gland (P<.006), and multinodularity (P<.011) were significant predictors of recurrence.
Higher preoperative thyroid volume, histopathologic characteristics of nodules, and multinodular disease are associated with an increased risk of recurrence in patients with unilateral nodular goiter. Unilateral lobectomy is an effective therapeutic option with low reoperation rates in unilateral benign thyroid disease.
Does unilateral lobectomy suffice to manage unilateral nontoxic goiter?
In Singapore, as strict laws are a strong deterrent against armed violence, little is known about the epidemiology of penetrating stab wound injuries. Our study aimed to investigate the epidemiology of stab wound injuries at a major trauma centre in Singapore and determine if there was a difference in severity between self-inflicted stab wound (SI) injuries and those inflicted by others (IO). We retrospectively reviewed all penetrating injuries at Tan Tock Seng Hospital, and identified and categorised all stab wound injuries as SI or IO. Basic demographic information, injury severity characteristics and outcome data were compared between these two groups. A review of all mortalities was performed, including recording the causes of death. Between 2005 and 2010, there were a total of 149 stab wound injuries, of which 24 (16.1%) were SI and 125 (83.9%) were IO injuries. Patients tended to be young (mean age 34.1 ± 14.2 years). The mean Injury Severity Score was significantly different between the SI and IO groups (8.8 ± 6.5 vs. 12.3 ± 8.1; p = 0.03). In both groups, the majority underwent an operative procedure (83.3% vs. 85.6%) and had an average hospital stay of four days.
The study confirms our hypothesis that SI injuries tend to be less severe than IO injuries and are more likely to occur at home rather than at a public area. This finding may be useful in the triage of patients with stab wound injuries.
A review of stab wound injuries at a tertiary trauma centre in Singapore: are self-inflicted ones less severe?
The increasing number of patients of more advanced age undergoing cardiac surgery means the number of those with previous curative (relapse free) mastectomy and irradiation of the chest is also increasing. A higher incidence of postoperative complications such as sternal infection in these patients is considered possible. Furthermore the question of whether mediastinal irradiation leads to a relevant internal thoracic artery (ITA) gaft damage remains unclear. In this context the benefit of arterial revascularization (CABG) using one or both ITAs is not sufficiently proven by data available from clinical studies. 70 patients (49-85 years) with previous mastectomy or Hodgkin/non-Hodgkin's disease and mediastinal irradiation underwent CABG (n = 59) or an aortic valve replacement (AVR, n = 11). 20 patients received bilateral internal thoracic artery grafts, 34 a single internal thoracic artery graft, and in 16 patients an internal thoracic artery was not used. Perioperative data and data concerning postoperative complications such as mortality, myocardial infarction, and sternal infection or refixation was gathered and compared with all other patients receiving CABG (n = 5102). An histological investigation of ITA segments was done in 12 patients. There was no significant enhancement of the perioperative risk in comparison with other patients of a corresponding age group. Internal thoracic artery damage induced by irradiation was not present. There was no increased incidence of sternal instability requiring refixation observed.
In the patient cohort investigated there is in general no need for restrictive use of the ITA in CABG.
Cardiac surgery in patients with previous carcinoma of the breast and mediastinal irradiation: is the internal thoracic artery graft obsolete?
Although a wide variety of recognized pathogens can cause community-acquired pneumonia, in many patients the etiology remains unknown after routine diagnostic workup. The aim of this study was to identify the causal agent in these patients by obtaining lung aspirates with transthoracic needle aspiration. During a 15-month period, all consecutive patients with community-acquired pneumonia who were eligible for transthoracic needle aspiration were enrolled in the study. In addition to conventional microbial methods (culture of blood and sputum, serologic studies), we performed cultures and genetic and antigen tests for common respiratory pathogens in lung aspirates. The study group consisted of 109 patients. Conventional microbial studies identified an etiology in 54 patients (50%), including Mycoplasma pneumoniae in 19 patients, Chlamydia pneumoniae in 9 patients, and Streptococcus pneumoniae in 9 patients. Among the remaining 55 patients, study of the lung aspiration provided evidence of the causal agent in 36 (65%). In 4 additional patients with a single microbial diagnosis by conventional methods, the lung sample provided evidence of an additional microorganism. The new pathogens detected by lung aspiration were S. pneumoniae in 18 patients, Haemophilus influenzae in 6 patients, Pneumocystis carinii in 4 patients, and C. pneumoniae in 3 patients; other organisms were identified in 4 patients.
In our study, S. pneumoniae was the leading cause of community-acquired pneumonia, accounting for 25% of all cases, including about one-third of the cases the cause of which could not be ascertained with routine diagnostic methods.
Is Streptococcus pneumoniae the leading cause of pneumonia of unknown etiology?
The burden of alcohol-related diseases differs widely among countries. Since the 1980s, a band of countries in Central and Eastern Europe have experienced a steep rise in deaths from chronic liver diseases and cirrhosis. A possible risk factor is the consumption of illegally produced home-made spirits in these countries containing varying amounts of aliphatic alcohols and which may be hepatotoxic. However, little is known about the composition of such beverages. To compare the concentration of short-chain aliphatic alcohols in spirits from illegal and legal sources in Hungary. Samples taken from commercial retailers and illegal sources were collected and their aliphatic patterns and alcohol concentrations were determined by gas chromatographic/mass spectrometric (GC/MS) analysis. The concentrations of methanol, isobutanol, 1-propanol, 1-butanol, 2-butanol and isoamyl alcohol were significantly higher in home-made spirits than those of from commercial sources.
The results suggest that the consumption of home-made spirits is an additional risk factor for the development of alcohol-induced cirrhosis and may have contributed to high level of liver cirrhosis mortality in Central and Eastern Europe. Restrictions on supply and sale of alcohol from illicit sources are needed urgently to reduce significantly the mortality from chronic liver disease.
Could the high level of cirrhosis in central and eastern Europe be due partly to the quality of alcohol consumed?
We aimed to examine whether women who adhered to Institute of Medicine (IOM) guidelines for gestational weight gain (GWG) had improved perinatal outcomes. This is a population-based retrospective cohort study of nulliparous women with term singleton vertex births in the United States from 2011 through 2012. Women with medical or obstetric complications were excluded. Prepregnancy body mass index was calculated using reported weight and height. Women were categorized into 4 groups based on GWG and prepregnancy body mass index: (1) weight gain less than, (2) weight gain within, (3) weight gain 1-19 lb in excess of, and (4) weight gain ≥20 lb in excess of the IOM guidelines. The χ(2) test and multivariable logistic regression analysis were used for statistical comparisons. Compared to women who had GWG within the IOM guidelines, women with excessive weight gain, particularly ≥20 lb, were more likely to have adverse maternal outcomes (preeclampsia: adjusted odds ratio [aOR], 2.78; 95% confidence interval [CI], 2.82-2.93; eclampsia: aOR, 2.51; 95% CI, 2.27-2.78; cesarean: aOR, 2.1; 95% CI, 2.14-2.19), blood transfusion (aOR, 1.22; 95% CI, 1.11-1.33), and neonatal outcomes (5-minute Apgar<4: aOR, 1.22; 95% CI, 1.14-1.31; ventilation use>6 hours: aOR, 1.24; 95% CI, 1.15-1.33; seizure: aOR, 1.53; 95% CI, 1.24-1.89). Women who gained less than IOM guidelines had lower risks of hypertensive disorders of pregnancy and obstetric interventions but were more likely to have small-for-gestational-age neonates (aOR, 1.55; 95% CI, 1.52-1.59).
Women whose GWG is in excess of IOM guidelines have higher risk of adverse maternal and neonatal outcomes, particularly in women with ≥20 lb excess weight gain above guidelines while women who had weight gain below the IOM guidelines were less likely to have maternal morbidity but had higher odds of small for gestational age.
Weight gain in pregnancy: does the Institute of Medicine have it right?
Male urethritis may be caused by mycoplasmas. Since Mycoplasma genitalium has previously been isolated from the urethra of two men with non-gonococcal urethritis (NGU), it was the aim of the study further to elucidate its role by measuring the prevalence of this organism in men with NGU. The polymerase chain reaction was used. Two different sequences of the gene coding for the main adhesin MgPa were amplified. Urethral, rectal, and throat samples from 99 male sexually transmitted disease (STD) patients with and without urethritis were studied. M genitalium DNA was demonstrated in 17/99 (17%) of the urethral swabs, but in none of the rectal and throat swabs. Significantly more patients with urethritis (13/52) were positive for M genitalium DNA than were patients without urethritis (4/47) (p<0.03). In those with urethritis M genitalium DNA was found more often in Chlamydia trachomatis negative NGU (12/34) than in those with chlamydial NGU (1/14) (p = 0.05). Attempts to culture M genitalium from the PCR positive specimens were unsuccessful.
M genitalium DNA was found significantly more often in male STD patients with non-chlamydial NGU than in men with chlamydial urethritis (p = 0.05) and in men without urethritis (p = 0.003), suggesting that M genitalium may be a cause of NGU. M genitalium DNA was not demonstrated in any of the throat or rectal swabsindicating that the urogenital tract is probably the primary site of infection or colonisation of this species.
Mycoplasma genitalium: a cause of male urethritis?
Between 1986 and 1996, the overall mean overnight length of stay for all diagnoses in Sweden decreased from 20.8 to 7.1 days. The study describes changes in surgical technique, from mastectomy to breast-conserving surgery, in treatment of female breast cancer and the parallel change in average length of hospital stay, and discusses the possible link between the trends. The study was performed as a descriptive register study on hospital admission data from the Swedish Hospital Discharge Register over a 16-year period (1980-95). During the study period, the mean length of stay for surgical curative breast cancer treatment in Sweden decreased by 56%. In 1980, the proportion of women receiving conservative surgery was 7%. At the end of the period, this share had increased to 51%. Breast-conserving surgery had an approximately 30% shorter mean length of stay compared with mastectomy. The gap was remarkably stable during the study period. The shift from mastectomy to breast-conserving surgery had a limited effect on the share of patients that went through lymph node dissection. Neither age nor the number of operations per woman could, to any significant extent, explain the decrease in mean length of stay. Approximately 14% of the overall decline can be attributed to the changes in technique.
Clinical practice style, in this case the surgical technique, has had an effect on length of stay, but the surgical technique can only to some extent explain the trend.
Do changes in surgical procedures for breast cancer have consequences for hospital mean length of stay?
The pathogenesis of cerebral arteriovenous malformations (cAVMs) is still not well understood. Generally, cAVMs are thought to be congenital lesions originating prenatally. We report a 7-year-old boy diagnosed with a de novo cAVM after 3 years of recurrent epileptic seizures. MR imaging at 4 years of age was normal. Follow-up MR imaging 3 years later demonstrated a de novo 2-cm cAVM in the right occipital lobe, confirmed by conventional angiography. We reviewed five previously reported cases of de novo cAVMs who did not have a previous neurovascular abnormality. Including our case, recurrent epileptic seizures are the major presentation (83.3 %) before de novo cAVM occurrence.
We suggest that epileptic seizure is a potential trigger of de novo cAVMs.
De novo cerebral arteriovenous malformations: is epileptic seizure a potential trigger?
Laboratory studies have shown that performance assessment judgments can be biased by "contrast effects." Assessors' scores become more positive, for example, when the assessed performance is preceded by relatively weak candidates. The authors queried whether this effect occurs in real, high-stakes performance assessments despite increased formality and behavioral descriptors. Data were obtained for the 2011 United Kingdom Foundational Programme clinical assessment and the 2008 University of Alberta Multiple Mini Interview. Candidate scores were compared with scores for immediately preceding candidates and progressively distant candidates. In addition, average scores for the preceding three candidates were calculated. Relationships between these variables were examined using linear regression. Negative relationships were observed between index scores and both immediately preceding and recent scores for all exam formats. Relationships were greater between index scores and the average of the three preceding scores. These effects persisted even when examiners had judged several performances, explaining up to 11% of observed variance on some occasions.
These findings suggest that contrast effects do influence examiner judgments in high-stakes performance-based assessments. Although the observed effect was smaller than observed in experimentally controlled laboratory studies, this is to be expected given that real-world data lessen the strength of the intervention by virtue of less distinct differences between candidates. Although it is possible that the format of circuital exams reduces examiners' susceptibility to these influences, the finding of a persistent effect after examiners had judged several candidates suggests that the potential influence on candidate scores should not be ignored.
Are Examiners' Judgments in OSCE-Style Assessments Influenced by Contrast Effects?
The evidence on whether there is work stress related dysregulation of the hypothalamic-pituitary-adrenal axis is equivocal. This study assessed the relation between work stress and diurnal cortisol rhythm in a large-scale occupational cohort, the Whitehall II study. Work stress was assessed in two ways, using the job-demand-control (JDC) and the effort-reward-imbalance (ERI) models. Salivary cortisol samples were collected six times over a normal day in 2002-2004. The cortisol awakening response (CAR) and diurnal cortisol decline (slope) were calculated. In this large occupational cohort (N = 2,126, mean age 57.1), modest differences in cortisol patterns were found for ERI models only, showing lower reward (β = -0.001, P-value = 0.04) and higher ERI (β = 0.002, P-value = 0.05) were related to a flatter slope in cortisol across the day. Meanwhile, moderate gender interactions were observed regarding CAR and JDC model.
We conclude that the associations of work stress with cortisol are modest, with associations apparent for ERI model rather than JDC model.
Is there an association between work stress and diurnal cortisol patterns?
Laparoscopic colorectal resection may induce bladder and sexual dysfunction secondary to injury to the autonomic nervous system. The aim of this study was to evaluate urinary and sexual function in male patients after laparoscopic colorectal resection for diverticular disease. From January 1997 to March 2002, we performed a retrospective analysis of urinary and sexual function in 56 consecutive male patients who had undergone laparoscopic colorectal resection for diverticular disease. Preoperative and 6-month postoperative assessment was carried out using data collected via standardized postal questionnaires. Three patients were excluded (one had a prior prostatectomy, one had Peyronie's disease, and one was treated with neuroleptics). Fifty-three patients with a mean age of 54 A+/- 2 years were included in the study. There were no conversions. The morbidity rate was 9.4%. Mean follow-up was 27 A+/- 2 months. There was no significant difference in preoperative and postoperative urinary function. Fifty-one patients (96%) were sexually active preoperatively and were still sexually active postoperatively. Compared with the preoperative period, postoperative impairment of libido, erection, ejaculation, and orgasm were not significant. Every patient was able to achieve ejaculation after the intervention, and no retrograde ejaculations were reported. One patient was unable to have an erection after the intervention.
Laparoscopic colorectal resection for diverticular disease does not significantly impair urinary and sexual function.
Does laparoscopic colorectal resection for diverticular disease impair male urinary and sexual function?
The study aims to reconcile conflicting published reports regarding the clinical efficacy of a single intraoperative dose of dexamethasone in reducing post-tonsillectomy morbidity. Systematic overview (meta-analysis). To critically evaluate the existing evidence, we performed a formal meta-analysis of eight double-blinded, randomized, placebo-controlled studies of dexamethasone in pediatric patients undergoing tonsillectomy or adenotonsillectomy. Reduction in postoperative emesis and pain, as well as early return to soft or solid diet, were studied as distinct end points. Children being given a single intraoperative dose of dexamethasone (dosing, 0.15-1.0 mg/kg; maximum dose, 8-25 mg) were two times less likely to vomit in the first 24 hours than children being given placebo (relative risk [RR] = 0.55; 95% confidence interval [CI], 0.41-0.74; P<.0001). Routine use in four children would be expected to result in one less patient having post-tonsillectomy emesis (risk difference [RD]= -0.24; 95% CI, -0.38 to -0.10; P = .0006). In addition, children being given dexamethasone were more likely to advance to a soft or solid diet on post-tonsillectomy day 1 (RR = 1.69; 95% CI, 1.02-2.79; P = .04) than those being given placebo. Because of missing data and varied outcome measures, pain could not be meaningfully analyzed as a distinct end point.
Given the frequency of tonsillectomy, relative safety and low cost of dexamethasone, and the reduction in postoperative morbidity, we recommend routine use of a single intravenous dose during pediatric tonsillectomy.
Do steroids reduce morbidity of tonsillectomy?
This article examines outcomes following repeated breast reduction using vertical scar reduction mammaplasty. The results of performing repeated breast reduction in patients for whom operative records were available for the previous breast reduction were compared with those for whom these records could not be obtained. A retrospective review of all patients who underwent repeated breast reduction for recurrent symptomatic mammary hypertrophy, inadequate volume reduction during the primary operation, and significant postoperative breast volume asymmetry was performed. Twenty-five patients had repeated breast reduction. The initial technique was known in 13 patients and unknown in 12 patients. The average total reduction per breast (including liposuction) was 658 g (range, 30 to 1150 g). Liposuction was used more often in cases for which the initial technique was unknown (p = 0.000). No patients experienced necrosis of the nipple-areola complex, and there was no significant difference in the complication rates between patients for whom the previous pedicle was known versus those in whom it was unknown (p = 0.220).
Using vertical scar reduction mammaplasty, repeated breast reduction is a safe procedure, even when the initial technique is unknown. A vertically oriented, inferior wedge excision of tissue can be safely excised, irrespective of the initial pedicle. For patients with ptosis in whom the nipple-areola complex needs to be transposed superiorly, a carefully planned and de-epithelialized superior pedicle should be used. In addition, liposuction is an important adjunct to achieve volume reduction, while limiting the amount of dissection during repeated breast reduction.CLINICAL QUESTION/
Does knowledge of the initial technique affect outcomes after repeated breast reduction?
Many published accounts of clinical trials report no differences between the treatment arms, while being underpowered to find differences. This study determined how the authors of these reports interpreted their findings. We examined 54 reports of surgical trials chosen randomly from a database of 110 influential trials conducted in 2008. Seven that reported having adequate statistical power (β ≥ 0.9) were excluded from further analysis, as were the 32 that reported significant differences between the treatment arms. We examined the remaining 15 to see whether the authors interpreted their negative findings appropriately. Appropriate interpretations discussed the lack of power and/or called for larger studies. Three of the 7 trials that did not report an a priori power calculation offered inappropriate interpretations, as did 3 of the 8 trials that reported an a priori power<0.90. However, we examined only a modest number of trial reports from 1 year.
Negative findings in underpowered trials were often interpreted as showing the equivalence of the treatment arms with no discussion of the issue of being underpowered. This may lead clinicians to accept new treatments that have not been validated.
Are surgical trials with negative results being interpreted correctly?
Helicobacter pylori infection is associated with gastric cancer, but the effect of eradication treatment on gastric cancer risk is not well defined. To determine whether H. pylori eradication treatment can reduce the risk for gastric cancer. PubMed, EMBASE, Cochrane Library, Google Scholar, and online clinical trial registers through 31 January 2009, without language restrictions. Randomized trials that compared eradication treatment with no treatment in H. pylori-positive patients and that assessed gastric cancer or progression of preneoplastic lesions during follow-up. Two authors independently reviewed articles and extracted data. Seven studies met inclusion criteria, 1 of which was excluded from pooled analysis because of clinical and methodological heterogeneity. All studies were performed in areas with high incidence of gastric cancer, mostly in Asia. Overall, 37 of 3388 (1.1%) treated patients developed gastric cancer compared with 56 of 3307 (1.7%) untreated (control) participants. In a pooled analysis of 6 studies with a total of 6695 participants followed from 4 to 10 years, the relative risk for gastric cancer was 0.65 (95% CI, 0.43 to 0.98). All studies but 1 were performed in Asia. Only 2 assessed gastric cancer incidence, and only 2 were double-blinded.
Helicobacter pylori eradication treatment seems to reduce gastric cancer risk.
Meta-analysis: can Helicobacter pylori eradication treatment reduce the risk for gastric cancer?
The Instability Severity Index Score (ISIS) includes preoperative clinical and radiological risk factors to select patients who can benefit from an arthroscopic Bankart procedure with a low rate of recurrence. Patients who underwent an arthroscopic Bankart for anterior shoulder instability with an ISIS lower than or equal to four were assessed after a minimum of 5-year follow-up. Forty-five shoulders were assessed at a mean of 79 months (range 60-118 months). Average age was 29.4 years (range 17-58 years) at the time of surgery. Postoperative functions were assessed by the Walch and Duplay and the Rowe scores for 26 patients; an adapted telephonic interview was performed for the 19 remaining patients who could not be reassessed clinically. A failure was defined by the recurrence of an anterior dislocation or subluxation. Patients were asked whether they were finally very satisfied, satisfied or unhappy. The mean Walch and Duplay score at last follow-up was 84.3 (range 35-100). The final result for these patients was excellent in 14 patients (53.8 %), good in seven cases (26.9 %), poor in three patients (11.5 %) and bad in two patients (7.7 %). The mean Rowe score was 82.6 (range 35-100). Thirty-nine patients (86.7 %) were subjectively very satisfied or satisfied, and six (13.3 %) were unhappy. Four patients (8.9 %) had a recurrence of frank dislocation with a mean delay of 34 months (range 12-72 months). Three of them had a Hill-Sachs lesion preoperatively. Two patients had a preoperative ISIS at 4 points and two patients at 3 points.
The selection based on the ISIS allows a low rate of failure after an average term of 5 years. Lowering the limit for indication to 3 points allows to avoid the association between two major risk factors for recurrence, which are valued at 2 points. The existence of a Hill-Sachs lesion is a stronger indicator for the outcome of instability repair.
Results of 45 arthroscopic Bankart procedures: Does the ISIS remain a reliable prognostic assessment after 5 years?
Pancreas graft thrombosis remains the leading non-immunologic cause of graft loss after pancreas transplantation. We studied the role of hypercoagulable states (HCS) in pancreas graft thrombosis (pthx). Between January 1, 1994, and January 1, 2003, 131 pancreas transplant recipients experienced a pthx (n = 67) or other thrombotic events. Fifty-six recipients consented to have their blood drawn and tested for the HCS. These results were compared with a control group of pancreas transplant recipients who did not experience a thrombotic event. Fisher's exact test was used to compare the groups. We found 18% of the recipients with pancreas thrombosis to have a HCS. Factor V Leiden (FVL) was found in 15% vs. 4% in the control group (p = ns) vs. 3-5% in the general white population. We found 3% of the pancreas thrombosis patients to have a prothrombin gene mutation (PGM) vs. 0% in the control group (p = ns) vs. 1-2% in the general white population.
Of pancreas transplant recipients with thrombosis, 18% had one or more of the most common factors associated with a HCS (FVL or PGM). This can be compared with 4% in a control group and 4-7% in the general white population, respectively. Although the differences are not statistically significant due to small numbers, we feel that the findings may be clinically relevant. While this is only a pilot study, it may be reasonable to screen select pancreas transplant candidates for HCS, especially FVL and PGM, until more data become available.
Do inherited hypercoagulable states play a role in thrombotic events affecting kidney/pancreas transplant recipients?
Determining neurological level of injury (NLI) is of paramount importance after spinal cord injury (SCI), although its accuracy depends upon the reliability of the neurologic examination. Here, we determine if anatomic location of cervical cord injury by MRI (MRI level of injury) can predict NLI in the acute traumatic setting. A retrospective review was undertaken of SCI patients with macroscopic evidence of cervical cord injury from non-penetrating trauma, all of whom had undergone cervical spine MRI and complete neurologic testing. The recorded MRI information included cord lesion type (intra-axial edema, hemorrhage) and MRI locations of upper and lower lesion boundary, as well as lesion epicenter. Pearson correlation and Bland-Altman analyses were used to assess the relationship between MRI levels of injury and NLI. All five MRI parameters, namely (1) upper and (2) lower boundaries of cord edema, (3) lesion epicenter, and (4) upper and (5) lower boundaries of cord hemorrhage demonstrated statistically significant, positive correlations with NLI. The MRI locations of upper and lower boundary of hemorrhage were found to have the strongest correlation with NLI (r = 0.72 and 0.61, respectively; p<0.01). A weaker (low to moderate) correlation existed between lower boundary of cord edema and NLI (r = 0.30; p<0.01). Upper boundary of cord hemorrhage on MRI demonstrated the best agreement with NLI (mean difference 0.03 ± 0.73; p<0.01) by Bland-Altman analysis.
MRI level of injury has the potential to serve as a surrogate for NLI in instances where the neurologic examination is either unavailable or unreliable.
Can anatomic level of injury on MRI predict neurological level in acute cervical spinal cord injury?
Keloid disease is a benign, quasineoplastic disease with a high recurrence rate. Mesenchymal-like stem cells (MLSC) have previously been demonstrated in keloid scars and may be involved in keloid pathobiology. However, as these cells have only been examined by single colour fluorescence activated cell sorting (FACS) alone, they need to be more comprehensively characterized so that the key cellular contributors to keloid scars can be better understood. To identify and characterize MLSC in intralesional and extralesional keloid, and to distinguish haematopoietic stem cells (HSC) from mesenchymal stem cells (MSC). Punch biopsies from intralesional (top, middle and margin) and extralesional keloid scar sites were obtained from 17 patients with a keloid. Multicolour FACS analysis using antibodies specific for HSC markers CD34 and CD117 and MSC markers CD13, CD29, CD44 and CD90 was performed on freshly isolated keloid scar cells and on passage 0 and 1 cells. This was complemented by real-time quantitative polymerase chain reaction (PCR) and immunohistological in situ analyses. Keloid scars contain distinct subpopulations of MLSCs. Cells positive for CD13, CD29, CD44 and CD90 were found to be significantly (P<0·05) higher in the top and middle compartments of keloid scars compared with extralesional skin, where cells positive for CD34, CD90 and CD117 (representing HSCs) predominated. A unique population of CD34+ cells (cells positive for CD13, CD29, CD34, CD44 and CD90) were found in keloid scars and in extralesional skin. FACS and quantitative PCR analysis showed that many of the MSC markers were progressively downregulated and all HSC markers were lost during extended keloid fibroblast culture up to passage 1.
We have found distinct subpopulations of haematopoietic and nonhaematopoietic MSC in keloid scars, whereby HSC accumulate extralesionally, while keloids seem to provide a niche environment for nonhaematopoietic MSC. Future therapy of keloids may have to target differentially both stem cell populations in order to deprive these tumours of their regenerative cell pools.
Differential distribution of haematopoietic and nonhaematopoietic progenitor cells in intralesional and extralesional keloid: do keloid scars provide a niche for nonhaematopoietic mesenchymal stem cells?
The cognitive subscale of the Alzheimer's Disease Assessment Scale (ADAS-Cog) has been established internationally as an instrument for the assessment of treatment efficacy and cognitive performance in clinical trials. There is no data about the validity and characteristics of ADAS-Cog in Hungarian sample. This study is a part of the Hungarian standardization process of ADAS-Cog. It is crucial to examine the cognitive performance of patients with pseudodementia caused by depression (D) because of its' similarities with Alzheimer's disease (AK). The objective of the study was to analyze the characteristics of the cognitive subscale for further validation purposes. The study aimed at analyzing the ADAS-Cog performance of patients with D and AK in a Hungarian sample to make future studies more accurate through more exact differentiation between the two diseases. Fourty-seven normal elderly control (KNT) subjects, 66 AK patients and 39 patients with D participated in the study. The mental state and the severity of depressive symptoms of the participants were assessed by the means of ADAS-Cog, Mini Mental State Examination (MMSE) and Beck Depression Inventory. The ADAS-Cog is sensitive to the cognitive decline of the depressed group (sensitivity=69.2%, specificity=89.4%, AUC=0.868, p>0.001). While the performance of the two patient groups differed from the KNT, the groups are overlapping and the characteristic of the ROC curve and the optimal cut-off point (D:11.8; AK:12.1) indicates that the differentiation is mediocre.
The results suggest that pseudodementia should be considered during the design of studies using ADASCog. Because the cognitive subscale can't accurately differentiate between AK and pseudodementia additional measures like BDI should be administered.
Is it pseudo-dementia?
Determine rates of local excision (LE) over time, and test the hypothesis that LE carries increased oncologic risks but reduced perioperative morbidity when compared with standard resection (SR). Despite the lack of level I/level II evidence supporting its oncologic adequacy, LE is performed for stage I rectal cancer. Surgical therapy for 35,179 patients with stage I rectal cancer diagnosed in 1989 to 2003 was examined over time, utilizing the National Cancer Database. A special study then analyzed perioperative outcomes, local recurrence and survival in 2124 patients diagnosed between 1994 and 1996, including 765 (T1, 601; T2, 164) treated by LE and 1359 (T1, 493; T2, 866) treated by SR. From 1989 to 2003, the use of LE has increased (T1, 26.6-43.7%; T2, 5.8-16.8%; P<0.001 both). The special study demonstrated significantly lower 30-day morbidity after LE versus SR (5.6% vs. 14.6%; P<0.001). After adjusting for patient and tumor characteristics, the 5-year local recurrence after LE versus SR was 12.5 versus 6.9% (P = 0.003; hazard ratio = 0.38; 95% CI, 0.23-0.62) for T1 tumors, and 22.1 versus 15.1% (P = 0.01; hazard ratio = 0.69; 95% CI, 0.44-1.07) for T2 tumors. The 5-year overall survival (T1, 77.4% vs. 81.7%, P = 0.09; T2, 67.6% vs. 76.5%, P = 0.01) was influenced by age and comorbidities but not the type of surgery.
This study provides the best evidence for both the increasing use and the associated risks of LE versus SR. For each individual patient, the benefits of LE must be balanced against the heightened risk of local failure.
Is the increasing rate of local excision for stage I rectal cancer in the United States justified?
Despite the advocated use of rehabilitation tools in clinical rehabilitation of with rheumatoid arthritis (RA) patients, little is known about the representation of the patient perspective in these tools.AIM: Aim of the study was to explore the experiences of RA patients with rehabilitation and the coverage by rehabilitation tools. Qualitative focus group study with RA patients about experiences with rehabilitation. Rheumatology rehabilitation clinic of a Dutch university hospital. Focus groups were tape recorded and transcribed verbatim. From the meaningful units, concepts were extracted and linked to the International Classification of Functioning, Disability and Health (ICF). Rehabilitation tools validated for RA were identified using a structured literature search. Using the ICF as common framework, we determined for each concept identified in the focus groups the coverage by each rehabilitation tool. Nineteen patients participated in 4 focus groups. Fifty-one concepts were identified in 368 meaningful units derived from the transcribed data. From the literature the ICF Core Sets for RA, Canadian Occupational Performance Measure, Rehabilitation Activities Profile and WHO Disability Assessment Schedule II were elected. The concepts from the focus groups were best covered by the ICF Core Sets (44 out of 51; 86%), followed by the WHODAS II (39%), RAP (35%) and COPM (16%).
With the exception of the ICF Core Sets for RA, current rehabilitation tools poorly cover the RA patients' perception on rehabilitation.
Do rehabilitation tools cover the perspective of patients with rheumatoid arthritis?
Arthrocentesis and hydraulic distention of the temporomandibular joint (TMJ) has proven to be an effective modality in treating patients exhibiting clinical findings consistent with the diagnosis of disc displacement without reduction. The purpose of this study was to investigate whether the magnetic resonance imaging (MRI) variables of effusion and/or bone marrow edema may predict treatment outcomes of arthrocentesis and hydraulic distention of the TMJ. The study group comprised 37 consecutive patients with TMJ pain, who were assigned a unilateral clinical TMJ disorder of TMJ pain associated with an internal derangement (ID) type III (disc displacement without reduction) and a TMJ pain side-related MRI diagnosis of disc displacement without reduction associated with osteoarthrosis (OA). Bilateral sagittal and coronal MRI images were obtained immediately before the operation to establish the presence or absence of ID, OA, TMJ effusion, and bone marrow edema. Pain level and mandibular range of motion (ROM) were assessed preoperatively and compared with the respective 2-month follow-up findings. Outcome criteria for success were a ROM>or=35 mm and pain reduction>50%. A logistic regression analysis was used to compute the odds ratio for TMJ effusion and bone marrow edema for successful outcomes (n = 21) versus unsuccessful (n = 16) outcomes. At the 2-month follow-up, clinical evaluation showed a significant reduction in TMJ pain during function (P = .000), a significant reduction in clinical diagnoses of TMJ disorders (P = .016), and a significant increase in ROM (P = .000). A significant increase in the risk of an unsuccessful outcome of ROM<35 mm and/or pain reduction>or=50% occurred with MRI findings of effusion (odds ratio 1:10.8 = 0.09; P = .007).
TMJ effusion may prove to be an important prognostic determinant of successful arthrocentesis. However, the data re-emphasize the concept that the prediction of a specific outcome is not a matter of simple linearity, in which the presence of 1 factor may equate with predictive ability, but rather is a function of a complex interaction among different biological variables.
Temporomandibular joint internal derangement and osteoarthrosis: are effusion and bone marrow edema prognostic indicators for arthrocentesis and hydraulic distention?
The New Zealand non-ST elevation acute coronary syndrome (NSTEACS) guideline recommends that clinically appropriate patients with combined high risk features (positive troponin and ischaemic ECG and a GRACE score>140) have coronary angiography within the first hospital day. All other ACS patients referred for angiography should be studied within 72 hours. We evaluated the relationship between risk criteria, and both the incidence and timing of angiography in our practice. 2868 consecutive patients (2007 to 2010) with NSTEACS admitted to Middlemore, Waikato and Taranaki Hospitals. Individual patient demographic, risk factor, diagnostic, investigation and in-hospital outcome data was collected prospectively using Acute PREDICT software. 391 (13.6%) patients met the combined high risk criteria. Compared with lower risk patients they were older and more likely to have known cardiac disease, diabetes, renal impairment, left ventricular failure, left ventricular systolic dysfunction and more likely to die in hospital. Patients with combined high risk were less likely than others to undergo coronary angiography (61.6% vs 75%, p<0.0001). Only a fifth of combined high risk patients referred had coronary angiography within 1 day. Only just over half of those referred for angiography were studied within 3 days.
The New Zealand guidelines high risk criteria identify one in seven patients with NSTEACS as potentially appropriate for angiography within the first day. For those referred this was infrequently achieved, and only half of all NSTEACS patients referred met the 3-day target. Implementation of a national ACS registry to support more appropriate and timely management is appropriate.
Risk stratification and timing of coronary angiography in acute coronary syndromes: are we targeting the right patients in a timely manner?
To determine serum concentrations of nitric oxide metabolites (NOX) in patients with severe ulcerative colitis and to assess whether these concentrations predict clinical outcome. Twenty-six patients (16 men and 10 women, mean age 46 years) with severe ulcerative colitis requiring hospitalization for parenteral steroid therapy. Thirteen patients had a complete clinical response and symptoms resolved after 5 days of parenteral steroid administration; 13 made an incomplete recovery and needed further treatment (six cyclosporin, seven colectomy). Serum concentrations of NOX and C-reactive protein (CRP) were measured daily for 3 days in all patients and as clinically indicated thereafter. The normal range for NOX was established by measuring the concentration in 25 healthy controls. Mean serum NOX and CRP concentrations were significantly elevated in both the patients with a complete and those with an incomplete response compared with controls (P<0.001) on day 1 and fell during the first 3 days of therapy. On day 3, mean serum concentrations of NOX and CRP were lower in the patients with a complete response, but only the difference in CRP attained statistical significance (P = 0.02). There was no correlation between NOX and CRP concentrations.
In the majority of patients with severe ulcerative colitis, circulating concentrations of NOX are increased at presentation and fall promptly during parenteral steroid therapy, irrespective of clinical outcome. However, in a small number of patients NOX concentrations do not fall during steroid treatment and such patients will probably require additional medical therapy or surgery.
Are serum concentrations of nitric oxide metabolites useful for predicting the clinical outcome of severe ulcerative colitis?
Polyetheretherketone (PEEK) has a wide range of clinical applications but does not directly bond to bone. Bulk incorporation of osteoconductive materials including hydroxyapatite (HA) into the PEEK matrix is a potential solution to address the formation of a fibrous tissue layer between PEEK and bone and has not been tested.QUESTIONS/ Using in vivo ovine animal models, we asked: (1) Does PEEK-HA improve cortical and cancellous bone ongrowth compared with PEEK? (2) Does PEEK-HA improve bone ongrowth and fusion outcome in a more challenging functional ovine cervical fusion model? The in vivo responses of PEEK-HA Enhanced and PEEK-OPTIMA®Natural were evaluated for bone ongrowth in the form of dowels implanted in the cancellous and cortical bone of adult sheep and examined at 4 and 12 weeks as well as interbody cervical fusion at 6, 12, and 26 weeks. The bone-implant interface was evaluated with radiographic and histologic endpoints for a qualitative assessment of direct bone contact of an intervening fibrous tissue later. Gamma-irradiated cortical allograft cages were evaluated as well. Incorporating HA into the PEEK matrix resulted in more direct bone apposition as opposed to the fibrous tissue interface with PEEK alone in the bone ongrowth as well as interbody cervical fusions. No adverse reactions were found at the implant-bone interface for either material. Radiography and histology revealed resorption and fracture of the allograft devices in vivo.
Incorporating HA into PEEK provides a more favorable environment than PEEK alone for bone ongrowth. Cervical fusion was improved with PEEK-HA compared with PEEK alone as well as allograft bone interbody devices.
Does PEEK/HA Enhance Bone Formation Compared With PEEK in a Sheep Cervical Fusion Model?
Most incident hemodialysis (HD) patients who initiate dialysis therapy with anemia usually can achieve a hemoglobin (Hb) level of 11 g/dL or greater (>or =110 g/L) within a few months of the initiation of recombinant human erythropoietin (EPO) therapy. However, patients unable to achieve this level may be at greater risk for adverse outcomes. Whether intractable anemia is a modifiable problem or a marker for other conditions is unclear. This question was addressed in a cohort of 130,544 incident HD patients from 1996 to 2000 who were administered EPO regularly. Medicare claims data were used to determine demographic characteristics, comorbidities, hospitalizations, and related events. Patients who did not achieve an Hb level of 11 g/dL or greater (>or =110 g/L; n = 19,096; 14.6%) during months 4 to 9 after dialysis therapy initiation were compared with those who did. Patients unable to achieve an Hb level of 11 g/dL (110 g/L) were younger and more often of nonwhite race. In addition, these patients had more comorbid conditions; experienced more hospitalizations with longer stays, more infectious hospitalizations, and more catheter insertions; and were administered more blood transfusions. EPO was administered in higher and increasing doses during the years of study among patients with intractable anemia compared with those with an Hb level of 11 g/dL or greater (>or =110 g/L), likely denoting increasing attempts to correct anemia over the years.
It is apparent that incident HD patients unable to achieve an Hb level of 11 g/dL or greater (>or =110 g/L) have a greater disease burden. The independent association of intractable anemia with such future outcomes as cardiovascular events and hospitalizations remains to be determined.
Intractable anemia among hemodialysis patients: a sign of suboptimal management or a marker of disease?
Near-infrared spectroscopy (NIRS) is a useful non-invasive tool for monitoring infants undergoing cardiac surgery. In this study, we aimed to determine the NIRS values in cyanotic and acyanotic patients who underwent corrective cardiac surgery for congenital heart diseases. Thirty consecutive infants who were operated on with the diagnosis of ventricular septal defect (n=15) and tetralogy of Fallot (n=15) were evaluated retrospectively. A definitive repair of the underlying cardiac pathology was achieved in all cases. A total of six measurements of cerebral and renal NIRS were performed at different stages of the perioperative period. The laboratory data, mean urine output and serum lactate levels were evaluated along with NIRS values in each group. The NIRS values differ in both groups, even after the corrective surgical procedure is performed. The recovery of renal NIRS values is delayed in the cyanotic patients.
Even though definitive surgical repair is performed in cyanotic infants, recovery of the renal vasculature may be delayed by up to two days, which is suggestive of a vulnerable period for renal dysfunction.
Are perioperative near-infrared spectroscopy values correlated with clinical and biochemical parameters in cyanotic and acyanotic infants following corrective cardiac surgery?
Cardiovascular disease (CVD) is the main cause of death in hemodialysis (HD) patients. Vascular calcification (VC) is common in these patients. The main objective of this study was to evaluate if a semiquantitative radiographic method is able to detect VC progression in a prospective cohort of patients and predict the risk of cardiovascular events. Secondarily, we intend to identify predictors of the presence and progression of VC. 49 patients undergoing HD for ≥ 90 days were included. At the beginning and after 12 months, the VC score (VCS) was determined by the Kauppila method, and clinical, nutritional, and laboratory markers were measured. The rates of fatal and nonfatal cardiovascular events were analyzed from months 13 to 24. Of 49 patients, 55.1% were male, 46.9% diabetic, and the mean age was 59.5 ± 14.4 years. At the beginning of the follow-up, 65.3% of the patients exhibited VC with a median VCS of 4 points. The intracellular water was negatively associated with VC and its intensity. The presence of VC was the only independent predictor of VC progression. Among patients with VC, 17 showed rapid progression, and 15 showed slow progression. The VCS was independently associated with rapid progression, while ΔCS (final VCS - initial VCS) was an independent predictor of cardiovascular events.
The Kauppila method was able to detect VC, its progression, and predict cardiovascular events. These results suggest an association of VC with nutritional status.
Is Kauppila method able to detect the progression of vascular calcification and predict cardiovascular events in patients undergoing hemodialysis?
There is a gap in the current breast cancer survivorship literature identifying potential sample biases that may result from recruiting participants via different methods. To document whether participant recruitment method influences baseline demographic or psychosocial variables and trial participation among breast cancer survivors recruited for a physical activity intervention trial. Participants were recruited for the trial via either a reactive method (letters mailed through their oncologist's office inviting them to contact the research staff) or a proactive method (referred in person by their oncologist at a clinic appointment). The groups of participants recruited via the two methods were compared based on baseline sociodemographic characteristics, weight, time since diagnosis, stage of disease, treatment, motivational readiness for physical activity, level of physical activity, self-reported physical and mental health, willingness to receive the intervention, and study retention. Participants recruited proactively were closer to the point of diagnosis (mean = 2.5 years, standard deviation (SD) = 1.9 years) than participants recruited reactively via letter mailings (mean = 3.4 years, SD = 2.3 years; p<.05). The two groups were similar with respect to all other baseline characteristics and retention. Recruitment via the two methods was not concurrent. Also, proactive recruitment occurred at a single hospital site. Mailings were made by the oncologists; we are unable to estimate how many letters were mailed. Similarly, we have no information for the patients who were not referred to the study during proactive recruitment.
Despite the potential for differences in characteristics and degree of trial participation between trial participants recruited proactively and reactively, in this investigation, the two groups were similar. Information from other trials in other conditions may confirm or modify our conclusion.
Proactive versus reactive recruitment to a physical activity intervention for breast cancer survivors: does it matter?
To establish whether laparoscopic incidental appendectomy in gynecological diseases is related to postoperative intraabdominal infection and complications. This study was performed prospectively in 443 patients who underwent laparoscopic surgery without appendectomy (n = 222) or with appendectomy (n = 221). On postoperative day 1, drain fluid was cultured in all patients. All data were compared using Student's t test and χ2 test. Bacteria grew in cultures of 93 patients (21.0%): 38 (17.1%) in the nonappendectomy group and 55 (24.9%) in the appendectomy group (p<0.01). There were statistical differences in the incidence of bacterial growth, postoperative complications, and post-operative laboratory changes for percentage of neutrophils (p<0.01) and C-reactive protein (p<0.01). Thirteen genera of bacteria grew in the drain culture. The 9 commensal organisms of the human intestine were identified in all patients, each 8 genera of bacteria in both groups. The surgical type did not affect the postoperative drain culture results.
Postoperative bacterial growth and complications were increased in the laparoscopic incidental appendectomy group. Infections with bacteria from the intestine in both groups were also related to contamination during surgery and postoperative care.
Is Laparoscopic Incidental Appendectomy in Gynecological Diseases Related to Postoperative Intraabdominal Infection and Complications?
In this study, we investigated the association of positive biopsy core percent (PBCP), as well as other preoperative factors, with prostate cancer outcomes in a cohort of consecutive patients with clinically localized prostate cancer who underwent RRP. Data from 203 patients who underwent RRP from March 1993 to May 2004 for clinically organ confined prostate cancer was analysed. The correlation of preoperative serum prostate specific antigen (PSA) level, biopsy Gleason score, total number of positive biopsies and PBCP with the extent of disease at final pathology and biochemical progression were analyzed. The mean PBCP was 29.8+/-21.1 (median 25). Histopathological examination of the RRP specimens revealed ECE in 66 (32.5 %), SVI in 43 (21.2 %), LNI in 8 (4 %), and positive SM in 59 (29.1 %). Overall, only 9% of patients ( 18 of 203) had biochemical progression at a median postoperative follow-up of 22 months. Univariate analysis revealed serum PSA, biopsy Gleason Score, the number of positive cores and PBCP as predictive factors for extra-prostatic disease in RRP specimens. However, multivariate analysis revealed that biopsy Gleason score and serum PSA were the strongest independent predictive factors for extra-prostatic disease while percent positive biopsy cores carried significance in the prediction of ECE and SM positivity. The number of positive cores was not a predictor of non-organ confined disease. Preoperative serum PSA was the only prognostic factor for determination of biochemical failure.
Gleason score is the most important and independent predictive factor for extra-prostatic disease. The percentage of cores positive for cancer has significance only in the prediction of ECE and SM positivity. Further studies are needed before routine use of PBCP as one of the important preoperative prognostic factors.
Is the positive biopsy core percent really predictive of non-organ confined prostate carcinoma?
To investigate whether the aetiology for hearing impairment in neonates with unilateral auditory neuropathy spectrum disorder could be explained by structural abnormalities such as cochlear nerve aplasia, a cerebellopontine angle tumour or another identifiable lesion. In this prospective case series, 17 neonates were diagnosed with unilateral auditory neuropathy spectrum disorder on electrophysiological testing. Diagnostic audiology testing, including auditory brainstem response testing, was supplemented with computed tomography and/or magnetic resonance imaging. Ten of the neonates (59 per cent) showed evidence for cochlear nerve aplasia. Of the remaining seven, four were shown to have another abnormality of the temporal bone on imaging. Only three neonates (18 per cent) were not diagnosed with cochlear nerve aplasia or another lesion. Three computed tomography scans were reported as normal, but subsequent magnetic resonance imaging revealed cochlear nerve aplasia.
Auditory neuropathy spectrum disorder as a unilateral condition mandates further investigation for a definitive diagnosis. This series demonstrates that most neonates with unilateral auditory neuropathy spectrum disorder had pathology as visualised on computed tomography and/or magnetic resonance imaging scans. Magnetic resonance imaging is an appropriate first-line imaging modality.
Unilateral auditory neuropathy spectrum disorder: retrocochlear lesion in disguise?
To evaluate acute toxic effects of Euphorbia helioscopia in order to assure the safety and usefulness of herbal remedy. The Organization for Economic Cooperation and Development (OECD) for chemical testing guidelines No. 425 for acute oral toxicity testing were followed in this study. Mice were divided into three groups (n = 5). Group I served as control. Groups II and III were administered methanol extract of E. helioscopia leaves and latex orally at dose of 2000 mg/kg, respectively. Then, all the animals were observed for two weeks. Blood sampling was done by cardiac puncture after 14 days from each group for biochemical analysis. Histopathology was performed to find out any microscopic lesion in vital organs. LD50 was found greater than 2000 mg/kg. There was decrease in cholesterol, triglycerides, LDL and VLDL levels of latex and leaves with methanol extract-treated animals, with respect to control indicating plant's hypolipidemic effect. On macroscopic examination, no lesions were found on vital organs, such as liver, heart and kidney; and normal architecture was observed on microscopic examination.
On the basis of results, it was concluded that methanol extract of E. helioscopia leaves and latex were devoid of toxic effects in acute toxicity study.
Is folklore use of Euphorbia helioscopia devoid of toxic effects?
The purpose of our study was to determine whether adding oblique bilateral rib radiography to the skeletal survey for child abuse significantly increases detection of the number of rib fractures. We identified all patients under 2 years old who underwent a skeletal survey for suspected child abuse from January 2003 through July 2011 and who had at least one rib fracture. These patients were age-matched with control subjects without fractures. Two randomized radiographic series of the ribs were performed, one containing two views (anteroposterior and lateral) and another with four views (added right and left oblique). Three fellowship-trained radiologists (two in pediatrics and one in trauma) blinded to original reports independently evaluated the series using a Likert scale of 1 (no fracture) to 5 (definite fracture). We analyzed the following: sensitivity and specificity of the two-view series for detection of any rib fracture and for location (using the four-view series as the reference standard), interobserver variability, and confidence level. We identified 212 patients (106 with one or more fractures and 106 without). The sensitivity and specificity of the two-view series were 81% and 91%, respectively. Sensitivity and specificity for detection of posterior rib fractures were 74% and 92%, respectively. There was good agreement between observers for detection of rib fractures in both series (average kappa values of 0.70 and 0.78 for two-views and four-views, respectively). Confidence significantly increased for four-views.
Adding bilateral oblique rib radiographs to the skeletal survey results in increased rib fracture detection and increased confidence of readers.
Is the new ACR-SPR practice guideline for addition of oblique views of the ribs to the skeletal survey for child abuse justified?
This study investigates issues associated with the United States Orphan Drug Act. A comprehensive orphan drug database was compiled from FDA data and corporate annual reports of major pharmaceutical companies. Analysis allowed the generation of a descriptive orphan drug portrait as well as documentation of orphan drugs along their lifecycle. Currently, 2002 products have obtained orphan drug designation with 352 drugs obtaining FDA approval. Approximately 33% of orphan drugs are oncology products. On average, products obtain 1.7 orphan designations with approximately 70% obtaining a single designation. At least 9% of orphan drugs have reached blockbuster status with two-thirds having two or more designations. An additional 25 orphan drugs had sales exceeding US$ 100 million in 2008 alone. Since 1983, at least 14 previously discontinued products have been recycled as orphan drugs.
The United States Orphan Drug Act has created issues which, in some cases, have led to commercial and ethical abuses. Orphan Drug Act reform is necessary but current incentives, including 7 year market exclusivity, should be maintained in order to favour patients as well as economic prosperity. Suggested reforms include price regulation, subsidy paybacks for profitable drugs and the establishment of an International Orphan Drug Office.
The US Orphan Drug Act: rare disease research stimulator or commercial opportunity?
The aim of this study was to determine the positional stability and success rate of palatally placed length-reduced temporary anchorage devices (LRTADs) (length, 4 or 6 mm). Twenty-two patients (ages, 21-62 years; 14 women, 8 men) were enrolled in the study. Each received 1 LRTAD (Orthosystem, Straumann, Switzerland) placed in the midsagittal palate for multifunctional anchorage tasks. Standardized cephalograms were taken directly after implant placement and at the end of treatment to analyze any implant movements. The cephalometric tracings were superimposed on anterior nasal spine to posterior nasal spine in posterior nasal spine to analyze changes in implant angulation and position during treatment. The LRTADs were also evaluated clinically for mobility. Two of 22 implants showed mobility during the healing period (first 10-12 weeks after placement). Thus, the success rate was 91%. The remaining 20 palatally placed LRTADs had no mobility during healing (10-12 weeks) or the loading period (18 months 1 week) and were evaluated radiographically. The mean differences between the initial and final cephalometric evaluations (n = 20) were 0.5 degrees for changes in implant angulation and -0.6 mm for changes in implant position. These changes were most likely due to inaccuracies in cephalometric landmark identification rather than to LRTAD movements because no mobility was recorded.
One palatally placed LRTAD was sufficient for multifunctional stationary anchorage tasks in the maxilla under clinical loading conditions. The success rate was 91%. Implant loss occurred during the healing period.
Do palatal implants remain positionally stable under orthodontic load?
Significant pulmonary regurgitation, declining right-sided ejection fraction, increased right ventricular (RV) volumes as well as left ventricular (LV) dysfunction have all been identified as predictors of poor outcomes in patients with congenital heart disease (CHD). The prognostic value of the cardiac output (CO) in these patients however has never been studied. All consecutive ambulatory adult patients with CHD referred for magnetic resonance imaging (MRI) at the Montreal Children's Hospital between June 2007 and May 2009 were included. Right ventricular (RV) and left ventricular (LV) variables including end diastolic and end systolic volumes (EDV, ESV respectively), ejection fractions (EF) and regurgitant volumes were obtained. Cardiac index (CI) was calculated. Patients were followed for cardiac-related hospitalizations and cardiac interventions. Ninety-six patients were included. Median follow up was 3.9 ± 1.4 years. Nineteen percent of patients had a systemic CI<2.4 l/min/m(2). LVEDV, LVEF and RVEF were significantly diminished in the low CI group with a significant increase in RVESV and total regurgitant volume. Best predictors of low CI were LVEF (AUC=0.74), RVEF (AUC=0.71), total RV regurgitant volume (AUC=0.64) and RVESV (AUC=0.563). Low systemic CI was the best predictor of cardiac-related hospitalizations (hazard ratio 3.5, 95% confidence interval 1.5-8.5) and cardiac interventions (hazard ratio 2.2, 95% confidence interval 1.3-4.0), superior to LVEF, RVEF, total regurgitant volume and RVESV parameters.
In patients with congenital heart disease, cardiac index is the best predictor of cardiac hospitalizations and cardiac interventions.
Cardiac output as a predictor in congenital heart disease: Are we stating the obvious?
Shunts, the main treatment for hydrocephalus, are problematic as they frequently malfunction. Identifying shunt malfunction requires parents to recognize its symptoms and health professionals to integrate parents' information about the child's symptoms within the clinical assessment to reach a diagnosis.AIM: To investigate parent-professional shared decision making during the diagnosis of suspected shunt malfunction in acute hospital admissions. A mixed method study involving audio recordings of admission consultations, a shared decision making questionnaire and interviews 1-week post-consultation, was undertaken. Twenty-eight family members and fourteen health professionals participated. The interactions were analysed using conversational analysis, framework approach for the interview data and descriptive statistics for questionnaire responses. Both parents and professionals focussed on establishing a diagnosis and ruling out shunt malfunction when a child with hydrocephalus was ill. Participants' perceived effective collaboration as central to this task: parents wanted to contribute to the process of diagnosis by providing information about the likely cause of symptoms. Professionals were satisfied with the level of involvement by parents, although parent satisfaction was more variable. The challenge for professionals was to integrate parents' expertise of their child's presenting symptoms within clinical decision making processes.
In this context, both parents' and professionals' perceived their interactions to be about problem-solving, rather than making decisions about treatments. Although the shared decision-making model can help patients to make better decisions between treatment options, it is unclear how best to support collaboration between professionals and parents to ensure a good problem-solving process.
Are parents and professionals making shared decisions about a child's care on presentation of a suspected shunt malfunction: a mixed method study?
To examine patients' perspectives on ease of use and pain with the MediSense alternate site blood glucose testing device (Soft-Sense) compared with their current glucose testing method, and to evaluate the analytical performance of the MediSense device with the laboratory reference method. Study participants were shown how to use the Soft-Sense glucose device and asked to perform two tests on their forearm. A capillary sample was collected from their finger and tested on the external port of the Soft-Sense meter and a laboratory method (YSI Glucose Analyser). Finally, one drop of blood was also directly tested from the finger onto the external port. Patients completed a questionnaire comparing ease of use and associated pain of their current testing method with the Soft-Sense meter. Patients preferred the Soft-Sense device to their own for ease of use and for less pain (93% found it easier to use and 96% less painful; P<0.001). Glucose results correlated closely with the laboratory method (mean absolute percentage bias for the forearm 11.0%, finger 6.0%, and collected capillary sample 5.7%). Error grid analysis showed that all Soft-Sense results were clinically acceptable.
Patients prefer the Soft-Sense alternate site testing device to their existing measuring method. The device accurately measures whole blood glucose.
Alternate site blood glucose testing: do patients prefer it?
The potential influence of lunar phases on the occurrence of myocardial infarction is still controversial. The purpose of the present study was to investigate the association of the lunar cycle on the occurrence of fatal and non-fatal myocardial infarction based on a myocardial infarction registry. We studied 15,985 patients consecutively hospitalised with an acute myocardial infarction (AMI) between 1 January 1985 and 31 December 2007 with a known date of symptom onset who were recruited from a population-based myocardial infarction registry. The exact hour of AMI onset was known for 9813 events. Poisson regression analysis was performed to examine the relation between the lunar cycle and the occurrence of AMI. There was no association between new moon, full moon, waning moon and waxing moon and the occurrence of AMI. However, we observed that the three days after a new moon may be significantly protective for the occurrence of AMI, rate ratio (RR) 0.94 (95% CI 0.91-0.98), and the day before a new moon had a slightly negative effect (RR 1.06, 95% CI 1.00-1.12). Stratified analysis did not reveal any susceptible subgroups.
The moon phases did not show any apparent association with AMI occurrence. However, there might be a 'cardioprotective' time three days after a new moon.
The influence of lunar phases on the occurrence of myocardial infarction: fact or myth?
Studies from the amnesia literature suggest that errorless learning can produce superior results to errorful learning. However, it was found in a previous investigation by the present authors that errorless and errorful therapy produced equivalent results for patients with aphasic word-finding difficulties. A study in the academic literature of phoneme discrimination learning found that errorful learning produced equivalent results to errorless learning when feedback was given. In the authors' previous study, feedback was available to the participants in the errorful therapy. It is possible, therefore, that this feedback may have improved the results from errorless learning -- thereby reducing an underlying difference between the two techniques. Generally, feedback is thought to aid learning, however, there is little information in the speech therapy literature about this factor. The present investigation was conducted as a follow-up to authors' original study to compare errorless and errorful therapy for the amelioration of aphasic word-finding difficulties. The second aim was to replicate key findings from the original study: namely, that recognition memory, executive/problem-solving skills and monitoring ability predict immediate and long-term naming improvements but not the participants' remaining language ability.METHODS & Seven of the original 11 participants took part in a multiple baseline, crossover, case series design.OUTCOMES & The previous results were replicated: errorless and errorful therapy produced equivalent results immediately post-therapy and at follow-up. There was no effect of omitting feedback - the participants learnt equally well without therapist's feedback about whether or not their response was correct. In addition, executive/problem-solving skills and monitoring ability again predicted immediate naming improvements not language ability.
The findings support the view that cognitive abilities and in particular executive function are important contributors to rehabilitation.
Treatment of anomia using errorless versus errorful learning: are frontal executive skills and feedback important?
Early implantation of centrifugal devices in patients with postcardiotomy cardiogenic shock may provide a bridge to recovery and allow subsequent long-term survival. Since January 1989, 62 patients were supported with centrifugal pumps because of failure to wean from cardiopulmonary bypass. Indications were postcardiotomy cardiogenic shock (PCCS) (n = 60), bridge to cardiac retransplantation (n = 1), and right ventricular failure (n = 1). Patients' ages ranged from 23 to 78 years; 40 were men (65%), and 22 were women (35%). Twenty-two patients (35%) had a left ventricular assist device; 9 patients (15%) had a right ventricular assist device; and 31 patients (50%) had a biventricular assist device. Length of support ranged from 1 day to 19 days. Forty-two patients (68%) were weaned successfully; 27 patients survived to discharge (44%). Complications included bleeding (n = 41, 66%), renal failure (n = 28, 45%), and respiratory failure (n = 26, 42%). Currently, 23 patients survived 10 or more years (n = 1), 6 to 10 years (n = 7), 1 to 5 years (n = 10), and less than 1 year (n = 5).
Centrifugal pumps are available, easy to use, and relatively inexpensive. Our experience justifies their continued use as a bridge to recovery for patients with postcardiotomy cardiogenic shock, despite the availability and increasing use of more expensive devices.
Bridge to recovery for postcardiotomy failure: is there still a role for centrifugal pumps?
General surgery in Canada varies from single system subspecialty practice in large centres to multisystem broad-based practice in smaller communities. We have attempted to determine whether Canadian training programs in general surgery are appropriate for these varied practices. A questionnaire was circulated to members of the Canadian Association of General Surgeons to collect demographic data and information about community size and patterns of practice. We also sought the source of training for general surgical subspecialties and other surgical specialties if applicable. Surgeons in smaller communities performed significantly more subspecialty and other specialty surgical practice than do surgeons in larger communities. Much of the training for this practice comes not from the primary fellowship but from senior colleagues in the community. Surgeons in smaller communities feel less well prepared than their colleagues in larger communities and are less likely to take additional fellowship training.
These results have important implications for surgical educators and manpower planners.
Training of Canadian general surgeons: are they really prepared?
Increasing use of diagnostic radiography has led to concern about the malignant potential of ionizing radiation. We aimed to quantify the cumulative effective dose (CED) from diagnostic medical imaging in children with inflammatory bowel disease (IBD) and to identify which children are at greatest risk for high amounts of image-related radiation exposure. A retrospective chart review of pediatric IBD patients seen between January 1 and May 30, 2008 was conducted. The effective dose of radiation received from all of the radiology tests performed during the course of each patient's treatment was estimated using typical effective doses and our institution's computed tomography dose index. A CED ≥50 mSv was considered high. Complete records were available for 257 of 372 screened subjects. One hundred seventy-one had Crohn disease (CD) and 86 had ulcerative colitis (UC). The mean CED was 17.56 ± 15.91 mSv and was greater for children with CD than for those with UC (20.5 ± 17.5 vs 11.7 ± 9.9 mSv, P<0.0001). Fifteen children (5.8%) had a CED ≥50 mSv, including 14 of 171 (8.2%) with CD and 1 of 86 (1.2%) with UC (P = 0.02). In children with CD, factors associated with high CED per multivariate analysis were any IBD-related surgery (odds ratio 42, 95% confidence interval 8-223, P<0.0001) and platelet count (odds ratio 16, 95% confidence interval 1.5-175, P = 0.02).
Although all doses of ionizing radiation have some malignancy-inducing potential, a small but important percentage of children with IBD are exposed to particularly high doses of ionizing radiation from diagnostic tests and procedures. Physicians caring for such patients must seek to limit radiation exposure whenever possible to lessen the lifetime risk of malignancy.
Pediatric inflammatory bowel disease and imaging-related radiation: are we increasing the likelihood of malignancy?
To identify the extent to which meta-analyses currently include unpublished data and whether editors, meta-analysts, and methodologists believe unpublished material should be included. This article describes two related studies: a literature review and a cross-sectional survey. For the literature review, we identified all articles indexed by the key word meta-analysis from January 1989 to February 1991 and determined whether unpublished material had been searched for, obtained, and included in the meta-analyses. For the cross-sectional survey, we surveyed authors of these meta-analyses, authors of articles addressing methodological issues in meta-analysis published during the same period, and editors of journals in which both types of articles were published. We asked the respondents about their attitudes concerning inclusion of unpublished data and publication of articles from which data had previously been included in a scientific overview. Inclusion of unpublished data and opinions about whether unpublished material should be included in overviews and whether prior inclusion of data in an overview should bear on publication. Of 150 meta-analyses, 46 (30.7%) included unpublished data in their primary analysis. Of authors surveyed, 85% responded. Of the meta-analysts and methodologists, 77.7% felt that unpublished material should definitely or probably be included in scientific overviews; this was true of 46.9% of the editors. A total of 86.4% of the meta-analysts and methodologists and 53.2% of the editors felt that inclusion of data in a prior overview should have no bearing on full publication of original research.
While inclusion of unpublished data in scientific overviews remains controversial, most investigators directly involved in meta-analysis believe that unpublished data should not be systematically excluded. The most valid synthesis of available information will result when meta-analysts subject published and unpublished material to the same rigorous methodological evaluation and present results with and without unpublished sources of data.
Should unpublished data be included in meta-analyses?
To determine whether premenopausal daughters of women with postmenopausal osteoporosis have lower peak bone mass than the daughters of normal women the same age, and to analyze the related risk factors affecting bone mass variation. 126 pairs of mother with postmenopausal osteoporosis and her premenopausal daughter, and 136 pairs of normal postmenopausal mother and her premenopausal daughter selected for 410 core families including one healthy premenopausal daughter aged 20 - 40, all of Han ethnicity living in Shanghai recruited by advertisement and lectures. A questionnaire survey was conducted to investigate their dietary custom, Dual-energy X-ray absorptiometry at lumber spine 1 - 4 (L(1 - 4)) and proximal femur was conducted to measure the values of bone mineral density (BMD). The BMD values in L(1 - 4), femoral neck, and greater trochanter of the daughters of mothers with osteoporosis were 0.68 g/cm(2) +/- 0.07 g/cm(2), 0.59 g/cm(2) +/- 0.08 g/cm(2), and 0.47 g/cm(2) +/- 0.07 g/cm(2) respectively, all significantly lower than those of the daughters of normal mothers (0.86 g/cm(2) +/- 0.14 g/cm(2), 0.70 g/cm(2) +/- 0.11 g/cm(2), and 0.57 g/cm(2) +/- 0.10 g/cm(2) respectively, all P<0.001). The average body weight of the daughters of mothers with osteoporosis was lighter then that of the daughters of normal mothers by 4.8% (P<0.05). Multivariate regression analysis showed that age, body height, age of menarche, and milk intake were not influencing factors of BMD value, however, body weight was most significantly associated with BMD of the premenopausal daughters, contributing to the BMD variation at L(1 - 4), femoral neck, and greater trochanter by 9.4%, 16.5%, and 16.6% respectively. When body weight was excluded in the model, lower BMD of mother became the most important factors affecting the BMD variation, contributing to the BMD variation at L(1 - 4), femoral neck, and greater trochanter by 5.1%, 5.3%, and 4.2% respectively.
The daughters of mothers with osteoporosis have reduced peak bone mass. It is likely due to the lower body weight of the daughter and the lower bone mass of the mother.
Do the premenopausal daughters of women with postmenopausal osteoporosis have lower peak bone mass?
Cardiovascular diseases remain the leading cause of death in women and there is a need for more accurate risk assessment scores. The aims of our study were to compare the accuracy of several widely used cardiac risk assessment scores in predicting the likelihood of obstructive coronary artery disease (CAD) on CT coronary angiography (CTCA) in symptomatic women and to explore which female-specific risk factors were independent predictors of obstructive CAD on CTCA and whether adding these risk factors to pre-test probability scores would improve their predictive value. Data were obtained from a cohort of 228 consecutively included symptomatic women undergoing evaluation for CAD and referred for CTCA. Obstructive CAD was defined as ≥50% luminal stenosis on CTCA. Pre-test probability for CAD was calculated according to the Diamond and Forrester score, New score, Duke clinical score, and an updated Diamond and Forrester score. Female-specific factors were obtained by a written questionnaire. Pre-test probability scores were compared with ROC analysis and showed that only the New score and the updated Diamond and Forrester score were significant predictive scores for obstructive CAD on CTCA (area under the curve, AUC, 0.67, p < 0.01; AUC 0.61, p = 0.04, respectively). Multivariable logistic regression analysis identified that gestational diabetes mellitus (GDM) and oestrogen status were independent predictors of obstructive CAD when adjusted for the pre-test probability scores. The updated Diamond and Forrester score was used for net reclassification improvement (NRI) analysis, since the New score already accounts for oestrogen status. Adding GDM and oestrogen status to the updated Diamond and Forrester score resulted in a significant NRI (p = 0.04).
There is a large variability in prediction of obstructive CAD using different pre-test probability risk scores in symptomatic women. Logistic regression analysis revealed that oestrogen status and GDM were independently associated with the occurrence of obstructive stenosis on CTCA. The predictive ability of cardiac pre-test probability scores improved significantly with the addition of oestrogen status and GDM.
Comparison of different cardiac risk scores for coronary artery disease in symptomatic women: do female-specific risk factors matter?
We evaluate the international diffusion of the Ottawa Ankle and Knee Rules and determine emergency physicians' attitudes toward clinical decision rules in general. We conducted a cross-sectional, self-administered mail survey of random samples of 500 members each of the American College of Emergency Physicians, Canadian Association of Emergency Physicians, British Association for Accident and Emergency Medicine, Spanish Society for Emergency Medicine, and all members (n=1,350) of the French Speaking Society of Emergency Physicians, France. Main outcome measures were awareness of the Ottawa Ankle and Knee Rules, reported use of these rules, and attitudes toward clinical decision rules in general. A total of 1,769 (57%) emergency physicians responded, with country-specific response rates between 49% (United States and France) and 79% (Canada). More than 69% of physicians in all countries, except Spain, were aware of the Ottawa Ankle Rules. Use of the Ottawa Ankle Rules differed by country with more than 70% of all responding Canadian and United Kingdom physicians reporting frequent use of the rules compared with fewer than one third of US, French, and Spanish physicians. The Ottawa Knee Rule was less well known and less used by physicians in all countries. Most physicians in all countries viewed decision rules as intended to improve the quality of health care (>78%), a convenient source of advice (>67%), and good educational tools (>61%). Of all physicians, those from the United States held the least positive attitudes toward decision rules.
This constitutes the largest international survey of emergency physicians' attitudes toward and use of clinical decision rules. Striking differences were apparent among countries with regard to knowledge and use of decision rules. Despite similar awareness in the United States, Canada, and the United Kingdom, US physicians appeared much less likely to use the Ottawa Ankle Rules. Future research should investigate factors leading to differences in rates of diffusion among countries and address strategies to enhance dissemination and implementation of such rules in the emergency department.
Awareness and use of the Ottawa ankle and knee rules in 5 countries: can publication alone be enough to change practice?
A young people's clinic has been running at the Archway Sexual Health Clinic since 1997. This offers a weekly walk-in service to both young men and women under 20 years old. To review the audit data of over 7 years' experience from the dedicated young people's clinic at the Archway Sexual Health Clinic, the "Arch." These data were collected by retrospective notes review of a consecutive series of all male attendees over a 7 year period from 1997 to the end of 2003 with collection of data in Microsoft Excel database. These figures show an overall increase in attendee numbers, but also a relative rise in the numbers of young men using the service. A high number of bacterial sexually transmitted infections were detected. In 2001 and 2003 respectively, 14.5% and 17.8% of the young men using the service were diagnosed with Chlamydia trachomatis. A user survey aimed to identify factors that may be encouraging the young men to access the service.
Responding to the views of young men using the service has played a part in service development. Initiatives at the "Arch" such as the condom policy and choice of gender of staff may be factors encouraging increasing attendances. Continuing to work with other agencies to develop ways to engage young men are recommended. Word of mouth recommendation cannot be underestimated in publicising the service.
Are we getting the message across?
The purpose of this study was to localize the cell bodies of palisade endings that are associated with the myotendinous junctions of the extraocular muscles. Rhesus monkeys received tract-tracer injections (tetramethylrhodamine dextran [TMR-DA] or choleratoxin subunit B [CTB]) into the oculomotor and trochlear nuclei, which contain the motoneurons of extraocular muscles. All extraocular muscles were processed for the combined immunocytochemical detection of the tracer and SNAP-25 or synaptophysin for the visualization of the complete muscle innervation. In all muscles--except the lateral rectus--en plaque and en grappe motor endings, but also palisade endings, were anterogradely labeled. In addition a few tracer-labeled tendon organs were found. One group of tracer-negative nerve fibers was identified as thin tyrosine hydroxylase-positive sympathetic fibers, and a second less numerous group of tracer-negative fibers may originate from the trigeminal ganglia. No cellular or terminal tracer labeling was present within the mesencephalic trigeminal nucleus or the trigeminal ganglia.
These results confirm those of earlier studies and furthermore suggest that the somata of palisade endings are located close to the extraocular motor nuclei--in this case, probably within the C and S groups around the periphery of the oculomotor nucleus. The multiple en grappe endings have also been shown to arise from these cells groups, but it is not possible to distinguish different populations in these experiments.
Do palisade endings in extraocular muscles arise from neurons in the motor nuclei?
The goal of this study was to use a well-described system of quantifying levator ani defect (LAD) severity using magnetic resonance imaging (MRI) to examine the relationship between defect severity and the presence or absence of prolapse. This is a secondary analysis of two case-control studies comparing 284 cases (with prolapse) to 219 controls (normal support) defined by using Pelvic Organ Prolapse Quantification (POP-Q) exams. LAD were assessed on MRI, with scores from 0 (no defects) to 6 (complete, bilateral defects). The number of cases and controls at each score were compared. Logistic regression and receiver operating characteristic (ROC) analyses were used to quantify relationships between LAD and prolapse. The proportion of cases exceeds the overall prolapse rate in this study at LAD scores ≥3, with higher rates of prolapse at higher LAD scores (p < 0.0000001). Prolapse risk stratifies into low risk at LAD scores 0-2, moderate at 3-5, and high at 6. ROC analysis for classification of prolapse based on LAD scores has an area under the curve of 69.9% (p < 0.001), suggesting LAD alone can discriminate between normal support and prolapse for nearly 70% of patients. Logistic regression identified higher parity and higher LAD scores as independent predictors of prolapse.
There are three clusters of prolapse risk: low (0-2), moderate (3-5), and high (6). Although LAD have a dose-response-like effect for prolapse, other factors are clearly involved.
Levator ani defect scores and pelvic organ prolapse: is there a threshold effect?
Study objectives are to test for differences (1) in rates of negative traffic outcomes between DUI offenders who have lifetime drug use disorders (DUD) and those with no lifetime DUD; and (2) by drug class for those with a DUD. The study sample of 379 male and 74 female repeat DUI offenders was interviewed using the Composite International Diagnostic Interview. Compared to those with an alcohol use disorder (AUD) only, the relative risk of being involved in a vehicular crash was greater for those with a central nervous system (CNS) depressant use disorder. The risk of being convicted of a traffic offense was higher for those with any DUD and for those with a CNS stimulant use disorder. Differences by class of drug used, after adjusting for demographics, were a 47 percent (confidence interval: 6-103%) greater risk of being in a crash with a CNS depressant use disorder and 28 percent (confidence interval: 11-48%) greater risk of a traffic conviction with a stimulant use disorder.
Results underscore the increased risk of negative traffic outcomes among repeat DUI offenders diagnosed with DUD, particularly CNS depressant disorders, supporting the call to establish policies that include comprehensive evaluation and treatment for this population.
Repeat DUI offenders who have had a drug diagnosis: are they more prone to traffic crashes and violations?
Age at first and last birth and the number of children are known to influence the risk of endometrial and ovarian cancers. However, it remains unknown whether the difference in years between first and last childbirth plays a role. The Swedish Family-Cancer Database allowed us to carry out the largest study ever on reproductive factors in these cancers. We selected over 5.7 million women from the database. We estimated the effect of number of children, age at birth and difference between age at first and last birth by Poisson regression adjusted for age, period, region and socioeconomic status. The risk for endometrial cancer is negatively associated with increasing number of children and increasing age at first as well as age at last birth. Weaker associations are found for ovarian cancer. Age at last birth is the factor that shows highest influence. A large difference in first and last childbirth shows a protective effect on the risk of endometrial cancer.
Our findings suggest that the risk of endometrial cancer is significantly decreased for women having at least a difference of 10 years between their first and last birth. Ovarian cancer does not seem to be influenced by the time interval between first and last birth.
Does the time interval between first and last birth influence the risk of endometrial and ovarian cancer?
Intrinsic sphincter deficiency may cause disabling stress urinary incontinence. While some pelvic operations are implicated as a cause of this condition, simple hysterectomy for benign disease is not recognized as one of them. We evaluated the association of simple hysterectomy with intrinsic sphincter deficiency. We performed a case control study to assess the association of simple hysterectomy with intrinsic sphincter deficiency in a consecutive group of 387 incontinent women. From 1995 to 1997 we identified 67 patients with and 67 controls without a history of hysterectomy. Further comparison was done after forming a subgroup at low risk for intrinsic sphincter deficiency. All patients were evaluated by a fluoroscopic urodynamic technique and abdominal leak point pressure was determined. Intrinsic sphincter deficiency was present in 48% of the 67 patients and 24% of the 67 controls. In the lower risk subgroup we noted this condition in 29 patients (52%) and 53 controls (21%).
In this population of incontinent women intrinsic sphincter deficiency, as diagnosed by low abdominal leak point pressure, appears to be a complication of simple hysterectomy.
Is intrinsic sphincter deficiency a complication of simple hysterectomy?
Despite the huge cost of the program, the Centers for Medicare and Medicaid Services (CMS) has maintained a policy that cost-effectiveness is not considered in national coverage determinations (NCDs). To assess whether an implicit cost-effectiveness threshold exists and to determine if economic evidence has been considered in previous NCDs. A literature search was conducted to identify estimates of cost-effectiveness relevant to each NCD from 1999-2007 (n = 103). The economic evaluation that best represented each coverage decision was included in a review of the cost-effectiveness of medical interventions considered in NCDs. Of the 64 coverage decisions determined to have a corresponding cost-effectiveness estimate, 49 were associated with a positive coverage decision and 15 with a noncoverage decision. Of the positive decisions, 20 were associated with an economic evaluation that estimated the intervention to be dominant (costs less and was more effective than the alternative), 12 with an incremental cost-effectiveness ratio (ICER) of less than $50,000, 8 with an ICER greater than $50,000 but less than $100,000, and 9 with an ICER greater than $100,000. Fourteen of the sample of 64 decision memos cited or discussed cost-effectiveness information.
CMS is covering a number of interventions that do not appear to be cost-effective, suggesting that resources could be allocated more efficiently. Although the authors identified several instances where cost-effectiveness evidence was cited in NCDs, they found no clear evidence of an implicit threshold.
Does Medicare have an implicit cost-effectiveness threshold?
Diagnosis of lymphatic filariasis using serum has been established but the utility of hydrocele fluid for the purpose is not exactly known. Since, hydrocele is a chronic form of the disease manifestation in a variety of situations and often poses difficulty in diagnosing its origin, we have evaluated the usefulness usage of hydrocele fluid for diagnosis of filarial origin of hydrocele in this study. Paired samples of serum and hydrocele fluid from 51 individuals with hydrocele, living in an endemic area of Wuchereria bancrofti were assessed. Circulating filarial antigen, filarial specific antibody and cytokine assay were performed in both serum and hydrocele fluid of patients. Og4C3 assay detected circulating filarial antigen (CFA) in serum and corresponding hydrocele fluids. The level of IgG, IFN-γ and IL-10 were found to be high in CFA-negative, while IgM and IgE were high in CFApositive hydrocele fluid and serum samples associated with hydrocele. On the other hand neither CFA-positive nor CFA-negative hydrocele fluid and serum samples associated with hydrocele showed any difference in IgG4 level.INTERPRETATION &
This study showed that the filaria related antigens and antibodies found in serum can be detected with equal sensitivity in hydrocele fluid. Therefore, it can be used as an alternative to serum for immunodiagnosis of filariasis, and help monitoring the filarisis elimination programme.
Hydrocele fluid: can it be used for immunodiagnosis of lymphatic filariasis?
To retrospectively determine if pixel histogram analysis of unenhanced computed tomographic (CT) images can be used to distinguish angiomyolipomas (AMLs) with minimal fat from clear cell renal cell carcinomas (CCRCCs). The human studies committee approved this HIPAA-complaint study, with waiver of informed consent. Patients with pathologically proved AMLs lacking visible macroscopic fat at CT and patients with pathologically proved CCRCCs were included. Lesions were measured, and a histogram (number of pixels with each attenuation) was calculated electronically within a central region of interest. The percentage of pixels below the attenuation thresholds -20 HU and 10 HU was calculated in both cohorts. The unpaired Student t test was used to compare the average percentage of subthreshold pixels at each threshold. P<.05 indicated a significant difference. The number of lesions with more than the selected percentage of subthreshold pixels was calculated in both groups, and the chi(2) test was used to test the significance of differences between cohorts. The area under the receiver operating characteristic (ROC) curve was used to determine if any percentage of subthreshold pixels could be used to differentiate between the two cohorts. There were 22 patients with pathologically proved AMLs lacking visible macroscopic fat on CT images. Tuberous sclerosis affected three of these patients. Mean maximal transverse lesion diameter was 20 mm (range, 11-38 mm). There were 28 patients in the CCRCC comparison group. Mean maximal transverse lesion diameter was 26 mm (range, 15-36 mm). Neither the Student t test (P>.2 for all thresholds<0 HU) nor the chi(2) test (P>.15 for all thresholds<0 HU) revealed a significant difference between cohorts. A lesion with more low-attenuation pixels was significantly more likely to be characterized as CCRCC than as AML with ROC curve analysis.
Once AMLs with visible fat on CT images are excluded, pixel histogram analysis cannot be used to distinguish between AMLs and CCRCCs.
Pixel distribution analysis: can it be used to distinguish clear cell carcinomas from angiomyolipomas with minimal fat?
There is some evidence that maternal smoking increases susceptibility to personal smoking's detrimental effects. One might question whether early life disadvantage might influence susceptibility to occupational exposure. In this cross-sectional study we investigated respiratory symptoms, asthma and self-reported chronic obstructive pulmonary disease (COPD) as related to working as a cleaner in Northern European populations, and whether early life factors influenced susceptibility to occupational cleaning's unhealthy effects. The RHINE III questionnaire study assessed occupational cleaning in 13,499 participants. Associations with respiratory symptoms, asthma and self-reported COPD were analysed with multiple logistic regressions, adjusting for sex, age, smoking, educational level, parent´s educational level, BMI and participating centre. Interaction of occupational cleaning with early life disadvantage (maternal smoking, severe respiratory infection<5 years, born during winter months, maternal age at birth>35 years) was investigated. Among 2138 ever-cleaners the risks of wheeze (OR 1.4, 95% CI 1.3-1.6), adult-onset asthma (1.5 [1.2-1.8]) and self-reported COPD (1.7 [1.3-2.2]) were increased. The risk increased with years in occupational cleaning (adult-onset asthma: ≤1 year 0.9 [0.7-1.3]; 1-4 years 1.5 [1.1-2.0]; ≥4 years 1.6 [1.2-2.1]). The association of wheeze with cleaning activity ≥4 years was significantly stronger for those with early life disadvantage than in those without (1.8 [1.5-2.3]vs. 1.3 [0.96-1.8]; pinteraction 0.035).
Occupational cleaners had increased risk of asthma and self-reported COPD. Respiratory symptom risk was particularly increased in persons with factors suggestive of early life disadvantage. We hypothesize that early life disadvantage may increase airway vulnerability to harmful exposure from cleaning agents later in life.
Respiratory Health in Cleaners in Northern Europe: Is Susceptibility Established in Early Life?
Evolution of periarticular implant technology has led to stiffer, more stable fixation constructs. However, as plate options increase, comparisons between different sized constructs have not been performed. The purpose of this study is to biomechanically assess any significant differences between 3.5- and 4.5-mm locked tibial plateau plates in a simple bicondylar fracture model. A total of 24 synthetic composite bone models (12 Schatzker V and 12 Schatzker VI) specimens were tested. In each group, six specimens were fixed with a 3.5-mm locked proximal tibia plate and six specimens were fixed with a 4.5-mm locking plate. Testing measures included axial ramp loading to 500 N, cyclic loading to 10,000 cycles and axial load to failure. In the Schatzker V comparison model, there were no significant differences in inferior displacement or plastic deformation after 10, 100, 1,000 and 10,000 cycles. In regards to axial load, the 4.5-mm plate exhibited a significantly higher load to failure (P = 0.05). In the Schatzker VI comparison model, there were significant differences in inferior displacement or elastic deformation after 10, 100, 1,000, and 10,000 cycles. In regards to axial load, the 4.5-mm plate again exhibited a higher load to failure, but this was not statistically significant (P = 0.21).
In the advent of technological advancement, periarticular locking plate technology has offered an invaluable option in treating bicondylar tibial plateau fractures. Comparing the biomechanical properties of 3.5- and 4.5-mm locking plates yielded no significant differences in cyclic loading, even in regards to elastic and plastic deformation. Not surprisingly, the 4.5-mm plate was more robust in axial load to failure, but only in the Schatzker V model. In our testing construct, overall, without significant differences, the smaller, lower-profile 3.5-mm plate seems to be a biomechanically sound option in the reconstruction of bicondylar plateau fractures.
A biomechanical comparison between locked 3.5-mm plates and 4.5-mm plates for the treatment of simple bicondylar tibial plateau fractures: is bigger necessarily better?
The use of stented bioprostheses for aortic valve replacement (AVR) in elderly patients with a small aortic annulus may result in unsatisfactory hemodynamic performance of the prosthesis. To overcome this limitation, new bioprostheses have been designed for complete supra-annular implantation, but the actual hemodynamic advantage of the supra-annular implant over the intra-annular has not been fully investigated. Accordingly, the hemodynamic performance of the same stented bioprosthesis (except for sewing ring design) implanted in the supra-annular and conventional intra-annular seating was compared. Twenty-two patients received an intra-annular implant, and 38 a supra-annular implant. Age (74 +/- 5 versus 76 +/- 5 years, p = 0.54), gender (55% versus 50% males, p = 0.79) and body surface area (1.74 +/- 0.2 versus 1.81 +/- 0.2 m2, p = 0.13) were similar in both subgroups, who underwent echocardiography at 8 +/- 2 and 6 +/- 2 months after surgery, respectively (p = 0.09). The two patient subgroups had similar preoperative left ventricular outflow tract diameters (2.06 +/- 0.2 and 2.1 +/- 0.2 cm; p = 0.62), average size of implanted prosthesis (21.0 and 21.3 mm; p = 0.44) and mean transprosthetic flow rate (246 +/- 70 and 218 +/- 58 ml/s; p = 0.12). Mean (8 +/- 3 and 19 +/- 8 mmHg, p<0.0001), and peak (17 +/- 6 and 40 +/- 13 mmHg; p<0.0001) transprosthetic gradients were lower, and mean effective orifice area (EOA) (1.78 +/- 0.4 and 1.45 +/- 0.5 cm2, p = 0.006) was higher in patients with supra-annular implants than in those with intraannular. The incidence of patient-prosthesis mismatch (EOA index<0.85 cm2/m2) decreased from 50% to 34% (p<0.0001), with no case of severe mismatch using the supra-annular implant. During follow up, a left ventricular mass reduction occurred in patients with supra-annular implants (from 225 +/- 110 to 173 +/- 59 g/m2; p<0.03), but not in patients with intra-annular implants (173 +/- 62 and 186 +/- 64 g/m2; p = 0.87)
The study results showed that, compared to intra-annular implantation, supra-annular implantation of bioprosthetic stented valves in the aortic position was associated with a significantly better hemodynamic performance of the prosthesis and significant regression of left ventricular hypertrophy.
Stented bioprosthetic valve hemodynamics: is the supra-annular implant better than the intra-annular?
To determine whether Alzheimer's disease (AD) is associated with preferential atrophy of either the left or right hippocampus. We examined right-left asymmetry in hippocampal volume and atrophy rates in 32 subjects with probable AD and 50 age-matched controls. Hippocampi were measured on two serial volumetric MRI scans using a technique that minimizes laterality bias. We found a non-significant trend for right>left (R>L) asymmetry in controls at both time points (R>L: 1.7%; CI: -0.3-3.7%; p = 0.1). AD subjects showed a similar non-significant trend for R>L asymmetry at baseline (R>L: 1.8%; CI: -1.9-5.5%; p = 0.32), but not at repeat (p = 0.739). Change in R/L ratio between visits in AD patients was significant (p = 0.02). The AD group had significantly higher variance in these ratios than the controls at baseline (p = 0.02), but not repeat (p = 0.06). AD patients had higher atrophy rates than controls (p<0.001). Mean (CI) annualized atrophy rates for left and right hippocampi were 1.2% (0.5-1.8%) and 1.1% (0.5-1.8%) for the controls, and 4.6% (3.3-6.0%) and 6.3% (4.9-7.8%) for AD subjects. There was no significant asymmetry in atrophy rates in controls (p = 0.9), but borderline significantly higher atrophy rates in the right hippocampus of the AD group (p = 0.05) compared to the left. Presence of an APOEepsilon4 allele had no significant effect on the size, asymmetry or atrophy rates in AD (p>0.20).
We report minor R>L asymmetry in hippocampal volumes in controls and present some evidence to suggest that there is a change in the natural R>L asymmetry during the progression of AD.
Does Alzheimer's disease affect hippocampal asymmetry?
The importance of a safety culture to maximize safety is no longer questioned. However, achieving sustainable culture improvements are less evident. Evidence is growing for a multifaceted approach, where multiple safety interventions are combined. Lean management is such an integral approach to improve safety, quality and efficiency and therefore, could be expected to improve the safety culture. This paper presents the effects of lean management activities on the patient safety culture in a radiotherapy institute. Patient safety culture was evaluated over a three year period using triangulation of methodologies. Two surveys were distributed three times, workshops were performed twice, data from an incident reporting system (IRS) was monitored and results were explored using structured interviews with professionals. Averages, chi-square, logistical and multi-level regression were used for analysis. The workshops showed no changes in safety culture, whereas the surveys showed improvements on six out of twelve dimensions of safety climate. The intention to report incidents not reaching patient-level decreased in accordance with the decreasing number of reports in the IRS. However, the intention to take action in order to prevent future incidents improved (factorial survey presented β: 1.19 with p: 0.01).
Due to increased problem solving and improvements in equipment, the number of incidents decreased. Although the intention to report incidents not reaching patient-level decreased, employees experienced sustained safety awareness and an increased intention to structurally improve. The patient safety culture improved due to the lean activities combined with an organizational restructure, and actual patient safety outcomes might have improved as well.
Does lean management improve patient safety culture?
Although studies have shown that pollen addition and/or removal decreases floral longevity, less attention has been paid to the relationship between reproductive costs and floral longevity. In addition, the influence of reproductive costs on floral longevity responses to pollen addition and/or removal has not yet been evaluated. Here, the orchid Cohniella ascendens is used to answer the following questions. (a) Does experimental removal of flower buds in C. ascendens increase flower longevity? (b) Does pollen addition and/or removal decrease floral longevity, and does this response depend on plant reproductive resource status? To study the effect of reproductive costs on floral longevity 21 plants were selected from which we removed 50 % of the developing flower buds on a marked inflorescence. Another 21 plants were not manipulated (controls). One month later, one of four flowers on each marked inflorescence received one of the following pollen manipulation treatments: control, pollinia removal, pollination without pollinia removal or pollination with pollinia removal. The response variable measured was the number of days each flower remained open (i.e. longevity). The results showed significant flower bud removal and pollen manipulation effects on floral longevity; the interaction between these two factors was not significant. Flowers on inflorescences with previously removed flower buds remained open significantly longer than flowers on control inflorescences. On the other hand, pollinated flowers closed much faster than control and removed-pollinia flowers, the latter not closing significantly faster than control flowers, although this result was marginal.
The results emphasize the strong relationship between floral longevity and pollination in orchids, as well as the influence of reproductive costs on the former.
Is floral longevity influenced by reproductive costs and pollination success in Cohniella ascendens (Orchidaceae)?
Inhalation anaesthetics and anthracycline chemotherapeutic drugs may both prolong the QT interval of the electrocardiogram. We investigated whether isoflurane may induce or augment QTc prolongation in patients who had previously received cancer chemotherapy including anthracycline drugs. Forty women undergoing surgery for breast cancer were included in the study. They were divided into two groups: (A) women previously treated with anthracyclines (n=20); and (B) women not treated with antineoplastic drugs (n=20). All patients received a standardized balanced anaesthetic in which isoflurane 0.5 vol% was used. The QT and corrected QT intervals were measured before anaesthesia, after induction and tracheal intubation, after 1, 5, 15, 30, 60 and 90 min of anaesthesia, and during recovery. In both groups we observed a tendency to QTc prolongation, but statistically significant differences among baseline values and values observed during isoflurane-containing anaesthesia were seen only in group A. During anaesthesia, significant differences in QTc values between the two groups were observed.
In female patients pretreated with anthracyclines for breast cancer, the tendency to QTc prolongation during isoflurane-containing general anaesthesia was more strongly expressed than in patients without previous chemotherapy.
Is prolongation of the QTc interval during isoflurane anaesthesia more prominent in women pretreated with anthracyclines for breast cancer?
The 'single-item measure' was developed as a short self-report tool for assessing physical activity. The aim of this study was to test the criterion validity of the single-item measure against accelerometry. Participants (n=66, 65% female, age: 39±11 years) wore an accelerometer (ActiGraph GT3X) over a 7-day period and on day 8, completed the single-item measure. The number of days of ≥30 min of accelerometer-determined moderate to vigorous intensity physical activity (MVPA) were calculated using two approaches; first by including all minutes of MVPA and second by including only MVPA accumulated in bouts of ≥10 min (counts/min ≥1952). Associations between the single-item measure and accelerometer were examined using Spearman correlations and 95% limits of agreement. Percent agreement and κ statistic were used to assess agreement between the tools in classifying participants as sufficiently/insufficiently active. Correlations between the number of days of ≥30 min MVPA recorded by the single-item and accelerometer ranged from 0.46 to 0.57. Participants underreported their activity on the single-item measure (-1.59 days) when compared with all objectively measured MVPA, but stronger congruence was observed when compared with MVPA accumulated in bouts of ≥10 min (0.38 days). Overall agreement between the single-item and accelerometry in classifying participants as sufficiently/insufficiently active was 58% (k=0.23, 95% CI 0.05 to 0.41) when including all MVPA and 76% (k=0.39, 95% CI 0.14 to 0.64) when including activity undertaken in bouts of ≥10 min.
The single-item measure is a valid screening tool to determine whether respondents are sufficiently active to benefit their health.
Can a single question provide an accurate measure of physical activity?
Although biofeedback treatment ameliorates symptoms in patients with fecal incontinence, whether it improves anorectal function is unclear.AIM: To examine prospectively whether biofeedback therapy influences objective and subjective parameters of anorectal function and whether it improves outcome. Nineteen consecutive patients (females = 17) with fecal incontinence for over 1 year, underwent a three-phase outpatient treatment program. This program consisted of pelvic muscle strengthening exercises (phase 1), 1-h biofeedback therapy sessions twice a week (phase 2), and reinforcement sessions at 6 wk, 3, 6, and 12 months (phase 3). The number of sessions in phase 2 were customized for each patient. Anorectal manometry, saline continence test, prospective stool diaries, and bowel satisfaction scores were used to assess the changes in bowel function before and 1 year after therapy. After biofeedback therapy, the anal squeeze sphincter pressure (p<0.05), the duration of squeeze (p<0.001), and the capacity to retain liquids (p<0.05) increased. Rectoanal coordination also improved with a reduction (p<0.001) in rectal pressure and an increase (p<0.05) in the continence index. Threshold volumes for first perception and desire to defecate decreased (p<0.05). The number of therapy sessions varied, mean (range) = 7 (4-13). Stool frequency and the number of incontinence episodes decreased (p<0.001). Bowel satisfaction score improved (p<0.001).
Biofeedback therapy is effective and improves objective and subjective parameters of anorectal function in patients with fecal incontinence. Customizing the number of sessions and providing periodic reinforcement may improve the success rate.
Can biofeedback therapy improve anorectal function in fecal incontinence?
We investigated the effect of fracture surgery on serum procalcitonin levels and the value of procalcitonin in differentiating inflammatory reaction caused by fracture surgery from postoperative infective complications. Twenty-one patients (8 women, 13 men; mean age 72.5 years; range 50 to 105 years) who underwent surgery for pertrochanteric hip fractures were evaluated according to the procedures employed, namely osteosynthesis, and hemiarthroplasty. Procalcitonin (PCT), C-reactive protein (CRP), white blood cell count (WBC), and body temperature were measured before surgery and for five days postoperatively. No postoperative wound infections occurred. Seven patients developed complications. The mean preoperative CRP level was five times above the normal. It made a peak on the second day and then began to decrease, but still was four times higher than the preoperative level on the fifth day. Preoperatively, the mean PCT level was lower than the normal in all the patients. It made a peak on the first postoperative day without exceeding the normal range and returned to the preoperative level on the fifth day. In contrast to CRP levels which were above the normal in all the patients, PCT levels were higher than the normal only in patients who developed complications. Taking the cut-off value as = or>0.5 ng/ml, the sensitivity and specificity of PCT to determine systemic complications were 100% and 100% on the first day, and 100% and 50% on the second day, respectively.
Procalcitonin may prove to be a useful parameter to identify early postoperative systemic complications after fracture surgery.
Can procalcitonin be used for the diagnosis and follow-up of postoperative complications after fracture surgery?
Short term morbidity, functional outcome, recurrence and quality of life outcomes after robotic assisted ventral mesh rectopexy (RVMR) and laparoscopic ventral mesh rectopexy (LVMR) were compared. This study includes 51 consecutive patients having operations for external rectal prolapse (ERP) in a tertiary centre between October 2009 and December 2012. Of these, 17 patients had RVMR and 34 underwent LVMR. The groups were matched for age, gender, body mass index (BMI), and American Society of Anesthesiologists (ASA) grades. The same operative technique and mesh was used and follow up was 12 months. Data was collected on patient demographics, surgery duration, blood loss, duration of hospital stay and operative complications. Functional outcomes were measured using the faecal incontinence severity index (FISI) and Wexner faecal incontinence scoring. Quality of life was scored using SF36 questionnaires pre and postoperatively. All patients were female except three (median 59, range 25-89). There was one laparoscopic converted to open procedure. RVMR procedures were longer in duration (p = 0.013) but with no difference in blood loss between the groups. The average duration of stay was 2 days in both groups. There were six minor postoperative complications in LVMR procedures and none in the RVMR group. Pre and postoperative Wexner and FISI scoring were significantly lower in the RVMR group (p = 0.042 and p = 0.024, respectively). SF-36 questionnaires showed better scoring in physical and emotional component in RVMR group (p = 0.015). There was no recurrence in either group during follow-up.
Both LVMR and RVMR are similar in terms of safety and efficacy. Although not randomized, this data may suggest a better functional outcome and quality of life in patients having RVMR for ERP.
Short-term outcome of laparoscopic versus robotic ventral mesh rectopexy for full-thickness rectal prolapse. Is robotic superior?
The putaminal abnormalities detected on 1.5 T magnetic resonance imaging (MRI), such as putaminal atrophy, slit-like hyperintense rim, and hypointensity in the putamen on T2-weighted (T2W) imaging are important signs on differentiating multiple system atrophy with parkinsonism (MSA-P) from Parkinson's disease (PD). However, the putaminal abnormalities may have different manifestations on 3.0 T from those on 1.5 T. To investigate the diagnostic value of putaminal abnormalities on 3.0 T MRI for differentiating MSA-P from PD. The study included a MSA-P group (9 men, 9 women), a PD group (12 men, 14 women), and a control group (11 men, 13 women). All subjects were examined with 3.0 T MRI using the conventional protocol. Putaminal atrophy, T2-hypointensity in the dorsolateral putamenat, and a slit-like hyperintense rim on the lateral putamen were evaluated in each subject. There were no significant differences in the slit-like hyperintense rim (P = 0.782) or T2-hypointensity in the dorsolateral putamen (P = 0.338) among the three groups. Bilateral putaminal atrophy was found in 44.4% (8 of 18) of the MSA-P patients, in only 7.7% (2 of 26) of the PD patients, and in none of the controls. The proportion of subjects with putaminal atrophy was significantly higher in the MAS-P group (P = 0.008) and control group (P < 0.001). The specificity and sensitivity of putaminal atrophy for distinguishing MSA-P from PD was 92.3% and 44.4%, respectively.
The signal changes in the putamen on T2W imaging on 3.0 T MRI, including slit-like hyperintense rim and putaminal hypointensity, are not specific signs for MSA-P. Putaminal atrophy is highly specific for differentiating MSA-P from PD and healthy controls, but its insufficient sensitivity limits its diagnostic value.
The putaminal abnormalities on 3.0T magnetic resonance imaging: can they separate parkinsonism-predominant multiple system atrophy from Parkinson's disease?
Computed tomography (CT) gantry rotation time is one factor influencing image quality. Until now, there has been no report investigating the influence of gantry rotation time on chest CT image quality. To investigate the influence of faster gantry rotation time on image quality and subjective and objective image parameters in chest CT imaging. Chest CT scans from 160 patients were examined in this study. All scans were performed using a single-source mode (collimation, 128 × 0.6 mm; pitch, 1.2) on a dual-source CT scanner. Only gantry rotation time was modified, while other CT parameters were kept stable for each scan (120 kV/110 reference mAs). Patients were divided into four groups based on rotation time: group 1, 1 s/ rotation (rot); group 2, 0.5 s/rot; group 3, 0.33 s/rot; group 4, 0.28 s/rot. Two blinded radiologists subjectively compared CT image quality, noise, and artifacts, as well as radiation exposure, from all groups. For objective comparison, all image datasets were analyzed by a radiologist with 5 years of experience concerning objective measurements as well as signal-to-noise ratio (SNR). We found that faster gantry rotation times (0.28 s/rot and 0.33 s/rot) resulted in more streak artifacts, image noise, and decreased image quality. However, there was no significant difference in radiation exposure between faster and slower rotation times (P>0.7).
Faster CT gantry rotation reduces scan time and motion artifacts. However, accelerating rotation time increases image noise and streak artifacts. Therefore, a slower CT gantry rotation speed is still recommended for higher image quality in some cases.
CT chest and gantry rotation time: does the rotation time influence image quality?
Plastic surgeons are often asked to assist with the reconstruction of lower extremity wounds. These patients many times require free tissue transfer for coverage given paucity of soft tissue. Anecdotally, many orthopedic surgeons prefer muscle coverage--particularly in the setting of potentially infected bone. Today's surgeons now easily harvest and transfer fasciocutaneous flaps--a versatile option with less donor-site morbidity. We hypothesized that there would be no difference in outcomes between these 2 types of reconstruction. We performed a single-institution retrospective review of lower extremity free flap reconstructions in the last 10 years. Demographics, preoperative and postoperative course, and the documented time to weight-bearing and bony union were collected. Major cohorts compared were muscle free flaps and fasciocutaneous free flaps, further divided into subgroups including acute trauma, tumor resection, osteomyelitis, and nonunion. Data comparisons were made using paired t test and Fischer exact tests. There were 121 patients who met inclusion criteria--86 in the muscle flap group, and 35 in the fasciocutaneous group and demographics were equal. Total complication rates were higher in smokers than nonsmokers (P<0.03). There was no significant difference in major or minor complication rates between muscle and fasciocutaneous flaps in any subgroup. In both the acute fracture group and the infected nonunion group, there was a significantly faster return to weight bearing in the fasciocutaneous group (P<0.03) although there was no difference in documented time to bony union. Patients who underwent fasciocutaneous reconstruction were more likely to require revisionary surgery for improved aesthetics (P<0.001).
Our data suggest that in essentially all clinical parameters, there is no difference between free flap type used for soft tissue coverage of the lower extremity. Patients undergoing reconstruction with a fasciocutaneous flap may return to weight bearing earlier--although they are more likely to require elective flap revisions. These results imply essentially equivalent outcomes regardless of flap type or operative indication, in contrast with some of the biases in the orthopedic community. The particular flap chosen for any reconstruction should remain solely at the discretion of the plastic surgeon.
Comparing Muscle and Fasciocutaneous Free Flaps in Lower Extremity Reconstruction--Does It Matter?
To evaluate whether the "black geode" sign is a characteristic magnetic resonance imaging (MRI) finding for extracranial schwannomas. Forty-three patients with pathologically confirmed extracranial schwannomas underwent preoperative gadolinium-enhanced MRI. The black geode sign was defined as the appearance of enhanced outer and inner rings. MR images were retrospectively reviewed for size, configuration, and signal intensity of the lesions in addition to the presence of the black geode sign. Gadolinium-enhanced T1-weighted images revealed the black geode sign in seven of 43 patients (16%). The thickness of inner rings (mean 0.6 cm, range 0.3-0.8 cm) was significantly greater than that of outer rings (mean 0.2 cm, range 0.1-0.3 cm) (P<0.01). While outer rings were circular or elliptical in shape with smooth contours, inner rings had a lobular configuration with irregular thickness and contours. The degrees of enhancement were significantly stronger with inner rings than with outer rings (P<0.01). In histopathological correlation of five patients who underwent total excision, inner and outer rings corresponded to peridegenerative areas and fibrous capsules, respectively.
The black geode sign may be fairly specific to extracranial schwannomas on gadolinium-enhanced MR images.
Is "black geode" sign a characteristic MRI finding for extracranial schwannomas?
Biobanks may play a pivotal role in lung cancer patients' management, research, and health policy. The Nancy "Centre of Biologic Resources" analyzed the evolving profiles of operated lung cancer patients and their management over 20 years. A total of 1259 consecutive patients operated upon from 1988 till 2007 were included. Survival rates were statistically compared before and after 1997. The parameters associated with a significant improvement of survival were determined. After 1997, lung cancer was diagnosed at an earlier stage. For Squamous Cell Lung Cancer (SQCLC), stages IA increased from 5.4 to 19.5% and for Adenocarcinoma (ADC), stage IA increased from 9.9 to 24.7%. More women with stage I ADC were operated upon after 1997 (p = 0.01). More patients with Large Cell Lung Cancer were diagnosed recently. Recent patients received more adjuvant or neo-adjuvant chemotherapy (p<0.001) and less radiotherapy (stage p = 0.019, stage p<0.001). A longer overall patients' survival was observed after 1997 (chi test for SQLC and ADC independently p<or = 0.002). Among SQCLC long survivors, those at stage I-II, below 50 years, were more numerous. A longer survival was associated with early stage in ADC patients. Stage was the single constant factor for overall outcome.
Overall and stage-adjusted survival of operated lung cancer patients has been improved in the last decade due mainly to earlier diagnosis. The generalized use of computed tomography scan, chemotherapy, and a collegial management improved patients' survival.
Do evolving practices improve survival in operated lung cancer patients?
To compare the accuracy of endoscopic ultrasonography (EUS) imaging with histopathology in the diagnosis of upper gastrointestinal subepithelial lesions. Thirty-seven patients (21 female; mean age: 55 y) underwent endoscopic submucosal resection (ESMR) of upper gastro intestinal subepithelial lesions at a tertiary care facility. All patients underwent EUS before ESMR of the lesion. Information regarding location, size, echogenecity, layer of origin, presumptive diagnosis based on EUS imaging, and histopathology diagnosis after ESMR of the subepithelial lesion was recorded. Twenty-seven subepithelial lesions were resected from the stomach, 5 from the esophagus, and 5 from the duodenum. The mean size of the lesions was 9 mm (range, 6-18 mm). Thirty-six lesions originated from the submucosa, and 1 from the muscularis propria. Using histopathology as the gold standard, the overall diagnostic accuracy of EUS imaging was 49% (18 out of 37). The accuracy of EUS imaging for the diagnosis of esophageal, gastric, and duodenal subepithelial lesions was 20%, 56%, and 40%, respectively. One patient developed a microperforation, and 1 developed bleeding during the ESMR procedure. No complications were reported with the EUS procedure.
The diagnostic accuracy of EUS imaging is inferior to histopathology in the diagnosis of small upper gastrointestinal subepithelial lesions. Endoluminal resection is a relatively safe and noninvasive modality that not only provides tissue sample for accurate diagnostic interpretation, but also aids in the complete removal of small subepithelial lesions of the upper gastrointestinal tract.
Are endoscopic ultrasonography imaging characteristics reliable for the diagnosis of small upper gastrointestinal subepithelial lesions?
To determine if the Nuchal index (NIx) is increased in euploid fetuses with structural congenital heart defects (CHD). Euploid fetuses with CHD between 18 and 24 weeks gestation were identified. The next fetus meeting the same criteria with a normal fetal echocardiogram were selected as a control. The NIx [(mean nuchal thickness /mean biparietal diameter) x 100] and cardiac axis (CA; degrees) were calculated for each fetus. Standard descriptive tests and two-tailed t test were used. The NIx in the abnormal (n = 20) and control (n = 20) groups were 9.10 (2.35) and 7.54 (p = 0.04) and CA was 55.8 degrees and 48.6 degrees (p = 0.02), respectively.
The NIx and CA were significantly different in fetuses with CHD. A prospective study to confirm these findings and determine clinical utility is warranted.
Is the nuchal index increased in fetuses with congenital structural heart defects?
The aim of the present study was to evaluate the risks and benefits of concurrent prophylactic cholecystectomy (CPC) during laparoscopic Roux-en-Y gastric bypass (LRYGB). From December 2000 to November 2006, CPC during LRYGB was only performed in the presence of gallbladder pathology (n = 140). Beginning in December 2006, CPC was performed during all LRYGB procedures (n = 134). Exclusion criteria were open bypass procedure, previous bariatric surgery other than gastric banding, and previous cholecystectomy (CCE) or necessary concurrent CCE due to gallbladder pathology. During a median follow-up of 3.1 years, 26 (18.6%; 95% CI, 12.9-25.9%) of 140 patients without CPC subsequently required a CCE, leading to a gallbladder disease-free survival rate at 5 years of 77.4% (95% CI, 67.3-87.6%). Multivariate analysis identified a distal LRYGB and excess weight loss of>75% at 2 years to be significant risk factors for the development of biliary complications while a preoperative BMI > 50 m(2)/kg was protective. In the second series, prophylactic CCE was not associated with prolonged hospitalization or operative time. The postoperative complications were not related to the CPC.
The present data indicate that a substantial number of patients develop gallbladder complications after LRYGB. Furthermore, CPC can safely be performed during LRYGB. Based on these findings, CPC should be considered a reasonable approach in severely obese patients undergoing LRYGB.
Is routine cholecystectomy justified in severely obese patients undergoing a laparoscopic Roux-en-Y gastric bypass procedure?
The objective of this study was to describe the characteristics and outcome of pediatric patients presenting to an emergency department (ED) following out-of-hospital primary cardiac arrest (OHPCA), to determine if long-term survival is influenced by specific resuscitation interventions. This was a prospective observational study of cases of OHPCA during sport or exertion in young patients presenting to an ED over a 5-year period. Cases were identified from a resuscitation database, which documented patient demographics, nature of event, emergency treatment, response times, and clinical progress. These data were analyzed to determine outcomes. Nine children were identified who presented following OHPCA during the study period. The mean age was 10.7 (±4.2) years. All were subsequently diagnosed with an underlying primary cardiac disorder. Six patients (66.6%) survived to make a full recovery. All patients who survived had received early chest compressions (within 5 minutes) and early defibrillation (within 10 minutes). The initial cardiac arrest rhythm in all survivors had been an electrically cardiovertable rhythm. Five (83%) of the 6 survivors did not receive epinephrine during resuscitation.
The importance of early chest compressions and defibrillation in collapsed young athletes is highlighted in this report. These interventions can result in full long-term neurological recovery. Use of epinephrine in these patients may be dangerous. We suggest that special consideration should be given to this subgroup of patients in the development of future resuscitation guidelines.
Primary cardiac arrest following sport or exertion in children presenting to an emergency department: chest compressions and early defibrillation can save lives, but is intravenous epinephrine always appropriate?
At a time of increased need and demand for general internists in Canada, the attractiveness of generalist careers (including general internal medicine, GIM) has been falling as evidenced by the low number of residents choosing this specialty. One hypothesis for the lack of interest in a generalist career is lack of comfort with the skills needed to practice after training, and the mismatch between the tertiary care, inpatient training environment and "real life". This project was designed to determine perceived effectiveness of training for 10 years of graduates of Canadian GIM programs to assist in the development of curriculum and objectives for general internists that will meet the needs of graduates and ultimately society. Mailed survey designed to explore perceived importance of training for and preparation for various aspects of Canadian GIM practice. After extensive piloting of the survey, including a pilot survey of two universities to improve the questionnaire, all graduates of the 16 universities over the previous ten years were surveyed. Gaps (difference between importance and preparation) were demonstrated in many of the CanMEDS 2000/2005 competencies. Medical problems of pregnancy, perioperative care, pain management, chronic care, ambulatory care and community GIM rotations were the medical expert areas with the largest gaps. Exposure to procedural skills was perceived to be lacking. Some procedural skills valued as important for current GIM trainees and performed frequently (example ambulatory ECG interpretation) had low preparation ratings by trainees. Other areas of perceived discrepancy between training and practice included: manager role (set up of an office), health advocate (counseling for prevention, for example smoking cessation), and professional (end of life issues, ethics).
Graduates of Canadian GIM training programs over the last ten years have identified perceived gaps between training and important areas for practice. They have identified competencies that should be emphasized in Canadian GIM programs. Ongoing review of graduate's perceptions of training programs as it applies to their current practice is important to ensure ongoing appropriateness of training programs. This information will be used to strengthen GIM training programs in Canada.
Are Canadian General Internal Medicine training program graduates well prepared for their future careers?
The purpose of this study was to examine how the cultural factors, stigma, being strong, and religiosity influence symptom distress in African American cancer survivors. This descriptive correlational study was designed using the Sociocultural Stress and Coping Framework. Seventy-seven African American cancer survivors, recruited from oncology clinics and the community in North Carolina, completed a questionnaire that consisted of measures of demographic and illness characteristics, the Perceived Stigma Scale, the Ways of Helping Questionnaire, the Religious Involvement Scale, and the Symptom Distress Scale. The two cultural factors that were significantly associated with symptom distress were stigma (β = .23, p<.05) and organized religion (β = -.50, p<.05). No significant associations were found between being strong or nonorganized religiosity and symptom distress. The most commonly reported symptoms were fatigue (M = 2.44, SD = 1.20), pain (M = 2.26, SD = 1.43), and insomnia (M = 1.95, SD = 1.25).
The findings of this study indicate that the cultural factors, stigma, and organized religiosity were significantly associated with symptom distress.
African American cancer survivors: do cultural factors influence symptom distress?
The Food Neophobia Scale (FNS) is widely used in different countries, however appropriate psychometric analyses are required to allow cross-cultural comparisons. To our knowledge, most studies have been conducted among children and adult populations, with no reference to pregnant women. The objective of this study was to translate and test the psychometric properties of a Portuguese version of the FNS, and to identify clusters of food neophobia during pregnancy. The FNS was translated into Portuguese by three health researchers, and back-translated into English by an independent native English speaker and professional translator. The scale was self-administered in a sample of 219 women from the baseline evaluation of the Taste intervention study (HabEat project: http://www.habeat.eu/), who attended medical visits in two hospitals from Porto, Portugal, reporting food neophobia during the last trimester of pregnancy. The FNS consists of 10 items with a 7-point rating scale. An exploratory analysis was performed to evaluate the scale's dimensionality, followed by a confirmatory factor analysis to test the fit of the previous model by using different indexes. Cronbach's alpha coefficient was calculated to evaluate the internal reliability of the scale. The construct validity was assessed by comparing the FNS scores by categories of education, age and fruit and vegetables intake by ANOVA. A Model-based clustering was used to identify patterns of food neophobia; the number of latent classes was defined according to the Bayesian information criterion. A two-factor model solution was obtained (after excluding item 8 with a factor loading<0.4), explaining 51% of the total variance. Cronbach's alpha was 0.75 for factor 1 (5 items) and 0.71 for factor 2 (4 items). Items 1, 4, 6, 9 and 10 loaded into the first factor (i.e. more willingness to try new foods; less neophobic traits) and items 2, 3, 5 and 7 loaded into a second factor (i.e. more neophobic traits). A good global of fitness of the model was confirmed by fit indexes: TLI=0.876, CFI=0.911, RMSEA=0.088 and SRMR=0.051. The higher the education, age, and fruit and vegetables intake the lower the neophobic tendency, measured by the Portuguese FNS. Three patterns (i.e. clusters) of food neophobia, characterizing neophobia traits of pregnant women were identified: Moderate Neophilic, Moderate Neophobic, and Extreme Neophilic (cut-off points were provided).
The Portuguese version of the FNS has the basic requirements of a valid and reliable measure of food neophobia and permits the identification of clusters of neophobic traits during pregnancy.
Could the Food Neophobia Scale be adapted to pregnant women?
Differences in the management of women and men with acute coronary symptoms are well documented, but relatively little is known about practices for patients with ischemic stroke. We sought to determine whether there are sex-associated differences in the utilization of diagnostic tests for ischemic stroke patients treated at academic hospitals in the United States. Medical records were abstracted for consecutive ischemic stroke patients admitted to 32 US academic medical centers from January through June, 2004, as part of the University HealthSystem Consortium Ischemic Stroke Benchmarking Project. We compared the utilization rates of diagnostic tests including neuroimaging (CT or MRI), electrocardiogram (ECG), ultrasound of the carotid arteries, and echocardiography (transthoracic or transesophageal) for women and men. Multivariate logistic regression was used to test for sex differences with adjustment for potential confounders. The study included 1,256 ischemic stroke patients (611 women; 645 men; mean age 66.6 +/- 14.6 years; 56% white). There were no differences between women and men in the use of neuroimaging (odds ratio, OR = 1.37; 95% confidence interval, CI = 0.58-3.24), ECG (OR = 1.00, 95% CI = 0.70-1.44), carotid artery ultrasound (OR = 0.93, 95% CI = 0.72-1.21) or echocardiography (OR = 0.70, 95% CI = 0.70-1.22). The results were similar after covariate adjustment.
Women and men admitted to US academic hospitals receive comparable diagnostic evaluations, even after adjusting for sociodemographic and clinical factors.
Diagnostic evaluation for patients with ischemic stroke: are there sex differences?
This paper proposes the use of decision trees as the basis for automatically extracting information from published randomized controlled trial (RCT) reports. An exploratory analysis of RCT abstracts is undertaken to investigate the feasibility of using decision trees as a semantic structure. Quality-of-paper measures are also examined. A subset of 455 abstracts (randomly selected from a set of 7620 retrieved from Medline from 1998 - 2006) are examined for the quality of RCT reporting, the identifiability of RCTs from abstracts, and the completeness and complexity of RCT abstracts with respect to key decision tree elements. Abstracts were manually assigned to 6 sub-groups distinguishing whether they were primary RCTs versus other design types. For primary RCT studies, we analyzed and annotated the reporting of intervention comparison, population assignment and outcome values. To measure completeness, the frequencies by which complete intervention, population and outcome information are reported in abstracts were measured. A qualitative examination of the reporting language was conducted. Decision tree elements are manually identifiable in the majority of primary RCT abstracts. 73.8% of a random subset was primary studies with a single population assigned to two or more interventions. 68% of these primary RCT abstracts were structured. 63% contained pharmaceutical interventions. 84% reported the total number of study subjects. In a subset of 21 abstracts examined, 71% reported numerical outcome values.
The manual identifiability of decision tree elements in the abstract suggests that decision trees could be a suitable construct to guide machine summarisation of RCTs. The presence of decision tree elements could also act as an indicator for RCT report quality in terms of completeness and uniformity.
Are decision trees a feasible knowledge representation to guide extraction of critical information from randomized controlled trial reports?
Hyperuricemia is commonly associated with obesity, glucose intolerance, hypertension, dyslipidemia, and atherosclerotic cardiovascular disease. The resemblance of the metabolic syndrome and hyperuricemia has led to the suggestion that hyperuricemia is a part of the metabolic syndrome. The purpose of this study is to examine the contribution of uric acid (UA) as an additional component of the metabolic syndrome in middle-aged men. In total, 393 male participants, aged 45-60 years, were recruited from a professional health evaluation program. Anthropometric measurements and blood pressure (BP) were taken after an overnight fast. Fasting blood samples were collected for the measurements of glucose, UA, and lipid profile. Logistic regression models were fitted to examine the relationship between UA and the diagnosis of metabolic syndrome. Factor analysis was performed to explore the relationship between UA and the components of the metabolic syndrome. The diagnosis of the metabolic syndrome was significantly associated with waist circumference (WC), glucose, triglycerides (TG), high-density lipoprotein cholesterol (HDL-C), systolic BP, and liver enzyme levels, but not associated with UA levels. The sensitivity of hyperuricemia (serum UA>or = 7.0 mg/dL) for the diagnosis of the metabolic syndrome was 58.0% and the specificity was 55.3%. In factor analysis, UA aggregated with body mass index, WC, glucose, log TG, and HDL-C as a metabolic factor. Systolic and diastolic BP were loaded on a second factor separately. The model loaded with UA explained a similar proportion of the total variance (56.9%), as did the model loaded without UA (62.5%).
Our results suggest that the contribution of UA as an additional component of the syndrome seems to be insignificant. We propose that hyperuricemia might not be an important facet for the understanding of the underlying structure of the metabolic syndrome.
Is hyperuricemia another facet of the metabolic syndrome?
To determine whether postdating delayed antibiotic prescriptions results in a further decrease (over usual delayed prescriptions) in antibiotic use. Randomized controlled trial. A small rural town in Newfoundland and Labrador. A total of 149 consecutive adult primary care patients who presented with acute upper respiratory tract infections. Delayed prescriptions for patients who might require antibiotics were randomly dated either the day of the office visit (ie, the usual group) or 2 days later (ie, the postdated group). Whether or not the prescriptions were filled and the time it took for the patients to fill the prescriptions were noted by the 4 local pharmacies and relayed to the investigators. In total, 149 delayed antibiotic prescriptions were written, 1 per patient. Of the 74 usual delayed prescriptions given out, 32 (43.2%) were filled; of the 75 postdated delayed prescriptions given out, 33 (44.0%) were filled. Sixteen patients from each group filled their delayed prescriptions earlier than the recommended 48 hours. Statistical analyses-χ² tests to compare the rates of antibiotic use between the 2 groups and t tests to compare the mean time to fill the prescription between the 2 groups-indicated that these results were not significant (P>.05).
Although delayed prescriptions reduce the rate of antibiotic use, postdating the delayed prescription does not seem to lead to further reduction in use.
Postdated versus usual delayed antibiotic prescriptions in primary care: Reduction in antibiotic use for acute respiratory infections?
To compare the effects of isoenergetic amounts of milk, cheese and butter (adjusted to the same content of lactose and casein) on fasting and postprandial blood lipids and lipoproteins, and on postprandial glucose and insulin response. The experiments were designed to provide 20% of total energy from dairy fat, as either whole milk, mean (+/-SD) 2164 (+/-97) g, butter 93 (+/-4) g, and hard cheese 305 (+/-45) g, which were served to 14 healthy young men for three periods of three weeks each, separated by washout periods, in a randomized, cross-over study with strictly controlled dietary intake. Fasting blood samples were taken at the end of the study periods. Measurements of the postprandial effect of the three different dairy test products (0.7 g of milk fat/kg body weight) were carried out on day 4 of each intervention period. Blood samples were taken before and at 2, 4, 6 and 8 hours following intake of the meals. Fasting LDL cholesterol concentration was significantly higher after butter than cheese diet (p = 0.037), with a borderline significant difference in total cholesterol (p = 0.054) after the experimental periods of three weeks. Postprandial glucose showed a higher response after cheese diet than after milk diet (p = 0.010, diet x time interaction).
A different effect of fat in milk and butter could not be confirmed in this study. The moderately lower LDL cholesterol after cheese diet compared to butter diet should be investigated further.
Does fat in milk, butter and cheese affect blood lipids and cholesterol differently?
With increasing resources being spent on nutritional supplements, this study sought to evaluate the effect of introducing guidelines on prescribing of supplements, by auditing practice, prior to, and after the implementation of guidelines. Prescribing practice was evaluated from patient interviews, and knowledge of health professionals examined from questionnaires from 50 GP practices. Training on the use of guidelines on prescribing supplements was implemented, incorporating a Nutritional Screening Tool and practical application of high-energy dietary advice, targeting GPs and Community Nurses. Education to GPs and Community Nurses significantly reduced total prescribing by 15% and reduced the levels of inappropriate prescribing from 77% to 59% due to an improvement in monitoring of patients prescribed supplements. Although knowledge regarding high-energy dietary advice for nutritionally 'at risk' patients did improve as a result of the training, this was not demonstrated in practice. This lack of relevant dietary advice remained the main reason that inappropriate supplement prescriptions remained high.
Education on guidelines incorporating a Nutritional Screening Tool has proved to be an effective method of achieving more appropriate prescribing of supplements, suggesting the need for ongoing training of health professionals in Primary Care.
Prescribing of oral nutritional supplements in Primary Care: can guidelines supported by education improve prescribing practice?