context
stringlengths 151
5.1k
| answer
stringlengths 3
2.87k
| question
stringlengths 1
25.7k
|
---|---|---|
To evaluate the effectiveness of cytology and colposcopy in the follow up of women treated by laser CO(2) ablation for cervical intraepithelial neoplasia (CIN). A retrospective analysis of 1784 patients with CIN treated with laser CO(2) ablation from 1998 to 2003 at the Royal Women's Hospital, Melbourne. Data were collected prospectively in a computerised colposcopy database utilised in the Dysplasia Unit. There was equal distribution in the study population treated for low-grade intraepithelial neoplasia (LSIL) and high-grade intraepithelial neoplasia (HSIL), with no significant age difference in both groups (mean 27.6 years). The success rate of treatment of LSIL (94%) was similar to that of HSIL (92%). Fifty-seven per cent of all failures occurred within the first year of treatment. Colposcopy was more sensitive in detecting intraepithelial neoplasia (SIL) than cytology, whereas cytology was more specific. No case of cancer was diagnosed on follow up. | CO(2) laser ablation was equally effective in treating different grades of CIN. There is a continuing incidence of recurrent CIN, even up to five years after initial treatment, which suggests that it is necessary for patients to adhere to long-term follow up. The combination of cytology and colposcopic assessment is essential for surveillance of SIL in the initial two years after treatment. | Is colposcopy needed following laser ablation for dysplasia? |
CRC data were collected from surgery/pathology registers, and polyp data from colonoscopy reports. Patients who met the criteria for familial adenomatous polyposis, hereditary non-polyposis colorectal cancer syndrome or inflammatory bowel disease were excluded from the study. Overlap of patients between the two groups (cancers and polyps) was carefully avoided. The χ² statistical test and a regression analysis were performed. Data from a total of 768 patients (352 and 416 patients, respectively, in periods A and B) who underwent surgery for cancer were collected. During the same time periods, a total of 1693 polyps were analyzed from 978 patients with complete colonoscopies (428 polyps from 273 patients during period A and 1265 polyps from 705 patients during period B). A proximal shift in cancer occurred during the latter years for both sexes, but particularly in males. Proximal cancer increased>3-fold in period B compared to period A in males [odds ratio (OR) 3.31, 95%CI: 2.00-5.47; P<0.0001). A similar proximal shift was observed for polyps, particularly in males (OR 1.87, 95%CI: 1.23-2.87; P<0.0038), but also in females (OR 1.62, 95%CI: 0.96-2.73; P<0.07). | The prevalence of proximal proliferative colonic lesions seems to have increased over the last decade, particularly in males. | Is proliferative colonic disease presentation changing? |
Users of ±3,4-methylenedioxymethamphetamine (MDMA), "ecstasy," report that the drug produces unusual psychological effects, including increased empathy and prosocial feelings. These "empathogenic" effects are cited as reasons for recreational ecstasy use and also form the basis for the proposed use of MDMA in psychotherapy. However, they have yet to be characterized in controlled studies. Here, we investigate effects of MDMA on an important social cognitive capacity, the identification of emotional expression in others, and on socially relevant mood states. Over four sessions, healthy ecstasy-using volunteers (n = 21) received MDMA (.75, 1.5 mg/kg), methamphetamine (METH) (20 mg), and placebo under double-blind, randomized conditions. They completed self-report ratings of relevant affective states and undertook tasks in which they identified emotions from images of faces, pictures of eyes, and vocal cues. MDMA (1.5 mg/kg) significantly increased ratings of feeling "loving" and "friendly", and MDMA (.75 mg/kg) increased "loneliness". Both MDMA (1.5 mg/kg) and METH increased "playfulness"; only METH increased "sociability". MDMA (1.5 mg/kg) robustly decreased accuracy of facial fear recognition relative to placebo. | The drug MDMA increased "empathogenic" feelings but reduced accurate identification of threat-related facial emotional signals in others, findings consistent with increased social approach behavior rather than empathy. This effect of MDMA on social cognition has implications for both recreational and therapeutic use. In recreational users, acute drug effects might alter social risk-taking while intoxicated. Socioemotional processing alterations such as those documented here might underlie possible psychotherapeutic benefits of this drug; further investigation of such mechanisms could inform treatment design to maximize active components of MDMA-assisted psychotherapy. | Is ecstasy an "empathogen"? |
Irreversible electroporation (IRE) is (virtually) always called non-thermal despite many reports showing that significant Joule heating occurs. Our first aim is to validate with mathematical simulations that IRE as currently practiced has a non-negligible thermal response. Our second aim is to present a method that allows simple temperature estimation to aid IRE treatment planning. We derived an approximate analytical solution of the bio-heat equation for multiple 2-needle IRE pulses in an electrically conducting medium, with and without a blood vessel, and incorporated published observations that an electric pulse increases the medium's electric conductance. IRE simulation in prostate-resembling tissue shows thermal lesions with 67-92°C temperatures, which match the positions of the coagulative necrotic lesions seen in an experimental study. Simulation of IRE around a blood vessel when blood flow removes the heated blood between pulses confirms clinical observations that the perivascular tissue is thermally injured without affecting vascular patency. | The demonstration that significant Joule heating surrounds current multiple-pulsed IRE practice may contribute to future in-depth discussions on this thermal issue. This is an important subject because it has long been under-exposed in literature. Its awareness pleads for preventing IRE from calling "non-thermal" in future publications, in order to provide IRE-users with the most accurate information possible. The prospect of thermal treatment planning as outlined in this paper likely aids to the important further successful dissemination of IRE in interventional medicine. Prostate 75:332-335, 2015. © 2014 The Authors. The Prostate Published by Wiley Periodicals, Inc. | Irreversible electroporation: just another form of thermal therapy? |
Most estimated associations of posttraumatic stress disorder (PTSD) with DSM-IV drug dependence and abuse are from cross-sectional studies or from prospective studies of adults that generally do not take into account suspected causal determinants measured in early childhood. To estimate risk for incident drug disorders associated with prior DSM-IV PTSD. Multiwave longitudinal study of an epidemiologic sample of young adults first assessed at entry to first grade of primary school in the fall semesters of 1985 and 1986, with 2 young adult follow-up assessments. Mid-Atlantic US urban community. Young adults (n = 988; aged 19-24 years) free of clinical features of DSM-IV drug use disorders at the first young adult assessment and therefore at risk for newly incident drug use disorders during the 1-year follow-up period. During the 12-month interval between the 2 young adult follow-up assessments, newly incident (1) DSM-IV drug abuse or dependence; (2) DSM-IV drug abuse; (3) DSM-IV drug dependence; and (4) emerging dependence problems (1 or 2 newly incident clinical features of DSM-IV drug dependence), among subjects with no prior clinical features of drug use disorders. Prior PTSD (but not trauma only) was associated with excess risk for drug abuse or dependence (adjusted relative risk, 4.9; 95% confidence interval, 1.6-15.2) and emerging dependence problems (adjusted relative risk, 4.9; 95% confidence interval, 1.2-20.1) compared with the no-trauma group controlling for childhood factors. Subjects with PTSD also had a greater adjusted relative risk for drug abuse or dependence compared with subjects exposed to trauma only (adjusted relative risk, 2.0; 95% confidence interval, 1.1-3.8) controlling for childhood factors. | Association of PTSD with subsequent incident drug use disorders remained substantial after statistical adjustment for early life experiences and predispositions reported in previous studies as carrying elevated risk for both disorders. Posttraumatic stress disorder might be a causal determinant of drug use disorders, possibly representing complications such as attempts to self-medicate troubling trauma-associated memories, nightmares, or painful hyperarousal symptoms. | Incidence of drug problems in young adults exposed to trauma and posttraumatic stress disorder: do early life experiences and predispositions matter? |
To compare the function in aortic position of cryopreserved pulmonary homografts subjected to pulmonary hypertension with that of normal cryopreserved pulmonary homografts. Pulmonary valves (52) were implanted in aortic position in different cardiothoracic centres. The valves were classified as follows: Group I-pulmonary hypertension (procured from recipients of heart/heart-lung transplantation, 31 valves), Group II-normal pulmonary pressure (procured from cadavers and multiorgan donors, 21 valves). Regular echocardiographic follow-up was obtained by the implanting centers. Significant echo changes were defined as insufficiency>2+ and/or stenosis producing a delta P>30 mm Hg. Pulmonary homografts showed the following significant echo changes: in the Pulmonary Hypertension Group, 7, 27 and 33% at 12, 24 and 36 months, respectively; in the normal PA Group 10, 37.5 and 80% at 12, 24 and 36 months, respectively. In both groups the most common echocardiographic alteration was homograft insufficiency rather than stenosis. Thus, pulmonary homografts subjected to long-term pulmonary hypertension have significantly less echo changes than normal pulmonary homografts, especially after 12 months (chi 2: P<0.036). | These findings suggest that pulmonary valves subjected to pulmonary hypertension might be more appropriate than normal pulmonary homograft for aortic valve replacement, constituting a possible alternative in case of lack of aortic valve homografts. However, the failure of two out of five valves in the longer term must dictate caution while waiting further long-term results. | Are pulmonary homografts subjected to pulmonary hypertension more appropriate for aortic valve replacement than normal pulmonary homografts? |
The aim of this study was to investigate the efficacy of tranexamic acid in an irrigant fluid in decreasing intraoperative blood loss during orthognathic surgery. This was a prospective, randomized, double-blind, placebo-controlled trial of elective bimaxillary osteotomy. Forty patients were included in the study and 20 were randomly assigned to each group. Drawing of random lots determined whether 0.05% tranexamic acid in normal saline solution or normal saline was used as an irrigant fluid during surgery. All patients underwent hypotensive anesthesia and surgery according to standard protocol. Intraoperative blood loss, operative and hypotensive times, preoperative and postoperative hematocrit levels, transfusion of blood product, and amount of irrigant fluid were recorded. Parametric data were reported as mean ± standard deviation and nonparametric data were counted. Changes in parametric variables were analyzed using unpaired Student t test. Two-sided significance tests were used. P<.05 was accepted as statistically significant. Blood loss during bimaxillary surgery was not decreased significantly in the tranexamic acid group compared with the control group (832.5 ± 315.5 vs 917.5 ± 424.0 mL, respectively, P = .47). | Tranexamic acid in an irrigant fluid does not significantly decrease intraoperative blood loss compared with placebo during orthognathic surgery. | Does tranexamic acid in an irrigating fluid reduce intraoperative blood loss in orthognathic surgery? |
To investigate reflux development and changes in resting venous diameters in the DVT and the non-DVT lower limbs. Twenty subjects (40 limbs) with acute unilateral proximal DVT diagnosed by ultrasound, who were treated with low-molecular-weight-heparin (LMWH), followed by at least three months of oral warfarin therapy, were enrolled in the study. The limbs were classified according to CEAP (clinical, aetiologic, anatomic, pathophysiology) clinical classification on a scale of 0-6. Duplex ultrasound (DUS) was employed to assess DVT resolution, vein diameter and venous reflux in both limbs at intervals of zero, three, six and 12 months. Venous reflux was defined as a valve closure time more than 1 s. There were 13 men and seven women, average age was 40.8 years and average body mass index 27.7 kg/m2. In the DVT limbs at three, six and 12 months, deep veins were non-occluded in 40%, 60% and 85%, respectively. By 12 months, 16 (80%) had developed venous reflux, mostly in the femoral (FV) and popliteal veins (PV); eight limbs (40%) were in clinical classes 4-6. In the contralateral 20 non-DVT limbs, four limbs developed borderline reflux at the sapheno-femoral junction (SFJ) after six months and mean diameters of SFJ, FV and PV increased significantly. | Venous reflux is highly likely to occur in DVT limbs within a year follow-up period. Venous dilatation can occur in the contralateral unaffected lower limb, consistent with a systemic effect. Our results are suggestive and further studies are needed. | Is venous reflux a common disease in post-thrombotic patients with unilateral deep vein thrombosis episode? |
As a symptom of multiple sclerosis (MS), fatigue is difficult to manage because of its unknown etiology, the lack of efficacy of the drugs tested to date and the absence of consensus about which would be the ideal measure to assess fatigue. Our aim was to assess the frequency of fatigue in a sample of MS patients and healthy controls (HC) using two fatigue scales, the Fatigue Severity Scale (FSS) and the Modified Fatigue Impact Scale (MFIS) with physical, cognitive and psychosocial subscales. We also studied the relationship fatigue has with depression, disability and interferon beta. Three hundred and fifty-four individuals (231 MS patients and 123 HC) were included in this cross-sectional study. Fatigue was assessed using the FSS and MFIS. Depression was measured by the Beck Depression Inventory (BDI), and disability by the Expanded Disability Status Scale (EDSS). A status of fatigue was considered when the FSS>or =5, of non-fatigue when the FSS<or =4, and scores between 4.1 and 4.9 were considered doubtful fatigue cases. Fifty-five percent of MS patients and 13% of HC were fatigued. The global MFIS score positively correlated with the FSS in MS and HC (r =0.68 for MS and r =0.59 for HC, p<0.0001). Nonetheless, the MFIS physical subscale showed the strongest correlation score with the FSS (r =0.75, p<0.0001). In addition, a prediction analysis showed the physical MFIS subscale to be the only independent predictor of FSS score (p<0.0001), suggesting other aspects of fatigue, as cognition and psychosocial functions, may be explored by the FSS to a lesser extent. Depression also correlated with fatigue (r =0.48 for the FSS and r =0.7 for the MFIS, p<0.0001) and, although EDSS correlated with fatigue as well, the scores decreased after correcting for depression. Interferon beta showed no relationship with fatigue. | Fatigue is a frequent symptom found in MS patients and clearly related with depression. Each fatigue scale correlates with one another, indicating that they are measuring similar constructs. Nevertheless, spheres of fatigue as cognition and psychosocial functions are probably better measured by the MFIS, although this hypothesis will need to be confirmed with appropriate psychometrical testing. | Does the Modified Fatigue Impact Scale offer a more comprehensive assessment of fatigue in MS? |
To examine the value of a new screening instrument in a visual-analogue format. We report the design and validation of a new five-dimensional tool called the Emotion Thermometers (ET). This is a combination of five visual-analogue scales in the form of four predictor domains (distress, anxiety, depression, anger) and one outcome domain (need for help). Between March and August 2007, 130 patients attending the chemotherapy suite for their first chemotherapy treatment were asked to complete several questionnaires with validation for distress, anxiety and depression. Of 81 with low distress on the Distress Thermometer (DT), 51% recorded emotional difficulties on the new ET tool, suggesting added value beyond the DT alone. Of those with a broadly defined emotional complication, 93.3% could be identified using the Anxiety Thermometer (AnxT) alone, compared with 54.4% who would be recognized using the DT alone. Using a cut-off of 3v4 on all thermometers against the total Hospital Anxiety and Depression Scale (HADS) score (cut-off 14v15), the optimal thermometer was the Anger Thermometer (sensitivity 61%, specificity 92%). Against HADS anxiety scale, the optimal thermometer was AnxT (sensitivity 92%, specificity 61%) and against the HADS depression scale, the optimal thermometer was the Depression Thermometer (DepT; sensitivity 60%, specificity 78%). Finally, against DSM-IV major depression, the optimal thermometer was the DepT (sensitivity 80%, specificity 79%). Further improvements may be possible by using a combination of thermometers or by repeating the screen. | The diagnostic accuracy of the DT can be improved by the inclusion of simple addition linear domains without substantially increasing the time needed to apply the test. | Can the Distress Thermometer be improved by additional mood domains? |
To compare the short-term outcomes of acute knee injuries treated by specialists and generalists. Using patient logs, 168 adults with acute knee injuries were identified; 131 (78%) completed a questionnaire 3 months after initial presentation. The mean age of the 77 male and 54 female responders was 34.6 years (range 18-73 years). The injuries were classified as mild (n = 35), moderate (n = 75), or severe (n = 21). Most responders were satisfied with their care and outcome, but 22% noted some functional limitations. The 59 patients seeing an orthopedist were more likely to have had a severe injury, more physician visits, activity limitations, lost time from work or recreation, and more pain when compared with the 72 patients who never saw an orthopedist. Excluding surgical patients, however, satisfaction was not significantly different by provider. After multivariate modeling (adjusting for age, sex, injury severity, and diagnosis), there was no significant association between having seen an orthopedist and either treatment success or satisfaction. | With the exception of time lost for recuperation in our community there is little difference in short-term outcome for patients with acute knee injury not undergoing surgery, regardless of the specialty of the treating physician. | Short-term outcomes of acute knee injuries: does the provider make a difference? |
The objective of the study was to determine the integrity of flat square knots. Three sutures were used in both 0 and 2-0 suture gauges: poliglecaprone 25 (Monocryl), polyglactin 910 (Vicryl), and silk. For each, flat square knots were tied with either 3 or 5 throws. Knots were tested to failure. The major outcome measured was the proportion of 3 throw knots untying, compared with that of 5 throw knots. There were high rates of untying for the poliglecaprone 25 and for the polyglactin 910 with both suture gauges when tied with only 3 throws. The failure rate decreased significantly when the throws were increased. There was no statistical benefit to increasing the number of throws for silk. | Knot failure is decreased by increasing the number of throws for poliglecaprone 25 and polyglactin 910. However, there is no advantage to increasing the number of throws from 3 to 5 for silk. | Flat square knots: are 3 throws enough? |
Prospective study of children aged 15 days to 2 years admitted with a diagnosis of first convulsion over a 2-year period (May 2004-May 2006). All infants were examined by an experienced ophthalmologist using indirect ophthalmoscopy within 72 h of admission. 182 of 389 children seen in the accident and emergency department were admitted and two were found to have retinal haemorrhages. Both children were eventually diagnosed as being abused. | Convulsions alone are unlikely to cause retinal haemorrhages in children under 2 years of age. | Do retinal haemorrhages occur in infants with convulsions? |
To examine condom efficacy and the stages of change in explaining condom usage among a sample of 278 injected-drug users (IDUs). Data were collected as part of a statewide behavioral surveillance study. Participants reported multiple risk behaviors, including having unprotected sex with multiple partners and exchanging sex. Approximately half the sample was in the precontemplation or contemplation stage for condom use, whereas 21% were in maintenance. Mediational analyses indicated that condom efficacy partially mediated the association between stage of change and self-reported condom use. | The findings are discussed within the context of targeting HIV prevention interventions. | Does efficacy mediate stage of change and condom use in injected-drug users? |
Gold markers are frequently used for a better daily repositioning of the prostate before irradiation. The purpose of this work was to analyze if the combination of an androgen deprivation with the external irradiation could modify the position of the gold markers in the prostate. Ten patients have been treated for a prostate cancer, using three implanted gold markers. The variations of the intermarker distances in the prostate were measured and collected on daily OBI(®) kilovoltage images acquired at 0° and 90°. Five patients had a 6-month androgen deprivation started before the external irradiation (H group) and five did not (NH group). A total number of 1062 distances were calculated. No distance variation greater than 3.7mm was seen between two markers, in any of the two groups. The median standard deviations of the daily intermarker distance differences were 0.7mm (range 0.3-1.2mm) for the H group and 0.6mm (range 0.2-1.2mm) for the NH group. The intermarker distances variations were noted as greater than -2mm, between -2mm and 2mm and greater than 2mm in 16.4, 83.4 and 0.2% for the H group and 1.3, 98.5 and 0.2% for the NH group, respectively. | The distance variations remained less than 4mm in both groups and for all the measurements. In the NH group, the variation of the distance between two markers remained below 2mm in 98.5%. In the H group, the presence of a reduction of distance above 2mm in 16.4% of measurements could indicate the shrinkage of the prostate volume. | Does hormone therapy modify the position of the gold markers in the prostate during irradiation? |
To compare the efficacy of letrozole with clomiphene citrate for ovarian stimulation prior to intrauterine insemination. A total of 272 patients underwent 362 cycles of stimulation prior to IUI from January 2009 to May 2011. One hundred and twenty-six patients (170 cycles) received letrozole 5 mg and 146 patients (192 cycles) received clomiphene citrate 100 mg for 5 days starting on day 3 of the cycle. Intrauterine insemination was done 36 ± 4 h after hCG injection in 99 patients (131 cycles) in the letrozole group and 115 patients (136 cycles) in the clomiphene group. The number of mature follicles and serum E2 levels was monitored on the day of hCG. Endometrial thickness was noted on the day of hCG and IUI as well. The pregnancy rate, the miscarriage rate, and other parameters were evaluated only for patients who underwent IUI. The mean number of mature follicles was significantly more in the clomiphene group (1.66 ± 1.69 vs. 1.41 ± 1.53) than in the letrozole group. No significant difference in endometrial thickness on the day of hCG administration was found; however, endometrial thickness was statistically significantly higher in the clomiphene group on the day of IUI (9.28 ± 1.43 vs. 8.77 ± 1.34). Serum E(2) levels on the day of hCG were statistically significantly higher in the clomiphene group (501.09 ± 2.45 vs. 214.79 ± 2.43 pg/ml). Pregnancy rate was 9.2 %/cycle and 12.12 %/patient in the letrozole group, and 8.1 %/cycle and 9.5 %/cycle in the clomiphene group; these differences were not statistically significant. | This study found no superiority of letrozole over clomiphene when used for ovarian stimulation and IUI. | Is letrozole superior to clomiphene for ovarian stimulation prior to intrauterine insemination? |
Deficits in cognitive functioning have been associated with bipolar disorder during episodes of depression and mania, as well as during periods of symptomatic remission. Separate evidence suggests that patients may lack awareness of these deficits and may even be overly confident with self-appraisals. The extent to which these separately or together represent prodromes of the disorder versus a consequence of the disorder remains unclear. The present study sought to test whether risk for bipolar disorder in a younger, college-aged cohort of individuals would be associated with lower performance in cognitive ability yet higher self-appraisal of cognitive functioning. Participants (N=128) completed an objective measure of working memory, a self-report measure of everyday cognitive deficits, and a measure associated with risk for bipolar disorder. Contrary to expectation, risk for bipolar disorder did not significantly predict poorer working memory. However, a person's risk for bipolar disorder was associated with higher self-appraisal of cognitive functioning relative to those with lower risk despite there being no indication of a difference in ability on the working memory task. Participant recruitment relied on an analog sample; moreover, assessment of cognitive functioning was limited to working memory. | Results add to a growing body of evidence indicating that overconfidence may be part of the cognitive profile of individuals at risk for bipolar disorder. | Does risk for bipolar disorder heighten the disconnect between objective and subjective appraisals of cognition? |
To report the incidence of massive fetomaternal hemorrhage (FMH) associated with fetal death and to test the hypothesis that FMH is more likely to occur in those with risk factors for FMH. All cases of fetal death of infants weighing>500 gm between January 1, 1990 and December 31, 1994 were reviewed for evidence of massive FMH (>or = 2% fetal cells in the maternal circulation as measured by the Betke-Kleihauer test). Women with risk factors were compared with those without risk factors with respect to the occurrence of massive FMH. The prevalence of massive FMH was 14 of 319 (4.4%) cases, occurring in 4 of 102 (3.9%) of those with risk factors and 10 of 217 (4.6%) of patients without risk factors (p = 0.78). Otherwise unexplained fetal death was associated with massive FMH in 5 of 141 (3.5%). Major fetal anomalies were present in 5 of 14 (35.7%) cases of massive FMH. | Clinical risk factors do not predict an increased likelihood of massive FMH. Massive FMH is associated with fetal anomalies. Betke-Kleihauer testing should be performed in all cases of fetal death, including those with anomalies regardless of the presence or absence of risk factors for FMH. | Massive fetomaternal hemorrhage and fetal death: are they predictable? |
The criterium defining the threshold size of adrenal incidentaloma (AI) is a size greater than 1 cm diameter. However, data concerning AI≤1 cm in diameter is scant. The aim of this study was to evaluate the function of adrenal masses≤1 cm and to compare them with adrenal masses>1 cm. The study included 130 consecutive patients with AI (38 and 92 AI at ≤ 1 cm and> 1 cm, respectively). The patients were evaluated according to demographic and hormonal characteristics. The prevalence of SCS was 5.3 and 12% in AI≤1 cm and> 1 cm diameter, respectively. Hyperaldosteronism was found only in patients with> 1 cm AI. Pheochromocytoma were not found in either group. Patients with> 1 cm AI had a higher prevalence of SCS and primary hyperaldosteronism than patients with ≤ 1 cm AI, but the difference was not significant. The prevalence of diabetes and hypertension was high both in non-functional AI with ≤ 1 cm and> 1 cm patients and showed no significant difference between the 2 groups. | Our study is the first to focus on the clinical and hormonal characteristics of patients with ≤ 1 cm AI. Those with AI≤1 cm harboured SCS, as was the case for AI>1 cm. Similar to AI>1 cm, non-functional AI≤1 cm also had a higher prevalence of diabetes and hypertension. | Are small adrenal incidentalomas solely a radiological finding? |
To compare urodynamic findings after holmium laser enucleation of the prostate (HoLEP) and transurethral resection of the prostate (TURP) for the treatment of benign prostatic hyperplasia-related bladder outlet obstruction. From January to October 2002, 100 consecutive patients with benign prostatic hyperplasia with obstructive lower urinary tract symptoms were randomized to surgical treatment with either HoLEP (group 1, n = 52) or TURP (group 2, n = 48). All patients were preoperatively assessed using the International Prostate Symptom Score and quality-of-life question, total serum prostate-specific antigen measurement, transrectal ultrasonography, and complete urodynamic study. The operative time, catheterization time, and overall hospital stay were also recorded for both groups. All patients were assessed at 1, 6, and 12 months postoperatively using a complete urodynamic evaluation. All patients were obstructed preoperatively (Schäfer grade greater than 2). Both groups were comparable in terms of age, total serum prostate-specific antigen level, International Prostate Symptom Score, and urodynamic results. At 1, 6, and 12 months of follow-up, no statistically significant differences were recorded in terms of detrussor pressure at maximal urinary flow rate, Schäfer grade (linear passive urethral resistance relation), maximal urinary flow rate, International Prostate Symptom Score, and quality-of-life score. In the HoLEP group, the catheterization time and hospital stay were significantly shorter. Transitory lower urinary tract symptoms after 3 months of follow-up and dysuria were more frequent in the HoLEP group than in the TURP group, although at 12 months of follow-up, the results were comparable. | Both HoLEP and TURP were equally effective in relieving bladder outlet obstruction. Although associated with greater early self-resolving irritative symptoms, HoLEP can guarantee a shorter catheterization time and hospital stay with longer operative times, proposing itself as an attractive alternative to standard TURP. | Urodynamics after TURP and HoLEP in urodynamically obstructed patients: are there any differences at 1 year of follow-up? |
Open transumbilical pyloromyotomy (UMBP) and laparoscopic pyloromyotomy (LAP) have been compared on different outcomes, but postoperative pain as a primary end point had never been assessed. The aim of this study was to compare the use of analgesia in UMBP and LAP patients. Infants with hypertrophic pyloric stenosis treated by UMBP in 2008-2009 were matched with LAP-treated infants. Demographics, type and use of analgesia, and length of stay were recorded. Statistical analysis was performed using the Fisher exact test. Each group contained 19 patients (N = 38) with comparable demographics and no comorbid condition. Bupivacaine was injected intraoperatively in all UMBP and 89% of LAP infants. There was a trend toward increased acetaminophen use in LAP infants (79% vs 58%, P = .61) in the recovery room. There was no difference in opiates use (3 UMBP vs 1 LAP, P = .60). In the ward, more UMBP patients received acetaminophen (78% vs 53%, P = .03). This difference was significant. Mean postoperative length of stay was similar in both groups. | Our study suggests that UMBP infants might experience more postoperative pain in the ward, without any impact on various outcomes. A prospective study with a larger sample size should be undertaken to verify these findings. | Open transumbilical pyloromyotomy: is it more painful than the laparoscopic approach? |
To describe AIDS and malaria geography in Brazil, highlighting the role of injecting drug users (IDUs) in malaria outbreaks occurring in malaria-free regions, and the potential clinical and public health implications of malaria/HIV co-infection. Review of the available literature and original analyses using geoprocessing and spatial analysis techniques. Both HIV/AIDS and malaria distribution are currently undergoing profound changes in Brazil, with mutual expansion to intersecting geographical regions and social networks. Very recent reports describe the first clinical case of AIDS in a remote Amazonian ethnic group, as well as malaria cases in Rio de Janeiro state (hitherto a malaria-free area for 20 years); in addition, two outbreaks of both infections occurred at the beginning of the 1990s in the most industrialized Brazilian state (São Paulo), due to the sharing of needles and syringes by drug users. Spatial data point to: (a) the expansion of HIV/AIDS towards malarigenic areas located in the centre-west and north of Brazil, along the main cocaine trafficking routes, with IDU networks apparently playing a core role; and (b) the possibility of new outbreaks of secondary malaria in urban settings where HIV/AIDS is still expanding, through the sharing of needles and syringes. | New outbreaks of cases of HIV and malaria are likely to occur among Brazilian IDUs, and might conceivably contribute to the development of treatment-resistant strains of malaria in this population. Health professionals should be alert to this possibility, which could also eventually occur in IDU networks in developed countries. | Co-infection with malaria and HIV in injecting drug users in Brazil: a new challenge to public health? |
The optimal management of congenital adenomatoid malformation of the lung remains controversial. Prenatal ultrasonographic analysis has increasingly discovered asymptomatic lesions, raising questions about the need for and timing of surgical treatment for asymptomatic congenital adenomatoid malformation. The aim of our study was to analyze the short-term postoperative outcome of symptomatic congenital adenomatoid malformations compared with asymptomatic malformations. All the data of patients presenting with congenital adenomatoid malformations histologically diagnosed and operated on between 1998 and 2005 at our institution were retrospectively reviewed. Patients were divided into 2 groups: group A comprised asymptomatic infants, and group B comprised symptomatic infants. Major outcomes considered were the length of ventilation, pleural drainage, and hospital stay. Postoperative morbidity and mortality were also evaluated. Asymptomatic patients were further stratified for age at the time of the operation to evaluate whether age at surgical intervention affects the outcome. The Fisher's exact and Mann-Whitney tests were used as appropriate. Fifty-seven patients were consecutively treated. Thirty-five patients were given diagnoses of asymptomatic lesions and were enrolled into group A, whereas 22 patients presenting with symptoms were entered into group B. The lengths of ventilation, pleural drainage, and hospital stay were significantly longer in patients with symptomatic congenital adenomatoid malformations. Moreover, symptomatic patients presented with a higher postoperative complication rate. The age-based stratification of asymptomatic children did not show any difference on either postoperative mortality or major outcome considered. | Children with congenital adenomatoid malformations operated on when asymptomatic present a better short-term outcome than symptomatic children. In addition, age at the time of the operation does not negatively affect the outcome. Our findings support early surgical treatment for asymptomatic congenital adenomatoid malformation. | Asymptomatic congenital cystic adenomatoid malformation of the lung: is it time to operate? |
Crohn's disease (CD) is a chronic inflammatory bowel disease characterised by a peculiar accumulation of mesenteric adipose tissue covering the inflamed intestinal wall. The authors characterised different adipose tissue compartments of patients with CD using morphological and molecular techniques and compared them to those of subjects with obesity (OB) and healthy subjects with normal weight (N). Adipose tissue samples were taken from subcutaneous adipose tissue, omental visceral adipose tissue (VAT) and healthy mesenteric depot (hMES), as well as from fat wrapping the affected (unhealthy) intestinal tracts (uhMES). Microarray analyses, validated by real-time quantitative PCR technique, were performed in whole adipose tissue and in isolated adipocytes. The morphology of subcutaneous adipose tissue was similar in subjects with CD and those with N. In patients with CD, VAT adipocytes were smaller than those derived from uhMES and hMES and were smaller than VAT adipocytes of subjects with N. The molecular profiles of CD, VAT and uhMES were characterised by upregulation of genes related to inflammation and downregulation of those involved in lipid metabolism. Adipocytes isolated from VAT of subjects with CD and those with OB exhibited similar upregulation of genes involved in inflammation and immunity. VAT adipocytes of patients with CD compared to those of patients with OB also showed a greater upregulation of several anti-inflammatory genes. | In patients with CD, VAT distant from uhMES is affected by inflammation and displays features similar to those of VAT of patients with severe OB. The small diameter of VAT adipocytes of CD, together with their high expression of anti-inflammatory genes, suggests a potentially protective role for this tissue. VAT adipocytes may play an important role in the pathophysiology and/or activity of CD. | Visceral adipocytes: old actors in obesity and new protagonists in Crohn's disease? |
A proliferation-inducing ligand (APRIL) is a new member of the tumour necrosis factor family which is intimately connected to the regulation of cellular pathways. The aim of this study was to assess serum concentrations of APRIL in systemic sclerosis patients, and to correlate them with the main clinical and serological features of the disease. Sera from 35 patients with systemic sclerosis, 25 had limited cutaneous and 10 had diffuse cutaneous subtypes, and 35 normal healthy subjects were assayed for APRIL by Enzyme Linked Immunosorbant Assay. Demographic, clinical, autoantibodies and serological data were prospectively assessed. Serum APRIL concentrations were higher in patients with systemic sclerosis and in both its subtypes compared to the healthy controls (p<0.0001 in all). Patients with elevated APRIL levels had significantly higher incidences of myositis than those with normal levels (p=0.04). We did not find significant differences in other organ involvement prevalence between systemic sclerosis patients with elevated vs. normal APRIL levels. In addition, the frequencies of autoantibodies (i.e., anti-topoisomerase I, anti-centromere) were comparable between both groups. Serum APRIL levels were correlated with serum γ-globulins concentrations (r=0.404, p=0.016) but not with C-reactive protein, skin score, nor pulmonary functions. Serum APRIL was also correlated with creatine kinase levels only in systemic sclerosis patients with myositis (r=0.786, p=0.02). | Our preliminary results suggest increased serum APRIL levels in systemic sclerosis patients, particularly in those associated with myositis and hypergammaglobinemia. To confirm our results, we propose that larger scale, multicentre studies with longer evaluation periods are needed. | Elevated serum levels of a proliferation-inducing ligand in patients with systemic sclerosis: possible association with myositis? |
Decreased cerebral blood volume is known to be a predictor for final infarct volume in acute cerebral artery occlusion. To evaluate the predictability of final infarct volume in patients with acute occlusion of the middle cerebral artery (MCA) or the distal internal carotid artery (ICA) and successful endovascular recanalization, pooled blood volume (PBV) was measured using flat-panel detector computed tomography (FPD CT). Twenty patients with acute unilateral occlusion of the MCA or distal ACI without demarcated infarction, as proven by CT at admission, and successful Thrombolysis in cerebral infarction score (TICI 2b or 3) endovascular thrombectomy were included. Cerebral PBV maps were acquired from each patient immediately before endovascular thrombectomy. Twenty-four hours after recanalization, each patient underwent multislice CT to visualize final infarct volume. Extent of the areas of decreased PBV was compared with the final infarct volume proven by follow-up CT the next day. In 15 of 20 patients, areas of distinct PBV decrease corresponded to final infarct volume. In 5 patients, areas of decreased PBV overestimated final extension of ischemia probably due to inappropriate timing of data acquisition and misery perfusion. | PBV mapping using FPD CT is a promising tool to predict areas of irrecoverable brain parenchyma in acute thromboembolic stroke. Further validation is necessary before routine use for decision making for interventional thrombectomy. | Does preinterventional flat-panel computer tomography pooled blood volume mapping predict final infarct volume after mechanical thrombectomy in acute cerebral artery occlusion? |
201 persons from a population-based cohort of Danish type 1 diabetic patients were examined at baseline and again 25 years later. At both examinations the patients were asked about their smoking habits. The level of retinopathy was evaluated by ophthalmoscopy at baseline and by nine 45-degree colour field fundus photos at the follow-up. In multivariate analyses there was a trend that current smokers at baseline were more likely to develop PDR at the follow-up (odds ratio 1.90, 95% confidence interval 0.88-4.11, p = 0.10). Neither smoking status at the follow-up nor pack-years of smoking were associated with PDR. | We found neither a beneficial nor a harmful effect of smoking on long-term incidence. Selective mortality among smokers and patients with PDR at baseline might provide at least part of the explanation for this. | Is smoking a risk factor for proliferative diabetic retinopathy in type 1 diabetes? |
Despite some reports on a potential link between parenthood and mental health, associations have not been systematically investigated yet. The present article provides prevalence rates of the most common mental disorders for parents and non-parents. Interactions between demographic and socio-economic variables, parental status and mental health are explored. Data from the 1998/99 German Health Survey (GHS) and its Mental Health Supplement (GHS-MHS) were analysed using logistic regression models. Analyses were restricted to participants in the age group 18 to 49 (N=2,801). Mental disorders and syndromes were assessed with a standardized diagnostic interview (M-CIDI). Parenthood was associated with lower rates of psychiatric morbidity in general, and depressive and substance use disorders, in particular. The association between parental status and mental health was more distinct in men than in women, whereas partnership status moderated this relationship: An absence of partnership was associated with increased rates of all common mental disorders. Among non-parents, such a difference could not be found. Full-time employment, compared to part-time employment or unemployment, was linked to lower rates of the common mental disorders among fathers but not among mothers and non-parents. Age, education and income had no effects on the association between parental status and mental health. | Parenthood is positively associated with mental health, particularly for men. Most differences can be found for depressive and substance use disorders. Partnership seems especially important for parents since it does not affect prevalence rates of mental disorders among non-parents. | Is parenthood associated with mental health? |
Phase 3 clinical trials performed primarily outside the US demonstrate that intravesical instillation of chemotherapy immediately after transurethral resection of the bladder (TURB) decreases cancer recurrence rates. The authors sought to determine whether US urologists have adopted this practice, and its potential effect on costs of bladder cancer (BC) care. By using 1997-2004 MEDSTAT claims data, the authors identified patients with newly diagnosed BC who underwent cystoscopic biopsy or TURB, and those who received intravesical chemotherapy within 1 day after TURB. Economic consequences of this treatment compared with TURB alone were modeled using published efficacy estimates and Medicare reimbursements. The authors used a time horizon of 3 years and assumed that this treatment was given for all newly diagnosed low-risk BC patients. Between 1997 and 2004, the authors identified 16,748 patients with newly diagnosed BC, of whom 14,677 underwent cystoscopic biopsy or TURB. Of these, only 49 (0.33%) received same-day intravesical instillation of chemotherapy. From 1997 through 2004, there has been little change in the use of this treatment. The authors estimated a 3-year savings of $538 to $690 (10% to 12%) per patient treated with TURB and immediate intravesical chemotherapy compared with TURB alone, reflecting a yearly national savings of $19.8 to $24.8 million. | Instillation of intravesical chemotherapy immediately after TURB has not been embraced in the US. Adopting this policy would significantly lower the cost of BC care. | Treatment of nonmuscle invading bladder cancer: do physicians in the United States practice evidence based medicine? |
Cuffed endotracheal tubes are used to prevent gas leak and also pulmonary aspiration in ventilated patients. The pressure exerted on the tracheal wall is similar to intracuff pressure. The perfusion pressure for the tracheal mucosa is 40 cm H(2)O. Cuff pressures greater than 40 cm H(2)O may cause various ischemic changes and complications. High cuff pressures have also been implicated in postoperative sore throat and nonischemic complications. Postintubation endotracheal tube cuff pressures are not routinely measured in the ED or prehospital setting. The time spent in these settings may be long enough for pressure-induced tracheal mucosal injury to occur. The purpose of this study is to assess cuff pressures in intubated patients before aeromedical transport. All intubated patients transported by an aeromedical transport program during a 3-month period were included in this study. Patients were intubated either by helicopter physicians or before helicopter arrival at the referring hospital or by ambulance personnel. Cuff pressure was measured using a manometer (Cuffpressure, Posey Co, USA). This measurement was recorded, and correction was performed, if necessary, to achieve a cuff pressure of 14 to 27 cm H(2)O while preventing an air leak. Data were analyzed for the distribution of intracuff pressures and incidence of elevated pressure on first measurement and the need for correction. There were 62 patients in this study. The mean first recorded pressure was 63 +/- 34 cm H(2)O. Initial cuff pressures were greater than 40 cm H(2)O in 36 (58%) patients and required correction. | In this study, most cuff pressures exceeded safe pressure and required correction. Measurement of intracuff pressure is a simple and inexpensive procedure and should be done whenever a patient is intubated, in either the prehospital or hospital setting, because this may reduce long-term morbidity. | Endotracheal intracuff pressures in the ED and prehospital setting: is there a problem? |
Sublobar resection for early-stage lung cancer is still a controversial issue. We sought to compare sublobar resection (segmentectomy or wedge resection) with lobectomy in the treatment of patients with a second primary lung cancer. From January 1995 to December 2010, 121 patients with second primary lung cancer, classified by the criteria proposed by Martini and Melamed, were treated at our Institution. We had 23 patients with a synchronous tumour and 98 with metachronous. As second treatment, we performed 61 lobectomies (17 of these were completion pneumonectomies), 38 atypical resections and 22 segmentectomies. Histology was adenocarcinoma in 49, squamous in 38, bronchoalveolar carcinomas in 14, adenosquamous in 8, large cells in 2, anaplastic in 5 and other histologies in 5. Overall 5-year survival from second surgery was 42%; overall operative mortality was 2.5% (3 patients), while morbidity was 19% (22 patients). Morbidity was comparable between the lobectomy group, sublobar resection and completion pneumonectomies (12.8, 27.7 and 30.8%, respectively, P = 0.21). Regarding the type of surgery, the lobectomy group showed a better 5-year survival than sublobar resection (57.5 and 36%, respectively, P = 0.016). Compared with lobectomies, completion pneumonectomies showed a significantly less-favourable survival (57.5 and 20%, respectively, P = 0.001). | From our experience, lobectomy should still be considered as the treatment of choice in the management of second primary lung cancer, but sublobar resection remains a valid option in high-risk patients with limited pulmonary function. Completion pneumonectomy was a negative prognostic factor in long-term survival. | Is lobectomy really more effective than sublobar resection in the surgical treatment of second primary lung cancer? |
This study aimed to compare outcomes for mastoidotympanoplasty and for tympanoplasty alone in cases of quiescent, tubotympanic, chronic, suppurative otitis media. Single-blinded, randomised, controlled study within a tertiary referral hospital. Sixty-eight cases were randomly allocated into two groups. In group one, 35 ears underwent type one tympanoplasty along with cortical mastoidectomy. In group two, 33 ears underwent type one tympanoplasty alone. Outcome measures were as follows: perforation closure and graft uptake, hearing improvement, disease eradication, and post-operative complications. There were no statistically significant differences in hearing improvement, tympanic perforation closure, graft uptake or disease eradication, comparing the two groups at three and six months post-operatively. | Mastoidotympanoplasty was not found to be superior to tympanoplasty alone over a short term follow-up period. Hence, it may not be necessary to undertake routine mastoid exploration at this stage of disease. | Cortical mastoidectomy in quiescent, tubotympanic, chronic otitis media: is it routinely necessary? |
To explore the possible relationship between Y-chromosome microdeletions and a rare spermatogenic disorder, globozoospermia. Twelve patients with 100% globozoospermia were evaluated. Each man was questioned about his medical and surgical history and underwent a thorough andrologic examination. Plasma follicle-stimulating hormone (FSH), luteinizing hormone (LH), testosterone and prolactin levels were measured. Routine sperm analysis and morphology with electron microscopy were done. A set of 17 Y-specific sequence-tagged sites spanning the 3 AZF regions and also RBM1, ZFY, SRY, sY78, CDY, BPY2 and PRY were tested to detect the existence of Y-chromosome microdeletions by polymerase chain reaction. The mean age of the patients was 36 years (range, 27-42). No patient had abnormal blood concentrations of FSH, LH, testosterone or prolactin. Semen analysis revealed normal values for volume (2.2 mL) and concentration (32 x 10(6)/mL) but subnormal values for motility (37%) and progressive motility (24%). On light and electron microscopy, all the spermatozoa were round headed, with abnormal morphologic features. Patients had normal 46,XY karyotyping. No microdeletion of the Y chromosome was detected in any patient. | Although this study did not find any Y-chromosome microdeletions in patients with globozoospermia, the exact genetic locus resulting in this pathology requires further study. | Globozoospermia: Do Y-chromosome microdeletions play a role in this rare spermatogenic disorder? |
Recently hair transplantation has been widely applied not only to correct androgenetic alopecia, but also to correct hair loss on other parts of the body such as the eyebrows and pubic area. It is believed that the transplanted hairs will maintain their integrity and characteristics after transplantation to new nonscalp sites. To evaluate whether the transplanted hairs maintain their hair growth characteristics after transplantation to a new anatomic site other than the scalp. Three study designs were used. Study I: Hair transplantation from the author's occipital scalp to his lower leg was performed and clinical evaluations were made at both 6 months and at 3 years after the transplantation. Study II: After finding changes in hair growth characteristics, transplanted hairs were harvested from the leg and retransplanted to the left side of the nape of the neck (group A). As a control study, occipital hairs were transplanted to the opposite side (group B). Observations were made at 6 months after the operation. Study III: An observational study was done in 12 patients with androgenetic alopecia about 1 year after transplantation of occipital hair to frontal scalp. At each step, survival rates were documented and the rate of growth and the diameter of the shafts were measured for both recipient and donor sites. Study I: Surviving hairs on the lower leg showed a lower growth rate (8.2 +/- 0.9 mm/month), but the same diameter (0.086 +/- 0.018 mm) compared with occipital hairs (16.0 +/- 1.1 mm/month, 0.088 +/- 0.016 mm). The survival rate 3 years after transplantation was 60.2%. Study II: There was no significant difference in the growth rate, shaft diameter, and survival rate between retransplanted hairs (group A) and controls (group B). Groups A and B showed a lower growth rate, but the same diameter, compared with occipital hairs. Study III: There was no significant difference in the growth rate and shaft diameter between the transplanted hairs on the frontal scalp and the occipital hairs. | These results strongly suggest that the recipient site affects some characteristics of transplanted hairs, such as their growth and survival rates. | Does the recipient site influence the hair growth characteristics in hair transplantation? |
In 10-30% of women, vaginal birth results in levator ani tears ('avulsion') that are associated with pelvic floor dysfunction in later life. We hypothesised that women notice reduced pelvic floor muscle strength after childbirth, especially those with avulsion. This is a secondary analysis of two perinatal studies. At 3-6 months postpartum, women were asked to estimate pelvic floor muscle strength relative to antepartum strength. Translabial ultrasound was performed to determine pelvic floor structure and function. Five hundred and thirteen primiparous women were seen at a median of 129 days after delivery of a singleton at a mean gestation of 40 weeks. At follow-up, 481 were able to rate pelvic floor strength (mean 89%). This reduction was associated with delivery mode (P < 0.001), episiotomy (P = 0.01), perineal tears (P = 0.025) and avulsion (n = 45, P = 0.04). | After the birth of a first child, women notice a significant reduction in pelvic floor muscle strength, which is associated with delivery mode as well as perineal and pelvic floor muscle trauma. | Do women notice the effect of childbirth-related pelvic floor trauma? |
In this study, we looked for evidence that octreotide, a drug used specifically in acromegaly and other digestive pathologies, can have a radioprotective effect on salivary glands. This effect has already been proven on the pituitary gland, which is why we postulated that octreotide could act the same way on rat parotid glands. A prospective randomized controlled study on animals was conducted. With a noninvasive technique, we collected saliva from the parotid glands of 18 anesthetized rats at time 0 (preirradiation) and 1 month (postirradiation). Each sampling technique lasted 40 minutes, with pilocarpine injection at time 0 and 20 minutes. Saliva was collected bilaterally. Eighteen rats, nine in the saline group and nine in the octreotide group, were randomized. The substance was injected 30 minutes before irradiation. Thirty gray were given with the gamma knife on the left parotid gland of each rat following a computerized targeting method. Each gland was examined after the last saliva collection to determine the percentage of five criteria: fibrosis, ducts, fat, vessels, and acini. Data are available for 17 rats (nine in the octreotide group and eight in the saline group). Statistical analysis was done with a t-test (independent and paired). We noted that the postirradiation secretion in the left (radiated) gland was diminished compared with the right (nonradiated) gland in the saline group (p = .014). Fibrosis was increased in the irradiated (left) gland in both groups (p = .024 in the octreotide group and p = .033 in the saline group). The percentage of duct cells was more important in the left (radiated) gland of the octreotide group (p = .046). A trend appeared for a decrease in acinic cells only in the control group (p = .063). | Octreotide acted as a radioprotective agent on rat parotid glands 1 month after irradiation with 30 Gy given with the gamma knife. | Radiation-induced xerostomia: is octreotide the solution? |
Students tend to rate university courses more positively if they do well. Greenwald and Gillmore (1997a) suggested that it is not students' absolute grades that are important but rather how these grades compare to their expectations. However, this hypothesis is difficult to evaluate because few studies have measured grade expectations at the beginning of courses.AIM: By measuring students' grade expectations and enjoyment at several stages during a course, we hoped to evaluate the extent to which expectations modulate the impact of grades on course enjoyment. Participants were 242 students in a university course in psychology. Students were asked what grades they expected, and how much they were enjoying the course, at four stages. The effect of grades and grade expectations on enjoyment were analysed using restricted maximum likelihood (REML) and regression analyses. The best predictor of course enjoyment varied somewhat at different stages, but in general it was the extent to which students' grades surpassed their expectations. Students' expectations at the beginning of the course proved particularly influential. | Grade expectations do influence how students react to course grades, but the prominent role of pre-course expectations suggests that it may be important to distinguish between grade aspirations and grade expectations. It appears to be students' aspirations--the grades they hope to achieve--that most strongly shape their emotional reactions, rather than the more realistic expectations they may form later in a course. | The effects of grades on course enjoyment: did you get the grade you wanted? |
Because perioperative complications of unrecognized obstructive sleep apnea (OSA) can be severe, many bariatric surgery programs routinely screen all patients. However, many obese non-bariatric surgery patients do not get screened. We wanted to evaluate the need for routine preoperative OSA screening. Morbidly obese patients with a body mass index (BMI) > 40 kg/m(2) undergoing bariatric surgery--all screened for OSA--were compared to morbidly obese orthopedic lower extremity total joint replacements (TJR) patients--not screened for OSA. Cardio-pulmonary complications were recorded. Eight hundred eighty-two morbidly obese patients undergoing either bariatric (n = 467) or orthopedic TJR surgery (n = 415) were compared. As a result of screening, 119 bariatric surgery patients (25.5 %) were newly diagnosed with OSA, bringing the incidence to 42.8 % (200/467). Orthopedic surgery group had 72 of 415 (17.3 %) patients with pre-existing OSA. The unscreened orthopedic patients had a 6.7 % (23/343) cardiopulmonary complications rate compared to 2.6 % (7/267) for screened bariatric surgery patients. This difference was not statistically significant when adjusted for age and comorbidity (p = 0.3383). | Sleep apnea screening prior to bariatric surgery identifies an additional 25 % of patients as having OSA. In this study, unscreened morbidly obese patients did not have an increased incidence of cardiopulmonary complications after surgery compared to screened patients. Prospective randomized studies should be conducted to definitively assess utility and cost effectiveness of routine OSA screening of all morbidly obese patients undergoing surgery. Preoperative OSA screening may be safely omitted when randomizing patients for such a trial. | Sleep apnea: is routine preoperative screening necessary? |
Proper timing of stabilization for spine injuries is discussed controversially. Whereas early repair of long bone fractures is known to reduce complications. We investigated retrospectively 48 patients who were stabilized in a ventrodorsal approach for fractures of the thoracic spine. Patients were divided into three groups. All patients in groups I and II presented radiological or clinical signs of lung contusion. Patients were stabilized in the prone position via single-step dorsal stabilization with internal transpedicular fixation and ventral fusion with titanium cage or autologous bone graft using a minimally invasive video-assisted thoracotomy. The average duration of the procedures in group I was 213+/-40 min, in group II 250+/-75 min, and in group III 255+/-65 min (p: n.s.). Intraoperative blood loss did not differ significantly between the three groups. The PaO(2)/FiO(2) ratio improved in groups I and III, whereas in group II an significant impairment of lung function occurred perioperatively. Postoperative ICU stay was comparable in groups I and II (I: 10+/-5 days; II: 9+/-7 days); overall ICU stay tended to be shorter in group I versus II. The postoperative dependence on ventilator support did not differ significantly among the three groups. The mortality rate was 0% in this series. | Our data provide further evidence that early stabilization of combined thoracic and thoracic spine injuries is safe, does not alter perioperative lung function, and results in a reduced overall ICU stay. | Does timing of thoracic spine stabilization influence perioperative lung function after trauma? |
Using frameworks, such as the long-term conditions pyramid of healthcare, primary care organizations (PCOs) in England and Wales are exploring ways of developing services for people with long-term respiratory disease. We aimed to explore the current and planned respiratory services and the roles of people responsible for change. A purposive sample of 30 PCOs in England and Wales. Semi-structured telephone interviews with the person responsible for driving the reconfiguration of respiratory services. Recorded interviews were transcribed and coded, and themes identified. The association of the composition of the team driving change with the breadth of services provided was explored using a matrix. All but two of the PCOs described clinical services developed to address the needs of people with respiratory conditions, usually with a focus on preventing admissions for chronic obstructive pulmonary disease (COPD). Although the majority identified the need to develop a strategic approach to service development and to meet educational needs of primary care professionals, relatively few described clearly developed plans for addressing these issues. Involvement of clinicians from both primary and secondary care was associated with a broad multifaceted approach to service development. Teamwork was often challenging, but could prove rewarding for participants and could result in a fruitful alignment of objectives. The imminent merger of PCOs and overriding financial constraints resulted in a 'fluid' context which challenged successful implementation of plans. | While the majority of PCOs are developing clinical services for people with complex needs (principally in order to reduce admissions), relatively few are addressing the broader strategic issues and providing for local educational needs. The presence of multidisciplinary teams, which integrated primary and secondary care clinicians with PCO management, was associated with more comprehensive service provision addressing the needs of all respiratory patients. Future research needs to provide insight into the structures, processes and inter-professional relationships that facilitate development of clinical, educational and policy initiatives which aim to enhance local delivery of respiratory care. | Is multidisciplinary teamwork the key? |
It is common clinical practice to instill a topical anaesthetic prior to the instillation of a mydriatic agent into the eye. The main rationale for using the anaesthetic is to increase corneal permeability, so that more of the mydriatic agent reaches the receptor sites within the anterior chamber. It addition, as mydriatics generally cause stinging, prior use of an anaesthetic should reduce the degree of discomfort. The aim of the present study was to determine whether the efficacy of mydriasis produced by an antimuscarinic agent is enhanced by prior instillation of a topical anaesthetic. The study was performed using a double-masked protocol on 20 healthy young subjects. One drop of either proparacaine (proxymetacaine) (0.5%) or isotonic saline was instilled into the eye, followed by one drop of tropicamide (0.5%). Pupil diameter was measured using a customized photographic device at 0, 10, 20, 30 and 60 min following drug instillation. Additionally, subjects were asked to rate the degree of discomfort following the instillation of each drop on a scale from 0 (no discomfort) to 10 (agony). There was no significant difference in either the rate of onset of mydriasis, or the maximum pupil diameter achieved between the two conditions. The mean change in pupil diameter produced by tropicamide after the instillation of saline or proparacaine was 2.31 and 2.28 mm, respectively. The mean discomfort scores following instillation of saline and proparacaine were 1.15 and 1.65, respectively, while mean discomfort scores following the instillation of tropicamide after saline or proparacaine were 4.00 and 0.85, respectively. | Instillation of a topical anaesthetic does significantly reduce the degree of discomfort produced by the instillation of tropicamide. However, it does not produce any significant increase in either the magnitude or rate of onset of mydriasis. | Does prior instillation of a topical anaesthetic alter the pupillary mydriasis produced by tropicamide (0.5%)? |
The role of preoperative bilateral breast MRI in breast cancer patients being considered for breast-conserving therapy has been controversial. We hypothesized that preoperative MRI, along with an active program in MRI-directed biopsies, would lead to a change in multidisciplinary treatment planning for patients being considered for breast-conserving cancer therapy, and it would be associated with reduced rates of margin-positive partial mastectomies. A retrospective review of a consecutive series of patients who were treated for breast cancer at a single center between January 2005 and July 2007 was conducted. Patients in the study were candidates for breast-conserving cancer therapy based on physical examination, mammography, and ultrasonography. All patients were evaluated by a preoperative breast MRI. Analysis included number and result of MRI-directed biopsies, impact of MRI on treatment planning, and incidence of margin-positive partial mastectomy within the series of patients. Seventy-nine female patients were analyzed. Median age was 57 years. MRI led to the performance of 25 MRI-directed biopsies for previously unrecognized suspicious lesions in 21 patients. Forty-four percent of MRI-directed biopsies were positive for cancer. MRI was associated with a change in management in 15 patients (19%) for multicentric ipsilateral cancer (n = 7), a more extensive primary lesion size (n = 6), or contralateral breast cancer (n = 2). Incidence of margin-positive partial mastectomy requiring additional resective operation was very low in this series (10%). | Bilateral breast MRI, when used in conjunction with MRI-directed biopsy procedures, can be helpful in planning multidisciplinary treatment of candidates for breast-conserving cancer therapy. By allowing more accurate local staging of tumors, MRI is a tool that can be used to help reduce high reexcision rates for margin-positive partial mastectomies. | Is there a role for routine use of MRI in selection of patients for breast-conserving cancer therapy? |
Over two-thirds of UK medical schools are augmenting their selection procedures for medical students by using the United Kingdom Clinical Aptitude Test (UKCAT), which employs tests of cognitive and non-cognitive personal qualities, but clear evidence of the tests' predictive validity is lacking. This study explores whether academic performance and professional behaviours that are important in a health professional context can be predicted by these measures, when taken before or very early in the medical course. This prospective cohort study follows the progress of the entire student cohort who entered Hull York Medical School in September 2007, having taken the UKCAT cognitive tests in 2006 and the non-cognitive tests a year later. This paper reports on the students' first and second academic years of study. The main outcome measures were regular, repeated tutor assessment of individual students' interpersonal skills and professional behaviour, and annual examination performance in the three domains of recall and application of knowledge, evaluation of data, and communication and practical clinical skills. The relationships between non-cognitive test scores, cognitive test scores, tutor assessments and examination results were explored using the Pearson product-moment correlations for each group of data; the data for students obtaining the top and bottom 20% of the summative examination results were compared using Analysis of Variance. Personal qualities measured by non-cognitive tests showed a number of statistically significant relationships with ratings of behaviour made by tutors, with performance in each year's objective structured clinical examinations (OSCEs), and with themed written summative examination marks in each year. Cognitive ability scores were also significantly related to each year's examination results, but seldom to professional behaviours. The top 20% of examination achievers could be differentiated from the bottom 20% on both non-cognitive and cognitive measures. | This study shows numerous significant relationships between both cognitive and non-cognitive test scores, academic examination scores and indicators of professional behaviours in medical students. This suggests that measurement of non-cognitive personal qualities in applicants to medical school could make a useful contribution to selection and admission decisions. Further research is required in larger representative groups, and with more refined predictor measures and behavioural assessment methods, to establish beyond doubt the incremental validity of such measures over conventional cognitive assessments. | Can personal qualities of medical students predict in-course examination success and professional behaviour? |
To provide an assessment and comparison of the demographics, medical school academic performance, United States Medical Licensing Examination (USMLE) performance, and research experience between American Medical Graduate (AMG) and United States International Medical Graduate (USIMG) candidates who applied for and successfully matched into categorical general surgery residency programs. Data were obtained through the Electronic Residency Application Service (ERAS) and a post-match survey distributed to all applicants. The study was conducted at a community-based, university-affiliated hospital. All United States citizen graduates of allopathic American medical schools or international medical schools, who were applying for a general surgery residency position at our institution. A total of 854 candidates applied, including 143 AMGs and 223 USIMGs. Seventy-two AMGs (50.3%) and 41 USIMGs (18.4%) were invited to interview (p<0.0001). Mean USMLE step 1 scores were higher among USIMG applicants overall (USIMG: 212.1 ± 14.9 vs AMG: 206.9 ± 15.5; p<0.0005) and among those invited to interview (USIMG: 227.8 ± 16.2 vs AMG: 215.5 ± 16.2; p<0.0001). Seventy percent of AMGs matched into a categorical surgery residency compared with 31.6% of USIMGs (p<0.001). Compared with AMGs, USIMGs applied to more programs (USIMG: 90.3 ± 42.8 vs AMG: 52.1 ± 26.4; p<0.002), were offered fewer interviews (USIMG: 9.0 ± 6.9 vs AMG: 20.9 ± 13.7; p<0.0001), and subsequently ranked fewer programs (USIMG: 7.5 ± 4.5 vs AMG: 12.5 ± 6.1; p<0.0008). | USIMGs require higher USMLE scores than their AMG counterparts to be considered for categorical general surgery residency positions. However, excellence on the USMLE neither ensures an invitation to interview nor categorical match success. A well-rounded application in conjunction with a practical application strategy is critical for USIMGs to achieve success in attaining a general surgery residency position. | An ERAS-based survey evaluating demographics, United States Medical Licensing Examination Performance, and research experience between American medical graduates and United States citizen international medical graduates: is the bar higher on the continent? |
In a multicentre, longitudinal cohort study, 142 children with Type 1 diabetes completed a scale that assessed adherence to treatment and a test of diabetes knowledge at the beginning of the study (T0) and four years later (T4). HbA1c and clinical data were collected at T0 and at T4. From T0 to T4, the mean HbA1c increased from 8.2 +/- 1.6 to 9.1 +/- 1.4% (P<0.001). Among patients at pubertal stages 1-4, adherence did not decline from T0 to T4, whereas the HbA1c level increased and a positive correlation between adherence and the knowledge score was noted (r = 0.32; P<0.02). Among adolescents at pubertal stage 5, the level of adherence decreased (P<0.01) from T0 to T4 and the HbA1c level increased despite an increase in the knowledge score (P<0.001), a negative correlation between HbA1c and adherence was found (r = -0.37; P = 0.001) and adherence at T4 significantly added to the prediction of HbA1c at T4. | In this longitudinal study, an initial worsening of glycaemic control as a result of puberty preceded worsening of adherence behaviours. Low levels of adherence become predictive of HbA1c degradation among pubertal stage 5 adolescents. These results suggest a potential bi-directional relationship between glycaemic control and adherence. | Is the relationship between adherence behaviours and glycaemic control bi-directional at adolescence? |
To test the hypothesis that the prevalence of specific musculoskeletal pain symptoms has increased over time in the northwest region of England. To meet this objective we have examined the difference in the prevalence of low back, shoulder and widespread pain between the 1950s and today using historical data collected by the Arthritis Research Campaign (arc). Two cross-sectional surveys conducted over 40 yr apart in the northwest region of England. The status of two regional pain sites and widespread pain was determined using interview and questionnaire responses, for the earlier and later studies respectively. Subjects were classified positively if they reported low back pain, shoulder pain or widespread pain on the day of the survey. Rates were standardized to the Greater Manchester population. There were large differences in the prevalence of musculoskeletal pain between the two surveys. For all three symptoms examined prevalence increased from 2- to 4-fold between the two surveys. In both surveys low back pain was more common in women. Shoulder and widespread pain was less prevalent in women than in men in the earlier survey but by the time of the later survey women reported more pain at these sites. | The prevalence of musculoskeletal pain is much higher than that reported over 40 yr ago. The change in prevalence is unlikely to be entirely due to the study design; other possible explanations such as the increased reporting or awareness of these symptoms is discussed. | Is musculoskeletal pain more common now than 40 years ago? |
To examine the concurrent and predictive associations between the number of steps taken per day and clinical outcomes in patients with fibromyalgia (FM). A total of 199 adults with FM (mean age 46.1 years, 95% women) who were enrolled in a randomized clinical trial wore a hip-mounted accelerometer for 1 week and completed self-report measures of physical function (Fibromyalgia Impact Questionnaire-Physical Impairment [FIQ-PI], Short Form 36 [SF-36]health survey physical component score [PCS], pain intensity and interference (Brief Pain Inventory [BPI]), and depressive symptoms (Patient Health Questionnaire-8 [PHQ-8]) as part of their baseline and followup assessments. Associations of steps per day with self-report clinical measures were evaluated from baseline to week 12 using multivariate regression models adjusted for demographic and baseline covariates. Study participants were primarily sedentary, averaging 4,019 ± 1,530 steps per day. Our findings demonstrate a linear relationship between the change in steps per day and improvement in health outcomes for FM. Incremental increases on the order of 1,000 steps per day were significantly associated with (and predictive of) improvements in FIQ-PI, SF-36 PCS, BPI pain interference, and PHQ-8 (all P<0.05). Although higher step counts were associated with lower FIQ and BPI pain intensity scores, these were not statistically significant. | Step count is an easily obtained and understood objective measure of daily physical activity. An exercise prescription that includes recommendations to gradually accumulate at least 5,000 additional steps per day may result in clinically significant improvements in outcomes relevant to patients with FM. Future studies are needed to elucidate the dose-response relationship between steps per day and patient outcomes in FM. | Does increasing steps per day predict improvement in physical function and pain interference in adults with fibromyalgia? |
The outcomes of 106 total knee arthroplasties implanted using a soft tissue balancing surgical technique at one surgical centre were used to assess the accuracy maintaining the knee's original joint line (JL). The aim of the study was to determine whether there is a shift of the presumed joint line after surgery. Preoperative and post-operative radiographs were compared to determine any changes in the articulation height. The preoperative distance of the fibular head to the natural joint line was measured and compared with the post-operative measurement of the fibular head to the femoral articulation line (measured on the radiograph and defined as Rxmm). Based on the actual medio-lateral dimension of the tibial metal back, the measured difference (RXmm) could be converted into true distances (in mm). The Blackburn-Peel index was assessed as an additional outcome prior to and following surgery. Preoperatively, the average distance from the fibular head to the joint line was 15.1 Rxmm (SD 4.3) while the post-surgical distance was 15.5 Rxmm (SD 5.6). The average deviation of the post-surgical JL in relation to the original JL amounted to 0.4 Rxmm (SD 3.7). The average deviation of the joint line converted into the true distance was -0.3 mm (with a range of -5.9 mm in distal direction to + 8.3 mm in the proximal direction). Valgus position appeared to generate rather a shift in proximal direction whereas varus deformity favours a shift in distal direction. Seven patients exhibited a deviation of more than 5 mm in either the distal or proximal direction. All of the patients of this subgroup had a preoperative anatomical abnormality including a severe malalignment, serious bone destruction or had previously undergone a high tibial osteotomy. | An exact reconstruction of the natural Joint Line is achievable when using the described soft tissue balancing surgical technique with the posterior cruciate ligament (PCL) retaining prosthesis design used in this series. | Does total knee joint replacement with the soft tissue balancing surgical technique maintain the natural joint line? |
Examination of the extent of offenders' engagement in change, and in rehabilitation programmes, is important to understanding success or failure following rehabilitation. In treatment programmes, the alliance between therapist and offender, and the therapy process itself appear central to progress offenders make that may reduce their criminal risk. But research with offenders seldom has measured therapeutic alliance and clinical writing suggests that it is difficult to form an alliance with those not ready to change their behaviour; especially with higher risk and psychopathic offenders. This study outlines the course of the therapeutic alliance in an 8-month treatment programme for high-risk, PCL-psychopathic violent prisoners. It examines relationships between early-treatment therapeutic alliance, therapists' global ratings of motivation to change, and initial stage of change on dynamic risk factors. In addition, it investigates which factors best predict who will complete treatment and change behaviourally during treatment. | In this challenging, high-needs client group, early-programme stage of change, therapists' perceptions of motivation, therapeutic alliance and psychopathy did not predict how much change prisoners made. Regardless of initial levels, prisoners whose alliance increased the most over the course of treatment made the most change. | Do early therapeutic alliance, motivation, and stages of change predict therapy change for high-risk, psychopathic violent prisoners? |
Anal incontinence affects approximately 10 percent of adult females. Damage to the anal sphincters has been considered as the cause of anal incontinence after childbirth in the sole prospective study so far available. The aims of the present study were to determine prospectively the incidence of anal incontinence and anal sphincter damage after childbirth and their relationship with obstetric parameters. We studied 259 consecutive females six weeks before and eight weeks after delivery. They were asked to fill in a questionnaire assessing fecal incontinence. Anal endosonography (7-10 MHz) was then performed. Two independent observers analyzed internal and external anal sphincters. A total of 233 patients (90 percent) were assessed, of whom 31 had cesarean section. De novo sphincter defects were observed in 16.7 percent (14 percent external, 1.7 percent internal, and 1 percent both) in the postpartum period only after vaginal delivery. These disruptions occurred with the same incidence after the first and the second childbirth. Independent risk factors (odds ratio; 95 percent confidence interval) for sphincter defect were forceps (12; 4-20), perineal tears (16; 9-25), episiotomy (6.6; 5-17), and parity (8.8; 4-19) as revealed by multivariate analyses. The overall rate of anal incontinence was 9 percent and independent risk factors (odds ratio; 95 percent confidence interval) involved forceps (4.5; 1.5-13), perineal tears (3.9; 1.4-10.9), sphincter defect (5.5; 5-15), and prolonged labor (3.4; 1-11). Among these patients only 45 percent had sphincter defects. | Anal incontinence after delivery is multifactorial, and anal sphincter defects account for only 45 percent of them. Primiparous and secundiparous patients have the same risk factors for sphincter disruption and anal incontinence. Because external anal sphincter disruptions are more frequent than internal anal sphincter damage, surgical repair should be discussed in symptomatic patients. | Are sphincter defects the cause of anal incontinence after vaginal delivery? |
The purpose of this study was to evaluate physician manpower and mobilization in an urban emergency department receiving patients after a major earthquake. Patient charts were reviewed. The workload of physicians was assessed semiquantitatively before and after a major earthquake. The physicians' mobilization in the postearthquake emergency response was assessed by using a confidential questionnaire. In the 3 days after the earthquake, 566 patients with earthquake-related illnesses or injuries were sent to the urban ED. Three hundred one (53.2%) patients arrived within the initial 10 hours. In the initial hours, there was no significant difference between the number of patients per physician per hour before and after the earthquake. Workloads of wound treatment and advanced life support procedures were significantly higher after the earthquake compared with before the earthquake, during the first to sixth hour and second to fifth hour, respectively. Sixty-five percent of the hospital's physicians did not assist in either the ED or in any other parts of the hospital in the initial 6 hours after the earthquake. | The number of physicians in the ED was insufficient in the initial hours after the earthquake because of the sudden influx of a large number of patients. Future disaster planning must address the issue of physicians' behavior with regard to their priorities immediately after a major earthquake and include greater provision for efficient mobilization of physicians. | Were there enough physicians in an emergency department in the affected area after a major earthquake? |
Multiple Endocrine Neoplasia Type 1 (MEN1) is an autosomal dominant inherited syndrome, related to mutations in the MEN1 gene. Controversial data suggest that the nonsynonymous p.Ala541Thr variant, usually considered as a non-pathogenic polymorphism, may be associated with an increased risk of MEN1-related lesions in carriers. The aim of this study was to evaluate the pathogenic influence of the p.Ala541Thr variant on clinical and functional outcomes. We analysed a series of 55 index patients carrying the p.Ala541Thr variant. Their clinical profile was compared to that of 117 MEN1 patients. The biological impact of the p.Ala541Thr variant on cell growth was additionally investigated on menin-deficient Leydig cell tumour (LCT)10 cells generated from Men1+/Men1- heterozygous knock-out mice, and compared with wild type (WT). The mean age at first appearance of endocrine lesions was similar in both p.Ala541Thr carriers and MEN1 patients, but no p.Ala541Thr patient had more than one cardinal MEN1 lesion at initial diagnosis. A second MEN1 lesion was diagnosed in 13% of MEN1 patients and in 7% of p.Ala541Thr carriers in the year following preliminary diagnosis. Functional studies on LCT10 cells showed that overexpression of the p.Ala541Thr variant did not inhibit cell growth, which is in direct contrast to results obtained from investigation of WT menin protein. | Taken together, these data raise the question of a potential pathogenicity of the p.Ala541Thr missense variant of menin that commonly occurs within the general population. Additional studies are required to investigate whether it may be involved in a low-penetrance MEN1 phenotype. | p.Ala541Thr variant of MEN1 gene: a non deleterious polymorphism or a pathogenic mutation? |
The global gene expression in Barrett's esophagus (BE) in comparison to adjacent or histologically similar tissues has not been extensively studied. To test the feasibility of conducting gene arrays in endoscopically obtained mucosal specimens. Cross-sectional feasibility study. The Houston Department of Veterans Affairs Medical Center. We collected endoscopic biopsies from BE, normal esophagus, antrum, duodenum, and sigmoid colon from 5 patients with BE. RNA was extracted and subjected to cDNA microarrays and gene expression was compared between BE and control tissues. Reverse transcription-PCR was conducted to confirm some of the findings. Gene expression profiles in BE tissues and 4 control sites: squamous esophagus, antrum, duodenum, sigmoid colon. On average, 2 biopsies by disposable jumbo biopsy forceps provided approximately 5 microg required for microarrays. From the original number of 22,283 gene probes, 13,805 genes had a quality score of P<.05 and were subjected to further comparison. BE gene expression clustered most closely with that of antrum and least closely with squamous esophagus. Of the 587 genes that had significantly different expression between BE and duodenum, 246 were upregulated and 341 were downregulated in BE. The expression of genes involved in apoptosis, negative regulation of apoptosis, and inflammatory response was significantly lower in BE compared to squamous esophagus. None of the gene groups were significantly overexpressed in BE compared to squamous esophagus or antrum. The reverse transcription-PCR confirmed the results of microarrays. Small sample size. | Microarray-based studies are feasible in endoscopically obtained tissues. Differences in gene expression could identify potential markers and shed light on the pathogenesis of BE. | Is genomic evaluation feasible in endoscopic studies of Barrett's esophagus? |
Cultured neonatal rat myocytes were exposed to 50 nm-charged polystyrene latex nanoparticles and examined using a combination of hopping probe scanning ion conductance microscopy, optical recording of action potential characteristics and patch clamp. Positively charged, amine-modified polystyrene latex nanoparticles showed cytotoxic effects and induced large-scale damage to cardiomyocyte membranes leading to calcium alternans and cell death. By contrast, negatively charged, carboxyl-modified polystyrene latex nanoparticles (NegNPs) were not overtly cytotoxic but triggered formation of 50-250-nm nanopores in the membrane. Cells exposed to NegNPs revealed pro-arrhythmic events, such as delayed afterdepolarizations, reduction in conduction velocity and pathological increment of action potential duration together with an increase in ionic current throughout the membrane, carried by the nanopores. | The utilization of charged nanoparticles is a novel concept for targeting cardiac excitability. However, this unique nanoscopic investigation reveals an altered electrophysiological substrate, which sensitized the heart cells towards arrhythmias. | Functional interaction between charged nanoparticles and cardiac tissue: a new paradigm for cardiac arrhythmia? |
The purpose of this study was to determine whether the orthodontic treatment provided by pediatric dentists reflects the orthodontic training received in pediatric dental residency programs. Five questions from a survey of the American Academy of Pediatric Dentistry (AAPD) diplomates in August 2002 and a survey of pediatric dental residency program directors in June 2002 were statistically analyzed to compare the orthodontic treatment provided by diplomates to that provided within pediatric dental residency programs. Patient populations differed financially between pediatric dental residencies and diplomates of the AAPD. Residents treated significantly more public assistance patients. The residents were more likely than diplomates to use most orthodontic appliances and treat most stages of dental development and most conditions/malocclusions with orthodontics. Diplomates anticipated a decrease in the amount of orthodontic treatment provided in the next 5 years, while program directors anticipated an increase. | The majority of the orthodontic treatment provided by pediatric dental residents and diplomates was similar, although the residents were exposed to more diverse orthodontic treatment modalities than those used by diplomates. The residencies were also more likely than the diplomates to increase the amount of orthodontic treatment provided in the next 5 years. | Do pediatric dentists practice the orthodontics they are taught? |
Recent events in our hospital, combined with international recommendations, catalyzed the need to move from mercury sphygmomanometry to automated blood pressure (BP) recording in pregnancy.AIM: To test the accuracy of the Omron T9P automated BP recorder in pregnant women, using mercury sphygmomanometry as the gold standard. Antenatal clinics and obstetric day assessment unit, St George Hospital, Sydney. Eighty-five pregnant women, 11% of whom were receiving antihypertensives. Differences in both systolic and diastolic BP between the T9P Omron device and mercury sphygmomanometer were obtained for each woman, using sequential automated and mercury BP recordings, as required by a modified British Hypertension Society (BHS) protocol. The accuracy of the device was graded according to the BHS and the Association for the Advancement of Medical Instrumentation (AAMI) standards. The Omron T9P device received an A/A grade according to this modified BHS and AAMI testing process, though the range of the 255 differences was 1-13 for systolic BP and 1-10 for diastolic BP. | The Omron T9P device is an accurate device for use predominantly in an outpatient antenatal clinical setting. Further studies are required solely within hypertensive pregnant women before its use can be recommended with certainty in this group. | Can we use the Omron T9P automated blood pressure monitor in pregnancy? |
Blood hemoglobin (Hb) can be continuously monitored utilizing noninvasive spectrophotometric finger sensors (Masimo SpHb). SpHb is not a consistently accurate guide to transfusion decisions when compared with laboratory Co-Oximetry (tHb). We evaluated whether a finger digital nerve block (DNB) would increase perfusion and, thereby, improve the accuracy of SpHb. Twenty adult patients undergoing spinal surgery received a DNB with lidocaine to the finger used for the monitoring of SpHb. SpHb-tHb differences were determined immediately following the DNB and approximately every hour thereafter. These differences were compared with those in our previously reported patients (N = 20) with no DNB. The SpHb-tHb difference was defined as "very accurate" if <0.5 g/dL and "inaccurate" if >2.0 g/dL. Perfusion index (PI) values at the time of each SpHb-tHb measurement were compared. There were 57 and 78 data points in this and our previous study, respectively. The presence of a DNB resulted in 37 % of measurements having SpHb values in the "very accurate group" versus 12 % in patients without a DNB. When the PI value was >2.0, only 1 of 57 DNB values was in the "inaccurate" group. The PI values were both higher and less variable in the patients who received a DNB. | A DNB significantly increased the number of "very accurate" SpHb values and decreased the number of "inaccurate" values. We conclude that a DNB may facilitate the use of SpHb as a guide to transfusion decisions, particularly when the PI is >2.0. | Does a digital regional nerve block improve the accuracy of noninvasive hemoglobin monitoring? |
Water supplied to the dental units must be of sufficient quality. The article presents the results of the microbiological analysis of cold municipal water which flows into a patient's disposable mouthwash cup, and demineralized water which flows through a waterline into the tool panel of a dental unit from the tank placed in the water group. In order to assess the degree of purity (impurities) of water used in dental units, 2 series of microbiological tests were carried out in 6 dental surgeries from April to June, 2013. The water samples for microbiological testing were collected into sterile microbiological bottles in accordance with the current methodology. The water for the tests was collected from a sterile cup-filling tap (municipal water) and from an air/water syringe (demineralized water). The bacteria were cultured according to the Polish Standards - PN-EN ISO 6222, PN-EN ISO 9308-1, and PN-EN ISO 16266. In the tested samples of water numerous psychrophilic bacteria (max 29 100 CFU/ml) and mesophilic bacteria (max 24 700 CFU/ml), including single coliforms, were found. | The results show that water delivered to a dental unit should be periodically tested bacteriologically and in terms of physical and chemical properties. Water systems of dental units should also be periodically disinfected to eliminate bacteria and biofilm. | Is water in dental units microbiologically safe? |
In 1996, the World Health Organization stated that 'childbirth is a natural process and in normal birth, there should be a valid reason to interfere with this natural process' and encouraged practices that are evidence-based. The practices encouraged included avoiding unnecessary augmentation of labour, facilitating upright position for birth and restricting the use of routine episiotomy. Many countries have been slow to fully implement evidence-based practice in maternity care. The aim of this study was to examine maternity hospital practices in Jordan and assess their consistency with evidence-based maternity care. An explorative research design with non-participant observation was used. Data were collected from low-risk women during labour and birth using a questionnaire for maternal characteristics and an observational checklist. A proportional stratified sample was selected to recruit from three major public hospitals in Jordan. Data were analysed using descriptive statistics. A total of 460 women were observed during labour and birth. The majority were multiparous (80%). A range of interventions were observed in women having a normal labour including augmentation (95%), continuous external fetal monitoring (77%), lithotomy position for birth (100%), and more than one third (37%) had an episiotomy with varying degrees of laceration (58%). | Childbirth practices were largely not in accordance with the World Health Organization evidence-based practices for normal birth. High levels of interventions were observed, many of which may not have been necessary in this low-risk population. Further work needs to occur to explore the reasons why evidence-based practice is not implemented in these hospitals. | Childbirth practices in Jordanian public hospitals: consistency with evidence-based maternity care? |
To find predictors of a difficult intubation in infants with an isolated or a syndromic cleft lip/palate. Retrospective review: single-blind trial. Tertiary care centre. A total of 145 infants born with cleft lip/palate were enrolled. Three clinical and seven lip/palate anatomic parameters were evaluated. The grade of intubation was determined by the anesthesiologist at the time of the labioplasty/staphylorrhaphy surgery at 3 and 10 months, respectively. Intubation grade. The relative risk of a difficult intubation in the cleft lip, cleft palate without the Pierre Robin sequence, cleft lip-palate, and cleft palate with Pierre Robin sequence groups was 0, 2.7, 10, and 23%, respectively. The infants born with the Pierre Robin sequence had a statistically significant higher intubation grade. The degree of difficulty was increased in cases with early airway and feeding problems (p<.0001). Within the group of cleft palate patients without any lip malformation, a wider cleft was associated with a higher intubation grade with statistical significance (p = .0323). | Infants born with Pierre Robin sequence have a statistically significantly higher risk of difficult intubation. Within this group, of all the studied factors, a clinical history of early airway and feeding problems was the best predictor of a difficult endotracheal intubation. In cleft palate patients without any cleft lip, larger width of the cleft is also a significant predictor. | Can we predict a difficult intubation in cleft lip/palate patients? |
The autonomic nervous system exerts important effects upon atrial fibrillation (AF) initiation. The strategy of anesthesia used during AF ablation may impact the provocation of AF triggers. We hypothesized that the use of general anesthesia (GA) would reduce the incidence of provokable AF triggers in patients undergoing AF ablation compared to patients studied while receiving only conscious sedation (CS). We performed a prospective, case control study comparing the incidence of provokable AF triggers in a consecutive series of patients undergoing AF ablation under GA using a standard trigger induction protocol. We compared the frequency and distribution of AF triggers to a second cohort of historical controls (matched for age, gender, left atrial dimension, and AF phenotype) who underwent ablation while receiving CS. We calculated that 44 total subjects (22 patients in each group) were required to detect a 50% reduction in the incidence of AF triggers in the GA cohort. There was no difference between the 2 groups in the rate of AF trigger inducibility (77% vs. 68%, P = 0.26) or the number of triggers provoked per patient (1.2 ± 0.8 vs. 1.3 ± 0.8, P = 0.38). Patients ablated under GA required higher doses of phenylephrine during the trigger induction protocol (408.3 mg [52-600] vs. 158.3 mg [0-75]; P = 0.003), and tended to require higher doses of isoproterenol to initiate triggers (92.8 mg [20-111] vs. 63.6 mg [6-103]; P = 0.25). | AF trigger induction during GA is both safe and efficacious. | Provocation of atrial fibrillation triggers during ablation: does the use of general anesthesia affect inducibility? |
Providing patients with instructions and equipment regarding self-removal of nonabsorbable sutures could represent a new efficiency in emergency department (ED) practice. The primary outcome was to compare the proportion of patients successfully removing their own sutures when provided with suture removal instructions and equipment versus the standard advice and follow-up care. Secondary outcomes included complication rates, number of physician visits, and patient comfort level. This prospective, controlled, single-blinded, pseudorandomized trial enrolled consecutive ED patients who met the eligibility criteria (age>19 years, simple lacerations, nonabsorbable sutures, immunocompetent). The study group was provided with wound care instructions, a suture removal kit, and instructions regarding suture self-removal. The control group received wound care instructions alone. Outcomes were assessed by telephone contact at least 14 days after suturing using a standardized questionnaire. Overall, 183 patients were enrolled (93 in the intervention group; 90 in the control group). Significantly more patients performed suture self-removal in the intervention group (91.5%; 95% CI 85.4-97.5) compared to the control group (62.8%; 95% CI 52.1-73.6) (p<0.001). Patients visited their physician less often in the intervention group (9.8%; 95% CI 3.3-16.2) compared to the control group (34.6%; 95% CI 24.1-45.2%) (p<0.001). Complication rates were similar in both groups. | Most patients are willing to remove, and capable of removing, their own sutures. Providing appropriate suture removal instructions and equipment to patients with simple lacerations in the ED appears to be both safe and acceptable. | Are patients willing to remove, and capable of removing, their own nonabsorbable sutures? |
We studied 67 cases of myocardial infarction, terminating with left ventricular rupture, between January 1988 and December 1996. The study was restricted to sudden death where, at coroner-directed autopsy, a ruptured myocardial infarction was determined as the cause of death. It was also restricted to patients who consulted a doctor within the two weeks prior to death. The report made to the coroner by attending police and the autopsy report was studied, and the requisite data were abstracted. Half of our study group did not present with chest pain. Of the atypical presentations: 15/67 cases (22%) were from referred pain (neck, arm, abdomen or back), 12/ 67 patients presented with "flu-like illness" (18%), 4/67 cases had respiratory presentations (cough or shortness of breath) (6%) and 2/67 falls (3%). Of those with chest pain, 16/34 (47%) were diagnosed or referred and 2/15 infarcts with atypical or referred pain were diagnosed. None of those presenting with "flu like illness" or respiratory symptoms was diagnosed or referred. | Fifty per cent of our patients had "silent" myocardial infarcts. A large proportion of this group complained of a flu-like illness, which is currently not considered a presentation of this disease. Patients at higher risk of a myocardial infarct, should be treated with a high index of suspicion when unwell, especially when complaining of a flu-like illness. Pathologically, posterior and lateral infarcts accounted for over half the cases. | Death due to unrecognised myocardial infarction causing left ventricular rupture: can we improve the diagnostic rate? |
In the draft proposal for DSM-5, the Work Group for Personality and Personality Disorders recommended that dimensional ratings of personality disorders replace DSM-IV's categorical approach toward classification. If a dimensional rating of personality disorder pathology is to be adopted, then the clinical significance of minimal levels of pathology should be established before they are formally incorporated into the diagnostic system because of the potential unforeseen consequences of such ratings. In the present report from the Rhode Island Methods to Improve Diagnostic Assessment and Services (MIDAS) project, we examined the low end of the severity dimension and compared psychiatric outpatients with 0 or 1 DSM-IV criterion for borderline personality disorder on various indices of psychosocial morbidity. Three thousand two hundred psychiatric outpatients were evaluated with semistructured diagnostic interviews for DSM-IV Axis I and Axis II disorders. The present report is based on the 1,976 patients meeting 0 or 1 DSM-IV criterion for borderline personality disorder. The reliability of determining if a patient was rated with 0 or 1 criterion for borderline personality disorder was good (κ = 0.70). Compared to patients with 0 borderline personality disorder criteria, patients with 1 criterion had significantly more current Axis I disorders (P<.001), suicide attempts (P<.01), suicidal ideation at the time of the evaluation (P<.001), psychiatric hospitalizations (P<.001), and time missed from work due to psychiatric illness (P<.001) and lower ratings on the Global Assessment of Functioning (P<.001). | Low-severity levels of borderline personality disorder pathology, defined as the presence of 1 criterion, can be determined reliably and have validity. | Does the presence of one feature of borderline personality disorder have clinical significance? |
Stage of Change constructs may be proxy markers of psychosocial health which, in turn, are related to oral health. To determine if Stage of Change constructs were associated with subjective oral health in a population at heightened risk of dental disease. Stage of Change constructs were developed from a validated 18-item scale and categorised into 'Pre-contemplative', 'Contemplative' and 'Active'. A convenience sample of 446 Australian non-Aboriginal women pregnant by an Aboriginal male (age range 14-43 years) provided data to evaluate the outcome variables (self-rated oral health and oral health impairment), the Stage of Change constructs and socio-demographic, behavioural and access-related factors. Factors significant at the p<0.05 level in bivariate analysis were entered into prevalence regression models. Approximately 54% of participants had fair/poor self-rated oral health and 34% had oral health impairment. Around 12% were 'Pre-contemplative', 46% 'Contemplative' and 42% 'Active'. Being either 'pre-contemplative' or 'contemplative' was associated with poor self-rated oral health after adjusting for socio-demographic factors. 'Pre-contemplative' ceased being significant after adjusting for dentate status and dental behaviour. 'Pre-contemplative' remained significant when adjusting for dental cost, but not 'Contemplative'. The Stages of Change constructs ceased being associated with self-rated oral health after adjusting for all confounders. Only 'Contemplative' (reference: 'Active') was a risk indicator in the null model for oral health impairment which persisted after adding dentate status, dental behaviour and dental cost variables, but not socio-demographics. When adjusting for all confounders, 'Contemplative' was not a risk indicator for oral health impairment. | Both the 'Pre-contemplative' and 'Contemplative' Stage of Change constructs were associated with poor self-rated oral health and oral health impairment after adjusting for some, but not all, covariates. When considered as a proxy marker of psychosocial health, Stage of Change constructs may have some relevance for subjective oral health. | Are Stage of Change constructs relevant for subjective oral health in a vulnerable population? |
The aim of this prospective study is to detect the efficiency of a previously described lymph node revealing solution (LNRS) in comparing with conventional lymph node dissection and re-dissection in colon, breast and urinary bladder carcinomas. Total 30 cases in which less than 10 lymph nodes were found by conventional method, were immersed for 6-8 hours in LNRS, dissected and processed. Control group, 12 cases, was first dissected then a second conventional dissection was performed. At the end, specimens were again immersed in LNRS for 6-8 hours and last dissections were done. In the first group, a total of 150 lymph nodes, 46 of them with metastasis were detected. After using LNRS, 26 additional lymph nodes among which 10 were positive were detected. When compared with the results of conventional dissection, the increase in number of total and metastatic lymph nodes with LNRS was significant (p<0.01). The mean size of the lymph nodes detected by the conventional and LNRS methods was 6.8 and 4.2 mm, respectively. The pathologic lymph node stage was changed in three bladder carcinoma cases, and one breast carcinoma. In the control group, 75 lymph nodes (11 with metastases), 19 lymph nodes (3 with metastases), 14 lymph nodes (one with metastases) were detected after first and second conventional dissection and LNRS methods, respectively. | LNRS enhanced the number of total and metastatic lymph nodes and is effective in detecting small lymph nodes. This method is useful for accurate staging where the number of detected lymph nodes is too small by the conventional method. | Lymph node revealing solution: is it effective on detecting minute lymph nodes? |
Work interruptions during patient care have been correlated with error. Task-switching is identified by the Accreditation Council for Graduate Medical Education (ACGME) as a core competency for emergency medicine (EM). Simulation has been suggested as a means of assessing EM core competencies. We assumed that senior EM residents had better task-switching abilities than junior EM residents. We hypothesized that this difference could be measured by observing the execution of patient care tasks in the simulation environment when a patient with a ST-elevation myocardial infarction (STEMI) interrupted the ongoing management of a septic shock case. This was a multi-site, prospective, observational, cohort study. The study population consisted of a convenience sample of EM residents in their first three years of training. Each subject performed a standardized simulated encounter by evaluating and treating a patient in septic shock. At a predetermined point in every sepsis case, the subject was given a STEMI electrocardiogram (ECG) for a separate chest pain patient in triage and required to verbalize an interpretation and action. We scored learner performance using a dichotomous checklist of critical actions covering sepsis care, ECG interpretation and triaging of the STEMI patient. Ninety-one subjects participated (30 postgraduate year [PGY]1s, 32 PGY2s, and 29 PGY3s). Of those, 87 properly managed the patient with septic shock (90.0% PGY1s, 100% PGY2, 96.6% PGY 3s; p=0.22). Of the 87 who successfully managed the septic shock, 80 correctly identified STEMI on the simulated STEMI patient (86.7% PGY1s, 96.9% PGY2s, 93.1% PGY3s; p=0.35). Of the 80 who successfully managed the septic shock patient and correctly identified the STEMI, 79 provided appropriate interventions for the STEMI patient (73.3% PGY1s, 93.8% PGY2s, 93.8% PGY3s; p=0.07). | When management of a septic shock patient was interrupted with a STEMI ECG in a simulated environment we were unable to measure a significant difference in the ability of EM residents to successfully task-switch when compared across PGY levels of training. This study may help refine the use of simulation to assess EM resident competencies. | Can Simulation Measure Differences in Task-Switching Ability Between Junior and Senior Emergency Medicine Residents? |
Mitral valve replacement is more frequently performed and perceived to be equivalent to repair in elderly patients, despite the superiority of repair in younger patients. Our objective was to compare mitral repair to replacement in elderly patients age 75 years or older. Patients younger than 75 years undergoing mitral valve surgery served as a reference population. Consecutive elderly patients undergoing operation for mitral regurgitation at our institution from 1998 to 2006 were reviewed. Elderly patients (mean age, 78.0 +/- 2.8 years) who underwent mitral repair (n = 70) or replacement (n = 47) were compared with cohorts of young patients (mean age, 58.9 +/- 9.3 years) who underwent repair (n = 100) or replacement (n = 98) during the same period. Patient details and outcomes were compared using univariate, multivariate, and Kaplan-Meier analyses. Mitral replacement in elderly patients had higher mortality than repair (23.4%, 11 of 47 versus 7.1%, 5 of 70; p = 0.01) or as compared with either operation in the reference group (p<0.0001). Postoperative stroke was higher in elderly replacement patients compared with repair (12.8%, 6 of 47 versus 0%; p = 0.003) or compared with either young cohort (p = 0.02). Compared with elderly repair patients, elderly replacement patients had more cerebrovascular disease (21.3%, 10 of 47 versus 4.3%, 3 of 70; p = 0.005) and rheumatic mitral valves (21.3%, 10 of 47 versus 0%; p = 0.0001). In the young group, overall complication and mortality were no different between replacement and repair. Long-term survival favored repair over replacement in elderly patients (p = 0.04). One elderly repair patient experienced late recurrence of persistent mitral regurgitation. | In patients age 75 years or older, mitral repair is associated with a lower risk of mortality, postoperative stroke, and prolonged intensive care unit and hospital stay compared with mitral replacement. Mitral repair can be performed in preference over replacement even in patients older than the age of 75. | Is mitral valve repair superior to replacement in elderly patients? |
The effectiveness of stereotactic radiosurgery (SRS) for cavernous malformation (CM) has not been fully assessed. Consequently, observation is usually recommended when a bleeding CM is initially discovered. Recurrent bleeding occurs with CMs, and these repeat hemorrhages can result in additional morbidity. From 1992 to 2011, 49 patients with brainstem CMs were treated with Gamma Knife radiosurgery (GKS). We classified patients into two groups: Group A (n = 31), patients who underwent GKS for a CM following a single symptomatic bleed, and group B (n = 18), patients who underwent GKS for a CM following two or more symptomatic bleeds. The mean marginal dose of radiation was 13.1 Gy (range 9.0-16.8 Gy): 12.8 Gy in group A and 13.7 Gy in group B. The mean follow-up period was 64.0 months (range 1-171 months). In group A, the annual hemorrhage rate (AHR) following GKS was 7.06 % within the first 2 years and 2.03 % after 2 years. In group B, four patients (22.2 %) developed new or worsening neurologic deterioration as a result of repeat hemorrhages. In group B, the AHR was 38.36 % prior to GKS, 9.84 % within the first two years, and 1.50 % after two years. There was no statistically significant difference in the AHRs at each follow-up period after GKS between the two groups. Adverse radiation effects (AREs) developed in a total of four patients (8.2 %); among them, one patient (2.0 %) developed a permanent case of diplopia. No mortality occurred in this series. | In this study, GKS was demonstrated to be a safe and effective alternative treatment for brain stem CMs that resulted in a reduction in the AHR. Consequently, we suggest that even CM patients who have suffered only a single bleed should not be contraindicated for SRS. | Gamma Knife radiosurgery for brainstem cavernous malformations: should a patient wait for the rebleed? |
Pediatric trauma centers (PTCs) were developed to improve the survival of injured children, but it is currently unknown if children admitted to PTCs are more likely to survive than those admitted to adult trauma centers (ATCs). Fifty-three thousand one hundred thirteen pediatric trauma cases from 22 PTCs and 31 ATCs included in the National Pediatric Trauma Registry were reviewed to evaluate survival rates at PTCs and ATCs. Overall, 1,259 children died. The raw mortality rate was lower at PTCs (1.81% of 32,554 children) than at ATCs (3.88% of 18,368 children). However, patients admitted to ATCs were more severely injured. When Injury Severity Score, Pediatric Trauma Score, mechanism (blunt or penetrating), gender, age, clustering, and American College of Surgeons (ACS) verification status were controlled for using a single logistic regression model, there was no statistically significant difference in survival between PTCs and ATCs (odds ratio, 1.02; 95% confidence interval, 0.83-1.26; p = 0.587). A similar comparison of the 12 ACS-verified trauma centers with the 41 nonverified centers showed verification to be associated with improved survival (odds ratio, 0.75; 95% confidence interval, 0.58-0.97; p = 0.013). | Although PTCs have higher overall survival rates than ATCs, this difference disappears when the analysis controls for Injury Severity Score, Pediatric Trauma Score, age, mechanism, and ACS verification status. ACS-verified centers have significantly higher survival rates than do unverified centers. | Do pediatric trauma centers have better survival rates than adult trauma centers? |
Multicentric cancer is present in a large proportion of mastectomies performed as treatment of breast cancer; it has been considered a contraindication to breast conservation. We reviewed the records of our patients with Stage I or II breast cancer treated with breast conserving surgery and radiation therapy over a 13-year period. Twenty-seven patients had two or more nodules of grossly visible cancer separated by histologically normal breast tissue. All patients had grossly negative margins of excision; however, four patients had microscopically positive margins. Nine patients had positive axillary nodes. All patients received radiation therapy to the breast postoperatively, with a median dose of 50.4 Gy in 28 fractions; 11 patients also received a boost dose of 6-20 Gy to the tumor bed. Eleven patients were given adjuvant chemotherapy and one patient was given adjuvant tamoxifen. With a median follow-up of 53 months, only one patient has relapsed in the breast (3.7%); that patient relapsed in multiple distant sites at the same time. Three patients have died of disseminated disease; the actuarial survival and disease-free survival rates at 4 years are 89%. | Breast conservation may be considered for patients with multicentric breast cancer discovered at the time of histologic examination. For patients with multicentric disease detected prior to surgery, breast conserving therapy may be appropriate as long as: (1) all clinically and radiographically apparent abnormalities are removed, (2) clear margins of resection are achieved, and (3) there is no extensive intraductal component. | Should multicentric disease be an absolute contraindication to the use of breast-conserving therapy? |
Stent-assisted coiling of intracranial aneurysms with self-expanding stents has widened the applicability of neuroendovascular therapies to those aneurysms previously considered "uncoilable" because of poor morphology. The Enterprise Vascular Reconstruction Device and Delivery System (Cordis) has demonstrated promising initial short-term results. However, the rates of delayed in-stent stenosis or thrombosis are not known. To report midterm results of the Enterprise stent system. A 10-center registry was created to provide a large volume of data on the safety and efficacy of the Enterprise stent system. Pooled data were compiled for consecutive patients undergoing Enterprise stent-assisted coiling at each institution. Available follow-up data were evaluated for the incidence of in-stent stenosis, thrombosis, and aneurysm occlusion. In total, 213 patients (176 females) with 219 aneurysms were treated with the Enterprise stent. One hundred ten patients had undergone delayed angiography (≥ 30 days from stent placement, mean follow-up 174.6 days). Forty percent of patients demonstrated total occlusion with 88% having ≥ 90% aneurysm occlusion. Six percent of patients had delayed (>30 days) angiographic findings, of which 3% demonstrated significant (≥ 50%) in-stent stenosis or occlusion. Seven delayed thrombotic events occurred (3%), along with 2 additional immediate periprocedural events. All 7 delayed events were concomitant to cessation of double-antiplatelet therapy. | Midterm occlusion rates are excellent, and stenosis and thrombosis rates are comparable to other available neurovascular stents. Interruption of antiplatelet therapy appears to be a factor in those developing delayed stenosis or thrombosis. | Delayed thrombosis or stenosis following enterprise-assisted stent-coiling: is it safe? |
Physical examination remains an important part of the initial evaluation of patients presenting with chest pain but little is known about the effect of patient gender on physician performance of the cardiovascular exam. To determine if resident physicians are less likely to perform five key components of the cardiovascular exam on female versus male standardized patients (SPs) presenting with acute chest pain. Videotape review of SP encounters during Objective Structured Clinical Examinations (OSCEs) administered by the Emory University Internal Medicine Residency Program in 2006 and 2007. Encounters were reviewed to assess residents' performance of five cardiac exam skills: auscultation of the aortic, pulmonic, tricuspid, and mitral valve areas and palpation for the apical impulse. One hundred forty-nine incoming residents. Residents' performance for each skill was classified as correct, incorrect, or unknown. One hundred ten of 149 (74 %) of encounters were available for review. Residents were less likely to correctly perform each of the five skills on female versus male SPs. This difference was statistically significant for auscultation of the tricuspid (p = 0.004, RR = 0.62, 95 % CI 0.46-0.83) and mitral (p = 0.007, RR = 0.58, 95 % CI = 0.41-0.83) valve regions and palpation for the apical impulse (p < 0.001, RR = 0.27, 95 % CI = 0.16-0.47). Male residents were less likely than female residents to correctly perform each maneuver on female versus male SPs. The interaction of SP gender and resident gender was statistically significant for auscultation of the mitral valve region (p = 0.006) and palpation for the apical impulse (p = 0.01). | We observed significant differences in the performance of key elements of the cardiac exam for female versus male SPs presenting with chest pain. This observation represents a previously unidentified but potentially important source of gender bias in the evaluation of patients presenting with cardiovascular complaints. | Does patient gender impact resident physicians' approach to the cardiac exam? |
Previous studies have indicated that propofol anaesthesia may reduce the incidence of postoperative nausea and vomiting after strabismus surgery in children. This study was designed to investigate the incidence of vomiting after strabismus surgery at two different levels of propofol anaesthesia compared to thiopental/isoflurane anaesthesia. Ninety ASA class I or II children, aged 5-14 yrs were randomly assigned to one of three groups: Group T/I (n = 30) induction with 5 mg kg-1 of thiopental and maintenance with isoflurane, group P5 (n = 31) induction with propofol 2 mg kg-1, maintenance with propofol infusion 5 mg kg-1 h-1 or group P10 (n = 29) induction with propofol 2 mg kg-1, maintenance with propofol 10 mg kg-1 h-1. All received glycopyrrolate, vecuronium, fentanyl and controlled ventilation with O2/N2O 30/70. Ketorolac i.v. was given to prevent postoperative pain. If additional analgesia was needed, ibuprofen/acetaminophen or buprenorphine was given according to clinical need. There were no differences between study groups with respect to age, weight, history of previous anaesthesia or emesis after previous anaesthesia, duration of anaesthesia, surgery or sleep after anaesthesia, or number of muscles operated. The incidence of vomiting was 37%, 29% and 28% in groups T/I, P5 and P10, respectively. There were no statistically significant differences between the three groups in the incidence of vomiting. The median age of patients who vomited was 7.5 (range 5.0-13.7) yrs while the median age of the patients who did not vomit was 9.1 (range 5.0-14.0) yrs (P<0.01). | In the present study, propofol anaesthesia compared to thiopental/isoflurane anaesthesia did not reduce the incidence of vomiting following strabismus surgery in children. | Does propofol reduce vomiting after strabismus surgery in children? |
To compare the prognosis of upper urinary tract (UUT)-urothelial carcinoma (UC) and UC of the bladder (UCB) by pathological staging in patients treated with radical surgeries. The study population comprised 335 and 302 consecutive radical surgery cases performed between 1991 and 2010 for UUT-UC and UCB, respectively. Five-year recurrence-free survival (RFS) and cancer-specific survival (CSS) rates were analyzed. The median follow-up period of all subjects was 59.3 months (range, 0.1-261.0 months). No difference was observed in median patient age, distribution of pathologic T stage, or rates of positive surgical margin between the two groups. The UUT-UC group had significantly more frequent hydronephrosis than the USB group (48.1% vs. 20.2%, p < 0.001). However, the UUT-UC group showed significantly less frequent grade III tumors (28.1% vs. 58.6%, p < 0.001), lymphovascular invasion (18.8% vs. 35.8%, p < 0.001), and associated carcinoma in situ (9.0% vs. 21.9%, p < 0.001) than the UCB group. Five year RFS rates in the UUT-UC and UCB groups were 77.0% and 75.9%, respectively (p = 0.546). No significant difference in RFS rate was observed between pathological T stage subgroups. Five year CSS rates in the UUT-UC and UCB groups were 76.1% and 76.2%, respectively (p = 0.462). No significant difference was observed in CSS rate between the pathologic T stage subgroups. | UUT-UC and UCB showed comparable prognosis at identical stages. However, our results should be verified in a prospective study due to the retrospective study design in this study. | Are urothelial carcinomas of the upper urinary tract a distinct entity from urothelial carcinomas of the urinary bladder? |
To determine the effects of age and work status on whether and where cardiovascular specialists would place hypothetical patients in the queue for coronary surgery. Mailed survey presenting a set of clinical scenarios either to be rated on a scale with 7 time frames for urgency of need or to be designated as questionable/inappropriate for intervention. The basic scenario was a patient with mild-moderate stable angina, good left ventricular function, and limited coronary disease; operative risks and stress test results were varied. Three identifiers were used: "45-year-old civil servant gainfully employed"; "45-year-old laborer disabled by angina, faces job loss"; and "75-year-old retiree, angina limits golf." Cardiologists and cardiac surgeons practicing in five Ontario medical centers (n = 120). There was a 59% response rate (120 usable responses). Large shifts in willingness to intervene occurred in favor of the disabled laborer (p less than 0.0001) and against the retiree (p-value ranges from 0.04 to less than 0.0001, depending on operative risk and stress test results), but not for the employed civil servant. Striking effects (p less than 0.0001) were also evident in ratings of waiting time, with the order of priority being the disabled laborer first, the civil servant second, and the retiree last. The overall mean shift due to work status or age was equal to, or larger than, the mean shift due to clinical factors, such as stress test results, changes in severity of stable angina, and extent of coronary disease. | Cardiovascular specialists may place considerable weight on age and work status in determining urgency and appropriateness of coronary revascularization. Risk-benefit concerns may partly explain shifting thresholds for intervention, but not differential waiting times. The influence of these factors should be sought in utilization audits and addressed from an ethical perspective. | Placing patients in the queue for coronary surgery: do age and work status alter Canadian specialists' decisions? |
To assess the association between self-rated health, obesity, and self-reported health behaviors of Latino immigrants Two hundred two Latino immigrants (mean age=31.63, SD=8.30, 54% female) participated in a 15-minute interview and height and weight measurements. Participants reporting good to excellent health reported engaging in physical activity during the past month (P<.05), eating more fruits and vegetables (P<.001 and P<.01 respectively), and watching less television (P<.01) than did those who reported fair to poor health. Self-rated health was not associated with BMI. | Greater attention to Latinos' self-perception of health in relation to weight is needed to develop interventions to improve health status. | Do Latino immigrants link self-rated health with BMI and health behaviors? |
Minimally invasive surgery offers many advantages, but its correct practice is associated with a steep learning curve. Telesurgery allows a surgeon at a remote site to guide and teach surgeons at a primary site by utilizing robotic devices, telecommunications, and video technology, thereby reducing complications. From September 1998 to July 2000, 17 procedures were telementored between two sites 9230 km apart: a primary operating room at the Policlinico Casilino "Tor Vergata" University of Rome and a remote site at the Johns Hopkins Medical Institutions in Baltimore. Of these procedures, 14 were laparoscopic cases: 8 spermatic vein ligations, 2 retroperitoneal renal biopsies, 3 simple nephrectomies, and 1 pyeloplasty. Three procedures were carried out to obtain percutaneous renal access. All procedures were performed with the help of two robots: the first robot, AESOP, for the orientation of the laparoscope, and the second one, PAKY, to perform the percutaneous renal access. In addition to the robotic device, the system provided four ISDN lines, a PC with dedicated software to manage the connection, audio and video connections, an external video camera with a panoramic view of the operating room, and remote control of the electrocautery and the Telestrator. All the procedures were accomplished with an uneventful postoperative course. Ten operative cases were telementored successfully. In five cases, it was not possible to establish a connection to the remote site, and two procedures were converted to open surgery because of intraoperative complications. The time delay of the image transmission was<1 second. | This preliminary experience has demonstrated the feasibility of international telementoring. It could provide education to surgeons and decrease the likelihood of complications attributable to inexperience with new techniques. | Is telesurgery a new reality? |
(1) To evaluate the frequency of visualisation and measurements of the normal appendix. (2) To correlate Body Mass Index (BMI) and gender with visualisation of the normal appendix. (3) To correlate age, gender and body length with appendiceal length. A retrospective review of 186 patients undergoing abdominal CT without suspicion of acute appendicitis was done. Frequency of visualisation and measurements (including maximal outer diameter, wall thickness, length, content, location of base and tip) of normal appendices were recorded. Prevalence of appendectomy was 34.4%. Sensitivity, specificity, positive predictive value, negative predictive value, and accuracy of visualisation of the normal appendix were 76%, 94%, 96%, 67%, and 82% respectively. The mean maximal diameter of the appendix was 8.19 mm±1.6 (SD) (range, 4.2-12.8 mm). The mean length of the appendix was 81.11 mm±28.44 (SD) (range, 7.2-158.8 mm). The mean wall thickness of the appendix was 2.22 mm±0.56 (SD) (range, 1.15-3.85 mm). The most common location of the appendiceal tip was pelvic in 66% appendices. The most common location of the appendiceal base was inferior, medial, and posterior in 37%. The normal appendix contained high-density material in 2.2%. There was a significant correlation between gender and appendiceal length, with men having longer appendices than women. | Most normal appendices are seen at multislice CT using i.v. contrast. The maximal outer diameter of the normal appendix overlaps with values currently used to diagnose appendicitis on CT. | The normal appendix on CT: does size matter? |
Deterioration in pulmonary function is a common complication following coronary artery bypass graft surgery and there is still speculation to the precise causative factors thereof. Cardioplegia solution not drained by the atriocaval cannula enters the lung parenchyma unless removed by a pulmonary artery (PA) vent. The hypothesis of the present study was that cold blood cardioplegia solution damages the lung parenchyma, resulting in an observed deterioration of clinical lung function. A prospective, double-blind, randomised trial was conducted on 142 patients. The study group of 71 patients had a PA vent inserted at the time of cannulation, preventing cardioplegia from going through the lungs. In addition, positive end expiratory pressure (PEEP) was applied and low-volume lung ventilation carried out during cardiopulmonary bypass (CPB). The control group (n =71) had cardioplegia enter the lung parenchyma during cardiopulmonary bypass. Clinical parameters of arterial blood gases, including estimated shunt fraction, spirometry tests and radiographic analysis was made preoperatively and at set times through the postoperative period. Baseline demographics and intraoperative and postoperative management was the same in both groups, thus, yielding a homogenous sample for analysis. Significant changes were noted in arterial blood gases, spirometry, and radiographic analysis of effusion and atelectasis over the time periods studied (p<0.001). There was, however, no significant difference between the study and control groups at any point (p>0.05). | The data, therefore, suggest that allowing cold blood cardioplegia solution to circulate the lungs during cardiopulmonary bypass does not have any (beneficial or detrimental) effect on clinical lung function postoperatively. | Does cold blood cardioplegia solution cause deterioration in clinical pulmonary function following coronary artery bypass graft surgery? |
An exploratory analysis of a prospective study of risk factors for acute pancreatitis after ERCP combined with endoscopic sphincterotomy showed that the frequency of acute pancreatitis was lower in patients who received heparin compared with patients not treated with heparin. The study was continued to further analyze the effect of heparin on the frequency of acute pancreatitis. Potential risk factors for acute pancreatitis and outcomes were evaluated prospectively for all ERCP procedures with endoscopic sphincterotomy performed between September 1994 and December 1998. The results were analyzed by univariate and multivariate methods to determine risk factors for complications. Heparin was administered to 32.9% of the patients (heparin group [HEP group], n = 268) for various clinical reasons (low-molecular-weight heparin, n = 208, unfractionated heparin n = 60). A group of 547 patients who did not receive heparin served as control patients (CON group). Eight hundred fifteen patients underwent ERCP with endoscopic sphincterotomy; acute pancreatitis occurred in 6.4% (n = 52). The frequency of acute pancreatitis was significantly lower in the HEP group versus the CON group in the final multivariate model, which included significant risk factors for acute pancreatitis (HEP group: 3.4%, 9/268 vs. CON group: 7.9%, 43/547; p = 0.005). HEP did not increase the risk of hemorrhage (HEP group: 1.1%, 3/268, 2 severe, none fatal vs. CON group: 2.0%, 11/547, 3 severe, 2 fatal). HEP (p = 0.005; OR 0.3: 95% CI [0.16, 0.73]) and the number of risk factors present (p = 0.0001; OR 2.5: 95% CI [1.80, 3.50]) influenced the frequency of acute pancreatitis independently. | Heparin was significantly associated with an extremely low frequency of post-ERCP pancreatitis without increasing the risk of hemorrhage after endoscopic sphincterotomy. Because this effect could not be attributed to other known or suspected confounders, our conclusion was that heparin administration before ERCP reduces the risk of pancreatitis. | Complications of endoscopic sphincterotomy: can heparin prevent acute pancreatitis after ERCP? |
The role local neighbourhood environments play in influencing purpose-specific walking behaviors has not been well-explored in prospective studies. This study aimed to cross-sectionally and prospectively examine whether local physical and social environments were associated with mothers' walking for leisure and for transport. In 2004, 357 mothers from Melbourne, Australia, provided information on their local physical and social neighbourhood environments, and in 2004 and 2006 reported weekly time spent walking for leisure and for transport. Environmental predictors of high levels of walking and increases in walking were examined using log binomial regression. Public transport accessibility and trusting many people in the neighbourhood were predictive of increases in walking for leisure, while connectivity, pedestrian crossings, a local traffic speed were predictive of increases in transport-related walking. Satisfaction with local facilities was associated with increasing both types of walking, and the social environment was important for maintaining high levels of both leisure- and transport-related walking. | The findings provide evidence of a longitudinal relationship between physical and social environments and walking behaviors amongst mothers. Enhancing satisfaction with local facilities and giving consideration to 'walkability', safety and public transport accessibility during environment planning processes may help mothers to increase walking. | Are perceptions of the physical and social environment associated with mothers' walking for leisure and for transport? |
Thirty-nine children, eight females and 31 males, were studied, with a mean chronological age of 12.37 +/- 2.24 years (range 8-15 years), mean bone age 9.58 +/- 2.21 years, Tanner stage 1-2, with height 134.12 +/- 11.27 cm (Ht-SDS -2.24 +/- 0.95), growth velocity (GV) 4.2 cm/yr (GV SDS -2.12 +/- 1.32) and an inadequate response to an initial GH stimulation test with clonidine (peak GH<10 microg/l). A second stimulation test with clonidine was performed in the same patients after gonadal steroid priming: 100 mg i.m. testosterone enanthate 5-8 days prior to GH stimulation in the males, and 1 mg estradiol valerate daily for 3 days in the females. GH, IGF-I, and testosterone/estradiol were measured before and after priming. Twenty-one of 39 children (53.8%) increased their GH response to a level of>10 microg/l following priming with gonadal steroids. Mean peak GH after priming was 12.32 +/- 8.7 microg/l compared to a peak GH level of 4.87 +/- 2.72 microg/l prior to gonadal steroid priming (peak GH 17.42 +/- 8.46 microg/l in the responders versus 5.95 +/- 2.76 microg/l in the non-responders). Although a significant increase in GH and IGF-I concentrations was noted following priming, we were unable to find a correlation between IGF-I concentrations and peak GH following priming. IGF-I levels were not different in the responders and non-responders to clonidine following priming. There was no correlation between pubertal staging and testosterone/estradiol concentrations before priming with the peak GH after priming. | Priming with gonadal steroids significantly improves GH secretion following GH stimulation with clonidine and diminishes the possibility of a false diagnosis of GH deficiency. Gonadal steroid priming should therefore be considered in the evaluation of the GH status of short children close to or during the early stages of puberty. | Is testosterone and estrogen priming prior to clonidine useful in the evaluation of the growth hormone status of short peripubertal children? |
As oral contraceptives (OCs) suppress anti-Müllerian hormone (AMH), and hormonal contraceptives (HCs), likely, suppress functional ovarian reserve, this study was initiated to determine whether HC affect oocyte yields. We investigated in a retrospective cohort study 43 oocyte donors in 71 in vitro fertilization (IVF) cycles, evaluating anti-Müllerian hormone (AMH) and oocyte yields as reflections of functional ovarian reserve (OR). In 25 IVF cycles egg donors were on HC within one month prior to IVF, and in 46 cycles they were not. Donors, based on their HCs, were further subdivided into 12 with less, and 13 with more androgenic progestins. While the three groups did not differ in age, age at menarche, BMI and AMH, oocyte yields among donors who utilized estrane- and gonane-derived (higher androgenic) HCs were lower 11.3 (95% CI 8.3 - 14.3) than either donors using no HCs 16.6 (95% CI 14.7 -18.4) (P<0.05) or those using anti-androgenic HCs 19.0 (95% CI 12.2-25.8) (P<0.01). Significance was maintained after adjustments for the donor age and total FSH dose used in ovulation induction. | Even in young oocyte donors, high androgenic OC exposure appears to suppress functional ovarian reserve and oocyte yields. Since OCs are often routinely used in preparation for IVF, such practice may require reevaluation. Especially in women with diminished ovarian reserve OCs, and especially high androgenic progestin HCs, should, likely, be avoided. | Does hormonal contraception prior to in vitro fertilization (IVF) negatively affect oocyte yields? |
By assessing current surgical outcome and symptomatic relief, this study attempts to answer whether atrial septal defects in adults should be closed. Thirty-nine adult patients aged 35.2 +/- 13.6 years underwent operation for an atrial septal defect between June 1988 and June 1994. Indications for closure were symptoms (33 patients) or a significant left-to-right atrial shunt (6 patients). Data were obtained from hospital records, and the latest status of the patients was determined by a written questionnaire. There were no deaths. Pulmonary embolism in 1 patient was the only complication observed. The QRS duration on the surface electrocardiogram decreased immediately (p<0.001), and the cardiothoracic ratio on chest radiographs was significantly lower 3 to 6 months after operation (p<0.001), both findings reflecting improved hemodynamics. No residual shunts were seen on follow-up (mean follow-up, 3.3 +/- 2.2 years). Twenty-seven (81.8%) of the 33 symptomatic patients improved clinically in terms of exercise performance, atrial arrhythmias, or both. Three (50%) of the 6 previously asymptomatic patients reported improved functional capacity post-operatively. | Today, operation for atrial septal defects in adults can be performed with no mortality and low morbidity and results in symptomatic improvement in the majority of patients. Clinical improvement was seen even in patients who considered themselves asymptomatic preoperatively. We advocate closure of atrial septal defects in adult patients with symptoms or significant atrial shunts. | Should atrial septal defects in adults be closed? |
Most studies of the prevalence of psychoactive substances in injured emergency department patients have excluded those who arrive more than 6h after injury. This may cause a selection bias. The aim of this study was: (1) to describe the characteristics of patients who arrive more than 6h after injury, compared to patients who arrive sooner (2) to examine whether self-report can add to the assessment of alcohol use when the patient is assessed more than 6h after injury. Blood sample analysis and self-report data were used to assess the prevalence of psychoactive substances in injured patients admitted to an emergency department within 48 h of injury (n=1611). Discriminant function analysis was used to assess group differences. The patients who arrived more than 6h after injury differed significantly from those who arrived earlier in several respects. They more often screened positive for hypnotics; they were older, they were more likely to have had a fall and they were more often injured at home and at night. Self reported use of alcohol showed good consistency with blood sample screening within 6h of injury and could therefore be used to assess alcohol use more than 6h after injury. | Patients who arrive more than 6h after injury differ significantly from those who arrive earlier. Future studies on the prevalence of psychoactive substances in emergency departments could expand the inclusion window. | Studying psychoactive substance use in injured patients: does exclusion of late arriving patients bias the results? |
social support is an important form of external support to patients and families. Assessment of postoperative external support provided by staff to patients and family members at discharge from hospital and related factors. Quantitative descriptive study conducted with surgical patients treated for disc herniation or spinal stenosis (N = 92) and family members (N = 55) in a central hospital in Finland in 2008-2010 to measure the importance of various forms of support and their association with respondents' overall postoperative coping. Patient education atmosphere was the most important source of external support for both patients and family members. Better overall coping was reported by both groups if the patient's behavior and intrafamilial emotions had changed in a positive way. Patients' overall coping was promoted if they received adequate information from staff. | Nurses' role and competence are crucial in supporting the coping of patients and families. | Staff support for back surgical patients and family members: does it improve coping at home? |
A clinical, observational, transversal and controlled study was conducted in 384 women: 49 nulliparous, 103 primigravid pregnant, 92 primiparous postpartum (vaginal delivery: n = 43; cesarean section delivery: n = 49), 22 climacteric, 65 postmenopausal, and 53 women identified as being unable to perform voluntary maximum contraction. All subjects were evaluated with digital palpation and PFM surface electromyography (sEMG) and completed the questionnaires: International Consultation on Incontinence Urinary Incontinence Short Form (ICIQ IU-SF) and International Consultation on Incontinence Questionnaire Overactive Bladder (ICIQ-OAB). Spearman's Correlation Coefficient and ANOVA were used to analyze the variables. The nulliparous women had higher PFM electromyographic activity than the other groups. The primigravid pregnant, cesarean section and vaginal delivery groups had higher electromyographic activity than the postmenopausal group. Studying PFM electromyographic activity with the factors evaluated, a negative correlation between age, parity, and the presence and severity of urinary symptoms was observed. There was no correlation between PFM electromyographic activity and BMI. | Fourteen percent of women participating were not able to perform active contraction of the PFM. PFM electromyographic activity changed during the female life cycle. PFM electromyographic activity correlated inversely with age, parity, and the presence and severity of urinary symptoms. | Electromyographic pelvic floor activity: Is there impact during the female life cycle? |
The objective was to examine the independent and gender-specific effects of WC and BMI on CVD risk factors, insulin resistance and β-cell function. A cross-sectional study of 2931 adults aged 20-79 years was carried out in Fujian province by multi-stratified sampling from July 2007 to May 2008. Gender-specific differences of WC and BMI on CVD risk factors, insulin resistance and β-cell function were displayed jointly by WC and BMI tertiles. The homeostasis model assessment of insulin resistance (HOMA-IR) index and the quantitative insulin-sensitivity check index (QUICKI): l/(log G0 ± log I0) were used to estimate insulin sensitivity; insulin secretion was assessed using the HOMA-β index; β-cell function was quantified as the ratio of the incremental insulin to glucose responses over the first 30 min during the OGTT (△I30/△G30). The oral disposition index (DIo) was calculated as ΔI(30) /ΔG(30) × 1/fasting insulin. The Matsuda index is defined as 10,000/sqrt (FBG × FPI × [G × I]) where FPG is fasting glucose, FPI is fasting insulin, G is mean glucose during the OGTT (calculated from glucose samples at 0, 30, and 120 min), and I is mean insulin during the OGTT (calculated from insulin samples at 0, 30, and 120 min). Waist circumference and BMI correlated with each other in both men (0.756, p<0.001) and women (0.728, p<0.001). The two indexes were independently associated with CVD risk factors (such as hypertension, metabolic syndrome, and dyslipidaemia) in both men and women. BMI was better than WC in assessing the risk of diabetes in men (p = 0.003 for BMI, and p = 0.234 for WC), while WC was better than BMI in predicting diabetes in women (p<0.001 for WC, and p = 0.831 for BMI). There were significant associations between BMI and insulin resistance or β-cell function even after adjustment for WC except for DIo in male subjects, but WC only associated with HOMA-IR positively or the Matsuda index and QUICKI negatively after adjustment for BMI. For women, associations between WC and insulin resistance or β-cell function remained strong even after adjustment for BMI besides DIo. However, there were no independent relations of BMI to insulin resistance and β-cell function except for Matsuda index with a significant negative association after adjustment for WC in women. | Body mass index and WC were independently associated with CVD risk factors. There were differences in the gender-specific relevance of measures of body fat distribution in assessing the insulin resistance, β-cell function and thus the risk of diabetes. Therefore, WC should be measured in addition to BMI to assess CVD risk accurately and implement efficient treatment strategies. | Body fat distribution and their associations with cardiovascular risk, insulin resistance and β-cell function: are there differences between men and women? |
Enhanced external counterpulsation (EECP) has been demonstrated to be an effective treatment for stable angina in patients with coronary disease. The hemodynamic effects of EECP are maximized when the ratio of diastolic to systolic pressure area is in the range of 1.5 to 2.0. It is hypothesized that patients undergoing EECP who are able to achieve higher diastolic augmentation (DA) ratios may derive greater clinical benefit. This study examines the relationship between the DA ratio and clinical outcomes in patients undergoing EECP. We analyzed demographic, noninvasive hemodynamic, and clinical outcome data on 1,004 patients enrolled in the International EECP Patient Registry (IEPR) for treatment of chronic angina between January 1998 and August 1999. Blood pressure waveforms were recorded from finger plethysmography. Six-month clinical outcomes were obtained by telephone interview. At the end of EECP treatment, 370 (37%) patients had a higher DA ratio (defined as>or = 1.5) and 634 (63%) had a lower DA ratio (defined as<1.5). Factors associated with a lower DA ratio included age>or =65 years (p<0.001), female gender (p<0.001), left ventricular ejection fraction<35% (p<0.05), hypertension (p<0.01), prior coronary bypass surgery (p<0.01), noncardiac vascular disease (p<0.001), multivessel disease (p<0.01), congestive heart failure (p<0.01), current smoking (p<0.01), unsuitability for further revascularization (p<0.001), and higher baseline angina class (p<0.001). There were no significant differences regarding diabetes mellitus, prior coronary angioplasty, prior myocardial infarction, or antianginal medication use between patients with higher or lower DA ratios. Based on a multiple logistic regression model, independent predictors of a DA ratio<1.5 at the end of EECP included current smoking (odds ratio 3.3; 95% confidence intervals 2.0-5.4); multivessel disease (1.7; 1.3-2.3); female gender (2.2; 1.7-3.0); no prior EECP (1.9; 1.1-3.3); noncardiac vascular disease (2.3; 1.7-2.9); age>or = 65 years (1.7; 1.4-2.2), and patients not suitable for revascularization (1.6; 1.2-2.0). By the end of therapy, there were no significant differences in myocardial infarction, revascularization rates, or nitroglycerin use with respect to higher DA ratios. At 6-month follow-up, patients with higher DA had a trend toward a greater reduction in angina class compared with those with lower DA (p = 0.069). There was a significantly higher rate of unstable angina and congestive heart failure in the group not achieving higher augmentation (p<0.05). | Patients who are younger, male, nonsmoking, and without multivessel coronary or noncardiac vascular disease are most likely to have higher DA with EECP. Patients with higher DA tended to have a greater reduction in angina class at 6-month follow-up compared with those with lower DA ratios. There is evidence that higher DA ratios are associated with improved short- or long-term clinical outcomes, suggesting that clinical benefit from EECP is associated with the magnitude of DA. | Does higher diastolic augmentation predict clinical benefit from enhanced external counterpulsation? |
To review the effectiveness of reduced-dose and restricted-volume radiation-only therapy in the treatment of intracranial germinoma and to assess the feasibility of reducing or eliminating the use of chemotherapy. Between January 1996 and March 2007, a retrospective analysis was performed that included 38 patients who received either reduced radiation alone (30 Gy for 26 patients) or reduced radiation with chemotherapy (n = 12 patients). All 38 patients received extended focal (including whole-ventricle) irradiation and were followed up until February 2008. Overall survival (OS) and relapse-free survival (RFS) rates were calculated. Variables associated with survival were evaluated by univariate Cox proportional hazards regression. Median follow-up was 62.4 months (range, 10.1-142.5 months). The total 5-year OS rate was 93.7%. The 5-year OS and RFS rates for patients receiving radiation only were 100% and 96.2%, respectively. The rates for those receiving radiation plus chemotherapy were 83.3 % and 91.7%, respectively (not statistically significant). No predictive factor was significantly associated with the OS or RFS rate. Chemotherapy had no significant effect on survival but was associated with a higher incidence of treatment-related toxicity. | A further decrease in the radiation dose to 30 Gy with whole-ventricle irradiation is sufficient to treat selected patients with intracranial germinoma. Wide-field irradiation or chemotherapy should be avoided as these methods are unnecessary. Thus, reduction of the radiation dose to 30 Gy may be feasible, even without chemotherapy. | Optimal treatment for intracranial germinoma: can we lower radiation dose without chemotherapy? |
Conversion of a surgically arthrodesed knee to total knee arthroplasty (TKA) is an option for a select group of patients who are not satisfied with their results. However, there is a paucity of literature on this topic. A systematic review of literature was performed to (1) describe the overall demographic characteristics; (2) evaluate the clinical outcomes; (3) determine the overall rate of complications; and (4) evaluate the overall satisfaction of patients who underwent conversion of an arthrodesed knee to TKA. A comprehensive literature search was systematically performed to evaluate all studies included in the literature until July 2015. The specific search terms used were "fusion knee" and "arthrodesis knee," which revealed a total of 2206 studies. A review and selection of these abstracts were then performed based on inclusion and/or exclusion criteria; a total of 10 articles were used for final review. There were a total of 98 surgically arthrodesed knees that subsequently underwent TKA. Patients had a mean age of 55 years and were followed up for a mean of 5 years. Using a random effects model, there was an overall complication rate of 47%, an overall revision rate of 25%, and an overall failure rate of 11%. However, most patients were overall satisfied with the procedure. | Fusion takedown is a challenging procedure that should only be performed by experienced surgeons after extensive discussion with the patients. The clinical outcomes are good with overall patient satisfaction, but complication rates are high including risk of repeat fusion or amputation. | Conversion of a Surgically Arthrodesed Knee to a Total Knee Arthroplasty-Is it Worth it? |
To qualitatively assess obstetricians' and neonatologists' responses to standardized patients (SPs) asking "What would you do?" during periviable counseling encounters. An exploratory single-center simulation study. SPs, portraying a pregnant woman presenting with ruptured membranes at 23 weeks, were instructed to ask, "What would you do?" if presented options regarding delivery management or resuscitation. Responses were independently reviewed and classified. We identified five response patterns: 'Disclose' (9/28), 'Don't Know' (11/28), 'Deflect' (23/28), 'Decline' (2/28), and 'Ignore' (2/28). Most physicians utilized more than one response pattern (22/28). Physicians 'deflected' the question by: restating or offering additional medical information; answering with a question; evoking a hypothetical patient; or redirecting the SP to other sources of support. When compared with neonatologists, obstetricians (40% vs. 15%) made personal or professional disclosures more often. Though both specialties readily acknowledged the importance of values in making a decision, only one physician attempted to elicit the patient's values. | "What would you do?" represented a missed opportunity for values elicitation. Interventions are needed to facilitate values elicitation and shared decision-making in periviable care. | "Doctor, what would you do? |
Controversy exists concerning the importance of operative time on patient outcomes. It is unclear whether faster is better or haste makes waste or similarly whether slower procedures represent a safe, meticulous approach or inexperienced dawdling. The objective of the present study was to determine the effect of operative time on 30-day outcomes in laparoscopic surgery. Patients who underwent laparoscopic general surgery procedures (colectomy, cholecystectomy, Nissen fundoplication, inguinal hernia, and gastric bypass) from the ACS-NSQIP 2005-2008 participant use file were identified. Exclusion criteria were defined a priori to identify same-day admission, elective procedures. Operative time was divided into deciles and summary statistics were analyzed. Univariate analyses using a Cochran-Armitage test for trend were completed. The effect of operative time on 30-day morbidity was further analyzed for each procedure type using multivariate regression controlling for case complexity and additional patient factors. Patients within the highest deciles were excluded to reduce outlier effect. A total of 76,748 elective general surgical patients who underwent laparoscopic procedures were analyzed. Univariate analyses of deciles of operative time demonstrated a statistically significant trend (p<0.0001) toward increasing odds of complications with increasing operative time for laparoscopic colectomy (n=10,135), cholecystectomy (n=37,407), Nissen fundoplication (n=4,934), and gastric bypass (n=17,842). The trend was not found to be significant for laparoscopic inguinal hernia repair (n=6,430; p=0.14). Multivariate modeling revealed the effect of operative time to remain significant after controlling for additional patient factors. | Increasing operative time was associated with increased odds of complications and, therefore, it appears that speed may matter in laparoscopic surgery. These analyses are limited in their inability to adjust for all patient factors, potential confounders, and case complexities. Additional hierarchical multivariate analyses at the surgeon level would be important to examine this relationship further. | Does speed matter? |
Little is known about the activity of the many new specialist outreach clinics in fundholding general practices that have emerged since the introduction of fundholding in 1991, though it has been claimed that specialist outreach clinics have shortened waiting times for fundholders' patients. This study describes the activity of specialist outreach clinics in fundholding practices in Sheffield, focusing on comparative waiting times between fundholding and non-fundholding practices. A descriptive study was carried out using routine out-patient activity data and a listing of outreach clinics obtained from fundholding practices. Thirty-seven specialist outreach clinics were established in fundholding practices by November 1994; 23 in surgical specialties. In 1994-1995, for gynaecology, orthopaedics and general surgery, the leading outreach specialties, 22.5 per cent of fundholders' first attendances were in outreach clinics. In those three specialties, 87.0 per cent of patients in specialist outreach clinics in fundholding practices vs 67.1 per cent in hospital clinics were routine appointments, and 17.4 per cent vs 9.4 per cent, respectively, were added to an in-patient waiting list. The proportion of first attendees seen in less than three months was 97.0 per cent in specialist outreach clinics in fundholding practices vs 88.1 per cent in hospital clinics; 90.4 per cent for the patients of fundholders who had outreach clinics vs 85.2 per cent for fundholders who did not; 88.1 per cent for all fundholders' patients vs 88.6 per cent for non-fundholders' patients. | The new specialist outreach service in fundholding practices in Sheffield is largely for surgical patients classified as routine patients. Although patients were seen more quickly in specialist outreach clinics, no overall inequality of waiting times between fundholding and non-fundholding practices was shown. | Specialist outreach clinics in Sheffield: a faster tier of out-patient provision for the patients of fundholding GPs? |
Forearm fractures are common skeletal injuries in childhood and can usually be treated nonoperatively with closed reduction and casting. Trends toward increasing operative treatment of these fractures have emerged. We aim to demonstrate the safety and efficacy of cast wedging for treatment of pediatric forearm fractures. We performed a prospective chart review of patients with forearm fractures, including distal radius (DR) fractures, treated with cast wedging at a single large pediatric hospital from June 2011 to September 2012. Inclusion criteria specified open distal radial physis, closed injury, loss of acceptable reduction, and availability of clinical and radiographic data from injury to cast removal. Exclusion criteria included pathologic fractures, neurovascular injury, fracture dislocations, open fractures, and closed DR physis. Reductions were performed and patients followed according to standard protocol at our institution, including placement into long-arm casts, initial follow-up visit within 5 to 10 days postinjury, and weekly visits for 2 weeks thereafter. If alignment were deemed unacceptable within 3 weeks of injury, cast wedging was utilized. Radiographic measurements of alignment included both radius and ulna on the injury film, postreduction, prewedge, postwedge, and final films. Radiographic technique was standardized, with repeatability testing demonstrating a precision of ±2 degrees. Over 15 months, our hospital treated 2124 forearm or DR fractures with closed reduction and casting. There were 60 fractures treated either with percutaneous fixation (36) or open treatment (24). A total of 79 forearm or DR fractures were treated with cast wedging secondary to loss of reduction, of which 70 patients had complete clinical and radiographic data. Average age was 8.4 years (range, 3 to 14 y), with 25 females and 45 males. Significant improvement in angulation for both-bone forearm fracture from prewedge to final films was seen in 69 children, with no major complications. One patient failed wedging and required surgical reduction and fixation. | Cast wedging is a simple, safe, noninvasive, and effective method for treatment of excessive angulation in pediatric forearm fractures. | Is there still a place for cast wedging in pediatric forearm fractures? |
Pubertal onset occurs earlier than in the past among U.S. girls. Early onset is associated with numerous deleterious outcomes across the life course, including overweight, breast cancer and cardiovascular health. Increases in childhood overweight have been implicated as a key reason for this secular trend. Scarce research, however, has examined how neighborhood environment may influence overweight and, in turn, pubertal timing. The current study prospectively examined associations between neighborhood environment and timing of pubertal onset in a multi-ethnic cohort of girls. Body mass index (BMI) was examined as a mediator of these associations. Participants were 213 girls, 6-8 years old at baseline, in an on-going longitudinal study. The current report is based on 5 time points (baseline and 4 annual follow-up visits). Neighborhood environment, assessed at baseline, used direct observation. Tanner stage and anthropometry were assessed annually in clinic. Survival analysis was utilized to investigate the influence of neighborhood factors on breast and pubic hair onset, with BMI as a mediator. We also examined the modifying role of girls' ethnicity. When adjusting for income, one neighborhood factor (Recreation) predicted delayed onset of breast and pubic hair development, but only for African American girls. BMI did not mediate the association between Recreation and pubertal onset; however, these associations persisted when BMI was included in the models. | For African American girls, but not girls from other ethnic groups, neighborhood availability of recreational outlets was associated with onset of breast and pubic hair. Given the documented risk for early puberty among African American girls, these findings have important potential implications for public health interventions related to timing of puberty and related health outcomes in adolescence and adulthood. | Does neighborhood environment influence girls' pubertal onset? |
To identify variables that increase the chance of neonatal scalp injury during vacuum extraction. We conducted a prospective observational study of 134 vacuum extraction-assisted deliveries at Olive View-UCLA Medical Center in 1995. Data collected included parity, gestational age, duration of first and second stages of labor, indication for operative delivery, station and position of fetal head, duration of vacuum application, number of "pop-offs," neonatal weight, and descriptions of scalp marks or injury. Cranial imaging studies were obtained if clinically indicated. There were 28 infants with scalp trauma, including 17 superficial lacerations, six large caputs, and 12 cephalohematomata; one infant had subgaleal, subdural, and subarachnoid hemorrhages. Logistic regression analysis showed duration of vacuum application to be the best predictor of scalp injury, followed by duration of second stage of labor and paramedian cup placement. Duration of vacuum application ranged from 0.5 to 26 minutes, with a median length of 3 minutes. The proportion of injuries was greater for applications exceeding 10 minutes (6 of 9) than for those 10 minutes or shorter (22 of 121, P<.01). We did not encounter any cases of clinically important scalp injury. | Cosmetic scalp trauma occurred in 21% of our newborns delivered by vacuum extraction and was more common after longer vacuum applications, longer second stages, and paramedian cup placement. | Vacuum extraction: does duration predict scalp injury? |
To evaluate whether formalized research training is associated with higher researcher productivity, academic rank, and acquisition of National Institutes of Health (NIH) grants within academic otolaryngology departments. Each of the 100 civilian otolaryngology program's departmental websites were analyzed to obtain a comprehensive list of faculty members credentials and characteristics, including academic rank, completion of a clinical fellowship, completion of a formal research fellowship, and attainment of a doctorate in philosophy (PhD) degree. We also recorded measures of scholarly impact and successful acquisition of NIH funding. A total of 1,495 academic physicians were included in our study. Of these, 14.1% had formal research training. Bivariate associations showed that formal research training was associated with a greater h-index, increased probability of acquiring NIH funding, and higher academic rank. Using a linear regression model, we found that otolaryngologists possessing a PhD had an associated h-index of 1.8 points higher, and those who completed a formal research fellowship had an h-index of 1.6 points higher. A PhD degree or completion of a research fellowship was not associated with a higher academic rank; however, a higher h-index and previous acquisition of an NIH grant were associated with a higher academic rank. The attainment of NIH funding was three times more likely for those with a formal research fellowship and 8.6 times more likely for otolaryngologists with a PhD degree. | Formalized research training is associated with academic success in otolaryngology. Such dedicated research training accompanies greater scholarly impact, acquisition of NIH funding, and a higher academic rank. | Does formal research training lead to academic success in otolaryngology? |
British guidelines on vascular disease prevention recommend adding a random (casual) blood glucose measurement to a lipid profile in those aged>or = 40 years. To assess this recommendation, we compared the predictive value of a risk model based on the Framingham risk score alone to one which additionally included information on fasting blood glucose, with respect to incident coronary heart disease (CHD) over 11 years. Men and women aged 40-63 years in Whitehall II were followed up for incident CHD: death/non-fatal myocardial infarction; angina confirmed by doctor diagnosis or electrocardiogram (ECG) and all first events. Fasting blood glucose was specified as a continuous variable or categorized by World Health Organization (WHO) 1999 glycaemic status (normal glucose tolerance, impaired fasting glucose or newly diagnosed diabetes). The hazard ratio for incident CHD was 1.10 (95%CI 1.09; 1.12) in men and 1.13 (1.10; 1.17) in women per percentage point increase in Framingham risk. The excess risk remained unchanged in models which added glycaemic status or continuous fasting glucose. The area under the receiver operating characteristic (ROC) curve for the Framingham score and incident coronary heart disease [0.70 (0.68; 0.73)] did not change when glycaemic status or fasting glucose was added to the prediction model. Reclassification with these modified models improved discrimination based on the Framingham score alone when glycaemic status was added, net reclassification improvement 2.4% (95% CI 0.2%; 4.6%), but not when fasting glucose was added. | Better detection of unrecognized diabetes is a valuable consequence of including a random blood glucose in a vascular risk profile. Our results suggest that this strategy is unlikely to improve risk stratification for CHD. | Do the Joint British Society (JBS2) guidelines on prevention of cardiovascular disease with respect to plasma glucose improve risk stratification in the general population? |