instruction
stringlengths
10
664
context
stringlengths
1
5.66k
response
stringlengths
1
3.34k
category
stringclasses
1 value
Do all children with suicidal ideation receive a significant psychiatric intervention?
In most physicians' minds, suicidal ideation carries with it an ominous prognosis requiring intensive evaluation and treatment. The aim of this study was therefore to determine the proportion of children identified with suicide ideation who received a significant psychiatric intervention. Medical records were reviewed for psychiatric interventions of all children presenting to the emergency department (ED) with behavioral disorders between 2004 and 2007, for whom a psychiatry consultation was obtained. Suicidal children were those who had expressed suicidal ideation. Significant psychiatric intervention was defined as one of the following: hospitalization in a psychiatric facility, period of observation in the ED (≥ 12 h), use of restraints, and prescription of psychiatric medication. Suicidal labeling was considered appropriate if one or more of the aforementioned interventions were recommended by a psychiatrist. The presence of psychiatric intervention was compared with that in children who presented with a behavioral disorder, not labeled as suicidal. Chi-squared or Fisher's exact test, whenever appropriate, was used to evaluate the association between suicide status and intervention. A total of 160 children (27.1%) were labeled as suicidal, and 431 (72.9%) with a behavioral disorder were classified as non-suicidal. A total of 244/431 (56.6%) in the non-suicidal group had a significant psychiatric intervention compared to 79/160 (49.4%) in the suicidal ideation group (P = 0.116). This 49.4% is significantly different from an a priori assumption that 100% of children with suicidal ideation would have a significant psychiatric intervention.
More than half (50.6%) of the children presenting with a label of suicidal ideation did not receive significant psychiatric intervention. This study calls into question the accuracy of suicide labeling in children referred to the ED.
closed_qa
Is chronic inhibition of phosphodiesterase type 5 cardioprotective and safe?
The myocardial effects of phosphodiesterase type 5 inhibitors (PDE5i) have recently received consideration in several preclinical studies. The risk/benefit ratio in humans remains unclear. We performed a meta-analysis of randomized, placebo-controlled trials (RCTs) to evaluate the efficacy and safety of PDE5i on cardiac morphology and function. From March 2012 to December 2013 (update: May 2014), we searched English-language studies from MEDLINE, EMBASE, Cochrane Central Register of Controlled Trials and SCOPUS-selecting RCTs of continuous PDE5i administration that reported cardiovascular outcomes: cardiac geometry and performance, afterload, endothelial function and safety. The pooled estimate of a weighted mean difference between treatment and placebo was obtained for all outcomes using a random effects model. A test for heterogeneity was performed and the I2 statistic calculated. Overall, 1,622 subjects were treated, with 954 randomized to PDE5i and 772 to placebo in 24 RCTs. According to our analysis, sustained PDE5 inhibition produced: (1) an anti-remodeling effect by reducing cardiac mass (-12.21 g/m2, 95% confidence interval (CI): -18.85; -5.57) in subjects with left ventricular hypertrophy (LVH) and by increasing end-diastolic volume (5.00 mL/m2; 95% CI: 3.29; 6.71) in non-LVH patients; (2) an improvement in cardiac performance by increasing cardiac index (0.30 L/min/m2, 95% CI: 0.202; 0.406) and ejection fraction (3.56%, 95% CI: 1.79; 5.33). These effects are parallel to a decline of N-terminal-pro brain natriuretic peptide (NT-proBNP) in subjects with severe LVH (-486.7 pg/ml, 95% CI: -712; -261). PDE5i administration also produced: (3) no changes in afterload parameters and (4) an improvement in flow-mediated vasodilation (3.31%, 95% CI: 0.53; 6.08). Flushing, headache, epistaxis and gastric symptoms were the commonest side effects.
This meta-analysis suggests for the first time that PDE5i have anti-remodeling properties and improve cardiac inotropism, independently of afterload changes, with a good safety profile. Given the reproducibility of the findings and tolerability across different populations, PDE5i could be reasonably offered to men with cardiac hypertrophy and early stage heart failure. Given the limited gender data, a larger trial on the sex-specific response to long-term PDE5i treatment is required.
closed_qa
Does obesity modify the relationship between exposure to occupational factors and musculoskeletal pain in men?
To analyze relationships between physical occupational exposures, post-retirement shoulder/knee pain, and obesity. 9 415 male participants (aged 63-73 in 2012) from the French GAZEL cohort answered self-administered questionnaires in 2006 and 2012. Occupational exposures retrospectively assessed in 2006 included arm elevation and squatting (never,<10 years, ≥10 years). "Severe" shoulder and knee pain were defined as ≥5 on an 8-point scale. BMI was self-reported. Mean BMI was 26.59 kg/m2 +/-3.5 in 2012. Long-term occupational exposure to arm elevation and squatting predicted severe shoulder and knee pain after retirement. Obesity (BMI≥30 kg/m2) was a risk factor for severe shoulder pain (adjusted OR 1.28; 95% CI 1.03, 1.90). Overweight (adjusted OR 1.71; 1.28,2.29) and obesity (adjusted OR 3.21; 1.90,5.41) were risk factors for severe knee pain. In stratified models, associations between long-term squatting and severe knee pain varied by BMI.
Obesity plays a role in relationships between occupational exposures and musculoskeletal pain. Further prospective studies should use BMI in analyses of musculoskeletal pain and occupational factors, and continue to clarify this relationship.
closed_qa
Is primary spontaneous pneumomediastinum a truly benign entity?
This study aimed to investigate the benignancy of primary spontaneous pneumomediastinum (PSP), and to establish an appropriate management strategy. Patients diagnosed with PSP between January 2003 and December 2013 were analysed retrospectively. From January 2013 onwards, a simplified protocol, with consensus for the management of PSP, was applied in our hospital. In total, 37 patients were identified as having PSP during the study period. Among them, 27 were enrolled prior to applying the new protocol. Among these patients, extra diagnostic tests, in addition to chest radiography (CXR) and chest computed tomography (CT), were performed in 15 patients (55.5%). In the pre-protocol decade, a total of 15 patients (55.5%) were initially fasted and 16 (59.3%) were administered antibiotics. Mean hospital stay was 2.9 days (range, 0-5 days). No patient developed complications during the hospital stay and outpatient follow up. Since the revised protocol was in practical use, 10 consecutive patients with PSP were enrolled and reviewed. No additional diagnostic imaging studies or procedures (except for CXR and chest CT) were performed in these patients; furthermore, diet was not restricted and prophylactic antibiotics were not prescribed. Mean hospital stay was 14.5 h (range, 1-34 h). No complications were observed in any of the patients.
Our management protocol (i.e. routine check of chest CT without any additional diagnostic tests, no special treatment, and early discharge with short-term follow up) may be safe and feasible for the treatment of PSP.
closed_qa
Does kidney transplantation to iliac artery deteriorate ischemia in the ipsilateral lower extremity with peripheral arterial disease?
This study was conducted to investigate the progression of lower extremity ischemia following kidney transplantation to iliac artery in patients with peripheral arterial disease. A retrospective chart review of all renal transplant patients completed at a university teaching medical center from January 2006 to December of 2011 was performed. A total of 219 patients underwent successful kidney transplantation to the common, external, or internal iliac artery. Pre- and post-transplantation ischemic changes in the ipsilateral lower extremity were reviewed and analyzed. Thirty-eight of the 219 patients had ipsilateral peripheral arterial disease and seven of them were symptomatic. Six of the seven symptomatic patients remained stable and one patient's rest pain improved, postoperatively. Eight patients developed new symptoms of ischemia 12 months later, including four with claudication, two with ischemic ulcers, and two with gangrene toes. The ulcers were healed following superficial femoral artery stenting and wound care. Toe amputation was performed in two patients with gangrene. No major amputation was required up to 48 months of follow-up.
Transplanted kidney to iliac artery does not significantly deteriorate ischemia in adults with ipsilateral lower extremity peripheral arterial disease. Late developed ischemic complications may be due to the progression of underlying arterial disease.
closed_qa
Are there preoperative factors related to a "soft pancreas" and are they predictive of pancreatic fistulas after pancreatic resection?
Soft pancreatic parenchyma is the most widely recognized risk factor for pancreatic fistula. We conducted this study to clarify if there are preoperative factors related to a soft pancreatic remnant and to establish if they are useful for predicting pancreatic fistula. This was a retrospective study of patients who underwent pancreatic resections at the Department of Surgical Sciences of the S. Orsola-Malpighi Hospital, Bologna, Italy. The factors considered were sex, age, co-morbidities, body mass index, American Society of Anesthesiologists score, characteristics of the pancreatic remnant, and preoperative diagnosis. Data from 208 patients were recorded. The risk factors predictive of a soft pancreatic remnant were BMI>24 kg/m(2) (P = 0.011), a Wirsung duct size ≤3 mm (P<0.001), and coexisting periampullary diseases (P<0.001). Using these factors, we developed a risk score model that was validated by considering the pancreatic fistula rate. The overall and clinically relevant pancreatic fistula rate increased with increasing score values (P = 0.002 and P = 0.028, respectively). Using a score cut-off value of six points, patients with a score ≥6 were considered to be at high risk.
Body mass index>24 kg/m(2), a Wirsung duct size<3 mm, and preoperative diagnosis represented the preoperative factors related to a soft pancreas. These factors proved useful in the building of a risk score model to predict the incidence of pancreatic fistula.
closed_qa
Failure of deactivation in the default mode network: a trait marker for schizophrenia?
Functional imaging studies in relatives of schizophrenic patients have had inconsistent findings, particularly with respect to altered dorsolateral prefrontal cortex activation. Some recent studies have also suggested that failure of deactivation may be seen. A total of 28 patients with schizophrenia, 28 of their siblings and 56 healthy controls underwent functional magnetic resonance imaging during performance of the n-back working memory task. An analysis of variance was fitted to individual whole-brain maps from each set of patient-relative-matched pair of controls. Clusters of significant difference among the groups were then used as regions of interest to compare mean activations and deactivations among the groups. In all, five clusters of significant differences were found. The schizophrenic patients, but not the relatives, showed reduced activation compared with the controls in the lateral frontal cortex bilaterally, the left basal ganglia and the cerebellum. In contrast, both the patients and the relatives showed significant failure of deactivation compared with the healthy controls in the medial frontal cortex, with the relatives also showing less failure than the patients. Failure of deactivation was not associated with schizotypy scores or presence of psychotic-like experiences in the relatives.
Both schizophrenic patients and their relatives show altered task-related deactivation in the medial frontal cortex. This in turn suggests that default mode network dysfunction may function as a trait marker for schizophrenia.
closed_qa
Are trajectories of self-regulation abilities from ages 2-3 to 6-7 associated with academic achievement in the early school years?
The aim of this study was to estimate the association between two key aspects of self-regulation, 'task attentiveness' and 'emotional regulation' assessed from ages 2-3 to 6-7 years, and academic achievement when children were aged 6-7 years. Participants (n = 3410) were children in the Longitudinal Study of Australian Children. Parents rated children's task attentiveness and emotional regulation abilities when children were aged 2-3, 4-5 and 6-7. Academic achievement was assessed using the Academic Rating Scale completed by teachers. Linear regression models were used to estimate the association between developmental trajectories (i.e. rate of change per year) of task attentiveness and emotional regulation, and academic achievement at 6-7 years. Improvements in task attentiveness between 2-3 and 6-7 years, adjusted for baseline levels of task attentiveness, child and family confounders, and children's receptive vocabulary and non-verbal reasoning skills at age 6-7 were associated with greater teacher-rated literacy [B = 0.05, 95% confidence interval (CI) = 0.04-0.06] and maths achievement (B = 0.04, 95% CI = 0.03-0.06) at 6-7 years. Improvements in emotional regulation, adjusting for baseline levels and covariates, were also associated with better teacher-rated literacy (B = 0.02, 95% CI = 0.01-0.04) but not with maths achievement (B = 0.01, 95% CI = -0.01-0.02) at 6-7 years. For literacy, improvements in task attentiveness had a stronger association with achievement at 6-7 years than improvements in emotional regulation.
Our study shows that improved trajectories of task attentiveness from ages 2-3 to 6-7 years are associated with improved literacy and maths achievement during the early school years. Trajectories of improving emotional regulation showed smaller effects on academic outcomes. Results suggest that interventions that improve task attentiveness when children are aged 2-3 to 6-7 years have the potential to improve literacy and maths achievement during the early school years.
closed_qa
Wall Enhancement of the Intracranial Aneurysms Revealed by Magnetic Resonance Vessel Wall Imaging Using Three-Dimensional Turbo Spin-Echo Sequence with Motion-Sensitized Driven-Equilibrium: A Sign of Ruptured Aneurysm?
Wall enhancement of saccular cerebral aneurysms has not been researched sufficiently. Our purpose of this study was to investigate the incidence of aneurysmal wall enhancement by the three-dimensional turbo spin-echo sequence with motion-sensitized driven equilibrium (MSDE-3D-TSE) imaging after gadolinium injection. We retrospectively reviewed the pre- and postcontrast MSDE-3D-TSE images of 117 consecutive patients with intracranial aneurysms from September 2011 to July 2013. A total of 61 ruptured and 83 unruptured aneurysms of 61 patients with subarachnoid hemorrhage (SAH) and 56 non-SAH patients were enrolled in this study. We evaluated the wall enhancement of each aneurysm on postcontrast MSDE-3D-TSE images compared with precontrast images. We classified the aneurysmal wall enhancement into three groups as "Strong enhancement," "Faint enhancement," and "No enhancement." "Strong/Faint enhancement" of the aneurysm was detected in 73.8/24.6 % of the ruptured aneurysms and 4.8/13.3 % of the unruptured aneurysms. "No enhancement" was found in 1.6 % of the ruptured aneurysms and 81.9 % of the unruptured aneurysms.
By magnetic resonance vessel wall imaging using the MSDE-3D-TSE sequence, wall enhancement was frequently observed on ruptured aneurysms. Therefore, aneurysmal wall enhancement may be an indicator of the ruptured condition, which is useful information for managing patients with SAH.
closed_qa
Feeding practices and child weight: is the association bidirectional in preschool children?
Parental feeding practices are associated with children's body mass index (BMI). It has been generally assumed that parental feeding determines children's eating behaviors and weight gain, but feeding practices could equally be a parent's response to child weight. In longitudinal analyses, we assessed the directionality in the relation between selected controlling feeding practices and BMI in early childhood. Participants were 4166 children from the population-based Generation R Study. BMI was measured at ages 2 and 6 y. With the use of the Child Feeding Questionnaire, parents reported on restriction, monitoring, and pressure to eat (child age: 4 y). BMI and feeding-behavior scales were transformed to SD scores. With the use of linear regression analyses, there was an indication that a higher BMI at age 2 y predicted higher levels of parental restriction (adjusted β = 0.07; 95% CI: 0.04, 0.10) and lower levels of pressure to eat (adjusted β = -0.20; 95% CI: -0.23, -0.17) 2 y later. Restriction at age 4 y positively predicted child BMI at 6 y of age, although this association attenuated to statistical nonsignificance after accounting for BMI at age 4 y (β = 0.01; 95% CI: -0.01, 0.03). Pressure to eat predicted lower BMI independently of BMI at age 4 y (β = -0.02; 95% CI: -0.04, -0.01). For both restriction and pressure to eat, the relation from BMI to parenting was stronger than the reverse (Wald's test for comparison: P = 0.03 and<0.001, respectively). Monitoring predicted a lower child BMI, but this relation was explained by confounding factors.
Although the feeding-BMI relation is bidirectional, the main direction of observed effects suggests that parents tend to adapt their controlling feeding practices in response to their child's BMI rather than the reverse. Therefore, some components of current programs aimed at preventing or treating unhealthy child weight may need to be carefully scrutinized, especially those targeting parental food-related restriction and pressure to eat.
closed_qa
Late-life depression in Rural China: do village infrastructure and availability of community resources matter?
This study aimed to examine whether physical infrastructure and availability of three types of community resources (old-age income support, healthcare facilities, and elder activity centers) in rural villages are associated with depressive symptoms among older adults in rural China. Data were from the 2011 baseline survey of the Chinese Health and Retirement Longitudinal Study (CHARLS). The sample included 3824 older adults aged 60 years or older residing in 301 rural villages across China. A score of 12 on the 10-item Center for Epidemiologic Studies Depression Scale was used as the cutoff for depressed versus not depressed. Village infrastructure was indicated by an index summing deficiency in six areas: drinking water, fuel, road, sewage, waste management, and toilet facilities. Three dichotomous variables indicated whether income support, healthcare facility, and elder activity center were available in the village. Respondents' demographic characteristics (age, gender, marital status, and living arrangements), health status (chronic conditions and physical disability), and socioeconomic status (education, support from children, health insurance, household luxury items, and housing quality) were covariates. Multilevel logistic regression was conducted. Controlling for individuals' socioeconomic status, health status, and demographic characteristics, village infrastructure deficiency was positively associated with the odds of being depressed among rural older Chinese, whereas the provision of income support and healthcare facilities in rural villages was associated with lower odds.
Village infrastructure and availability of community resources matter for depressive symptoms in rural older adults. Improving infrastructure, providing old-age income support, and establishing healthcare facilities in villages could be effective strategies to prevent late-life depression in rural China.
closed_qa
Nitrate Supplementation, Exercise, and Kidney Function: Are There Detrimental Effects?
Recently, dietary supplementation with inorganic nitrate (NO3) has been proposed to endurance athletes to increase their performance. However, it has been suggested that an excess of NO3 might be harmful. The present study analyzed the effect of NO3 supplementation on kidney function. Thirteen young male subjects performed a 20-min cycling exercise at 85% of the maximal oxygen capacity. Seven days before exercise, the subjects ingested either a placebo (Pl) or 450 mg of potassium nitrate (PN) per day. Venous blood samples and urine collections were collected before and immediately after exercise and after 60 min of recovery. Glomerular filtration rates (GFR) and clearances (Cl) were calculated from serum content and urine output for creatinine (Crn), albumin (Alb), and urea. Under resting conditions, GFR and all clearance measures did not differ between Pl and PN. Immediately after exercise, GFR remained stable in both Pl and PN, whereas Cl-urea decreased significantly (P<0.05) in Pl (-44%) and PN (-49%). Alb urine outputs were enhanced by 18- to 20-fold in Pl and PN, respectively (P<0.05). After the recovery period, GFR remained enhanced under Pl conditions, whereas Cl-urea returned to initial values in placebo and nitrate supplementation. Alb output and Cl-Alb remained enhanced under PN conditions.
These results mainly indicate that dietary nitrate supplementation over a week does not induce any specific kidney function modifications either at rest or during sustained submaximal exercise as compared with Pl.
closed_qa
Biceps Femoris Aponeurosis Size: A Potential Risk Factor for Strain Injury?
A disproportionately small biceps femoris long head (BFlh) proximal aponeurosis has been suggested as a risk factor for hamstring strain injury by concentrating mechanical strain on the surrounding muscle tissue. However, the size of the BFlh aponeurosis relative to BFlh muscle size, or overall knee flexor strength, has not been investigated. This study aimed to examine the relationship of BFlh proximal aponeurosis area with muscle size (maximal anatomical cross-sectional area (ACSAmax)) and knee flexor strength (isometric and eccentric). Magnetic resonance images of the dominant thigh of 30 healthy young males were analyzed to measure BFlh proximal aponeurosis area and muscle ACSAmax. Participants performed maximum voluntary contractions to assess knee flexion maximal isometric and eccentric torque (at 50° s and 350° s). BFlh proximal aponeurosis area varied considerably between participants (more than fourfold, range = 7.5-33.5 cm, mean = 20.4 ± 5.4 cm, coefficient of variation = 26.6%) and was not related to BFlh ACSAmax (r = 0.04, P = 0.83). Consequently, the aponeurosis/muscle area ratio (defined as BFlh proximal aponeurosis area divided by BFlh ACSAmax) exhibited sixfold variability, being 83% smaller in one individual than another (0.53 to 3.09, coefficient of variation = 32.5%). Moreover, aponeurosis size was not related to isometric (r = 0.28, P = 0.13) or eccentric knee flexion strength (r = 0.24, P ≥ 0.20).
BFlh proximal aponeurosis size exhibits high variability between healthy young men, and it was not related to BFlh muscle size or knee flexor strength. Individuals with a relatively small aponeurosis may be at increased risk of hamstring strain injury.
closed_qa
A stiff price to pay: does joint stiffness predict disability in an older population?
To describe the prevalence of joint stiffness and associated comorbidities in community-living older adults and to determine whether joint stiffness, independent of pain, contributes to new and worsening disability. Population-based cohort. Urban and suburban communities in the Boston, Massachusetts, area. Adults aged 70 and older (N = 765) underwent a baseline home interview and clinic examination, 680 participants completed the 18-month follow-up. Morning joint stiffness on most days in the past month was assessed in the arms, back, hips, and knees. Mobility limitations were measured using self-reported difficulty and the Short Physical Performance Battery (SPPB). The home interview and clinic examination included extensive health measures. Four hundred one participants reported morning joint stiffness, half of these with one site of stiffness and the other half with multisite stiffness. Twenty percent of participants with multisite stiffness and 50% with single site stiffness did not have a major stiffness-associated condition. After adjustment for pain severity and other covariates, multisite stiffness was associated with a 64% greater risk of developing new or worsening mobility difficulty (relative risk = 1.64, 95% confidence interval = 1.05-2.79). Those with multisite stiffness had declined more quickly in physical performance over the 18-month follow up.
Older adults with multisite stiffness are more likely to be at risk of disability than those without joint stiffness after accounting for pain severity and the presence of stiffness-associated conditions. Better assessment, along with strategies to prevent and treat multisite joint stiffness is needed to prevent or slow the progression of disability in elderly adults.
closed_qa
Radiographic Follow-up of DDH in Infants: Are X-rays Necessary After a Normalized Ultrasound?
Concerns about radiation exposure have created a controversy over long-term radiographic follow-up of developmental dysplasia of the hip (DDH) in infants who achieve normal clinical and ultrasonographic examinations. The purpose of this study was to assess the importance of continued radiographic monitoring by contrasting the incidence of residual radiographic dysplasia to the risks of radiation exposure. We reviewed a consecutive series of infants with idiopathic DDH presenting to our institution over 4 years. Infants with "normalized DDH" had achieved a stable clinical examination with an ultrasound revealing no signs of either hip instability or acetabular dysplasia. We excluded infants with persistently abnormal ultrasonographic indices, clinical examinations, or both by 6 months of age, including those requiring surgical reduction. Anteroposterior pelvic radiographs at approximately 6 and 12 months of age were then evaluated for evidence of residual radiographic acetabular dysplasia. Radiation effective dose was calculated using PCXMC software. We identified 115 infants with DDH who had achieved both normal ultrasonographic and clinical examinations at 3.1±1.1 months of age. At the age of 6.6±0.8 months, 17% of all infants demonstrated radiographic signs of acetabular dysplasia. Of infants left untreated (n=106), 33% had dysplasia on subsequent radiographs at 12.5±1.2 months of age. No significant differences were evident in either the 6- or 12-month rates of dysplasia between infants successfully treated with a Pavlik harness and infants normalizing without treatment but with a history of risk factors (P>0.05). The radiation effective dose was<0.01 mSv for the combined 6- and 12-month single-view anteroposterior radiographs of the pelvis.
The notable incidences of radiographic dysplasia after previous DDH normalization in our study cohort appear to outweigh the risks of radiation exposure. Our findings may warrant radiographic follow-up in this population of infants through at least walking age to allow timely diagnosis and early intervention of residual acetabular dysplasia.
closed_qa
Does video gaming affect orthopaedic skills acquisition?
Previous studies have suggested that there is a positive correlation between the extent of video gaming and efficiency of surgical skill acquisition on laparoscopic and endovascular surgical simulators amongst trainees. However, the link between video gaming and orthopaedic trauma simulation remains unexamined, in particular dynamic hip screw (DHS) stimulation. To assess effect of prior video gaming experience on virtual-reality (VR) haptic-enabled DHS simulator performance. 38 medical students, naïve to VR surgical simulation, were recruited and stratified relative to their video gaming exposure. Group 1 (n = 19, video-gamers) were defined as those who play more than one hour per day in the last calendar year. Group 2 (n = 19, non-gamers) were defined as those who play video games less than one hour per calendar year. Both cohorts performed five attempts on completing a VR DHS procedure and repeated the task after a week. Metrics assessed included time taken for task, simulated flouroscopy time and screw position. Median and Bonett-Price 95% confidence intervals were calculated for seven real-time objective performance metrics. Data was confirmed as non-parametric by the Kolmogorov-Smirnov test. Analysis was performed using the Mann-Whitney U test for independent data whilst the Wilcoxon signed ranked test was used for paired data. A result was deemed significant when a two-tailed p-value was less than 0.05. All 38 subjects completed the study. The groups were not significantly different at baseline. After ten attempts, there was no difference between Group 1 and Group 2 in any of the metrics tested. These included time taken for task, simulated fluoroscopy time, number of retries, tip-apex distance, percentage cut-out and global score.
Contrary to previous literature findings, there was no correlation between video gaming experience and gaining competency on a VR DHS simulator.
closed_qa
Do we need a femoral artery route for transvenous PDA closure in children with ADO-I?
The standard procedure in percutaneous closure of patent ductus arteriosus (PDA) with Amplatzer duct occluder-I (ADO-I) is transvenous closure guided by aortic access through femoral artery. The current study aims to compare the procedures for PDA closure with ADO-I: only transvenous access with the standard procedure. This study was designed retrospectively and 101 pediatric patients were included. PDA closure was done by only femoral venous access in 19 of them (group 1), arterial and venous access used in 92 patients (group 2) between 2004 to 2012 years. The position of the device and residual shunt in group1 was evaluated by the guidance of the aortogram obtained during the return phase of the pulmonary artery injection and guidance of transthoracic echocardiography. Shapiro-Wilk's test, Mann-Whitney U, chi-squared tests were used for statistical comparison. The procedure was successful in 18 (95%) patients in group 1 and 90 (98%) patients in group 2. Complications including the pulmonary artery embolization (n=1), protrusion to pulmonary artery (n=1), inguinal hematoma (n=3), bleeding (n=2) were only detected in group 2. In other words, while complications were observed in 7 (7.2%) patients in group 2, no minor/major complication was observed in group 1. Complete closure in group 1 was: in catheterization room 14 (77.8%), at 24th hour in 2 (11.1%), at first month in 2 (11.1%). Complete closure in group 2 was: 66 (73.4%) patients in the catheterization room, 21 (23.3%) at 24th hour, 3 (3.3%) at first month, complete closure occurred at the end of first month.
In percutaneouse PDA closure via ADO-I, this technique can be a choice for patients whose femoral artery could not be accessed, or access is impossible/contraindicated. But for the reliability and validity of this method, randomized multicenter clinical studies are necessary.
closed_qa
The SPORT value compass: do the extra costs of undergoing spine surgery produce better health benefits?
The Spine Patient Outcomes Research Trial aimed to determine the comparative effectiveness of surgical care versus nonoperative care by measuring longitudinal values: outcomes, satisfaction, and costs. This paper aims to summarize available evidence from the Spine Patient Outcomes Research Trial by addressing 2 important questions about outcomes and costs for 3 types of spine problem: (1) how do outcomes and costs of spine patients differ depending on whether they are treated surgically compared with nonoperative care? (2) What is the incremental cost per quality adjusted life year for surgical care over nonoperative care? After 4 years of follow-up, patients with 3 spine conditions that may be treated surgically or nonoperatively have systematic differences in value endpoints. The average surgical patient enjoys better health outcomes and higher treatment satisfaction but incurs higher costs.
Spine care is preference sensitive and because outcomes, satisfaction, and costs vary over time and between patients, data on value can help patients make better-informed decisions and help payers know what their dollars are buying.
closed_qa
Is preoperative pain duration important in spinal cord stimulation?
Conflicting data have been published as to whether the success rate of spinal cord stimulation (SCS) is inversely proportional to the time interval from the initial onset of symptoms to implantation. Recently, a new stimulation design called burst stimulation has been developed that seems to exert its effect by modulating both the medial and lateral pain pathways and has a better effect than tonic stimulation on global pain, back pain, and limb pain. We analyzed the effect of preoperative pain duration on the degree of pain suppression by both tonic and burst stimulation in a group of patients (n = 49) who underwent both tonic and burst SCS. Using Pearson correlation analysis and controlling for age and duration of SCS, no correlation could be found between the preoperative pain duration and the success of SCS, either for tonic or for burst SCS, as defined by a numeric rating scale for pain. Using a different analysis method, dividing patients into groups according to preoperative pain duration, the same absence of influence was found. Pain was better suppressed by burst stimulation than tonic stimulation, irrespective of the preoperative pain duration.
These results suggest that the duration of pain is not an exclusion criterion for SCS and that similar success rates can be obtained for longstanding pain and pain of more recent onset.
closed_qa
The most critical question when reading a meta-analysis report: Is it comparing apples with apples or apples with oranges?
While the number of meta-analyses published has increased recently, most of them have problems in the design, analysis, and/or presentation. An example of meta-analyses with a study selection bias is a meta-analysis of over 160,000 patients in 20 clinical trials, published in Eur Heart J in 2012 by van Vark, which concluded that the significant effect of renin-angiotensin-aldosterone system (RAAS) inhibition on all-cause mortality was limited to the class of angiotensin-converting enzyme inhibitors (ACEIs), whereas no mortality reduction could be demonstrated with angiotensin receptor blockers (ARBs). Here, we aimed to discuss how to select studies for a meta-analysis and to present our results of a re-analysis of the van Vark data. The data were re-analyzed in three steps: firstly, only ACEI/ARB-based studies (4 ACEI and 12 ARB studies) were included; secondly, placebo-controlled studies were excluded, and 10 studies left were analyzed; and thirdly, 2 studies that were retracted after the manuscript of van Vark had been published were excluded. The final analysis included 8 studies with ~65,000 patients (3 ACEI and 5 ARB studies). The hazard ratios for all-cause mortality and cardiovascular mortality were 0.992 (95% CI 0.899-1.095; p=0.875) and 1.017 (0.932-1.110; p=0.703) for the ACEI versus control group and 1.007 (0.958-1.059; p=0.778) and 0.967 (0.911-1.025; p=0.258) for the ARB versus control group in the first step. The results were similar in the second and third steps.
The studies to be included in meta-analyses, particularly comparing ACEIs and ARBs, should be chosen carefully.
closed_qa
Is duration of postoperative fasting associated with infection and prolonged length of stay in surgical patients?
Verify whether the postoperative fasting period increases the risk for infection and prolonged length of stay. Prospective cohort study. Elective surgery patients were included. Excluded: those with no conditions for nutritional assessment, admitted in minimal care units, as well as with<72h in-hospital stay. Postoperative fasting was recorded from the days of no nutrition therapy. The length of stay was considered prolonged when above the average according to the specialty and type of surgery. Logistic regression was used to assess associations and adjust for confounding factors. 521 patients were analyzed, 44.1% were fasted for a period ≥1 day, 91% for ≥3 days and 5.6% for more than 5 days. Patients with more than 5 days fasting were more eutrophic, more admitted to intensive care units, and had more postoperative surgical complications. After adjustment for confounding variables, it was noted that ≥1 day of postoperative fasting increased the infection risk by 2.04 (CI95%: 1.20 to 3.50), ≥3 days 2.81 (CI95%: 1.4-5.8), and in fasting for more than 5 days the infection risk was 2.88 times higher (CI95%: 1.17 to 7.16). The risk for prolonged hospitalization was 2.4 (CI95%: 1.48 to 3.77) among patients who had ≥1 day fasting, 4.44 (CI95%: 2.0 to 9.8)and 4.43 times higher (CI95%: 1.73 to 11.3) among patients with ≥3 days fasting and more than 5 days, respectively.
The longer duration of postoperative fasting was an independent risk factor both for infection and for prolonged hospital stay.
closed_qa
Can highly purified collagen coating modulate polypropylene mesh immune-inflammatory and fibroblastic reactions?
Collagen has been proposed to be a useful biomaterial, but previous attempts to combine meshes with a collagen membrane have failed. The objective was to verify the effect of high-purified collagen gel coating in the immune-inflammatory response, host collagen metabolism, and angiogenesis around polypropylene mesh. In 20 female Wistar rats were implanted, at one side of the abdominal wall, a monofilament polypropylene mesh (PP), and, on the other side, a mesh coated with a new highly purified collagen gel (PPC). The animals were divided into sub-groups and euthanized at 7, 14, 21, and 90 days after implantation. Immunohistochemical analysis was performed using interleukin 1 (IL-1), matrix metalloproteinases (MMP-2, MMP-3), surface antigen CD-31, and tumor necrosis factor (TNF-α). Objective analysis (percent reactive area, average density, and vessels concentration) was performed using AxioVision Software. Comparative analysis showed: higher vessel density in the PPC group after 14 days (p = 0.002); a decrease in the average density of MMP-2 in the PPC group after 21 and 90 days (p = 0.046); more stability in the behavior of MMP-3 in the PPC group throughout the periods with the percentage reactive area for MMP-3 showing a significant decrease just in the PP group after 14 and 90 days (p = 0.017), and also for MMP-3 average density, in which reduction was significant after 21 days in the PP group, but not until after 90 days in PPC group (p < 0.001).
Highly purified collagen coating causes significant changes in angiogenesis and in the immune reaction of metalloproteinase around mesh implants in rats. These findings can be useful for improving mesh biocompatibility for pelvic floor surgery if such effects could be properly controlled.
closed_qa
Do variable rates of alcohol drinking alter the ability to use transdermal alcohol monitors to estimate peak breath alcohol and total number of drinks?
Transdermal alcohol monitoring is a noninvasive method that continuously gathers transdermal alcohol concentrations (TAC) in real time; thus, its use is becoming increasingly more common in alcohol research. In previous studies, we developed models that use TAC data to estimate peak breath alcohol concentration (BrAC) and standard units consumed when the rate of consumption was tightly controlled. Twenty-two healthy participants aged 21 to 52 who reported consuming alcohol on 1 to 4 days per week were recruited from the community. The final study sample included 11 men and 8 women. Both TAC and BrAC were monitored while each participant drank 1, 2, 3, 4, and 5 beers in the laboratory on 5 separate days. In contrast to previous studies, a self-paced alcohol administration procedure was used. While there was considerable variation in the times it took to consume each beer, key TAC parameters were not affected by pace of drinking. TAC data were then used in combination with the previously derived equations and estimated peak BrAC and standard units of alcohol consumed.
Transdermal alcohol monitoring can be used to reliably estimate peak BrAC and standard number of units consumed regardless of the rate of consumption, further demonstrating its usefulness in clinical research.
closed_qa
Cross-border policy effects on alcohol outcomes: drinking without thinking on the u.s.-Mexico border?
Rates of alcohol-related outcomes are sensitive to policy differences in politically distinct, adjacent territories. Factors that shape these cross-border effects, particularly when the policy differences are longstanding, remain poorly understood. We compared the ability of 2 classes of variables with theoretical relevance to the U.S.-Mexico border context-bar attendance and alcohol-related social-cognitive variables-to explain elevated drinking on the U.S. side of the border relative to other areas of the United States. Data were collected from multistage cluster samples of adult Mexican Americans on and off the U.S.-Mexico border (current drinker N = 1,351). Structural equation models were used to test drinking context (frequency of bar attendance) and 6 different social-cognitive variables (including alcohol-related attitudes, norms, motives, and beliefs) as mediators of border effects on a composite drinking index. The border effect on drinking varied by age (with younger adults showing a stronger effect), consistent with previous findings and known risk factors in the region. Contrary to theoretical expectations, 6 different social-cognitive variables-despite relating strongly with drinking-were comparable in border and nonborder areas (within and across age) and played no role in elevated drinking on the border. Conversely, elevated drinking among border youth was mediated by bar attendance. This mediated moderation effect held after adjusting for potential sociodemographic and neighborhood-level confounders.
Increased drinking among U.S.-Mexico border youth is explained by patterns of bar attendance, but not by more permissive alcohol-related social-cognitive variables in border areas: Border youth attend bars and drink more than their nonborder counterparts, despite having comparable alcohol-related beliefs, attitudes, norms, and motives for use. Alcohol's heightened availability and visibility on both sides of the border may create opportunities for border youth to drink that otherwise would not be considered.
closed_qa
Are sulfonylurea and insulin therapies associated with a larger risk of cancer than metformin therapy?
Several meta-analyses of observational studies suggested that metformin use reduces cancer risk in type 2 diabetes. However, this result was not confirmed by the few available randomized controlled trials (RCTs), and many observational studies on metformin and cancer were potentially afflicted with time-related bias. We aimed to avoid this bias when comparing cancer incidence in users of sulfonylurea, insulin, and other diabetes medications, respectively, with cancer incidence in metformin users. In a retrospective observational study, we used the German Disease Analyzer database with patient data from general practices throughout Germany. The study sample included 22,556 patients diagnosed with type 2 diabetes. During the median follow-up time of 4.8 years, 1,446 (6.4%) patients developed any cancer. In Cox regression analyses with either monotherapies or first diabetes medications as drug exposure, users of sulfonylurea (or insulin or other antidiabetes medications) were compared with metformin users. In multivariable adjusted models, hazard ratios were 1.09 (95% CI 0.87-1.36) for sulfonylurea monotherapy, 1.14 (95% CI 0.85-1.55) for insulin monotherapy, and 0.94 (95% CI 0.67-1.33) for other diabetes medications compared with metformin monotherapy. Results were similar for comparison of first diabetes medications.
In a retrospective database analysis, taking into account potential time-related biases, no reduced cancer risk was found in metformin users. To clarify the association between diabetes medication and cancer risk, further well-designed observational studies and RCTs are needed.
closed_qa
Reassessing the impact of smoking on preeclampsia/eclampsia: are there age and racial differences?
To investigate the association between cigarette use during pregnancy and pregnancy-induced hypertension/preeclampsia/eclampsia (PIH) by maternal race/ethnicity and age. This retrospective cohort study was based on the U.S. 2010 natality data. Our study sample included U.S. women who delivered singleton pregnancies between 20 and 44 weeks of gestation without major fetal anomalies in 2010 (n = 3,113,164). Multivariate logistic regression models were fit to estimate crude and adjusted odds ratios and the corresponding 95% confidence intervals. We observed that the association between maternal smoking and PIH varied by maternal race/ethnicity and age. Compared with non-smokers, reduced odds of PIH among pregnant smokers was only evident for non-Hispanic white and non-Hispanic American Indian women aged less than 35 years. Non-Hispanic Asian/Pacific Islander women who smoked during pregnancy had increased odds of PIH regardless of maternal age. Non-Hispanic white and non-Hispanic black women 35 years or older who smoked during pregnancy also had increased odds of PIH.
Our study findings suggest important differences by maternal race/ethnicity and age in the association between cigarette use during pregnancy and PIH. More research is needed to establish the biologic and social mechanisms that might explain the variations with maternal age and race/ethnicity that were observed in our study.
closed_qa
Can we use incubators with atmospheric oxygen tension in the first phase of in vitro fertilization?
Aim of the present study was to compare two culture strategies used in our routine in vitro fertilization program. This is a retrospective analysis. Two culture systems were used in parallel and analysed retrospectively: 1) Use of atmospheric oxygen tension (~20 %) until insemination followed by use of low (~5 %) oxygen concentration; 2) Exclusive use of low oxygen concentration. Main outcome was the utilization rate defined as the number of transferred + vitrified embryos per inseminated oocytes. Secondary outcomes were clinical pregnancy and live birth rates. A total of 402 in vitro fertilization cycles were analyzed. Demographic and clinical data of patients belonging to the two culture systems were not significantly different. Utilization rate, cumulative clinical pregnancy rate and cumulative live birth rate per cycle was similar using two different oxygen concentration compared to exclusive use of low oxygen tension (37 % versus 39 %; 30 % versus 30 %; 23 % versus 28 %, respectively).
The use of a culture system with atmospheric oxygen tension from recovery of oocytes until insemination followed by culture in low oxygen gives results similar to exclusive use of low oxygen concentration.
closed_qa
Do sex and BMI predict or does stem design prevent muscle damage in anterior supine minimally invasive THA?
Cadaveric and clinical studies have suggested that, despite being touted as muscle-sparing, the direct anterior approach is still associated with muscle damage, particularly to the tensor fascia lata (TFL). Patient body mass index (BMI) and/or sex may also influence this parameter.QUESTIONS/ The purposes of the study were to determine if using a shorter femoral component reduces TFL damage or if patient sex or increasing BMI increases intraoperative TFL damage in direct anterior THA. Over a 1-year period, 599 direct anterior THAs were performed by three experienced anterior hip surgeons; of those, 421 direct anterior hips had complete data (70%) and were included in the study. The amount of visible damage to the TFL was recorded before closure. Two stem types were used, a standard-length flat-wedge taper (standard) or a 3-cm shortened version of the same stem (short). Stem selection was based on timeframe of the surgery, surgeon preference, or matching a previous implant type. During the study period, the three surgeons performed an additional 225 primary THAs with other approaches such that the direct anterior approach represented 73% of the THAs performed. A member of the operating team, either a fellow or physician assistant, graded the extent of damage based on a 0 to 3 scale. On this scale, 0 represented no muscle fiber damage, 1 superficial tearing, 2 deep tearing or maceration, and 3 complete tear or severe damage. Patient sex and BMI were recorded and compared with stem type and muscle damage scores. An ordinal logistic regression model was used for statistical analysis. After controlling for relevant confounding variables using logistic regression, we found that mean muscle damage was associated with male sex (0.93, SD 0.76 versus 0.70, SD 0.68; p<0.001) and increasing BMI levels (p<0.001). As BMI increased, more muscle damage also was found in men compared with women (p=0.05; odds ratio [OR], 1.029; 95% confidence interval [CI], 1.000-1.060). There was no overall difference in mean muscle damage between short and standard-length stems (0.78, SD 0.77 versus 0.85, SD 0.69, p=0.32); however, as BMI increased, less damage was seen with a short stem (p=0.04; OR, 0.968; 95% CI, 0.931-0.997).
Visible muscle damage occurred in most hips during anterior supine intermuscular hip arthroplasty. The clinical importance of this muscle damage requires further study, because some evidence suggests earlier restoration of gait and cessation of walking aids with direct anterior THA despite this damage; however, this was not studied here. Surgeons performing this approach can expect more difficulty and as a result possibly more damage to the TFL in patients with male sex and increased BMI. The use of a short stem can be considered for patients with increased BMI to limit damage to the TFL.
closed_qa
Can femoral rotation be localized and quantified using standard CT measures?
The terms "femoral anteversion" and "femoral torsion" have often been used interchangeably in the orthopaedic literature, yet they represent distinct anatomical entities. Anteversion refers to anterior tilt of the femoral neck, whereas torsion describes rotation of the femoral shaft. Together, these and other transverse plane differences describe what may be considered rotational deformities of the femur. Assessment of femoral rotation is now routinely measured by multiple axial CT methods. The most widely used radiographic technique (in which only two CT-derived axes are made, one through the femoral neck and one at the distal femoral condyles) may not accurately quantify proximal femoral anatomy nor allow identification of the anatomic locus of rotation.QUESTIONS/ (1) What CT methodology (a two-axis CT-derived technique, a three-axis technique adding an intertrochanteric axis--the "Kim method," or a volumetric three-dimensional reconstruction of the proximal femur) most accurately quantifies transverse plane femoral morphology; (2) localizes those deformities; and (3) is most reproducible across different observers? We constructed a high-definition femoral sawbones model in which osteotomies were performed at either the intertrochanteric region or femoral shaft. Transverse plane deformity was randomly introduced and CT-derived rotational profiles were constructed using three different CT methods. Accuracy and consistency of measurements of femoral rotation were calculated using p values and Fisher's exact test and intraclass correlation coefficients (ICCs). All three CT methodologies accurately quantified overall transverse plane rotation (mean differences 0.69° ± 3.88°, 0.69° ± 3.88°, and -1.09° ± 4.44° for the two-plane, Kim, and volumetric methods, respectively). However, use of a single neck and single distal femoral axis does not reliably identify the anatomic locus of rotation, whereas the Kim and volumetric methods do (p<0.0001). All three methods were highly reproducible between observers (ICCs of 0.9569, 0.9569, and 0.9359 for the traditional two-plane, Kim, and volumetric methods, respectively).
Only the Kim and volumetric methods can identify the anatomic location of transverse plane rotation and we recommend using one of the two techniques. Accurate anatomic localization of transverse plane rotation enables using precise anatomic terminology ("femoral torsion" versus "femoral [ante]version").
closed_qa
Do patient race and sex change surgeon recommendations for TKA?
Prior investigations have suggested that physician-related factors may contribute to differential use of TKA among women and ethnic minorities. We sought to evaluate the effect of surgeon bias on recommendations for TKA.QUESTIONS/ Using an experimental approach with standardized patient scenarios, we sought to evaluate surgeon recommendations regarding TKA, specifically to determine whether recommendations for TKA are influenced by (1) patient race, and (2) patient sex. We developed four computerized scenarios for all combinations of race (white or black) and sex (male or female) for otherwise similar patients with end-stage knee osteoarthritis. Patients gave an orthopaedic history of 2 years worsening pain with decreased functional status and failure of oral antiinflammatory medications and corticosteroid intraarticular injections. Orthopaedic surgeons attending the 2012 annual meetings of the New York State Society of Orthopaedic Surgeons and American Association of Hip and Knee Surgeons were recruited for the study. Surgeons passing an open recruitment table at each meeting were asked to participate. Of the 1111 surgeons in attendance at either meeting, 113 (10.2%) participated in the study. All participants viewed the "control" patient's story (white male) and were randomized to view one of the three "experimental" scenarios (white female, black male, black female). After viewing each scenario, the participants were anonymously asked whether they would recommend TKA. An a priori power analysis showed that 112 participants were needed to detect a 15% difference in the likelihood of recommending surgery for white versus nonwhite patients in the test scenarios evaluated with 90% power at a level of significance of 0.05. Of the 39 surgeons who viewed the white male plus black female scenario, there were 33 (85%) concordant responses (TKA offered to both patients) and six discordant responses (TKA offered to only one of the patients), with no effect of patient race and sex (p = 0.99). Of the 37 surgeons who viewed the white male plus black male scenario, there were 33 (89%) concordant responses and four discordant responses, with no effect of patient race (p = 0.32). Of the 37 surgeons who viewed the white male plus white female scenario, there were 30 (77%) concordant responses and seven discordant responses, with no effect of patient sex (p = 0.71).
After orthopaedic surgeons viewed video scenarios of patients with end-stage knee osteoarthritis, patient race and sex were not associated with a different likelihood of a surgical recommendation. Our findings support the notion that patient race and sex may be less influential on decision making when there are strong clinical data to support a decision. Physician bias may have a greater effect on decision making in situations where the indications for surgery are less clear.
closed_qa
Does neuroanatomy account for superior temporal dysfunction in early psychosis?
Neuroimaging studies of ultra-high risk (UHR) and first-episode psychosis (FEP) have revealed widespread alterations in brain structure and function. Recent evidence suggests there is an intrinsic relationship between these 2 types of alterations; however, there is very little research linking these 2 modalities in the early stages of psychosis. To test the hypothesis that functional alteration in UHR and FEP articipants would be associated with corresponding structural alteration, we examined brain function and structure in these participants as well as in a group of healthy controls using multimodal MRI. The data were analyzed using statistical parametric mapping. We included 24 participants in the FEP group, 18 in the UHR group and 21 in the control group. Patients in the FEP group showed a reduction in functional activation in the left superior temporal gyrus relative to controls, and the UHR group showed intermediate values. The same region showed a corresponding reduction in grey matter volume in the FEP group relative to controls. However, while the difference in grey matter volume remained significant after including functional activation as a covariate of no interest, the reduction in functional activation was no longer evident after including grey matter volume as a covariate of no interest. Our sample size was relatively small. All participants in the FEP group and 2 in the UHR group had received antipsychotic medication, which may have impacted neurofunction and/or neuroanatomy.
Our results suggest that superior temporal dysfunction in early psychosis is accounted for by a corresponding alteration in grey matter volume. This finding has important implications for the interpretation of functional alteration in early psychosis.
closed_qa
Physical restraint in the ICU: does it prevent device removal?
Physical restraint is frequently used in the intensive care setting but little is known regarding its clinical scenario and effectiveness in preventing adverse events (AEs), defined as device removal. We carried out a prospective observational study in three Intensive Care Units on 120 adult high-risk patients. The effectiveness of physical restraint was evaluated using the propensity score methodology in order to obtain comparable groups. Physical restraint was applied in 1371 of 3256 (43%) nurse shifts accounting for 120 patients. Substantial agitation, the nurse's judgement of insufficient sedation and sedative drug reduction were positively associated with physical restraint, whereas the presence of analgesics at admission, increased disease gravity and the treating hospital as the most substantial variable showed a negative association. Eighty-six AEs were observed in 44 patients. Quiet (SAS=1-4), unrestrained patients accounted for 40 cases, and agitated (SAS≥5) but physically restrained patients for 17 cases. The presence of any type of physical restraint had a protective effect against any type of AE (OR=0.28; CI 0.16-0.51). The observed AEs showed a limited impact on the patients' course of illness. No physical harm related to physical restraint was reported.
Physical restraint efficiently averts AEs. Its application is mainly driven by local habits. Typically, the almost recovered, apparently calm and hence unrestrained patient is at greatest risk for undesirable device removal. The control/interpretation of the patient's analgo-sedation might be inappropriate.
closed_qa
Diuretics, first-line antihypertensive agents: are they always safe in the elderly?
Diuretics are frequently recommended as antihypertensive agents. Some of the main side effects of diuretic therapy are hypokalaemia and hyponatremia. The objective of the study was to describe the frequency of hyponatremia in a group of elderly hypertensive patients treated with diuretics. The study included 202 elderly hypertensive patients (over 65 years old), treated with diuretics at least 4 weeks before hospitalization, consecutivly admitted in the Internal Medicine Clinic of the Clinical Emergency Hospital of Bucharest during a period of 4 months. The distribution by sex: 103 (52.28%) men and 94 (47.71%) women. The mean age of the patients was 72 ± 8 years. Incidence of hyponatremia was 24.87% (49 patients) in the whole group. From the 49 hyponatremic patients, 31 (63.26%) were women and 18 (36.73%) men. The distribution by age of hypertensive hyponatremic female patients was: between 65-70 years old--4 patients (12.90%), between 70-75 years old--7 patients (22.58%), over 75 years old--20 patients (64.51%). The distribution by age of hypertensive hyponatremic male patients was: between 65-70 years old--3 patients (16.66%), between 70-75 years old--2 patients(11.11%), over 75 years old--13 patients (72.22%). Most of the patients affected (73.46%) used a thiazide-type diuretic, the other 26.54% being on loop diuretics.
Elderly hypertensive patients were more likely to develop hyponatremia after age 75. Female patients had a higher frequency of hyponatremia than male patients. The main cause of hyponatremia in patients treated with diuretics was thiazide.
closed_qa
Does type of feeding in infancy influence lipid profile in later life?
To compare the lipid profile of exclusively breastfed and mixed-fed, term, appropriate for gestation age infants from 6 mo to 1 y of age. This prospective comparative study included one hundred ninety nine consecutive term healthy infants; 105 on exclusive breastfeeding (EBF) and 94 receiving mixed feeding (MF). These children were recruited at 6 mo of age and followed till 1 y of age. Serum lipid levels of babies were determined at recruitment (6 mo), 9 mo and 1 y of age. Statistical analysis was carried out using SPSS software. The mean total cholesterol (TC) at 6 mo in exclusively breastfed group (156.38 ± 50.42 mg/dl) was significantly higher than mixed fed group (139.5 ± 37.59 mg/dl). At 9 mo, high density lipoprotein cholesterol (HDL-C) and triglycerides (TG) levels were significantly different in EBF group than MF group. The lipid profile of both group of babies was comparable at 1 y of age. The HDL-C: LDL-C ratio was significantly different between the two groups (higher in breastfed group) at 1 y.
The present study highlights differential lipid profile of exclusively breastfed infants and mixed fed infants.
closed_qa
Does integrating research into the prosthetics and orthotics undergraduate curriculum enhance students' clinical practice?
Problem-based learning (where rather than feeding students the knowledge, they look for it themselves) has long been thought of as an ideal approach in teaching because it would encourage students to acquire knowledge from an undetermined medium of wrong and right answers. However, the effect of such approach in the learning experience of prosthetics and orthotics students has never been investigated. This study explores the implications of integrating problem-based learning into teaching on the students' learning experience via implementing a research-informed clinical practice module into the curriculum of last year prosthetics and orthotics undergraduate students at the University of Jordan (Amman, Jordan). Qualitative research pilot study. Grounded theory approach was used based on the data collected from interviewing a focus group of four students. Students have identified a number of arguments from their experience in the research-informed clinical practice where, generally speaking, students described research-informed clinical practice as a very good method of education.
Integrating problem-based learning into teaching has many positive implications. In particular, students pointed out that their learning experience and clinical practice have much improved after the research-informed clinical practice.
closed_qa
Nonresident parental influence on adolescent weight and weight-related behaviors: similar or different from resident parental influence?
Many parents do not live with, or have shared custody of, their adolescent children (i.e., nonresident parents). The degree of their influence on their children, as compared to parents who do live with their children the majority of the time (i.e. resident parents) has not been well-studied. The current study aimed to examine whether and how resident and nonresident parents' weight and weight-related behaviors are correlated with adolescents' weight and weight-related behaviors. Results will inform who may be important to include in adolescent obesity prevention interventions. Data from two linked population-based studies, EAT 2010 and F-EAT, were used for cross-sectional analyses. Resident parents (n = 200; 80% females; mean age =41.8), nonresident parents (n =200; 70% male; mean age =43.1), and adolescents (n =200; 60% girls; mean age =14.2 years) were socioeconomically and racially/ethnically diverse. Multiple regression models were fit to investigate the association between resident and nonresident parents' weight and weight-related behaviors and adolescents' weight and weight-related behaviors. Both resident and nonresident parents' BMI were significantly associated with adolescents' BMI percentile. Additionally, resident parents' sugar-sweetened beverage consumption and fruit and vegetable intake were significantly associated with adolescents' sugar-sweetened beverage intake and fruit and vegetable intake (p < 0.05), respectively. Furthermore, the association between nonresident parent physical activity and adolescent physical activity was marginally significant (p = 0.067). Neither resident nor nonresident parents' fast food consumption, breakfast frequency, or sedentary behaviors were significantly associated with adolescents' same behaviors.
These preliminary findings suggest that resident and nonresident parents may have slightly different influences on their adolescent children's weight-related behaviors. Longitudinal follow-up is needed to determine temporality of associations.
closed_qa
Is thermochemotherapy with the Synergo system a viable treatment option in patients with recurrent non-muscle-invasive bladder cancer?
To prospectively evaluate the outcome of combined microwave-induced bladder wall hyperthermia and intravesical mitomycin C instillation (thermochemotherapy) in patients with recurrent non-muscle-invasive bladder cancer. Between 2003 and 2009, 21 patients (median age 70 years, range 35-95 years) with recurrent non-muscle-invasive bladder cancer (pTaG1-2 n = 9; pTaG3 n = 3; pT1 n = 9; concurrent pTis n = 8) were prospectively enrolled. Of 21 patients, 15 (71%) had received previous intravesical instillations with bacillus Calmette-Guérin, mitomycin C and/or farmorubicin. Thermochemotherapy using the Synergo system was carried out in 11 of 21 patients (52%) with curative intent, and in 10 of 21 patients (48%) as prophylaxis against recurrence. The median number of thermochemotherapy cycles per patient was six (range 1-12). Adverse effects were frequent and severe: urinary urgency/frequency in 11 of 21 patients (52%), pain in eight of 21 patients (38%) and gross hematuria in five of 21 patients (24%). In eight of 21 patients (38%), thermochemotherapy had to be abandoned because of the severity of the adverse effects (pain in 3/8, severe bladder spasms in 2/8, allergic reaction in 2/8, urethral perforation in 1/8). Overall, six of 21 patients (29%) remained free of tumor after a median follow up of 50 months (range 1-120), six of 21 patients (29%) had to undergo cystectomy because of multifocal recurrences or cancer progression and seven of 21 patients (33%) died (2/7 of metastatic disease, 5/7 of non-cancer related causes).
Given the high rate of severe side-effects leading to treatment discontinuation, as well as the limited tumor response, thermochemotherapy should be offered only in highly selected cases of recurrent non-muscle-invasive bladder cancer.
closed_qa
Is endorectal ultrasound still useful for staging rectal cancer?
Staging in rectal carcinoma is important for planning treatment. Preoperative staging and treatment strategies have changed along with improvements in imaging techniques. The aim of this work is to evaluate the accuracy of endorectal ultrasound (ERUS) in rectal cancers, especially in low rectal cancers and stenotic cases. From January 2011 to December 2011, patients diagnosed with rectal cancer who were admitted to our endosonography unit for staging and who were operated on in our hospital were evaluated retrospectively. Patients who received neoadjuvant chemotherapy were excluded. Endosonographic staging was compared to postoperative pathological staging. In total, 38 patients (28 males, 10 females) were included. Their mean age was 57.6±11.3 years (27-75 years). Thirteen (34.2%) had stenotic lesions. The accuracy of ERUS for staging of lesions was evaluated according to pathology and was 73.7% overall (kappa coefficient = 0.317; p = 0.002). When patients were split into stenotic and non stenotic groups, the accuracy was 68% (kappa coefficient = 0.170; p = 0.125) for stenotic lesions and 84.6% (kappa coefficient = 0.606; p = 0.001) for non-stenotic lesions. Internal and external sphincter involvement were significantly correlated with the postoperative pathological evaluation in both groups.
Technological improvements in imaging methods have made the diagnosis and management of malignancies more precise. Low rectal tumours, have difficult characteristics for evaluation because of their unique location. Although ERUS has some disadvantages, it is still useful for T staging, evaluating sphincter involvement, and defining tumour size and distance from the anal verge. ERUS was less accurate for T staging of stenotic tumours, but the accuracy may still be within acceptable limits.
closed_qa
Capsule endoscopy capture rate: has 4 frames-per-second any impact over 2 frames-per-second?
One hundred patients with an indication for capsule endoscopy were included in the study. All procedures were performed with the new device (SB24). After an exhaustive evaluation of the SB24 videos, they were then converted to "SB2-like" videos for their evaluation. Findings, frames per finding, and clinical and therapeutic impact derived from video visualization were analyzed. Kappa index for interobserver agreement and χ (2) and Student's t tests for qualitative/quantitative variables, respectively, were used. Values of P under 0.05 were considered statistically significant. Eighty-nine out of 100 cases included in the study were ultimately included in the analysis. The SB24 videos detected the anatomical landmarks (Z-line and duodenal papilla) and lesions in more patients than the "SB2-like" videos. On the other hand, the SB24 videos detected more frames per landmark/lesion than the "SB2-like" videos. However, these differences were not statistically significant (P>0.05). Both clinical and therapeutic impacts were similar between SB24 and "SB2-like" videos (K = 0.954). The time spent by readers was significantly higher for SB24 videos visualization (P<0.05) than for "SB2-like" videos when all images captured by the capsule were considered. However, these differences become non-significant if we only take into account small bowel images (P>0.05).
More frames-per-second detect more landmarks, lesions, and frames per landmark/lesion, but is time consuming and has a very low impact on clinical and therapeutic management.
closed_qa
Quality of life research in endometrial cancer: what is needed to advance progress in this disease site?
Quality of life (QoL) in endometrial cancer (EC) is understudied. Incorporation of QoL questionnaires and patient-reported outcomes in clinical trials has been inconsistent, and the tools and interpretation of these measures are unfamiliar to most practitioners. In 2012, the Gynecologic Cancer InterGroup Symptom Benefit Working Group convened for a brainstorming collaborative session to address deficiencies and work toward improving the quality and quantity of QoL research in women with EC. Through literature review and international expert contributions, we compiled a comprehensive appraisal of current generic and disease site-specific QoL assessment tools, strengths and weaknesses of these measures, assessment of sexual health, statistical considerations, and an exploration of the unique array of histopathologic and clinical factors that may influence QoL outcomes in women with EC. This collaborative composition is the first publication specific to EC that addresses methodology in QoL research and the components necessary to achieve high quality QoL data in clinical trials. Future recommendations regarding (1) the incorporation of patient-reported outcomes in all clinical trials in EC, (2) definition of an a priori hypothesis, (3) utilization of validated tools and consideration of new tools corresponding to new therapies or specific symptoms, (4) publication within the same time frame as clinical outcome data, and (5) attempt to correct for disease site-specific potential confounders are presented.
Improved understanding of methodology in QoL research and an increased undertaking of EC-specific QoL research in clinical trials are imperative if we are to improve outcomes in women with EC.
closed_qa
Discordance between resident and faculty perceptions of resident autonomy: can self-determination theory help interpret differences and guide strategies for bridging the divide?
To identify and interpret differences between resident and faculty perceptions of resident autonomy and of faculty support of resident autonomy. Parallel questionnaires were sent to pediatric residents and faculty at the University of Rochester Medical Center in 2011. Items addressed self-determination theory (SDT) constructs (autonomy, competence, relatedness) and asked residents and faculty to rate and/or comment on their own and the other group's behaviors. Distributions of responses to 17 parallel Likert scale items were compared by Wilcoxon rank-sum tests. Written comments underwent qualitative content analysis. Respondents included 62/78 residents (79%) and 71/100 faculty (71%). The groups differed significantly on 15 of 17 parallel items but agreed that faculty sometimes provided too much direction. Written comments suggested that SDT constructs were closely interrelated in residency training. Residents expressed frustration that their care plans were changed without explanation. Faculty reported reluctance to give "passive" residents autonomy in patient care unless stakes were low. Many reported granting more independence to residents who displayed motivation and competence. Some described working to overcome residents' passivity by clarifying and reinforcing expectations.
Faculty and residents had discordant perceptions of resident autonomy and of faculty support for resident autonomy. When faculty restrict the independence of "passive" residents whose competence they question, residents may receive fewer opportunities for active learning. Strategies that support autonomy, such as scaffolding, may help residents gain confidence and competence, enhance residents' relatedness to team members and supervisors, and help programs adapt to accreditation requirements to foster residents' growth in independence.
closed_qa
Chest pain: if it hurts a lot, is heart attack more likely?
In previous studies including patients with suspected cardiac chest pain, those who had acute myocardial infarction (AMI) reported more severe chest pain than those without AMI. However, many patients with AMI present with very mild pain or discomfort. We aimed to investigate whether peak pain severity, as reported by patients in the Emergency Department, has any potential role in the risk stratification of patients with suspected cardiac chest pain. In this secondary analysis from a prospective diagnostic cohort study, we included patients presenting to the Emergency Department with suspected cardiac chest pain. Patients were asked to report their maximum pain severity using a 11-point numeric rating scale at the time of initial presentation. The primary outcome was a diagnosis of AMI, adjudicated by two independent investigators on the basis of reference standard (12 h) troponin testing. Of the 455 patients included in this analysis, 79 (17.4%) had AMI. Patients with AMI had marginally higher pain scores (eight, interquartile range 5-8) than those without AMI (seven, interquartile range 6-8, P=0.03). However, the area under the receiver operating characteristic curve for the numeric rating scale pain score was 0.58 (95% confidence interval 0.51-0.65), indicating poor overall diagnostic accuracy. AMI occurred in 12.1% of patients with pain score 0-3, 17.1% with pain score 4-6 and 18.8% with pain score 7-10. Among patients with AMI, pain score was not correlated with 12-h troponin levels (r=-0.001, P=0.99).
Pain score has limited diagnostic value for AMI. Scores should guide analgesia but shift the probability of AMI very little, and should not guide other clinical management.
closed_qa
Cyclophosphamide: As bad as its reputation?
Despite new treatment modalities, cyclophosphamide (CYC) remains a cornerstone in the treatment of organ or life-threatening vasculitides and connective tissue disorders. We aimed at analysing the short- and long-term side-effects of CYC treatment in patients with systemic autoimmune diseases. Chart review and phone interviews regarding side effects of CYC in patients with systemic autoimmune diseases treated between 1984 and 2011 in a single university centre. Adverse events were stratified according to the "Common Terminology Criteria for Adverse Events" version 4. A total of 168 patients were included. Cumulative CYC dose was 7.45 g (range 0.5-205 g). Gastro-intestinal side effects were seen in 68 events, hair loss occurred in 38 events. A total of 58 infections were diagnosed in 44/168 patients (26.2%) with 9/44 suffering multiple infections. Severity grading of infections was low in 37/58 cases (63.8%). One CYC-related infection-induced death (0.6%) was registered. Amenorrhoea occurred in 7/92 females (7.6%) with 5/7 remaining irreversible. In females with reversible amenorrhoea, prophylaxis with nafarelin had been administered. Malignancy was registered in 19 patients after 4.7 years (median, range 0.25-22.25) presenting as 4 premalignancies and 18 malignancies, 3 patients suffered 2 premalignancies/malignancies each. Patients with malignancies were older with a higher cumulative CYC dose. Death was registered in 28 patients (16.6%) with 2/28 probably related to CYC.
Considering the organ or life-threatening conditions which indicate the use of CYC, severe drug-induced health problems were rare. Our data confirm the necessity to follow-up patients long-term for timely diagnosis of malignancies. CYC side-effects do not per se justify prescription of newer drugs or biologic agents in the treatment of autoimmune diseases.
closed_qa
Is it all Lynch syndrome?
Mismatch repair-deficient (MMRD) colorectal cancer (CRC) and endometrial cancer (EC) may be suggestive of Lynch syndrome (LS). LS can be confirmed only by positive germ-line testing. It is unclear if individuals with MMRD tumors but no identifiable cause (MMRD+/germ-line-) have LS. Because LS is hereditary, individuals with LS are expected to have family histories of LS-related tumors. Our study compared the family histories of MMRD+/germ-line- CRC and/or EC patients with LS CRC and/or EC patients. A total of 253 individuals with an MMRD CRC or EC from one institution were included for analysis in one of four groups: LS; MMRD+/germ-line-; MMRD tumor with variant of uncertain significance (MMRD+/VUS); and sporadic MSI-H (MMRD tumor with MLH1 promoter hypermethylation or BRAF mutation). Family histories were analyzed utilizing MMRpro and PREMM1,2,6. Kruskal-Wallis tests were used to compare family history scores. MMRD+/germ-line- individuals had significantly lower median family history scores (MMRpro = 8.1, PREMM1,2,6 = 7.3) than did LS individuals (MMRpro = 89.8, PREMM1,2,6 = 26.1, P<0.0001).
MMRD+/germ-line- individuals have less suggestive family histories of LS than individuals with LS. These results imply that MMRD+/germ-line- individuals may not all have LS. This finding highlights the need to determine other causes of MMRD tumors so that these patients and their families can be accurately counseled regarding screening and management.Genet Med 17 6, 476-484.
closed_qa
Is the iPad suitable for image display at American Board of Radiology examinations?
The study aimed to determine the acceptability of the iPad 3 as a display option for American Board of Radiology (ABR) examinations. A set of 20 cases for each of nine specialties examined by the ABR was prepared. Each comprised between one and seven images and case information and had been used in previous ABR Initial Certification examinations. Examining radiologists (n = 119) at the ABR oral Initial Certification examinations reviewed sets from one or more specialties on both a 2 MP LED monitor and on the iPad 3 and rated the visibility of the salient image features for each case. The Wilcoxon signed rank test was performed to compare ratings. In addition, a thematic analysis of participants' opinions was undertaken. When all specialties were pooled, the iPad 3 ratings were significantly higher than the monitor ratings (p = 0.0217). The breast, gastrointestinal, genitourinary, and nuclear medicine specialties also returned significantly higher ratings for the visibility of relevant image features for the iPad 3. Monitor ratings were significantly higher for the vascular and interventional specialty, although no images were rated unacceptably poor on the iPad in this specialty.
The relevant image features were rated more visible on the iPad 3 than on the monitors overall. The iPad 3 was well accepted by a large majority of examiners and can be considered adequate for image display for examination in most or all specialties.
closed_qa
Does appendiceal diameter change with age?
The purposes of this study were to determine whether age-related changes in appendiceal diameter identified on CT and pathology are apparent on sonography and to assess the relationship between normal appendiceal diameter and patient-specific factors. Ultrasound examinations from 388 unique pediatric patients with normal appendixes, evenly distributed by age, were reviewed. Appendiceal diameter and wall thickness were correlated with patient age, sex, height, weight, and presence of enlarged lymph nodes. Mean (± SD) anteroposterior and transverse appendiceal diameters were 4.4 ± 0.9 and 5.1 ± 1.0 mm, respectively. Appendiceal diameter was normally distributed across the population but was not significantly associated with age. Centers for Disease Control and Prevention (CDC) weight percentile for age was the only statistically significant patient-specific predictor of transverse diameter (p = 0.001) and approached significance for anteroposterior diameter (p = 0.051). The presence of enlarged lymph nodes was a significant predictor of anteroposterior diameter (p = 0.029) and approached significance for transverse diameter (p = 0.07). Wall thickness was normally distributed across the population and was significantly associated with age (p = 0.011; effect size, -0.05 mm/y).
Appendiceal diameter measured on ultrasound is normally distributed in children and does not depend on age. Age-dependent diagnostic cutoffs for normal sonographic diameter are thus not needed. There is, however, a relationship between age and appendiceal wall thickness, suggesting the need for age-dependent diagnostic values if this criteria are to be used to diagnose appendicitis. Although the CDC weight percentile for age and the presence of enlarged lymph nodes affect appendiceal diameter on ultrasound, these effects are small and of doubtful clinical significance.
closed_qa
Does integrative medicine enhance balance in aging adults?
Postural balance and potentially fall risk increases among older adults living with neurological diseases, especially Parkinson's disease (PD). Since conventional therapies such as levodopa or deep brain stimulation may fail to alleviate or may even worsen balance, interest is growing in evaluating alternative PD therapies. The purpose of the current study was to assess improvement in postural balance in PD patients following electroacupuncture (EA) as an alternative therapy. 15 aging adults (71.2 ± 6.3 years) with idiopathic PD and 44 healthy age-matched participants (74.6 ± 6.5 years) were recruited. The PD participants were randomly assigned (at a ratio of 2:1) to an intervention (n = 10) or to a control group (n = 5). The intervention group received a 30-min EA treatment on a weekly basis for 3 weeks, while the control group received a sham treatment. Outcomes were assessed at baseline and after the final therapy. Measurements included balance assessment, specifically the ratio of medial-lateral (ML) center-of-gravity (COG) sway to anterior-posterior (AP) sway (COGML/AP) and ankle/hip sway during eyes-open, eyes-closed, and eyes-open dual-task trials, the Unified Parkinson's Disease Rating Scale (UPDRS), as well as quality of life, concerns for fall, and pain questionnaires. No difference was observed for the assessed parameters between the intervention and the control group at baseline. After treatment, an improvement in balance performance was observed in the intervention group. Compared with the healthy population, PD patients prior to treatment had larger COGML/AP sway with more dependency on upper-body movements for maintaining balance. Following EA therapy, COGML/AP sway was reduced by 31% and ankle/hip sway increased by 46% in the different conditions (p = 0.02 for the dual-task condition). The clinical rating revealed an overall improvement (p<0.01) in mentation, behavior, and mood (UPDRS part I, 49%), activities of daily living (UPDRS part II, 46%), and motor examination (UPDRS part III, 40%). There was a significant reduction (p<0.02) in the specific items regarding UPDRS fall status (67%) and rigidity (48%). Changes were small and nonsignificant in the controls (p>0.29).
This pilot study demonstrates improvement in rigidity and balance following EA. These preliminary results suggest EA could be a promising alternative treatment for balance disturbance in PD.
closed_qa
Can preoperative scoring systems be applied to Asian hip fracture populations?
Hip fractures in the elderly are a major cause of morbidity and mortality. Determining which patients will benefit from hip fracture surgery is crucial to reducing mortality and morbidity. Our objectives are: 1) to define the rate of index admission, 1-month and 1-year mortality in all hip fracture patients, and 2) to apply the Nottingham Hip Fracture Score (NHFS) to determine validity in an Asian population. This is a prospective cohort study of 212 patients with hip fractures above 60 years from September 2009 to April 2010 for 1-year. Sociodemographic, prefracture comorbidity and data on functional status was collected on admission, and at intervals after discharge. The main outcome measures were mortality on index admission, 1 month and 12 months after treatment. In our study, the overall mortality at 1-month and 1-year after surgery was 7.3% and 14.6% respectively. Surgically treated hip fracture patients had lower odds ratio (OR) for mortality as compared to conservatively treated ones. The OR was 0.17 during index admission, 0.17 at 1-month, and 0.18 at 12-months after discharge. These were statistically significant. Adjustments for age, gender, and duration to surgery were taken into account. The NHFS was found to be a good predictor of 1-month mortality after surgery.
Surgically treated hip fracture patients have a lower OR for mortality than conservatively managed ones even up to 1-year. The NHFS has shown to predict 1-month mortality accurately for surgically treated hip fracture patients, even for our Asian population. It can be used as a tool for clinicians at the individual patient level to communicate risk with patients and help plan care for fracture patients.
closed_qa
Is feeling extraneous from one's own body a core vulnerability feature in eating disorders?
To identify core vulnerability features capable of discriminating subjects who are more prone to develop eating disorders. A nonclinical group composed of 253 university students was studied by means of the Identity and Eating Disorders questionnaire (IDEA), exploring abnormal attitudes toward one's own body and difficulties in the definition of one's own identity, the Body Uneasiness Test (BUT) and different self-reported questionnaires evaluating the specific and general psychopathology of eating disorders. The results were compared with those of a clinical eating disorder group. In the student sample, a group composed of 35 subjects with abnormal eating patterns and a group (218 subjects) without such features were identified. The IDEA total and subscale scores were found to be significantly higher in subjects with abnormal eating patterns than in subjects without them (all p<0.001). Positive correlations between the IDEA total and subscale scores and the BUT global score were observed in both groups (all p<0.01). The comparison of the scores on the IDEA between the clinical group (patients with full-blown eating disorders) and the subjects with abnormal over-threshold eating patterns yields a significant difference in the 'feeling extraneous from one's own body' subscale of the IDEA.
The IDEA resulted in being a valid instrument to identify a vulnerability to eating disorders in subjects with abnormal eating patterns in the general population and to recognize the presence of a significant discomfort related to the body. Feeling extraneous from one's own body is the experience that discriminates most between clinical and nonclinical subjects.
closed_qa
Does a medial retraction blade transmit direct pressure to pharyngeal/esophageal wall during anterior cervical surgery?
A prospective study of 25 patients who underwent anterior cervical surgery. To assess retraction pressure and the exposure of pharyngeal/esophageal (P/E) wall to the medial retractor blade to clarify whether medial retraction causes direct pressure transmission to the P/E wall. Retraction pressure on P/E walls has been used to explain the relation between the retraction pressure and dysphagia or the efficacies of new retractor blades. However, it is doubtful whether the measured pressure represent real retraction pressure on the P/E wall because exposure of the P/E in the surgical field could be reduced by the shielding effect of thyroid cartilage. Epi- and endoesophageal pressures were serially measured using online pressure transducers 15 minutes before retraction, immediately after retraction, and 30 minutes after retraction. To measure the extent of P/E wall exposure to pressure transducer, we used posterior border of thyroid cartilage as a landmark. Intraoperative radiograph was used to mark the position of the posterior border of thyroid cartilage. We checked out the marked location on retractors by measuring the distance from distal retractor tip. The mean epiesophageal pressure significantly increased after retraction (0 mmHg: 88.7 ± 19.6 mmHg: 81.9 ± 15.3 mmHg). The mean endoesophageal pressure minimally changed after retraction (9.0 ± 6.6 mmHg: 15.7 ± 13.8 mmHg: 17.0 ± 14.3 mmHg). The mean location of the posterior border of thyroid cartilage was 7.3 ± 3.5 mm on the retractor blade from the tip, which means epiesophageal pressure was measured against the posterior border of thyroid cartilage, not against the P/E wall.
We suggest that a medial retraction blade does not transmit direct pressure on P/E wall due to minimal wall exposure and intervening thyroid cartilage. Our result should be considered when measuring retraction pressure during anterior cervical surgery or designing novel retractor systems.
closed_qa
Is organizational progress in the EFQM model related to employee satisfaction?
To determine whether there is greater employee satisfaction in organisations that have made more progress in implementation of the European Foundation for Quality Management (EFQM) model. A series of cross-sectional studies (one for each assessment cycle) comparing staff satisfaction survey results between groups of healthcare organisations by degree of implementation of the EFQM model (assessed in terms of external recognition of management quality in each organisation). 30 healthcare organisations including hospitals, primary care and mental health providers in Osakidetza, the Basque public health service. Employees of 30 Osakidetza organisations. Progress in implementation of EFQM model. Scores in 9 dimensions of employee satisfaction from questionnaires administered in healthcare organisations in 4 assessment cycles between 2001 and 2010. Comparing satisfaction results in organisations granted Gold or Silver Q Awards and those without this type of external recognition, we found statistically significant differences in the dimensions of training and internal communication. Then, comparing recipients of Gold Q Awards with those with no Q Certification, differences in leadership style and in policy and strategy also emerged as significant.
Progress of healthcare organisations in the implementation of the EFQM Excellence Model is associated with increases in their employee satisfaction in dimensions that can be managed at the level of each organisation, while dimensions in which no statistically significant differences were found represent common organisational elements with little scope for self-management.
closed_qa
Does medication abuse in patients with chronic migraine influence the effectiveness of preventive treatment with topiramate?
Patients with chronic migraine (CM) and medication abuse are difficult to treat, and have a greater tendency towards chronification and a poorer quality of life than those with other types of headache.AIM: To evaluate whether the presence of medication abuse lowers the effectiveness of topiramate. A series of patients with CM were grouped according to whether they met abuse criteria or not. They were advised to stop taking the drug that they were abusing. Treatment was adjusted to match their crises and preventive treatment with topiramate was established from the beginning. The number of days with headache and intense migraine in the previous month and at four months of treatment was evaluated. In all, 262 patients with CM criteria were selected and 167 (63.7%) of them fulfilled abuse criteria. In both groups there was a significant reduction in the number of days with headache/month and number of migraine attacks/month at the fourth month of treatment with topiramate. The percentage of reduction in the number of days with headache/month in CM without abuse was 59.3 ± 36.1%, and with abuse, 48.7 ± 41.7% (p = 0.0574). The percentage of reduction in the number of days with intense migraine/month in CM without abuse was 61.2%, and with abuse, 50% (p = 0.0224). Response rate according to the number of days with headache/month in CM without abuse was 69%, and with abuse, 57%. Response rate according to the number of intense migraines/month in CM without abuse was 76.8%, and in CM with abuse, 61% (p = 0.0097).
Topiramate was effective in patients with CM with and without medication abuse, although effectiveness is lower in the latter case.
closed_qa
Reliability of proliferation assessment by Ki-67 expression in neuroendocrine neoplasms: eyeballing or image analysis?
The latest WHO classification for neuroendocrine neoplasms (NEN) of the gastrointestinal tract defines grade according to Ki-67 and mitotic indices. Some have questioned the reproducibility and thus the reliability of Ki-67 assessment. We therefore investigated the accuracy of this proliferation marker in NEN. The Ki-67 index of tumor specimens of NEN (n = 73) was assessed by two pathologists as in routine practice with eyeballing and twice by image analysis using ImageJ freeware at different magnifications. RESULTS were correlated with overall survival. The intraclass correlation coefficient (ICC) between pathologists was 0.88. The ICC for the measurements using image analysis was 0.85. The ICC between all four measurements (pathologists and ImageJ) was 0.80. If the Ki-67 index was translated to grade as prescribed by the current WHO classification (<3% = grade 1, 3-20% = grade 2,>20% = grade 3), kappa was between 0.61 and 0.75. Grades based on pathologist scoring were often (16-29%) higher than grades assigned by image analysis (p<0.001). Grade was significantly correlated with survival (p<0.0001) irrespective of the way Ki-67 was assessed.
Assessment of the Ki-67 index by eyeballing correlates remarkably well with the Ki-67 index as calculated by image analysis and is therefore an accurate parameter. Moreover, it is significantly related to survival irrespective of the method used. Yet if the Ki-67 index is translated to grade, the grade should be interpreted with caution due to values around threshold levels.
closed_qa
The frontal branch of the facial nerve: can we define a safety zone?
The temporal branch of the facial nerve, a particularly important branch in facial expression, is commonly exposed to surgical trauma. The frontal branch is the most important branch of the temporal branch in the clinical point of view. However, it does not really define in the international nomenclature. The objective of this study was to clearly identify this branch, to perform a cartography of the crossing areas of this branch; and therefore to define statistically a zone of safety within the fronto-temporal region. We used 12 fresh cadavers to perform 24 facial nerve dissections. After the identification of the facial nerve, the branches of the temporofacial trunk were identified, dissected and followed till their penetration. We measured the relationship of the frontal branch with the zygomatic arch, temporal vessels and lateral border of the orbit. We conducted a statistical study to assess the risk of injury of this branch within the temporal region. We observed an important variability in the distribution of this branch in the temporal region. We defined three zones of decreasing safety at the level of three interest landmarks: at the level of the inferior part of the zygomatic arch, we estimated an elevated risk of nerve injury (>85%) from 22.6 to 26.06 mm in front of the tragus; at the level of the superior part of the zygomatic arch, we estimated an elevated risk of nerve injury (>85%) from 27.46 to 30.43 mm in front of the tragus; at the level of the lateral border of the orbit, we estimated an elevated risk of nerve injury (>85%) from 16.20 to 19.17 mm behind this landmark.
There exists no real area of anatomical safety in the temporal region. It seems, however, possible to define areas of relative safety that would be of great help for the surgeon or the morphologist wishing to approach pathologies of this region.
closed_qa
Rate and timing of spontaneous resolution in a vitreomacular traction group: should the role of watchful waiting be re-evaluated as an alternative to ocriplasmin therapy?
The incidence of spontaneous resolution of vitreomacular traction (VMT) is low in studies of Ocriplasmin that have had a limited follow-up. Previous studies did not look for morphological parameters in the natural history using spectral-domain ocular coherence tomography (SD-OCT) imaging. The purpose of this study was to investigate how often and when spontaneous VMT resolution occurs in candidates for Ocriplasmin therapy. The study is a retrospective chart review of patients who would have high chances of a benefit by an Ocriplasmin injection, without epiretinal membrane or vitreomacular adhesion of 1500 µm or more on SD-OCT. Main outcome measures were the frequency of complete VMT resolution and the best corrected visual acuity seen in the natural history. Out of the 46 patients that were included after screening 889 SD-OCT images, 20 were found to exhibit spontaneous resolution during the follow-up period (median: 594 days, 95% CI 567 to 719 days), the majority after 6-12 months of observation (95% CI 266 to 617 days). The group with spontaneous VMT resolution and a mean improvement of one line in best corrected visual acuity included a few patients losing vision by macular hole formation. In the absence of resolution, patients lost on average one early treatment diabetic retinopathy study letter per year. Younger age was found to increase the chance of spontaneous resolution.
A shorter follow-up might underestimate the incidence of spontaneous VMT resolution as the functional outcome of watchful waiting. The likelihood of resolution does not seem to decrease after 12 months.
closed_qa
Do environmental factors modify the genetic risk of prostate cancer?
Many SNPs influence prostate cancer risk. To what extent genetic risk can be reduced by environmental factors is unknown. We evaluated effect modification by environmental factors of the association between susceptibility SNPs and prostate cancer in 1,230 incident prostate cancer cases and 1,361 controls, all white and similar ages, nested in the Prostate, Lung, Colorectal, and Ovarian (PLCO) Cancer Trial. Genetic risk scores were calculated as number of risk alleles for 20 validated SNPs. We estimated the association between higher genetic risk (≥12 SNPs) and prostate cancer within environmental factor strata and tested for interaction. Men with ≥12 risk alleles had 1.98, 2.04, and 1.91 times the odds of total, advanced, and nonadvanced prostate cancer, respectively. These associations were attenuated with the use of selenium supplements, aspirin, ibuprofen, and higher vegetable intake. For selenium, the attenuation was most striking for advanced prostate cancer: compared with<12 alleles and no selenium, the OR for ≥12 alleles was 2.06 [95% confidence interval (CI), 1.67-2.55] in nonusers and 0.99 (0.38-2.58) in users (Pinteraction = 0.031). Aspirin had the most marked attenuation for nonadvanced prostate cancer: compared with<12 alleles and nonusers, the OR for ≥12 alleles was 2.25 (1.69-3.00) in nonusers and 1.70 (1.25-2.32) in users (Pinteraction = 0.009). This pattern was similar for ibuprofen (Pinteraction = 0.023) and vegetables (Pinteraction = 0.010).
This study suggests that selenium supplements may reduce genetic risk of advanced prostate cancer, whereas aspirin, ibuprofen, and vegetables may reduce genetic risk of nonadvanced prostate cancer.
closed_qa
Central venous oxygen saturation: a potential new marker for circulatory stress in haemodialysis patients?
Haemodialysis causes recurrent haemodynamic stress with subsequent ischaemic end-organ dysfunction. As dialysis prescriptions/schedules can be modified to lessen this circulatory stress, an easily applicable test to allow targeted interventions in vulnerable patients is urgently required. Intra-dialytic central venous oxygen saturation (ScvO2) and clinical markers (including ultrafiltration, blood pressure) were measured in 18 prevalent haemodialysis patients. Pre-dialysis ScvO2 was 63.5 ± 13% and fell significantly to 56.4 ± 8% at end dialysis (p = 0.046). Ultrafiltration volume, a key driver of dialysis-induced myocardial ischaemia, inversely correlated to ScvO2 (r = -0.680, p = 0.015).
This initial study demonstrates ScvO2 sampling is practical, with a potential clinical utility as an indicator of circulatory stress during dialysis.
closed_qa
Is local hypoperfusion the reason for transient neurological deficits after STA-MCA bypass for moyamoya disease?
Hyperperfusion is believed to be the cause of transient neurological events (TNEs) in patients with moyamoya disease (MMD) who have undergone an extracranial-to-intracranial (EC-IC) bypass between the superficial temporal artery (STA) and the middle cerebral artery (MCA). The objective of this study was to evaluate this possibility by analyzing cerebral blood flow (CBF) data obtained with thermal diffusion probes used at the authors' center. The authors examined postoperative cerebral perfusion in 31 patients with MMD who underwent a direct EC-IC STA-MCA bypass. A Hemedex Q500 flow probe was placed in the frontal lobe adjacent to the bypass and connected to a Bowman cerebral perfusion monitor, and CBF data were statistically analyzed using JMP 8.0.2 software. Seven patients experienced a TNE after surgery in the left hemisphere (that is, after left-sided surgery), manifesting as dysphasia approximately 24 hours postoperatively and which had improved by 48 hours. No TNEs were observed after right-sided surgeries. Operative and postoperative CBFs in the left side with the TNE were compared with those in the left side with no TNE and on the right side. A detailed analysis of 64,980 minute-by-minute flow observations showed that the initial postbypass CBF was higher on the left side where the TNEs occurred. This CBF increase was followed by a widely fluctuating pattern and a statistically significant and sharp drop in perfusion (p<0.001, mean difference of CBF between groups, paired t-test) associated with a TNE not observed in the other 2 groups.
On the basis of the authors' initial observations, an early-onset altered pattern of CBF was identified. These findings suggest local hypoperfusion as the cause of the TNEs. This hypoperfusion may originate from competing blood flows resulting from impaired cerebral autoregulation and a fluctuating flow in cerebral microcirculation.
closed_qa
Translating pharmacological findings from hypothyroid rodents to euthyroid humans: is there a functional role of endogenous 3,5-T2?
During the last two decades, it has become obvious that 3,5-diiodothyronine (3,5-T2), a well-known endogenous metabolite of the thyroid hormones thyroxine (T4) or triiodothyronine (T3), not only represents a simple degradation intermediate of the former but also exhibits specific metabolic activities. Administration of 3,5-T2 to hypothyroid rodents rapidly stimulated their basal metabolic rate, prevented high-fat diet-induced obesity as well as steatosis, and increased oxidation of long-chain fatty acids. The aim of the present study was to analyze associations between circulating 3,5-T2 in human serum and different epidemiological parameters, including age, sex, or smoking, as well as measures of anthropometry, glucose, and lipid metabolism. 3,5-T2 concentrations were measured by a recently developed immunoassay in sera of 761 euthyroid participants of the population-based Study of Health in Pomerania. Subsequently, analysis of variance and multivariate linear regression analysis were performed. Serum 3,5-T2 concentrations exhibited a right-skewed distribution, resulting in a median serum concentration of 0.24 nM (1st quartile: 0.20 nM; 3rd quartile: 0.37 nM). Significant associations between 3,5-T2 and serum fasting glucose, thyrotropin (TSH), as well as leptin concentrations were detected (p<0.05). Interestingly, the association to leptin concentrations seemed to be mediated by TSH. Age, sex, smoking, and blood lipid profile parameters did not show significant associations with circulating 3,5-T2.
Our findings from a healthy euthyroid population may point toward a physiological link between circulating 3,5-T2 and glucose metabolism.
closed_qa
Does rectum and bladder dose vary during the course of image-guided radiotherapy in the postprostatectomy setting?
To assess the variations in actual doses delivered to the rectum and bladder in the course of postprostatectomy radiotherapy using kilovoltage-cone-beam computed tomography datasets acquired during image-guided radiotherapy. Twenty consecutive patients treated with intensity-modulated or intensity-modulated arc therapy to the prostate bed were retrospectively evaluated. Both the planning tomography and kilovoltage-cone-beam computed tomography were acquired with an empty rectum and a half-full bladder. Target localization was performed on the basis of soft tissue matching using cone-beam computed tomography scans before each treatment fraction. A total of 16 cone-beam computed tomography scans per patient (acquired at the first 5 fractions and twice weekly thereafter) were used for the assessments. The bladder and rectum were re-contoured offline on each cone-beam computed tomography scan by a single physician, and the delivered doses were recalculated. The variations in certain dose-volume parameters for the rectum and bladder (BD2cc, RD 2cc, V40%, V50%, V60%, V65%) were analyzed using the paired t test. Most of the dose volume variations for rectum and bladder were significantly higher than predicted (P<0.05) for the 320 kilovoltage-cone-beam computed tomography sets, except for the doses received by 2 cc of the bladder and V50 and V60 of the rectum. The dose-volume parameters of the bladder did not meet our criteria of V65 ≤25% and V40 ≤50% in 10% and 20% of the patients, respectively. None of the dose-volume histograms showed rectal V65 ≥17%; however, the rectal V40 ≤35% dose constraint was not met in 11 patients. For all patients, the ANOVA test revealed no significant difference between the variations.
Actual doses delivered during treatment were found to be higher than predicted, but the majority of calculated bladder and rectal doses remained in the limits of our plan acceptance criteria. Interfraction variability of the rectum and bladder is a major concern in the postprostatectomy radiotherapy setting, even when patients are instructed about rectal and bladder preparation before the radiotherapy course. Image guidance with cone-beam computed tomography at each treatment fraction may offer a viable tool to account for interfraction variations of the rectum and bladder throughout the treatment course.
closed_qa
Blood alcohol concentration in intoxicated patients seen in the emergency department: does it influence discharge decisions?
The purpose of this study was to investigate whether blood alcohol concentration (BAC) measurement was routinely requested in emergency departments and whether the observation period in the emergency department allowed sufficient time for alcohol elimination before the patient was discharged. A retrospective review of medical records of all emergency alcohol-related admissions over a 12-month period from January 2012, in patients older than 18 years, was conducted. We estimated BAC at discharge for each patient by using the following formula: [BAC at admission--(length of stay × 15)]. Then, we focused on patients discharged from the emergency department with an estimated BAC greater than 50 mg/100 ml because of the risk of subsequent legal proceedings, because this is the legal limit for driving in France. A total of 907 patients admitted for acute alcohol intoxication (F10.0) were included, of whom 592 were male. Women were more likely to be admitted at night. The mean length of stay was 18.7 hours. BAC was measured in 893 patients. Patients ages 35-49 years had the highest measured BAC. No repeat BAC was taken before the discharge decision. Three hundred thirteen patients were discharged with an estimated BAC above 50 mg/100 ml.
Emergency physicians routinely requested BAC at admission but did not request alcohol kinetics while the patient was under observation. The discharge decision was based on clinical judgment. Doctors who do not advise patients appropriately before discharge may be guilty of negligence in their duty of care.
closed_qa
Do positive alcohol expectancies have a critical developmental period in pre-adolescents?
Positive outcome expectancies have been shown to predict initiation of alcohol use in children and to mediate and moderate the relationship between dispositional variables and drinking behavior. Negative outcome expectancies for alcohol appear to weaken as children progress to middle adolescence, but positive expectancies tend to increase during this time. Positive alcohol expectancies have been found to increase in children in third and fourth grades, indicating what some investigators have termed a possible critical period for the development of positive expectancies. In the present study, we assessed alcohol expectancies at baseline, 6, 12, and 18 months in 277 second-through sixth-grade students. Children completed the Alcohol Expectancy Questionnaire-Adolescent. Univariate analyses of covariance were conducted. There were significant main effects for grade on positive alcohol-expectancy change for Global Positive Transformations at 12 and 18 months, Social Behavior Enhancement or Impediment at 6 and 12 months, and Relaxation/Tension Reduction at 6 and 18 months, whereby a consistent pattern emerged in that lower grades did not differ from each other, but they differed significantly from the higher grades.
Data support a critical developmental period for positive alcohol expectancies, with the greatest change observed between third and fourth grade and between fourth and fifth grade, and only in those expectancies clearly describing positive outcomes (e.g., Relaxation/Tension Reduction) via positive or negative reinforcement versus those with either combined or ambiguous outcomes (e.g., Social Behavior Enhancement or Impediment).
closed_qa
Substance use disorders among first- and second- generation immigrant adults in the United States: evidence of an immigrant paradox?
A growing number of studies have examined the "immigrant paradox" with respect to the use of licit and illicit substances in the United States. However, there remains a need for a comprehensive examination of the multigenerational and global links between immigration and substance use disorders among adults in the United States. The present study, using data from the National Epidemiologic Survey on Alcohol and Related Conditions, aimed to address these gaps by comparing the prevalence of substance use disorders of first-generation (n = 3,338) and second-generation (n = 2,515) immigrants with native-born American adults (n = 15,733) in the United States. We also examined the prevalence of substance use disorders among first-generation emigrants from Asia, Africa, Europe, and Latin America in contrast to second-generation and native-born Americans. The prevalence of substance use disorders was highest among native-born Americans, slightly lower among second-generation immigrants, and markedly lower among first-generation immigrants. Adjusted risk ratios were largest among individuals who immigrated during adolescence (ages 12-17 years) and adulthood (age 18 years or older). Results were consistent among emigrants from major world regions.
Consistent with a broad body of literature examining the links between the immigrant paradox and health outcomes, results suggest that nativity and age at arrival are significant factors related to substance use disorders among first- and second-generation immigrants in the United States.
closed_qa
Does sexual self-concept ambiguity moderate relations among perceived peer norms for alcohol use, alcohol-dependence symptomatology, and HIV risk-taking behavior?
The current study examines the relation between peer descriptive norms for alcohol involvement and alcohol-dependence symptomatology and whether this relation differs as a function of sexual self-concept ambiguity (SSA). This study also examines the associations among peer descriptive norms for alcohol involvement, alcohol-dependence symptomatology, and lifetime HIV risk-taking behavior and how these relations are influenced by SSA. Women between ages 18 and 30 years (N = 351; M = 20.96, SD = 2.92) completed an online survey assessing sexual self-concept, peer descriptive norms, alcohol-dependence symptomatology, and HIV risk-taking behaviors. Structural equation modeling was used to test hypotheses of interest. There was a significant latent variable interaction between SSA and descriptive norms for peer alcohol use. There was a stronger positive relationship between peer descriptive norms for alcohol and alcohol-dependence symptomatology when SSA was higher compared with when SSA was lower. Both latent variables exhibited positive simple associations with alcohol-dependence symptoms. Peer descriptive norms for alcohol involvement directly and indirectly influenced HIV risk-taking behaviors, and the indirect influence was conditional based on SSA.
The current findings illustrate complex, nuanced associations between perceived norms, identity-related self-concepts, and risky health behaviors from various domains. Future intervention efforts may be warranted to address both problem alcohol use and HIV-risk engagement among individuals with greater sexual self-concept ambiguity.
closed_qa
"Do you think you have mental health problems?
Mental disorders are common among people presenting for the treatment of substance-related problems, and guidelines recommend systematic screening for these conditions. However, even short screens impose a considerable burden on clients and staff. In this study, we evaluated the performance of a single screening question (SSQ) relative to longer instruments. We analyzed a sample of 544 adults recruited from outpatient substance use disorder treatment centers in Ontario, Canada. Clients were asked a simple SSQ, followed by the Kessler Screening Scale for Psychological Distress (K6) and the Global Appraisal of Individual Needs Short Screener (GAIN-SS). Caseness was ascertained using the Structured Clinical Interview for DSMIV (SCID). We measured and compared performance using receiver operating characteristic (ROC) curve analysis and explored client characteristics associated with screen performance using ROC regression. Last, we used logistic regression to test whether SSQs can be usefully combined with other measures. The prevalence of past-month disorder was 71%. The performance of the SSQ (AUC = 0.77, 95% CI [0.72, 0.83]) was similar to, and not significantly different from, those of the K6 (AUC = 0.78) and the GAIN (AUC = 0.79). The K6 and the GAIN performed better than the SSQ among people with a psychotic disorder. The addition of the SSQ slightly improved the performance of the other measures.
SSQs can offer screening performance comparable to that of longer instruments. Reasons for caution include the small number of possible thresholds, lower accuracy than other measures in identifying psychotic disorder, and possible differential functioning in different populations. Performance of all three screens was also moderate; when prevalence is high and resources are available, offering full assessments may be preferable to screening. Nevertheless, SSQs offer an intriguing area for further evaluation.
closed_qa
Evaluation of an after-hours call center: are pediatric patients appropriately referred to the emergency department?
There is concern that after-hours nurse telephone triage systems are overwhelming the emergency department (ED) with nonemergent pediatric referrals. This study aimed to critically review a nonpediatric hospital-based call center with the aim of identifying the algorithms responsible for the majority of nonessential referrals. This is a retrospective observational study performed at a tertiary medical care facility over 1 year. Telephone triage forms of children and adolescents younger than 18 years, exclusively referred by triage nurses using the Barton Schmitt protocols, were reviewed, and their ED course was evaluated by consulting the electronic medical record. "Essential" referrals to the ED were classified as presentations warranting immediate evaluation or referrals requiring "essential interventions" such as serum laboratory tests, imaging, complex procedures, intravenous medications, subspecialty consultation, or admission. A total of 220 patients were included in this study. Of these, 73 (33%) were classified as nonessential, whereas 147 (67%) were classified as essential. Nonessential patients were significantly younger compared with essential referrals (P<0.05). They also had lower triage scores (P = 0.026) and shorter ED stays (P<0.0001). The algorithms for "fever-3 months or older" (12.3%), "vomiting without diarrhea" (8.2%), "trauma-head" (8.2%), "headache" (6.8%), and "sore throat" (5.5%) were determined most likely to result in a nonessential referral.
Our study identifies that a third of unnecessary pediatric visits to the ED occurred as a result of the nurse triage telephone system in question. We recommend review of the algorithms stated to reduce strain on local ED resources.
closed_qa
Does cruciate retention primary total knee arthroplasty affect proprioception, strength and clinical outcome?
It remains unclear what the contribution of the PCL is in total knee arthroplasty (TKA). The goal of this study was to investigate the influence of the PCL in TKA in relationship to clinical outcome, strength and proprioception. Two arthroplasty designs were compared: a posterior cruciate-substituting (PS) and a posterior cruciate-retaining (CR) TKA. A retrospective analysis was performed of 27 CR and 18 PS implants with a minimum of 1 year in vivo. Both groups were compared in terms of clinical outcome (range of motion, visual analogue scale for pain, Hospital for Special Surgery Knee Scoring system, Lysholm score and Knee Injury and Osteoarthritis Outcome Score), strength (Biodex System 3 Dynamometer(®)) and proprioception (balance and postural control using the Balance Master system(®)). Each design was also compared to the non-operated contralateral side in terms of strength and proprioception. There were no significant differences between both designs in terms of clinical outcome and strength. In terms of proprioception, only the rhythmic weight test at slow and moderate speed shifting from left to right was significant in favour of the CR design. None of the unilateral stance tests showed any significant difference between both designs. There was no difference in terms of strength and proprioception between the operated side and the non-operated side.
Retaining the PCL in TKA does not result in an improved performance in terms of clinical outcome and proprioception and does not show any difference in muscle strength.
closed_qa
Is isolated insert exchange a valuable choice for polyethylene wear in metal-backed unicompartmental knee arthroplasty?
The aim of this study was to evaluate the clinical outcome and survival rate after isolated liner exchange for polyethylene (PE) wear in well-fixed metal-backed fixed-bearing unicompartmental knee arthroplasty (UKA). Twenty medial UKAs in 19 patients [mean age 68.7 years ± 8.7 (range 48.5-81.5 years)] operated on for a direct PE liner exchange after isolated PE wear between 1996 and 2010 in two institutions were retrospectively reviewed. The mean delay between the index operation and revision was 8.2 years ± 2.6 (range 4.8-12.8 years). A four-level satisfaction questionnaire was used, and clinical outcomes were assessed using Knee Society scores (KSS) and range of motion (ROM) evaluation. Radiological evaluation analysed the position of the implants and progression of the disease. Survival rate of the implants was evaluated using Kaplan-Meier analysis with two different end-points. At the last follow-up [mean 6.8 years ± 5.2 (range 1.1-15.9 years)], 15 patients (79 %) were enthusiastic or satisfied. KSS improved from 73.4 to 86.4 points (p = 0.01) and function from 58.9 to 89.2 points (p < 0.001). ROM at last FU was 126.5° ± 10.3°. The survival rate at 12 years considering "revision for any reason" as the end-point was 71.3 ± 15.3 %, and the survival rate at 12 years considering "revision of UKA to TKA" as the end-point was 93.3 ± 6.4 %.
Isolated liner exchange for PE wear in well-fixed metal-backed fixed-bearing UKA represents a valuable treatment option in selective patients with durable improvement of clinical outcomes without compromising any future revision.
closed_qa
Compression of human primary cementoblasts leads to apoptosis: A possible cause of dental root resorption?
One of the most common side effects of orthodontic treatment is root resorption on the pressure side of tooth movement. This is usually repaired by cementoblasts, but 1-5 % of patients eventually experiences a marked reduction in root length because no repair has occurred. The reason why cementoblasts should lose their repair function in such cases is not well understood. There is evidence from genome-wide expression analysis (Illumina HumanHT-12 v4 Expression BeadChip Kit;> 30,000 genes) that apoptotic processes are upregulated after the compression of cementoblasts, which is particularly true of the pro-apoptotic gene AXUD1. Human primary cementoblasts (HPCBs) from two individuals were subjected to compressive loading at 30 g/cm(2) for 1/6/10 h. The cells were then evaluated for apoptosis by flow cytometry, for mRNA expression of putative genes (AXUD1, AXIN1, AXIN2) by quantitative PCR, and for involvement of c-Jun-N-terminal kinases (JNKs) in the regulation of AXUD1 via western blotting. In addition, platelet-derived growth factor receptor-β (PDGFRβ) was selectively inhibited by SU16f to analyze the effect of PDGFRβ-dependent signal transduction on AXUD1 and AXIN1 expression. The percentage of apoptotic HPCBs rose after only 6 h of compressive loading, and 18-20 % of cells were apoptotic after 10 h. Microarray data revealed significant upregulation of the pro-apoptotic gene AXUD1 after 6 h and quantitative PCR significant AXUD1 upregulation after 6 and 10 h of compression. AXIN1 and AXIN2 expression in HPCBs was significantly increased after compressive loading. Our tests also revealed that PDGFRβ signaling inhibition by SU16f augmented the expression of AXIN1 and AXUD1 in HPCBs under compression.
Increased apoptosis of compressed HPCBs might help explain why cementoblasts, rather than invariably repairing all cases of root resorption, sometimes allow the original root length to shorten. The pathway hypothesized to lead to cementoblast apoptosis involves PDGF signaling, with this signal transduction's inhibition augmenting the expression of pro-apoptotic genes. Thus activating PDGF signaling may modify the signaling pathway for the apoptosis of cementoblasts, which would reveal a protective role of PDGF for these cells. Further studies are needed to develop strategies of treatment capable of minimizing root resorption.
closed_qa
All-cause mortality and estimated renal function in type 2 diabetes mellitus outpatients: Is there a relationship with the equation used?
We investigated the relationship between serum creatinine (SCr) and estimated glomerular filtration rate (eGFR), evaluated by different formulae, and all-cause mortality (ACM) in type 2 diabetes mellitus (T2DM) outpatients. This observational cohort study considered 1365 T2DM outpatients, who had been followed up for a period of up to 11 years. eGFR was estimated using several equations. Seventy subjects (5.1%) died after a follow-up of 9.8 ± 3 years. Univariate analysis showed that diagnosis of nephropathy (odds ratio (OR): 2.554, 95% confidence interval (CI): 1.616-4.038, p<0.001) and microvascular complications (OR: 2.281, 95% CI: 1.449-3.593, p<0.001) were associated with ACM. Receiving operating characteristic (ROC) curves showed that the areas under the curve for ACM were similar using the different eGFR equations. eGFR values were predictors of ACM, and the hazard ratios (HRs) of the different equations for eGFR estimation were similar.
In our cohort of T2DM outpatients, different eGFR equations perform similarly in predicting ACM, whereas SCr did not.
closed_qa
Color vision in ADHD: part 2--does attention influence color perception?
To investigate the impact of exogenous covert attention on chromatic (blue and red) and achromatic visual perception in adults with and without Attention Deficit Hyperactivity Disorder (ADHD). Exogenous covert attention, which is a transient, automatic, stimulus-driven form of attention, is a key mechanism for selecting relevant information in visual arrays. 30 adults diagnosed with ADHD and 30 healthy adults, matched on age and gender, performed a psychophysical task designed to measure the effects of exogenous covert attention on perceived color saturation (blue, red) and contrast sensitivity. The effects of exogenous covert attention on perceived blue and red saturation levels and contrast sensitivity were similar in both groups, with no differences between males and females. Specifically, exogenous covert attention enhanced the perception of blue saturation and contrast sensitivity, but it had no effect on the perception of red saturation.
The findings suggest that exogenous covert attention is intact in adults with ADHD and does not account for the observed impairments in the perception of chromatic (blue and red) saturation.
closed_qa
Dental materials for primary dentition: are they suitable for occlusal restorations?
Eight specimens of each material were subjected to two-body wear test, using a chewing simulator. The wear region of each material was examined under a profilometer, measuring the vertical loss (μm) and the volume loss (mm(3)) of the materials. The results showed significant differences of vertical loss and volume loss of the test materials (p<0.001). Amalgam had the highest wear resistance. Twinky Star (compomer) had the lowest vertical loss and volume loss. There was no significant difference of vertical loss among compomers, Dyract Extra, Dyract Flow and Dyract Posterior. Riva Self Cure (GIC) had no statistically significant difference compared with the compomers (except Twinky Star). No statistically significant difference was found also between Equia (GIC) and Ketac Moral (GIC) with Dyract Extra (Compomer). RMGICs were found to have the lowest wear resistance. For the statistical analysis, the PASW 20.0 (SPSS Statistics, IBM, Chicago) package was used. Means and standard deviations were measured with descriptive statistics and analyzed using one-way ANOVA.
Compomers and some GICs, that have moderate wear resistance, may be sufficient for occlusal restorations in primary dentitions.
closed_qa
Thyroid fine needle aspiration biopsy: do nodule volume and cystic degeneration ratio affect specimen adequacy and cytological diagnosis time?
A fine needle aspiration biopsy (FNAB) of thyroid nodules - the least invasive and most accurate method used to investigate malignant lesions - may yield non-diagnostic specimens even under ultrasonographic guidance. To evaluate the effects of thyroid nodule volume and extent of cystic degeneration on both the non-diagnostic specimen ratio as well as cytopathologist's definitive cytological diagnosis time. In this single center study, FNAB was performed on 505 patients with single thyroid nodules greater than 10 mm. Nodule volume was calculated prior to FNAB and cystic degeneration ratio was recorded. All biopsies were performed by a single radiologist who also prepared specimen slides. Specimen adequacy and final diagnosis were made in the pathology laboratory by a single-blinded cytopathologist based on the Bethesda system. Definitive cytological diagnosis time was recorded upon reaching a definitive diagnosis. The specimen adequacy ratio was 85.3%. The mean nodule volume of adequate specimens was larger than those of non-diagnostic samples (6.00 mL vs. 3.05 mL; P = 0.001). There was no correlation between nodule volume and cytopathologist's definitive cytological diagnosis time (r = 0.042). Biopsy of predominantly solid nodules yielded better specimen adequacy ratios compared to predominantly cystic nodules (87.8% vs. 75.3%; P = 0.028). Definitive cytological diagnosis times were longer in predominantly cystic nodules compared to predominantly solid nodules (376 s vs. 294 s; P = 0.019).
Predominantly cystic nodules are likely to benefit from repeated nodular sampling until the specimen is declared adequate by an on-site cytopathologist. If a cytopathologist is not available, obtaining more specimens per nodule may achieve desired adequacy ratios.
closed_qa
Surgical fires in laser laryngeal surgery: are we safe enough?
Laser surgery of the larynx and airway remains high risk for the formation of operating room fire. Traditional methods of fire prevention have included use of "laser safe" tubes, inflation of a protective cuff with saline, and wet pledgets to protect the endotracheal tube from laser strikes. We tested a mechanical model of laser laryngeal surgery to evaluate the fire risk. Mechanical model. Laboratory. An intubation mannequin was positioned for suspension microlaryngoscopy. A Laser-Shield II cuffed endotracheal tube was placed through the larynx and the cuff inflated using saline. Wet pledgets covered the inflated cuff. A CO2 laser created an inadvertent cuff strike at varying oxygen concentrations. Risk reduction measures were implemented to discern any notable change in the outcome after fire. At 100% FiO2 an immediate fire with sustained flame was created and at 40% FiO2 a near immediate sustained flame was created. At 29% FiO2, a small nonsustained flame was noted. At room air, no fire was created. There was no discernible difference in the severity of laryngeal damage after the fire occurred whether the tube was immediately pulled from the mannequin or if saline was poured down the airway as a first response.
While "laser safe" tubes provide a layer of protection against fires, they are not fire proof. Inadvertent cuff perforation may result in fire formation in low-level oxygen enriched environments. Placement of wet pledgets do not provide absolute protection. Endotracheal tube (ETT) cuffs should be placed distally well away from an inadvertent laser strike while maintaining the minimum supplemental oxygen necessary.
closed_qa
Can a community health worker and a trained traditional birth attendant work as a team to deliver child health interventions in rural Zambia?
Teaming is an accepted approach in health care settings but rarely practiced at the community level in developing countries. Save the Children trained and deployed teams of volunteer community health workers (CHWs) and trained traditional birth attendants (TBAs) to provide essential newborn and curative care for children aged 0-59 months in rural Zambia. This paper assessed whether CHWs and trained TBAs can work as teams to deliver interventions and ensure a continuum of care for all children under-five, including newborns. We trained CHW-TBA teams in teaming concepts and assessed their level of teaming prospectively every six months for two years. The overall score was a function of both teamwork and taskwork. We also assessed personal, community and service factors likely to influence the level of teaming. We created forty-seven teams of predominantly younger, male CHWs and older, female trained TBAs. After two years of deployment, twenty-one teams scored "high", twelve scored "low," and fourteen were inactive. Teamwork was high for mutual trust, team cohesion, comprehension of team goals and objectives, and communication, but not for decision making/planning. Taskwork was high for joint behavior change communication and outreach services with local health workers, but not for intra-team referral. Teams with members residing within one hour's walking distance were more likely to score high.
It is feasible for a CHW and a trained TBA to work as a team. This may be an approach to provide a continuum of care for children under-five including newborns.
closed_qa
Is mix of care influenced by the provider environment?
In Australia, access to dental care has been available through several different pathways: (1) private practice; (2) public clinics; (3) Aboriginal Medical Services (AMS)-based clinics; and (4) until recently, the Chronic Disease Dental Scheme (CDDS). The aim of the present study was to compare the types of dental services most commonly delivered in the various clinical pathways based on the hypothesis that disease-driven care should lead to similar mixes of dental care provided. Data from a series of previously published sources was used to identify and compare the most commonly performed dental procedures in the different pathways. A comparison was also made with the available international data (US). There was a marked difference between service mixes provided through the four pathways. Patients obtaining dental care through AMS-based and public pathways had more extractions and less restorative and preventive care compared with private and CDDS pathways. Compared with the international data, dental service mixes in Australia were found to be not as evenly distributed. Value of care provided through private and CDDS pathways were two- to threefold higher than that of AMS-based and public pathways.
The data indicate that the original hypothesis that the disease-driven care should lead to similar mixes of provided dental care, is not supported.
closed_qa
Is periodontitis a risk factor for cognitive impairment and dementia?
Dementia is a multi-etiologic syndrome characterized by multiple cognitive deficits but not always by the presence of cognitive impairment. Cognitive impairment is associated with multiple non-modifiable risk factors but few modifiable factors. Epidemiologic studies have shown an association between periodontitis, a potentially modifiable risk factor, and cognitive impairment. The objective of this study is to determine whether clinical periodontitis is associated with the diagnosis of cognitive impairment/dementia after controlling for known risk factors, including age, sex, and education level. A case-control study was conducted in Granada, Spain, in two groups of dentate individuals aged>50 years: 1) cases with a firm diagnosis of mild cognitive impairment or dementia of any type or severity and 2) controls with no subjective memory loss complaints and a score>30 in the Phototest cognitive test (screening test for cognitive impairment). Periodontitis was evaluated by measuring tooth loss, plaque and bleeding indexes, probing depths, and clinical attachment loss (AL). The study included 409 dentate adults, 180 with cognitive impairment and 229 without. A moderate and statistically significant association was observed between AL and cognitive impairment after controlling for age, sex, education level, oral hygiene habits, and hyperlipidemia (P = 0.049). No significant association was found between tooth loss and cognitive impairment.
Periodontitis appears to be associated with cognitive impairment after controlling for confounders such as age, sex, and education level.
closed_qa
Can antiretroviral therapy modify the clinical course of HDV infection in HIV-positive patients?
Infection with hepatitis delta virus (HDV) affects approximately 6-14.5% of patients coinfected with HIV-1 and HBV, showing a more aggressive clinical course compared with an HIV-negative population. There is no universally approved treatment for chronic hepatitis D (CHD) in HIV-infected patients. Antiretroviral therapy (ART) containing tenofovir has been recently associated with HDV suppression. Our aim was to evaluate whether the outcome of CHD in HIV-infected patients can be favourably influenced by ART including reverse transcriptase inhibitors. The clinical course of four HBV/HDV/HIV-coinfected patients receiving ART were retrospectively examined. HDV RNA became undetectable in all patients after a variable period of ART along with the disappearance of hepatitis B surface antigen in two of them, and an increase in CD4(+) T-cell count. In all patients, virological changes were associated with improved liver function tests and clinical features.
We suggest that ART regimens including drugs active against HBV could have beneficial effects on the clinical course of CHD in patients with HIV-1 by favouring immunological reconstitution.
closed_qa
Dental health among older Israeli adults: is this a reflection of a medical care model inadequately addressing oral health?
Israel's health-care system is considered as one of the most efficient worldwide. The purpose of the present study was to assess oral health outcomes, dental care use and respective social inequalities among the older segment of the Israeli population. Secondary analyses were conducted of recently available data from the Survey of Health, Ageing, and Retirement in Europe (SHARE Israel, wave 2), which specifically includes information on chewing ability, denture wearing and dental care use obtained from more than 2,400 Israeli people, 50+ years of age. Multivariate logistic regressions and concentration indices were used to analyse determinants of oral health and dental care use. Seventy per cent of respondents reported being able to bite/chew on hard foods and 49% of respondents reported wearing dentures. Forty-three per cent of respondents had visited a dentist within the past 12 months, with about half of all dental visits being made for solely nonpreventive reasons. Significant income-related inequalities were identified, with higher income being associated with greater dental care use (particularly preventive dental visits), better chewing ability and less denture wearing.
For the older segment of the Israeli population and compared with other countries, the findings of the present study suggest a relatively low level of chewing ability, a high extent of nonpreventive dental visiting, as well as considerable inequalities in oral health and care. It seems that the Israeli health-care system may be improved even further by more comprehensive inclusion of dental care into universal health coverage.
closed_qa
Can the Perinatal Information System in Peru be used to measure the proportion of adverse birth outcomes attributable to maternal syphilis infection?
To describe the capacity of Peru's Perinatal Information System (Sistema Informático Perinatal, SIP) to provide estimates for monitoring the proportion of stillbirths and other adverse birth outcomes attributable to maternal syphilis. A descriptive study was conducted to assess the quality and completeness of SIP data from six Peruvian public hospitals that used the SIP continuously from 2000 - 2010 and had maternal syphilis prevalence of at least 0.5% during that period. In-depth interviews were conducted with Peruvian stakeholders about their experiences using the SIP. Information was found on 123 575 births from 2000 - 2010 and syphilis test results were available for 99 840 births. Among those 99 840 births, there were 1 075 maternal syphilis infections (1.1%) and 619 stillbirths (0.62%). Among women with syphilis infection in pregnancy, 1.7% had a stillbirth, compared to 0.6% of women without syphilis infection. Much of the information needed to estimate the proportion of stillbirths attributable to maternal syphilis was available in the SIP, with the exception of syphilis treatment information, which was not collected. However, SIP data collection is complex and time-consuming for clinicians. Data were unlinked across hospitals and not routinely used or quality-checked. Despite these limitations, the SIP data examined were complete and valid; in 98% of records, information on whether or not the infant was stillborn was the same in both the SIP and clinical charts. Nearly 89% of women had the same syphilis test result in clinical charts and the SIP.
The large number of syphilis infections reported in Peru's SIP and the ability to link maternal characteristics to newborn outcomes make the system potentially useful for monitoring the proportion of stillbirths attributable to congenital syphilis in Peru. To ensure good data quality and sustainability of Peru's SIP, data collection should be simplified and information should be continually quality-checked and used for the benefit of participating facilities.
closed_qa
Quality indicators for prostate radiotherapy: are patients disadvantaged by receiving treatment in a 'generalist' centre?
The purpose of this retrospective review was to evaluate concordance with evidence-based quality indicator guidelines for prostate cancer patients treated radically in a 'generalist' (as distinct from 'sub-specialist') centre. We were concerned that the quality of treatment may be lower in a generalist centre. If so, the findings could have relevance for many radiotherapy departments that treat prostate cancer. Two hundred fifteen consecutive patients received external beam radiotherapy (EBRT) and/or brachytherapy between 1.10.11 and 30.9.12. Treatment was deemed to be in line with evidence-based guidelines if the dose was: (i) 73.8-81 Gy at 1.8-2.0 Gy/fraction for EBRT alone (eviQ guidelines); (ii) 40-50 Gy (EBRT) for EBRT plus high-dose rate (HDR) brachytherapy boost (National Comprehensive Cancer Network (NCCN) guidelines); and (iii) 145 Gy for low dose rate (LDR) I-125 monotherapy (NCCN). Additionally, EBRT beam energy should be ≥6 MV using three-dimensional conformal RT (3D-CRT) or intensity-modulated RT (IMRT), and high-risk patients should receive neo-adjuvant androgen-deprivation therapy (ADT) (eviQ/NCCN). Treatment of pelvic nodes was also assessed. One hundred four high-risk, 84 intermediate-risk and 27 low-risk patients (NCCN criteria) were managed by eight of nine radiation oncologists. Concordance with guideline doses was confirmed in: (i) 125 of 136 patients (92%) treated with EBRT alone; (ii) 32 of 34 patients (94%) treated with EBRT + HDR BRT boost; and (iii) 45 of 45 patients (100%) treated with LDR BRT alone. All EBRT patients were treated with ≥6 MV beams using 3D-CRT (78%) or IMRT (22%). 84%, 21% and 0% of high-risk, intermediate-risk and low-risk patients received ADT, respectively. Overall treatment modality choice (including ADT use and duration where assessable) was concordant with guidelines for 176/207 (85%) of patients.
The vast majority of patients were treated concordant with evidence-based guidelines suggesting that, within the limits of the selected criteria, prostate cancer patients are unlikely to be disadvantaged by receiving radiotherapy in this 'generalist' centre.
closed_qa
Can hyperbaric oxygen be used to prevent deep infections in neuro-muscular scoliosis surgery?
The prevalence of postoperative wound infection in patients with neuromuscular scoliosis surgery is significantly higher than that in patients with other spinal surgery. Hyperbaric oxygen has been used as a supplement to treat postsurgical infections. Our aim was to determine beneficiary effects of hyperbaric oxygen treatment in terms of prevention of postoperative deep infection in this specific group of patients in a retrospective study. Forty two neuromuscular scoliosis cases, operated between 2006-2011 were retrospectively reviewed. Patients who had presence of scoliosis and/or kyphosis in addition to cerebral palsy or myelomeningocele, postoperative follow-up>1 year and posterior only surgery were the subjects of this study. Eighteen patients formed the Hyperbaric oxygen prophylaxis (P-HBO) group and 24, the control group. The P-HBO group received 30 sessions of HBO and standard antibiotic prophylaxis postoperative, and the control group (received standard antibiotic prophylaxis). In the P-HBO group of 18 patients, the etiology was cerebral palsy in 13 and myelomeningocele in 5 cases with a mean age of 16.7 (11-27 yrs). The average follow-up was 20.4 months (12-36mo). The etiology of patients in the control group was cerebral palsy in 17, and myelomeningocele in 7 cases. The average age was 15.3 years (8-32 yrs). The average follow-up was 38.7 months (18-66mo). The overall incidence of infection in the whole study group was 11.9% (5/42). The infection rate in the P-HBO and the control group were 5.5% (1/18), and 16.6% (4/24) respectively. The use of HBO was found to significantly decrease the incidence of postoperative infections in neuromuscular scoliosis patients.
In this study we found that hyperbaric oxygen has a possibility to reduce the rate of post-surgical deep infections in complex spine deformity in high risk neuromuscular patients.
closed_qa
Does cochleostomy location influence electrode trajectory and intracochlear trauma?
Trauma to intracochlear structures during cochlear implant insertion is associated with poorer hearing outcomes. One way surgeons can influence insertion trauma is by choosing the surgical approach. We seek to compare cochleostomy (CO), peri-round window (PRW), and round window (RW) approaches using a fresh frozen temporal bone model. Experiments using fresh frozen temporal bones. Cochlear implant insertions using the three aforementioned approaches were performed on 15 fresh frozen human temporal bones using a Cochlear 422 electrode. Insertions were evaluated by examining fluoroscopic recordings of histologic sections. Five cochlear implant insertions were performed using each of the three aforementioned approaches. Fluoroscopic examination revealed that none of the CO or PRW insertions contacted the modiolus during insertion, whereas three of five RW insertions did. RW insertions were less linear during insertion when compared to CO and PRW insertions (P < .05). CO insertions had significantly larger angular depth of insertion (487°) when compared to PRW (413°) and RW (375°) (P < .05). Histologic examination revealed one RW insertion resulted in osseous spiral lamina fracture, whereas the remaining insertions had no evidence of trauma. In the damaged specimen, the inserted electrode was observed to rest in the scala vestibuli, whereas the remaining electrodes rested in the scala tympani.
Due to variability in RW anatomy, a CO or PRW window surgical approach appears to minimize the risk for insertion trauma. However, with favorable anatomy, a Cochlear 422 electrode can be inserted with any of the three approaches.
closed_qa
Does mutual compensation of the cognitive effects induced by pain and opioids exist?
Studies have demonstrated that both pain and opioids have actions on the central nervous system that may interfere with cognitive function, but their effects have mainly been analysed separately and not as an integrated process. The objective of this study is to test two hypotheses: (1) the analgesic effect of opioids improves cognitive function by decreasing pain, and (2) pain antagonizes cognitive effects of opioids. Randomized, placebo-controlled, crossover study. Three experiments were conducted with 22 healthy males. Sustained attention, memory and motor function/attention/mental flexibility were evaluated by continuous reaction time (CRT), verbal fluency test (VFT) and trail making test-B (TMT-B), respectively. In the 1st experiment, the cognitive effects of experimental tonic pain of mild and moderate intensities produced by a computer-controlled pneumatic tourniquet cuff were assessed; in the 2nd, the effects of saline solution and remifentanil were assessed in the absence of pain; and in the 3rd experiment, the cognitive effects of moderate pain intensity relieved by remifentanil infusion were assessed followed by increasing pain to moderate intensity during a constant remifentanil infusion. The first two experiments demonstrated that pain and remifentanil impaired CRT. In the 3rd experiment, remifentanil infusion relieving pain significantly impaired CRT and further deterioration was noted following increasing pain intensity.
Pain and remifentanil seemed to have additive deleterious cognitive effects. This study represents an initial step to enhance our basic understanding of some of the cognitive effects following a painful stimulus and an opioid infusion separately and combined in a sequence comparable to clinical settings.
closed_qa
Is 3D technique superior to 2D in Down syndrome screening?
The objective of this article is to investigate whether in the clinical setting of second trimester ultrasound (US) investigations, 3D multiplanar correction prior to the measurement of Down syndrome (DS) facial markers (nasal bone length, prenasal thickness, fetal profile line, maxilla-nasion-mandible angle, prenasal thickness to nasal bone length ratio, and prefrontal space ratio) is superior to subjective judgment of a correct midsagittal plane by 2D technique. Measurements were performed on 2D images and 3D volumes (corrected to the midsagittal plane), acquired during the same scanning session. All six markers were measured in 105 datasets (75 of euploid fetuses and 30 of DS fetuses). The maxilla-nasion-mandible angle measured on 2D images was significantly larger than on 3D volumes (p < 0.01). In all other markers, there was no significant difference between measurements performed on 2D images or 3D volumes. No statistical difference was found for any marker between measurements performed on images acquired by either 2D or 3D US in their ability to discriminate between normal and DS fetuses.
Nasal bone length, prenasal thickness, fetal profile line, prenasal thickness to nasal bone length ratio, and prefrontal space ratio can be confidently used as DS markers in second trimester US examinations performed by 2D US. © 2014 John Wiley&Sons, Ltd.
closed_qa
Case-based learning and simulation: useful tools to enhance nurses' education?
To compare skills acquired by undergraduate nursing students enrolled in a medical-surgical course. To compare skills demonstrated by students with no previous clinical practice (undergraduates) and nurses with clinical experience enrolled in continuing professional education (CPE). In a nonrandomized clinical trial, 101 undergraduates enrolled in the "Adult Patients 1" course were assigned to the traditional lecture and discussion (n = 66) or lecture and discussion plus case-based learning (n = 35) arm of the study; 59 CPE nurses constituted a comparison group to assess the effects of previous clinical experience on learning outcomes. Scores on an objective structured clinical examination (OSCE), using a human patient simulator and cases validated by the National League for Nursing, were compared for the undergraduate control and intervention groups, and for CPE nurses (Student's t test). Controls scored lower than the intervention group on patient assessment (6.3 ± 2.3 vs 7.5 ± 1.4, p = .04, mean difference, -1.2 [95% confidence interval (CI) -2.4 to -0.03]) but the intervention group did not differ from CPE nurses (7.5 ± 1.4 vs 8.8 ± 1.5, p = .06, mean difference, -1.3 [95% CI -2.6 to 0.04]). The CPE nurses committed more "rules-based errors" than did undergraduates, specifically patient identifications (77.2% vs 55%, p = .7) and checking allergies before administering medication (68.2% vs 60%, p = .1).
The intervention group developed better patient assessment skills than the control group. Case-based learning helps to standardize the process, which can contribute to quality and consistency in practice: It is essential to correctly identify a problem in order to treat it. Clinical experience of CPE nurses was not associated with better adherence to safety protocols.
closed_qa
Can serum NGAL levels be used as an inflammation marker on hemodialysis patients with permanent catheter?
Neutrophil gelatinase-associated lipocalin (NGAL) is a member of lipocalin family and released from many tissues and cells. We aimed to investigate the relationship among serum NGAL levels, the inflammation markers (IL-6, hs-CRP, TNF-α) and different vascular access types used in dialysis patients. The study population included 90 patients and 30 healthy age-matched controls. The patients were divided into three groups (I, II, III) and group IV included the controls. In group I and II, the patients were with central venous permanent catheter and arterio-venous fistula, respectively. Group III included 30 patients with chronic renal failure. Hemogram, biochemical assays, ferritin, IL-6, hs-CRP, TNF-α, and NGAL were evaluated in all groups. Serum NGAL levels were markedly higher in group I than in group II (7645.80 ± 924.61 vs. 4131.20 ± 609.87 pg/mL; p < 0.05). Positive correlation was detected between NGAL levels and duration of catheter (r: 0.903, p: 0.000), hs-CRP (r: 0.796, p: 0.000), IL-6 (r: 0.687, p: 0.000), TNF-α (r: 0.568, p: 0.000) levels and ferritin (r: 0.318, p: 0.001), whereas NGAL levels were negatively correlated with serum albumin levels (r: -0.494, p: 0.000). In multiple regression analysis, duration of catheter hs-CRP and TNF-α were predictors of NGAL in hemodialysis patients.
Inflammation was observed in hemodialysis patients and increases with catheter. Our findings show that a strong relationship among serum NGAL levels, duration of catheter, hs-CRP and TNF-α. NGAL may be used as a new inflammation marker in hemodialysis patients.
closed_qa
Multicystic dysplastic kidney: is an initial voiding cystourethrogram necessary?
Traditionally, a voiding cystourethrogram (VCUG) has been obtained in patients diagnosed with multicystic dysplastic kidney (MCDK) because of published vesicoureteral reflux (VUR) rates between 10%-20%. However, with the diagnosis and treatment of low grade VUR undergoing significant changes, we questioned the utility of obtaining a VCUG in healthy patients with a MCDK. We reviewed our experience to see how many of the patients with documented VUR required surgical intervention. We performed a retrospective review of children diagnosed with unilateral MCDK from 2002 to 2012 who also underwent a VCUG. A total of 133 patients met our inclusion criteria. VUR was identified in 23 (17.3%) children. Four patients underwent ureteral reimplant (3.0%). Indications for surgical therapy included breakthrough urinary tract infections (2 patients), evidence of dysplasia/scarring (1 patient) and non-resolving reflux (1 patient). All patients with a history of VUR who are toilet trained, regardless of the grade or treatment, are currently being followed off antibiotic prophylaxis. To date, none have had a febrile urinary tract infection (UTI) since cessation of prophylactic antibiotics. Hydronephrosis in the contralateral kidney was not predictive of VUR (p = 0.99).
Routine VCUG in healthy children diagnosed with unilateral MCDK may not be warranted given the low incidence of clinically significant VUR. If a more conservative strategy is preferred, routine VCUG may be withheld in those children without normal kidney hydronephrosis and considered in patients with normal kidney hydronephrosis. If a VCUG is not performed the family should be instructed in signs and symptoms of urinary tract infection.
closed_qa
Is mycophenolate mofetil an alternative agent to corticosteroids in traumatic nerve paralysis?
The effects of an immunosuppressive agent, mycophenolate mofetil (MM), were investigated and compared with those of methylprednisolone (MP) and dexamethasone (DXM) on the traumatic nerve function. This is a randomized controlled animal study. This experimental study was performed on 84 male Wistar albino rats. The rats were assigned to 12 groups each consisting of 7 animals. The groups were formed according to application of normal-dose DXM (group 1A-B), high-dose MP (group 2A-B), normal-dose MP (group 3A-B), MM (group 4A-B), and MM with high-dose MP combination therapies (group VA-B). Right sciatic nerve dissection was performed, and compound muscle action potential thresholds were recorded. The nerve was traumatized with the compression of a Jeweller forceps for 20 seconds. Posttraumatic thresholds were also recorded. The compound muscle action potential thresholds were recorded in the first and fourth weeks for the assigned groups. Then, the nerve was transected and prepared for electron microscopic and histopathologic examinations. Nitric oxide and malondialdehyde assessments were performed on both tissue and blood samples. Only the MM and MP+MM groups had satisfactory electron microscopic findings and were about to reach the tissue characteristics of the control animals. Despite the electrophysiologic recovery, the DXM group was found to have poor electron microscopic scoring.
Mycophenolate mofetil has been found to be beneficial in the treatment of traumatic nerve paralysis. Although a complementary investigation is needed, this immunosuppressive agent may be an alternative to corticosteroids for the selected cases where steroid therapy is contraindicated.
closed_qa
Does the degree of ptosis predict the degree of psychological morbidity in bariatric patients undergoing reconstruction?
There is proven therapeutic benefit in bariatric surgery for obese patients. Successful bariatric surgery will result in massive weight loss and ptotic skin, which can cause significant functional and psychological problems. As the incidence of bariatric surgery increases, so will the demand for plastic surgery. Currently, there is no evidence-based indication for massive weight loss body contouring, and therefore there is no standardized provision. A prospective, multicenter, observational study of outcomes in 75 patients undergoing bariatric and plastic surgery procedures at two clinical sites was performed to determine whether the degree of ptosis can be determined by the type (malabsorptive or restrictive) of bariatric surgery and if the extent of disfigurement has an impact on psychological morbidity. Massive weight loss body contouring is not purely aesthetic surgery, but it leads to functional and psychosocial benefits. This study has given preliminary data on which anthropometric measurements and their thresholds lead to the greatest benefit from massive weight loss body contouring. From this study, the fourth quartiles of the anthropometric measurements xiphisternum to pubic symphysis (≥91 cm), umbilicus to pubic symphysis (≥38 cm), and hip circumference (≥143 cm) were statistically significant in crossing the psychometric tolerances from within the normal range to pathological psychology.
This study demonstrates that there is a statistically significant, quantifiable correlation among type of bariatric surgery, degree of ptosis, and psychological morbidity in patients who have undergone bariatric surgery. This pilot study could provide the basis for evidence-based guidelines for plastic surgery referral.
closed_qa
Do therapeutic imagery practices affect physiological and emotional indicators of threat in high self-critics?
Imagery is known to be a powerful means of stimulating various physiological processes and is increasingly used within standard psychological therapies. Compassion-focused imagery (CFI) has been used to stimulate affiliative emotion in people with mental health problems. However, evidence suggests that self-critical individuals may have particular difficulties in this domain with single trials. The aim of the present study was to further investigate the role of self-criticism in responsiveness to CFI by specifically pre-selecting participants based on trait self-criticism. Using the Forms of Self-Criticism/Self-Reassuring Scale, 29 individuals from a total sample of 139 were pre-selected to determine how self-criticism impacts upon an initial instance of imagery. All participants took part in three activities: a control imagery intervention (useable data N = 25), a standard CFI intervention (useable data N = 25), and a non-intervention control (useable data N = 24). Physiological measurements (alpha amylase) as well as questionnaire measures of emotional responding (i.e., the Positive and Negative Affect Schedule, the Types of Positive Affect Scale, and the State Adult Attachment Scale) were taken before and after the different interventions. Following both imagery interventions, repeated measures analyses revealed that alpha amylase increased significantly for high self-critics compared with low self-critics. High self-critics (HSC) also reported greater insecurity on entering the imagery session and more negative CFI experiences compared with low self-critics.
Data demonstrate that HSC respond negatively to imagery interventions in a single trial. This highlights that imagery focused therapies (e.g., CFI) need interventions that manage fears, blocks, and resistances to the techniques, particularly in HSC.
closed_qa
Is insomnia associated with deficits in neuropsychological functioning?
People with insomnia complain of cognitive deficits in daily life. Results from empirical studies examining associations between insomnia and cognitive impairment, however, are mixed. Research is needed that compares treatment-seeking and community-based insomnia study samples, measures subjective as well as objective cognitive functioning, and considers participants' pre-insomnia cognitive function. We used data from the Dunedin Study, a representative birth cohort of 1,037 individuals, to examine whether insomnia in early midlife was associated with subjective and objective cognitive functioning. We also tested whether individuals with insomnia who reported seeking treatment for their sleep problems (treatment-seekers) showed greater impairment than other individuals with insomnia (non-treatment-seekers). The role of key confounders, including childhood cognitive ability and comorbid health conditions, was evaluated. Insomnia was diagnosed at age 38 according to DSM-IV criteria. Objective neuropsychological assessments at age 38 included the WAIS-IV IQ test, the Wechsler Memory Scale, and the Trail-Making Test. Childhood cognitive functioning was assessed using the Wechsler Intelligence Scale for Children-Revised (WISC-R). A total of 949 cohort members were assessed for insomnia symptoms and other study measures at age 38. Although cohort members with insomnia (n = 186, 19.6%) had greater subjective cognitive impairment than their peers at age 38, they did not exhibit greater objective impairment on formal testing. Treatment-seekers, however, exhibited significant objective impairment compared to non-treatment-seekers. Controlling for comorbidity, daytime impairment, and medications slightly decreased this association. Childhood cognitive deficits antedated the adult cognitive deficits of treatment-seekers.
Links between insomnia and cognitive impairment may be strongest among individuals who seek clinical treatment. Clinicians should take into account the presence of complex health problems and lower premorbid cognitive function when planning treatment for insomnia patients.
closed_qa
Acute kidney injury after cardiac surgery: is minocycline protective?
Acute kidney injury (AKI) after cardiac bypass surgery (CABG) is common and carries a significant association with morbidity and mortality. Since minocycline therapy attenuates kidney injury in animal models of AKI, we tested its effects in patients undergoing CABG. This is a randomized, double-blinded, placebo-controlled, multi-center study. We screened high risk patients who were scheduled to undergo CABG in two medical centers between Jan 2008 and June 2011. 40 patients were randomized and 19 patients in each group completed the study. Minocycline prophylaxis was given twice daily, at least for four doses prior to CABG. Primary outcome was defined as AKI [0.3 mg/dl increase in creatinine (Cr)] within 5 days after surgery. Daily serum Cr for 5 days, various clinical and hemodynamic measures and length of stay were recorded. The two groups had similar baseline and intra-operative characteristics. The primary outcome occurred in 52.6% of patients in the minocycline group as compared to 36.8% of patients in the placebo group (p = 0.51). Peak Cr was 1.6 ± 0.7 vs. 1.5 ± 0.7 mg/dl (p = 0.45) in minocycline and placebo groups, respectively. Death at 30 days occurred in 0 vs. 10.5% in the minocycline and placebo groups, respectively (p = 0.48). There were no differences in post-operative length of stay, and cardiovascular events between the two groups. There was a trend towards lower diastolic pulmonary artery pressure [16.8 ± 4.7 vs. 20.7 ± 6.6 mmHg (p = 0.059)] and central venous pressure [11.8 ± 4.3 vs. 14.6 ± 5.6 mmHg (p = 0.13)]in the minocycline group compared to placebo on the first day after surgery.
Minocycline did not protect against AKI post-CABG.
closed_qa
Restless legs syndrome in non-dialysis renal patients: is it really that common?
Sleep disorders are frequent in chronic kidney disease (CKD). Among them, restless legs syndrome (RLS) may affect up to 60% of patients on dialysis, and it has been related to a poor quality of life and higher cardiovascular risk. Despite its high prevalence in advanced stages of renal disease, RLS frequency in non-dialysis CKD has not been clearly established. The aim of this study was to assess the frequency of RLS in non-dialysis CKD patients (stages 2 to 4) followed in a reference nephrology outpatient clinic. A standardized questionnaire following the international RLS study group diagnostic criteria was self-administered by 110 patients regularly followed in the nephrology clinic. The series comprised 69 men and 41 women, aged 68 ± 13.2 years, with mean serum creatinine of 1.7 ± 0.8 mg/dL. Subsequently, patients classified as probable RLS according to the questionnaire underwent a systematic neurological examination. The presence of peripheral artery disease was evaluated by the ankle-brachial index (ABI). The frequency of probable RLS according to the questionnaire results was 21% (17% for men and 27% for women). However, after thorough neurological examination, the diagnosis of RLS was confirmed in only 5 patients. Therefore, the overall definitive RLS frequency was 4.5% (within the prevalence reported for the general population) and was higher among women (9.7% vs 0.2%). In the remaining cases symptoms were due to leg discomfort related with other disorders. Patients with probable and improbable RLS were not significantly different in age, ABI, diabetes, and other comorbid circumstances, except for tricyclic antidepressant prescription, which was more frequent in the probable RLS group (17% vs 2%). Renal function was better in definitive RLS patients than cases classified as probable RLS by the questionnaire but not confirmed after neurological exam.
Although RLS can represent an early manifestation of CKD, its prevalence seems very close to that reported for the general population. Diagnostic confirmation of RLS dramatically falls after expert examination, raising the question whether, in the study of RLS cohorts, CKD has a potentially causal relationship or is a confounding factor associated with other causes of leg discomfort.
closed_qa
Is the chronotype associated with obstructive sleep apnea?
Chronotype and obstructive sleep apnea (OSA) appear to have a similar lifelong evolution, which could indicate a possible effect of morningness or eveningness in the apnea-hypopnea index (AHI). The present study aimed to examine the prevalence of chronotypes in a representative sample of São Paulo city residents and to investigate the effect of chronotypes on the severity of OSA. We performed a cross-sectional analysis using the São Paulo Epidemiologic Sleep Study (EPISONO). All participants underwent a full-night polysomnography and completed the Morningness-eveningness, Epworth Sleepiness Scale, and UNIFESP Sleep questionnaires. Chronotypes were classified as morning-type, evening-type, and intermediate. Morning-type individuals represented 52.1% of the sample, followed by intermediate (39.5%), and evening-type (8.4%) individuals. After stratifying the sample by body mass index (BMI) (>26.8 kg/m(2)) and age (>42 years), we observed increased AHI values in morning- and evening-type individuals.
We demonstrated, for the first time, an age- and BMI-related effect of morning- and evening-types in OSA severity, suggesting that the intermediate chronotype might play a role as a protective factor in older and overweight patients.
closed_qa
Bone marrow FDG-PET/CT in Hodgkin lymphoma revisited: do imaging and pathology match?
To directly compare visual and quantitative (18)F-fluoro-2-deoxy-D-glucose positron emission tomography/computed tomography (FDG-PET/CT) to bone marrow biopsy (BMB) findings in the right posterior iliac crest in patients with newly diagnosed Hodgkin lymphoma. This retrospective study included 26 patients with newly diagnosed Hodgkin lymphoma in whom FDG-PET/CT was performed before BMB of the right posterior iliac crest. The right posterior iliac crest was assessed for bone marrow involvement, both visually and semi-quantitatively [using maximum standardized uptake value (SUVmax) measurements]. BMB of right the posterior iliac crest was used as reference standard. BMB of the right posterior iliac crest was positive in 5 (19.2 %) of 26 patients. There was full agreement between visual FDG-PET/CT and BMB findings in the right posterior iliac crest (i.e. no false-positive or false-negative FDG-PET/CT findings). Accordingly, sensitivity, specificity, positive predictive value, and negative predictive value of visual FDG-PET/CT assessment for the detection of bone marrow involvement in the right posterior iliac crest were 100 % (5/5) (95 % CI 51.1-100 %), 100 % (21/21) (95 % CI 81.8-100 %), 100 % (5/5) (95 % CI 51.1-100 %), and 100 % (21/21) (95 % CI 81.8-100 %), respectively. SUVmax of BMB-positive cases (mean ± SD: 3.4 ± 0.85) was nearly significantly higher (P = 0.052) than that of BMB-negative cases (mean ± SD 2.7 ± 0.63).
This histopathological correlation study confirms the very high diagnostic value of FDG-PET/CT in the detection of bone marrow involvement in newly diagnosed Hodgkin lymphoma, and supports the substitution of BMB with FDG-PET/CT in this setting.
closed_qa
Is implementation of the 2013 Australian treatment guidelines for posttraumatic stress disorder cost-effective compared to current practice?
To assess, from a health sector perspective, the incremental cost-effectiveness of three treatment recommendations in the most recent Australian Clinical Practice Guidelines for posttraumatic stress disorder (PTSD). The interventions assessed are trauma-focused cognitive behavioural therapy (TF-CBT) and selective serotonin reuptake inhibitors (SSRIs) for the treatment of PTSD in adults and TF-CBT in children, compared to current practice in Australia. Economic modelling, using existing databases and published information, was used to assess cost-effectiveness. A cost-utility framework using both quality-adjusted life-years (QALYs) gained and disability-adjusted life-years (DALYs) averted was used. Costs were tracked for the duration of the respective interventions and applied to the estimated 12 months prevalent cases of PTSD in the Australian population of 2012. Simulation modelling was used to provide 95% uncertainty around the incremental cost-effectiveness ratios. Consideration was also given to factors not considered in the quantitative analysis but could determine the likely uptake of the proposed intervention guidelines. TF-CBT is highly cost-effective compared to current practice at $19,000/QALY, $16,000/DALY in adults and $8900/QALY, $8000/DALY in children. In adults, 100% of uncertainty iterations fell beneath the $50,000/QALY or DALY value-for-money threshold. Using SSRIs in people already on medications is cost-effective at $200/QALY, but has considerable uncertainty around the costs and benefits. While there is a 13% chance of health loss there is a 27% chance of the intervention dominating current practice by both saving dollars and improving health in adults.
The three Guideline recommended interventions evaluated in this study are likely to have a positive impact on the economic efficiency of the treatment of PTSD if adopted in full. While there are gaps in the evidence base, policy-makers can have considerable confidence that the recommendations assessed in the current study are likely to improve the efficiency of the mental health care sector.
closed_qa
Age-related cognitive task effects on gait characteristics: do different working memory components make a difference?
Though it is well recognized that gait characteristics are affected by concurrent cognitive tasks, how different working memory components contribute to dual task effects on gait is still unknown. The objective of the present study was to investigate dual-task effects on gait characteristics, specifically the application of cognitive tasks involving different working memory components. In addition, we also examined age-related differences in such dual-task effects. Three cognitive tasks (i.e. 'Random Digit Generation', 'Brooks' Spatial Memory', and 'Counting Backward') involving different working memory components were examined. Twelve young (6 males and 6 females, 20 ~ 25 years old) and 12 older participants (6 males and 6 females, 60 ~ 72 years old) took part in two phases of experiments. In the first phase, each cognitive task was defined at three difficulty levels, and perceived difficulty was compared across tasks. The cognitive tasks perceived to be equally difficult were selected for the second phase. In the second phase, four testing conditions were defined, corresponding to a baseline and the three equally difficult cognitive tasks. Participants walked on a treadmill at their self-selected comfortable speed in each testing condition. Body kinematics were collected during treadmill walking, and gait characteristics were assessed using spatial-temporal gait parameters. Application of the concurrent Brooks' Spatial Memory task led to longer step times compared to the baseline condition. Larger step width variability was observed in both the Brooks' Spatial Memory and Counting Backward dual-task conditions than in the baseline condition. In addition, cognitive task effects on step width variability differed between two age groups. In particular, the Brooks' Spatial Memory task led to significantly larger step width variability only among older adults.
These findings revealed that cognitive tasks involving the visuo-spatial sketchpad interfered with gait more severely in older versus young adults. Thus, dual-task training, in which a cognitive task involving the visuo-spatial sketchpad (e.g. the Brooks' Spatial Memory task) is concurrently performed with walking, could be beneficial to mitigate impairments in gait among older adults.
closed_qa
Osteodesis for hallux valgus correction: is it effective?
Although the etiology of hallux valgus is contested, in some patients it may be failure of the stabilizing soft tissue structures around the first ray of the foot. Because there is lack of effective soft tissue techniques, osteotomies have become the mainstream surgical approach to compensate for the underlying soft tissue deficiency; osteodesis, a soft tissue nonosteotomy technique, may be a third alternative, but its efficacy is unknown.QUESTIONS/ We asked: (1) Can an osteodesis, a distal soft tissue technique, correct hallux valgus satisfactorily in terms of deformity correction and improvement in American Orthopaedic Foot and Ankle Society (AOFAS) score? (2) Is the effectiveness of an osteodesis affected by the patient's age or deformity severity? (3) What complications are associated with this procedure? Between February and October 2010, we performed 126 operations to correct hallux valgus, of which 126 (100%) were osteodeses. Sixty-one patients (110 procedures) (87% of the total number of hallux valgus procedures) were available for followup at a minimum of 12 months (mean, 23 months; range, 12-38 months). This group formed our study cohort. During the study period, the general indications for this approach included failed conservative measures for pain relief and metatarsophalangeal angle greater than 20° or intermetatarsal angle greater than 9°. Intermetatarsal cerclage sutures were used to realign the first metatarsal and postoperative fibrosis was induced surgically between the first and second metatarsals to maintain its alignment. The radiologic first intermetatarsal angle, metatarsophalangeal angle, and medial sesamoid position were measured by Hardy and Clapham's methods for deformity and correction evaluation. Clinical results were assessed by the AOFAS score. The intermetatarsal angle was improved from a preoperative mean of 14° to 7° (p<0.001; Cohen's d=1.8) at followup, the metatarsophalangeal angle from 31° to 18° (p<0.001; Cohen's d=3.1), the medial sesamoid position from position 6 to 3 (p<0.001; Cohen's d=2.4), and AOFAS hallux score from 68 to 96 points (p<0.001). Neither patient age nor deformity severity affected the effectiveness of the osteodesis in correcting all three radiologic parameters; however, the deformities treated in this series generally were mild to moderate (mean intermetatarsal angle, 14°; range, 9°-22°). There were six stress fractures of the second metatarsal (5%), five temporary metatarsophalangeal joint medial subluxations all resolved in one month by the taping-reduction method without surgery, and six metatarsophalangeal joints with reduced dorsiflexion less than 60°.
The osteodesis is a soft tissue nonosteotomy technique, and provided adequate deformity correction and improvement in AOFAS scores for patients with mild to moderate hallux valgus deformities, although a small number of the patients had postoperative stress fractures of the second ray develop. Future prospective studies should compare this technique with osteotomy techniques in terms of effectiveness of the correction, restoration of hallux function, complications, and long-term recurrence.
closed_qa
Can 111In-RGD2 monitor response to therapy in head and neck tumor xenografts?
RGD (arginylglycylaspartic acid)-based imaging tracers allow specific imaging of integrin αvβ3 expression, proteins overexpressed during angiogenesis; however, few studies have investigated the potential of these tracers to monitor responses of antiangiogenic or radiation therapy. In the studies presented here, (111)In-RGD2 was assessed for its potential as an imaging tool to monitor such responses to therapies. DOTA-E-[c(RGDfK)]2 was radiolabeled with (111)In ((111)In-RGD2), and biodistribution studies were performed in mice with subcutaneous FaDu or SK-RC-52 xenografts after treatment with either antiangiogenic therapy (bevacizumab or sorafenib) or tumor irradiation (10 Gy). Micro-SPECT imaging studies and subsequent quantitative analysis were also performed. The effect of bevacizumab, sorafenib, or radiation therapy on tumor growth was determined. The uptake of (111)In-RGD2 in tumors, as determined from biodistribution studies, correlated well with that quantified from micro-SPECT images, and both showed that 15 d after irradiation (111)In-RGD2 uptake was enhanced. Specific or nonspecific uptake of (111)In-RGD2 in FaDu or SK-RC-52 xenografts was not affected after antiangiogenic therapy, except in head and neck squamous cell carcinoma 19 d after the start of sorafenib therapy (P<0.05). The uptake of (111)In-RGD2 followed tumor volume in studies featuring antiangiogenic therapy. However, the uptake of (111)In-RGD2 in FaDu xenografts was decreased as early as 4 h after tumor irradiation, despite nonspecific uptake remaining unaltered. Tumor growth was inhibited after antiangiogenic or radiation therapy.
Here, it is suggested that (111)In-RGD2 could allow in vivo monitoring of angiogenic responses after radiotherapy and may therefore prove a good clinical tool to monitor angiogenic responses early after the start of radiotherapy in patients with head and neck squamous cell carcinoma. Despite clear antitumor efficacy, antiangiogenic therapy did not alter tumor uptake of (111)In-RGD2, indicating that integrin expression was not altered.
closed_qa