Rohdewatts4244

Z Iurium Wiki

Verze z 23. 9. 2024, 14:28, kterou vytvořil Rohdewatts4244 (diskuse | příspěvky) (Založena nová stránka s textem „3; mean difference, 1.9; 95% CI, 1.7-2.0; P less then 0.001) and in an intensive care unit (mean, 2.1 versus 3.0; mean difference, 0.9; 95% CI, 0.8-0.9; P…“)
(rozdíl) ← Starší verze | zobrazit aktuální verzi (rozdíl) | Novější verze → (rozdíl)

3; mean difference, 1.9; 95% CI, 1.7-2.0; P less then 0.001) and in an intensive care unit (mean, 2.1 versus 3.0; mean difference, 0.9; 95% CI, 0.8-0.9; P less then 0.001); and less commonly received mechanical ventilation (15.5% versus 20.8%; P less then 0.001); hemodialysis (4.8% versus 7.7%; P less then 0.001); or cardiac catheterization (2.8% versus 4.6%; P less then 0.001). Female sex was independently associated with lower odds of in-hospital death (odds ratio, 0.88; 95% CI, 0.87-0.89). Mean (SD) 6-month direct healthcare cost was greater for in-hospital ($52 349 [$55 649]) than out-of-hospital ($35 998 [$31 900]) death. Conclusions Among decedents with heart failure, invasive care in the last 6 months increased in prevalence over time but was less common in women, who had lower odds of dying in a hospital.Background Soluble urokinase-type plasminogen activator receptor (suPAR) is associated with cardiovascular risks and poor renal outcomes. However, whether elevated suPAR levels are associated with 24-hour blood pressure patterns or kidney disease progression in patients with chronic kidney disease (CKD) is unclear. Methods and Results A total of 751 patients with CKD stage 1 to 5 were recruited from CMERC-HI (Cardiovascular and Metabolic Disease Etiology Research Center-High Risk) cohort study (2013-2018). The relationship of serum suPAR levels to 24-hour blood pressure parameters and CKD progression was analyzed. The median serum suPAR level was 1439.0 (interquartile range, 1026.2-2150.1) pg/mL, and the mean estimated glomerular filtration rate was 52.8±28.5 mL/min per 1.73 m2 at baseline. Patients with higher suPAR levels had significantly higher levels of office, 24-hour, daytime, and nighttime systolic blood pressure and nighttime diastolic blood pressure than those with lower suPAR levels. The highest suPAR tertile was associated with an increased risk of a reverse dipping pattern (odds ratio, 2.93; 95% CI, 1.27-6.76; P=0.01). this website During a follow-up of 43.2 (interquartile range, 27.0-55.6) months, the CKD progression occurred in 271 (36.1%) patients. The highest suPAR tertile was significantly associated with higher risk of CKD progression than the lowest tertile (hazard ratio [HR], 2.09; 95% CI, 1.37-3.21; P=0.001). When the relationship was reevaluated with respect to each dipping pattern (dipper, extreme dipper, nondipper, and reverse dipper), this association was consistent only in reverse dippers in whom the risk of CKD progression increased (HR, 1.43; 95% CI, 1.02-2.01; P=0.03) with every 1-unit increase in serum suPAR levels. Conclusions Elevated suPAR levels are independently associated with CKD progression, and this association is prominent in reverse dippers.Background Prevalence of cardiovascular disease risk factors and rates of atherosclerotic cardiovascular disease outcomes vary across racial/ethnic groups. This analysis examined the effects of evolocumab on LDL-C (low-density lipoprotein cholesterol) levels and LDL-C goals achievement by race/ethnicity. Methods and Results Data from 15 phase 2 and 3 studies of treatment with evolocumab versus placebo or ezetimibe were pooled (n=7669). Results were analyzed by participant clinical characteristics and by self-identified race/ethnicity. Key outcomes included percent change from baseline in LDL-C, achievement of LDL-C less then 70 mg/dL, and LDL-C reduction of ≥50% at 12 weeks and at 1 to 5 years. Across 12-week studies, mean percent change in LDL-C from baseline in evolocumab-treated participants was -52% to -59% for White and -46% to -67% for non-White participants, across clinical characteristics groups. LDL-C less then 70 mg/dL was achieved in 43% to 84% and 62% to 94% and LDL-C reduction of ≥50% in 63% to 78% and 58% to 86%, respectively. In 1- to 5-year studies, mean percent change in LDL-C was -46% to -52% for White and -49% to -55% for non-White participants. LDL-C less then 70 mg/dL was achieved in 53% to 84% and 66% to 77%, and LDL-C reduction of ≥50% in 53% to 67% and 58% to 68%, respectively. The treatment effect on mean percent change in LDL-C differed only in participants with type 2 diabetes mellitus, with a larger reduction in Asian participants. The qualitative interaction P values were nonsignificant, indicating consistent directionality of effect. Conclusions Similar reduction in LDL-C levels with evolocumab was observed across racial/ethnic groups in 12-week and 1- to 5-year studies. Among those with diabetes mellitus, Asian participants had greater LDL-C reduction.Background Contrast-associated acute kidney injury (CA-AKI) is associated with substantial morbidity and may be prevented by using less contrast during percutaneous coronary intervention (PCI). However, tools for determining safe contrast volumes are limited. We developed risk models to tailor safe contrast volume limits during PCI. Methods and Results Using data from all PCIs performed at 18 hospitals from January 2015 to March 2018, we developed logistic regression models for predicting CA-AKI, including simpler models ("pragmatic full," "pragmatic minimum") using only predictors easily derivable from electronic health records. We prospectively validated these models using PCI data from April 2018 to December 2018 and compared them to preexisting safe contrast models using the area under the receiver operating characteristic curve (AUC). The model derivation data set included 20 579 PCIs with 2102 CA-AKI cases. When applying models to the separate validation data set (5423 PCIs, 488 CA-AKI cases), prior safe contrast limits (5*Weight/Creatinine, 2*CreatinineClearance) were poor measures of safety with accuracies of 53.7% and 56.6% in predicting CA-AKI, respectively. The full, pragmatic full, and pragmatic minimum models performed significantly better (accuracy, 73.1%, 69.3%, 66.6%; AUC, 0.80, 0.76, 0.72 versus 0.59 for 5 * Weight/Creatinine, 0.61 for 2*CreatinineClearance). We found that applying safe contrast limits could meaningfully reduce CA-AKI risk in one-quarter of patients. Conclusions Compared with preexisting equations, new multivariate models for safe contrast limits were substantially more accurate in predicting CA-AKI and could help determine which patients benefit most from limiting contrast during PCI. Using readily available electronic health record data, these models could be implemented into electronic health records to provide actionable information for improving PCI safety.The influence of patient-level factors on palliative and hospice care is unclear. We conducted a retrospective review of 2321 patients aged ≥18 that died within 6 months of admission to our institution between 2012 and 2017. Patients were included for analysis if their chart was complete, their length of stay was ≥48 hours, and if based on their diagnoses, they would have benefited from palliative care consultation (PCC). Bayesian regression with a weakly informative prior was used to find the odds ratio (OR) and 99% credible interval (CrI) of receiving PCC based on race/ethnicity, education, language, insurance status, and income. 730 patients fit our inclusion criteria and 30% (n = 211) received PCC. The OR of receiving PCC was 1.26 (99% CrI, 0.73-2.12) for Blacks, 0.81 (99% CrI, 0.31-1.86) for Hispanics, and 0.69 (99% CrI, 0.19-2.46) for other minorities. Less than high school education was associated with greater odds of PCC (OR 2.28, 99% CrI, 1.09-4.93) compared to no schooling. Compared to English speakers, non-English speakers had higher odds of receiving PCC when cared for by medical services (OR 3.01 [99% CrI, 1.44-5.32]) but lower odds of PCC when cared for by surgical services (0.22 [99% CrI, less then 0.01-3.42]). Insurance status and income were not associated with differences in PCC. At our institution, we found no evidence of racial/ethnic, insurance, or income status affecting PCC while primary language spoken and educational status did. Further investigation is warranted to examine the system and provider-level factors influencing PCC's low utilization by medical and surgical specialties.Background Myocardial fibrosis is an important contributor for development of diastolic dysfunction. We investigated the impact of sirolimus as primary immunosuppression on diastolic dysfunction and fibrosis progression among heart transplantation recipients. Methods and Results In 100 heart transplantation recipients who were either treated with a calcineurin inhibitor (CNI) (n=51) or converted from CNI to sirolimus (n=49), diastolic function parameters were assessed using serial echocardiograms and right heart catheterizations. Myocardial fibrosis was quantified on serial myocardial biopsies. After 3 years, lateral e' increased within the sirolimus group but decreased in the CNI group (0.02±0.04 versus -0.02±0.04 m/s delta change; P=0.003, respectively). Both pulmonary capillary wedge pressure and diastolic pulmonary artery pressure significantly decreased in the sirolimus group but remained unchanged in the CNI group (-1.50±2.59 versus 0.20±2.20 mm Hg/year; P=0.02; and -1.72±3.39 versus 0.82±2.59 mm Hg/year; P=0.005, respectively). A trend for increased percentage of fibrosis was seen in the sirolimus group (8.48±3.17 to 10.10±3.0%; P=0.07) as compared with marginally significant progression in the CNI group (8.76±3.87 to 10.56±4.34%; P=0.04). The percent change in fibrosis did not differ significantly between the groups (1.62±4.67 versus 1.80±5.31%, respectively; P=0.88). Conclusions Early conversion to sirolimus is associated with improvement in diastolic dysfunction and filling pressures as compared with CNI therapy. Whether this could be attributed to attenuation of myocardial fibrosis progression with sirolimus treatment warrants further investigation.Background Detailed insights in temporal evolution of high-sensitivity cardiac troponin following acute coronary syndrome (ACS) are currently missing. We aimed to describe and compare the post-ACS kinetics of high-sensitivity cardiac troponin I (hs-cTnI) and high-sensitivity cardiac troponin T (hs-cTnT), and to determine their intra- and interindividual variation in clinically stable patients. Methods and Results We determined hs-cTnI (Abbott) and hs-cTnT (Roche) in 1507 repeated blood samples, derived from 191 patients with ACS (median, 8/patient) who remained free from adverse cardiac events during 1-year follow-up. Post-ACS kinetics were studied by linear mixed-effect models. Using the samples collected in the 6- to 12-month post-ACS time frame, patients were then considered to have chronic coronary syndrome. We determined (differences between) the average hs-cTnI and average hs-cTnT concentration, and the intra- and interindividual variation for both biomarkers. Compared with hs-cTnT, hs-cTnI peaked higher (median 3506 ng/L versus 494 ng/L; P less then 0.001) and was quicker below the biomarker-specific upper reference limit (16 versus 19 days; P less then 0.001). In the post-6-month samples, hs-cTnI and hs-cTnT showed modest correlation (rspearman=0.60), whereas the average hs-cTnT concentration was 5 times more likely to be above the upper reference limit than hs-cTnI. The intraindividual variations of hs-cTnI and hs-cTnT were 14.0% and 18.1%, while the interindividual variations were 94.1% and 75.9%. Conclusions Hs-cTnI peaked higher after ACS and was quicker below the upper reference limit. In the post-6-month samples, hs-cTnI and hs-cTnT were clearly not interchangeable, and average hs-cTnT concentrations were much more often above the upper reference limit than hs-cTnI. For both markers, the within-patient variation fell largely below beween-patient variation. Registration URL https//www.trialregister.nl; unique identifiers NTR1698 and NTR1106.

Autoři článku: Rohdewatts4244 (Logan Asmussen)