User login
Magnesium and Metabolic Syndrome: Any Connection?
TOPLINE:
Higher urinary magnesium loss, as indicated by an elevated magnesium depletion score (MDS), may be an independent risk factor for metabolic syndrome in US adults.
METHODOLOGY:
- Increasing evidence suggests that chronic hypomagnesemia may play a role in the pathogenesis of metabolic disorders, including overweight and obesity, insulin resistance, type 2 diabetes, hypertension, and dyslipidemia.
- Researchers examined the relationship between magnesium status and metabolic syndrome in 15,565 US adults (mean age, 47 years; half women) participating in the National Health and Nutrition Examination Survey (2003-2018), of whom 5438 had metabolic syndrome (mean age, 55 years).
- Magnesium deficiency was predicted by MDS, a four-factor score that aggregates diuretic use (one point), proton pump inhibitor (one point), kidney function (estimated glomerular filtration rate; one or two points), and heavy (one point).
- MDS was categorized into six levels (by scores 0-5), with a higher MDS indicating a more severe magnesium deficiency.
- Metabolic syndrome was defined according to the National Cholesterol Education Program’s Adult Treatment Panel III report.
TAKEAWAY:
- The proportion of patients with MDS ≥ 2 was higher in the group with vs without metabolic syndrome (P < .05).
- Even after adjusting for potential confounding factors, each 1-unit increase in the MDS increased the odds of metabolic syndrome by about 30% (adjusted odds ratio, 1.31; 95% CI, 1.17-1.45).
- A dose-response relationship was observed between MDS and metabolic syndrome, with MDS level 1 being associated with 1.28-fold higher odds of metabolic syndrome (95% CI, 1.06-1.55) than MDS level 0; further escalation in the odds was noted for MDS levels 2, 3, and 4.
- The association between metabolic syndrome and MDS remained consistent across all population subgroups defined by age, gender, race (except Mexican American), body mass index, drinking status, or smoking status.
IN PRACTICE:
“It is possible to prevent and reduce MetS [metabolic syndrome] by supplementing with magnesium supplements or encouraging higher magnesium intake diet because the diet is a factor that can be changed,” the authors wrote.
SOURCE:
The study was led by Xiaohao Wang, Department of Geriatrics, the First Affiliated Hospital, School of Medicine, Southern University of Science and Technology (Shenzhen People’s Hospital), Shenzhen, China. It was published online in the Journal of Clinical Endocrinology & Metabolism.
LIMITATIONS:
The study found no significant link between MDS level 5 and metabolic syndrome, likely due to the small sample size at this level. The study could not draw any causal relationship between metabolic syndrome and MDS owing to its cross-sectional nature. It also could not determine whether MDS was a better marker of magnesium deficiency than serum magnesium levels. MDS is a categorical, not continuous, variable.
DISCLOSURES:
This study was supported by grants from the National Natural Science Foundation of China and the Natural Science Foundation of Shenzhen City, China. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Higher urinary magnesium loss, as indicated by an elevated magnesium depletion score (MDS), may be an independent risk factor for metabolic syndrome in US adults.
METHODOLOGY:
- Increasing evidence suggests that chronic hypomagnesemia may play a role in the pathogenesis of metabolic disorders, including overweight and obesity, insulin resistance, type 2 diabetes, hypertension, and dyslipidemia.
- Researchers examined the relationship between magnesium status and metabolic syndrome in 15,565 US adults (mean age, 47 years; half women) participating in the National Health and Nutrition Examination Survey (2003-2018), of whom 5438 had metabolic syndrome (mean age, 55 years).
- Magnesium deficiency was predicted by MDS, a four-factor score that aggregates diuretic use (one point), proton pump inhibitor (one point), kidney function (estimated glomerular filtration rate; one or two points), and heavy (one point).
- MDS was categorized into six levels (by scores 0-5), with a higher MDS indicating a more severe magnesium deficiency.
- Metabolic syndrome was defined according to the National Cholesterol Education Program’s Adult Treatment Panel III report.
TAKEAWAY:
- The proportion of patients with MDS ≥ 2 was higher in the group with vs without metabolic syndrome (P < .05).
- Even after adjusting for potential confounding factors, each 1-unit increase in the MDS increased the odds of metabolic syndrome by about 30% (adjusted odds ratio, 1.31; 95% CI, 1.17-1.45).
- A dose-response relationship was observed between MDS and metabolic syndrome, with MDS level 1 being associated with 1.28-fold higher odds of metabolic syndrome (95% CI, 1.06-1.55) than MDS level 0; further escalation in the odds was noted for MDS levels 2, 3, and 4.
- The association between metabolic syndrome and MDS remained consistent across all population subgroups defined by age, gender, race (except Mexican American), body mass index, drinking status, or smoking status.
IN PRACTICE:
“It is possible to prevent and reduce MetS [metabolic syndrome] by supplementing with magnesium supplements or encouraging higher magnesium intake diet because the diet is a factor that can be changed,” the authors wrote.
SOURCE:
The study was led by Xiaohao Wang, Department of Geriatrics, the First Affiliated Hospital, School of Medicine, Southern University of Science and Technology (Shenzhen People’s Hospital), Shenzhen, China. It was published online in the Journal of Clinical Endocrinology & Metabolism.
LIMITATIONS:
The study found no significant link between MDS level 5 and metabolic syndrome, likely due to the small sample size at this level. The study could not draw any causal relationship between metabolic syndrome and MDS owing to its cross-sectional nature. It also could not determine whether MDS was a better marker of magnesium deficiency than serum magnesium levels. MDS is a categorical, not continuous, variable.
DISCLOSURES:
This study was supported by grants from the National Natural Science Foundation of China and the Natural Science Foundation of Shenzhen City, China. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Higher urinary magnesium loss, as indicated by an elevated magnesium depletion score (MDS), may be an independent risk factor for metabolic syndrome in US adults.
METHODOLOGY:
- Increasing evidence suggests that chronic hypomagnesemia may play a role in the pathogenesis of metabolic disorders, including overweight and obesity, insulin resistance, type 2 diabetes, hypertension, and dyslipidemia.
- Researchers examined the relationship between magnesium status and metabolic syndrome in 15,565 US adults (mean age, 47 years; half women) participating in the National Health and Nutrition Examination Survey (2003-2018), of whom 5438 had metabolic syndrome (mean age, 55 years).
- Magnesium deficiency was predicted by MDS, a four-factor score that aggregates diuretic use (one point), proton pump inhibitor (one point), kidney function (estimated glomerular filtration rate; one or two points), and heavy (one point).
- MDS was categorized into six levels (by scores 0-5), with a higher MDS indicating a more severe magnesium deficiency.
- Metabolic syndrome was defined according to the National Cholesterol Education Program’s Adult Treatment Panel III report.
TAKEAWAY:
- The proportion of patients with MDS ≥ 2 was higher in the group with vs without metabolic syndrome (P < .05).
- Even after adjusting for potential confounding factors, each 1-unit increase in the MDS increased the odds of metabolic syndrome by about 30% (adjusted odds ratio, 1.31; 95% CI, 1.17-1.45).
- A dose-response relationship was observed between MDS and metabolic syndrome, with MDS level 1 being associated with 1.28-fold higher odds of metabolic syndrome (95% CI, 1.06-1.55) than MDS level 0; further escalation in the odds was noted for MDS levels 2, 3, and 4.
- The association between metabolic syndrome and MDS remained consistent across all population subgroups defined by age, gender, race (except Mexican American), body mass index, drinking status, or smoking status.
IN PRACTICE:
“It is possible to prevent and reduce MetS [metabolic syndrome] by supplementing with magnesium supplements or encouraging higher magnesium intake diet because the diet is a factor that can be changed,” the authors wrote.
SOURCE:
The study was led by Xiaohao Wang, Department of Geriatrics, the First Affiliated Hospital, School of Medicine, Southern University of Science and Technology (Shenzhen People’s Hospital), Shenzhen, China. It was published online in the Journal of Clinical Endocrinology & Metabolism.
LIMITATIONS:
The study found no significant link between MDS level 5 and metabolic syndrome, likely due to the small sample size at this level. The study could not draw any causal relationship between metabolic syndrome and MDS owing to its cross-sectional nature. It also could not determine whether MDS was a better marker of magnesium deficiency than serum magnesium levels. MDS is a categorical, not continuous, variable.
DISCLOSURES:
This study was supported by grants from the National Natural Science Foundation of China and the Natural Science Foundation of Shenzhen City, China. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
Vitamin D Deficiency May Be Linked to Peripheral Neuropathy
TOPLINE:
Vitamin D deficiency is independently linked to the risk for diabetic peripheral neuropathy (DPN) by potentially affecting large nerve fibers in older patients with type 2 diabetes (T2D).
METHODOLOGY:
- Although previous research has shown that vitamin D deficiency is common in patients with diabetes and may increase the risk for peripheral neuropathy, its effects on large and small nerve fiber lesions have not been well explored yet.
- Researchers conducted a cross-sectional study to understand the association between vitamin D deficiency and DPN development in 230 older patients (mean age, 67 years) with T2D for about 15 years who were recruited from Beijing Hospital between 2020 and 2023.
- All patients were evaluated for DPN based on poor blood sugar control or symptoms such as pain and sensory abnormalities, of which 175 patients diagnosed with DPN were propensity-matched with 55 patients without DPN.
- Vitamin D deficiency, defined as serum 25-hydroxyvitamin D circulating levels below 20 ng/mL, was reported in 169 patients.
- Large nerve fiber lesions were evaluated using electromyography, and small nerve fiber lesions were assessed by measuring skin conductance.
TAKEAWAY:
- Vitamin D deficiency was more likely to affect large fiber lesions, suggested by longer median sensory nerve latency, minimum latency of the F-wave, and median nerve motor evoked potential latency than those in the vitamin D–sufficient group.
- Furthermore, vitamin D deficiency was linked to large fiber neuropathy with increased odds of prolongation of motor nerve latency (odds ratio, 1.362; P = .038).
- The electrochemical skin conductance, which indicates damage to small nerve fibers, was comparable between patients with and without vitamin D deficiency.
IN PRACTICE:
This study is too preliminary to have practice application.
SOURCE:
This study was led by Sijia Fei, Department of Endocrinology, Beijing Hospital, Beijing, People’s Republic of China, and was published online in Diabetes Research and Clinical Practice.
LIMITATIONS:
Skin biopsy, the “gold-standard” for quantifying intraepidermal nerve fiber density, was not used to assess small nerve fiber lesions. Additionally, a causal link between vitamin D deficiency and diabetic nerve damage was not established owing to the cross-sectional nature of the study. Some patients with T2D may have been receiving insulin therapy, which may have affected vitamin D levels.
DISCLOSURES:
The study was supported by grants from the National Natural Science Foundation of China and China National Key R&D Program. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Vitamin D deficiency is independently linked to the risk for diabetic peripheral neuropathy (DPN) by potentially affecting large nerve fibers in older patients with type 2 diabetes (T2D).
METHODOLOGY:
- Although previous research has shown that vitamin D deficiency is common in patients with diabetes and may increase the risk for peripheral neuropathy, its effects on large and small nerve fiber lesions have not been well explored yet.
- Researchers conducted a cross-sectional study to understand the association between vitamin D deficiency and DPN development in 230 older patients (mean age, 67 years) with T2D for about 15 years who were recruited from Beijing Hospital between 2020 and 2023.
- All patients were evaluated for DPN based on poor blood sugar control or symptoms such as pain and sensory abnormalities, of which 175 patients diagnosed with DPN were propensity-matched with 55 patients without DPN.
- Vitamin D deficiency, defined as serum 25-hydroxyvitamin D circulating levels below 20 ng/mL, was reported in 169 patients.
- Large nerve fiber lesions were evaluated using electromyography, and small nerve fiber lesions were assessed by measuring skin conductance.
TAKEAWAY:
- Vitamin D deficiency was more likely to affect large fiber lesions, suggested by longer median sensory nerve latency, minimum latency of the F-wave, and median nerve motor evoked potential latency than those in the vitamin D–sufficient group.
- Furthermore, vitamin D deficiency was linked to large fiber neuropathy with increased odds of prolongation of motor nerve latency (odds ratio, 1.362; P = .038).
- The electrochemical skin conductance, which indicates damage to small nerve fibers, was comparable between patients with and without vitamin D deficiency.
IN PRACTICE:
This study is too preliminary to have practice application.
SOURCE:
This study was led by Sijia Fei, Department of Endocrinology, Beijing Hospital, Beijing, People’s Republic of China, and was published online in Diabetes Research and Clinical Practice.
LIMITATIONS:
Skin biopsy, the “gold-standard” for quantifying intraepidermal nerve fiber density, was not used to assess small nerve fiber lesions. Additionally, a causal link between vitamin D deficiency and diabetic nerve damage was not established owing to the cross-sectional nature of the study. Some patients with T2D may have been receiving insulin therapy, which may have affected vitamin D levels.
DISCLOSURES:
The study was supported by grants from the National Natural Science Foundation of China and China National Key R&D Program. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Vitamin D deficiency is independently linked to the risk for diabetic peripheral neuropathy (DPN) by potentially affecting large nerve fibers in older patients with type 2 diabetes (T2D).
METHODOLOGY:
- Although previous research has shown that vitamin D deficiency is common in patients with diabetes and may increase the risk for peripheral neuropathy, its effects on large and small nerve fiber lesions have not been well explored yet.
- Researchers conducted a cross-sectional study to understand the association between vitamin D deficiency and DPN development in 230 older patients (mean age, 67 years) with T2D for about 15 years who were recruited from Beijing Hospital between 2020 and 2023.
- All patients were evaluated for DPN based on poor blood sugar control or symptoms such as pain and sensory abnormalities, of which 175 patients diagnosed with DPN were propensity-matched with 55 patients without DPN.
- Vitamin D deficiency, defined as serum 25-hydroxyvitamin D circulating levels below 20 ng/mL, was reported in 169 patients.
- Large nerve fiber lesions were evaluated using electromyography, and small nerve fiber lesions were assessed by measuring skin conductance.
TAKEAWAY:
- Vitamin D deficiency was more likely to affect large fiber lesions, suggested by longer median sensory nerve latency, minimum latency of the F-wave, and median nerve motor evoked potential latency than those in the vitamin D–sufficient group.
- Furthermore, vitamin D deficiency was linked to large fiber neuropathy with increased odds of prolongation of motor nerve latency (odds ratio, 1.362; P = .038).
- The electrochemical skin conductance, which indicates damage to small nerve fibers, was comparable between patients with and without vitamin D deficiency.
IN PRACTICE:
This study is too preliminary to have practice application.
SOURCE:
This study was led by Sijia Fei, Department of Endocrinology, Beijing Hospital, Beijing, People’s Republic of China, and was published online in Diabetes Research and Clinical Practice.
LIMITATIONS:
Skin biopsy, the “gold-standard” for quantifying intraepidermal nerve fiber density, was not used to assess small nerve fiber lesions. Additionally, a causal link between vitamin D deficiency and diabetic nerve damage was not established owing to the cross-sectional nature of the study. Some patients with T2D may have been receiving insulin therapy, which may have affected vitamin D levels.
DISCLOSURES:
The study was supported by grants from the National Natural Science Foundation of China and China National Key R&D Program. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
New CRC Risk Prediction Model Outperforms Polyp-Based Model
TOPLINE:
A comprehensive model considering patient age, diabetes, colonoscopy indications, and polyp findings can predict colorectal cancer (CRC) risk more accurately than the solely polyp-based model in patients with a first diagnosis of adenoma on colonoscopy.
METHODOLOGY:
- Because colonoscopy surveillance guidelines relying solely on previous polyp findings to assess CRC risk are imprecise, researchers developed and tested a comprehensive risk prediction model from a list of CRC-related predictors that included patient characteristics and clinical factors in addition to polyp findings.
- The comprehensive model included baseline colonoscopy indication, age group, diabetes diagnosis, and polyp findings (adenoma with advanced histology, polyp size ≥ 10 mm, and sessile serrated or traditional serrated adenoma).
- They randomly assigned 95,001 patients (mean age, 61.9 years; 45.5% women) who underwent colonoscopy with polypectomy to remove a conventional adenoma into two cohorts: Model development (66,500) and internal validation (28,501).
- In both cohorts, researchers compared the performance of the polyp findings-only method against the comprehensive model in predicting CRC, defined as an adenocarcinoma of the colon or rectum diagnosed a year after the baseline colonoscopy.
TAKEAWAY:
- During the follow-up period starting 1 year after colonoscopy, 495 patients were diagnosed with CRC; 354 were in the development cohort and 141 were in the validation cohort.
- The comprehensive model demonstrated better predictive performance than the traditional polyp-based model in the development cohort (area under the curve [AUC], 0.71 vs 0.61) and in the validation cohort (AUC, 0.7 vs 0.62).
- The difference in the Akaike Information Criterion values between the comprehensive and polyp models was 45.7, much above the threshold of 10, strongly indicating the superior performance of the comprehensive model.
IN PRACTICE:
“Improving the ability to accurately predict the patients at highest risk for CRC after polypectomy is critically important, given the considerable costs and resources associated with treating CRC and the better prognosis associated with early cancer detection. The current findings provide proof of concept that inclusion of CRC risk factors beyond prior polyp findings has the potential to improve post-colonoscopy risk stratification,” the authors wrote.
SOURCE:
The study, led by Jeffrey K. Lee, MD, MPH, Division of Research, Kaiser Permanente Northern California, Oakland, California, was published online in The American Journal of Gastroenterology.
LIMITATIONS:
External validation of the model’s performance is needed in different practice settings. The generalizability of the findings is limited because the study population did not include individuals without a prior adenoma or those with an isolated serrated polyp. Moreover, the examination of polyp size > 20 mm as a potential predictor of CRC was precluded due to incomplete data.
DISCLOSURES:
The study was conducted within the National Cancer Institute–funded Population-Based Research to Optimize the Screening Process II consortium and funded by a career development grant from the National Cancer Institute to Lee. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
A comprehensive model considering patient age, diabetes, colonoscopy indications, and polyp findings can predict colorectal cancer (CRC) risk more accurately than the solely polyp-based model in patients with a first diagnosis of adenoma on colonoscopy.
METHODOLOGY:
- Because colonoscopy surveillance guidelines relying solely on previous polyp findings to assess CRC risk are imprecise, researchers developed and tested a comprehensive risk prediction model from a list of CRC-related predictors that included patient characteristics and clinical factors in addition to polyp findings.
- The comprehensive model included baseline colonoscopy indication, age group, diabetes diagnosis, and polyp findings (adenoma with advanced histology, polyp size ≥ 10 mm, and sessile serrated or traditional serrated adenoma).
- They randomly assigned 95,001 patients (mean age, 61.9 years; 45.5% women) who underwent colonoscopy with polypectomy to remove a conventional adenoma into two cohorts: Model development (66,500) and internal validation (28,501).
- In both cohorts, researchers compared the performance of the polyp findings-only method against the comprehensive model in predicting CRC, defined as an adenocarcinoma of the colon or rectum diagnosed a year after the baseline colonoscopy.
TAKEAWAY:
- During the follow-up period starting 1 year after colonoscopy, 495 patients were diagnosed with CRC; 354 were in the development cohort and 141 were in the validation cohort.
- The comprehensive model demonstrated better predictive performance than the traditional polyp-based model in the development cohort (area under the curve [AUC], 0.71 vs 0.61) and in the validation cohort (AUC, 0.7 vs 0.62).
- The difference in the Akaike Information Criterion values between the comprehensive and polyp models was 45.7, much above the threshold of 10, strongly indicating the superior performance of the comprehensive model.
IN PRACTICE:
“Improving the ability to accurately predict the patients at highest risk for CRC after polypectomy is critically important, given the considerable costs and resources associated with treating CRC and the better prognosis associated with early cancer detection. The current findings provide proof of concept that inclusion of CRC risk factors beyond prior polyp findings has the potential to improve post-colonoscopy risk stratification,” the authors wrote.
SOURCE:
The study, led by Jeffrey K. Lee, MD, MPH, Division of Research, Kaiser Permanente Northern California, Oakland, California, was published online in The American Journal of Gastroenterology.
LIMITATIONS:
External validation of the model’s performance is needed in different practice settings. The generalizability of the findings is limited because the study population did not include individuals without a prior adenoma or those with an isolated serrated polyp. Moreover, the examination of polyp size > 20 mm as a potential predictor of CRC was precluded due to incomplete data.
DISCLOSURES:
The study was conducted within the National Cancer Institute–funded Population-Based Research to Optimize the Screening Process II consortium and funded by a career development grant from the National Cancer Institute to Lee. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
A comprehensive model considering patient age, diabetes, colonoscopy indications, and polyp findings can predict colorectal cancer (CRC) risk more accurately than the solely polyp-based model in patients with a first diagnosis of adenoma on colonoscopy.
METHODOLOGY:
- Because colonoscopy surveillance guidelines relying solely on previous polyp findings to assess CRC risk are imprecise, researchers developed and tested a comprehensive risk prediction model from a list of CRC-related predictors that included patient characteristics and clinical factors in addition to polyp findings.
- The comprehensive model included baseline colonoscopy indication, age group, diabetes diagnosis, and polyp findings (adenoma with advanced histology, polyp size ≥ 10 mm, and sessile serrated or traditional serrated adenoma).
- They randomly assigned 95,001 patients (mean age, 61.9 years; 45.5% women) who underwent colonoscopy with polypectomy to remove a conventional adenoma into two cohorts: Model development (66,500) and internal validation (28,501).
- In both cohorts, researchers compared the performance of the polyp findings-only method against the comprehensive model in predicting CRC, defined as an adenocarcinoma of the colon or rectum diagnosed a year after the baseline colonoscopy.
TAKEAWAY:
- During the follow-up period starting 1 year after colonoscopy, 495 patients were diagnosed with CRC; 354 were in the development cohort and 141 were in the validation cohort.
- The comprehensive model demonstrated better predictive performance than the traditional polyp-based model in the development cohort (area under the curve [AUC], 0.71 vs 0.61) and in the validation cohort (AUC, 0.7 vs 0.62).
- The difference in the Akaike Information Criterion values between the comprehensive and polyp models was 45.7, much above the threshold of 10, strongly indicating the superior performance of the comprehensive model.
IN PRACTICE:
“Improving the ability to accurately predict the patients at highest risk for CRC after polypectomy is critically important, given the considerable costs and resources associated with treating CRC and the better prognosis associated with early cancer detection. The current findings provide proof of concept that inclusion of CRC risk factors beyond prior polyp findings has the potential to improve post-colonoscopy risk stratification,” the authors wrote.
SOURCE:
The study, led by Jeffrey K. Lee, MD, MPH, Division of Research, Kaiser Permanente Northern California, Oakland, California, was published online in The American Journal of Gastroenterology.
LIMITATIONS:
External validation of the model’s performance is needed in different practice settings. The generalizability of the findings is limited because the study population did not include individuals without a prior adenoma or those with an isolated serrated polyp. Moreover, the examination of polyp size > 20 mm as a potential predictor of CRC was precluded due to incomplete data.
DISCLOSURES:
The study was conducted within the National Cancer Institute–funded Population-Based Research to Optimize the Screening Process II consortium and funded by a career development grant from the National Cancer Institute to Lee. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
Immunomodulators Do Not Affect COVID-19 Vaccine Efficacy
TOPLINE:
The results of a recent study suggest that biologics and small molecule inhibitors (SMIs) do not impair the protective effect of COVID-19 vaccine against hospitalization in patients with psoriasis and hidradenitis suppurativa (HS).
METHODOLOGY:
- It remains unknown whether immunomodulatory therapies impair COVID-19 vaccine efficacy and increase hospitalization rates linked to COVID-19 in patients with inflammatory skin conditions such as psoriasis or HS.
- Researchers conducted a cross-sectional study using data from the Epic Cosmos database from January 2020 to October 2023, identifying 30,845 patients with psoriasis or HS.
- Overall, 22,293 patients with documented completion of their primary COVID-19 vaccine series were included in the analysis.
- Of the vaccinated patients, they compared 7046 patients with psoriasis on SMIs and 2033 with psoriasis or HS on biologics with 13,214 patients who did not receive biologics or SMIs.
- The primary outcome was the COVID-19 hospitalization rate.
- Treatment with biologics did not increase COVID-19-related hospitalization rates in vaccinated patients with psoriasis or HS (hospitalization rate, 6.0% for both those taking and those not taking a biologic; P > .99).
- Similarly, hospitalization rates did not significantly differ between vaccinated patients who received SMIs vs those who did not (7.1% vs 6.0%; P = .0596).
IN PRACTICE:
These findings “encourage dermatologists to continue treating [psoriasis]/HS confidently despite the ongoing COVID-19 pandemic,” the authors concluded.
SOURCE:
The study led by Bella R. Lee from Ohio State University Wexner Medical Center, Columbus, was published online on March 13, 2024, in the Journal of the American Academy of Dermatology.
LIMITATIONS:
Multivariable adjustments could not be performed in this study due to unavailability of individual-level data, and hospital admissions that occurred outside the Epic system were not captured.
DISCLOSURES:
The study did not receive any funding. All authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
The results of a recent study suggest that biologics and small molecule inhibitors (SMIs) do not impair the protective effect of COVID-19 vaccine against hospitalization in patients with psoriasis and hidradenitis suppurativa (HS).
METHODOLOGY:
- It remains unknown whether immunomodulatory therapies impair COVID-19 vaccine efficacy and increase hospitalization rates linked to COVID-19 in patients with inflammatory skin conditions such as psoriasis or HS.
- Researchers conducted a cross-sectional study using data from the Epic Cosmos database from January 2020 to October 2023, identifying 30,845 patients with psoriasis or HS.
- Overall, 22,293 patients with documented completion of their primary COVID-19 vaccine series were included in the analysis.
- Of the vaccinated patients, they compared 7046 patients with psoriasis on SMIs and 2033 with psoriasis or HS on biologics with 13,214 patients who did not receive biologics or SMIs.
- The primary outcome was the COVID-19 hospitalization rate.
- Treatment with biologics did not increase COVID-19-related hospitalization rates in vaccinated patients with psoriasis or HS (hospitalization rate, 6.0% for both those taking and those not taking a biologic; P > .99).
- Similarly, hospitalization rates did not significantly differ between vaccinated patients who received SMIs vs those who did not (7.1% vs 6.0%; P = .0596).
IN PRACTICE:
These findings “encourage dermatologists to continue treating [psoriasis]/HS confidently despite the ongoing COVID-19 pandemic,” the authors concluded.
SOURCE:
The study led by Bella R. Lee from Ohio State University Wexner Medical Center, Columbus, was published online on March 13, 2024, in the Journal of the American Academy of Dermatology.
LIMITATIONS:
Multivariable adjustments could not be performed in this study due to unavailability of individual-level data, and hospital admissions that occurred outside the Epic system were not captured.
DISCLOSURES:
The study did not receive any funding. All authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
The results of a recent study suggest that biologics and small molecule inhibitors (SMIs) do not impair the protective effect of COVID-19 vaccine against hospitalization in patients with psoriasis and hidradenitis suppurativa (HS).
METHODOLOGY:
- It remains unknown whether immunomodulatory therapies impair COVID-19 vaccine efficacy and increase hospitalization rates linked to COVID-19 in patients with inflammatory skin conditions such as psoriasis or HS.
- Researchers conducted a cross-sectional study using data from the Epic Cosmos database from January 2020 to October 2023, identifying 30,845 patients with psoriasis or HS.
- Overall, 22,293 patients with documented completion of their primary COVID-19 vaccine series were included in the analysis.
- Of the vaccinated patients, they compared 7046 patients with psoriasis on SMIs and 2033 with psoriasis or HS on biologics with 13,214 patients who did not receive biologics or SMIs.
- The primary outcome was the COVID-19 hospitalization rate.
- Treatment with biologics did not increase COVID-19-related hospitalization rates in vaccinated patients with psoriasis or HS (hospitalization rate, 6.0% for both those taking and those not taking a biologic; P > .99).
- Similarly, hospitalization rates did not significantly differ between vaccinated patients who received SMIs vs those who did not (7.1% vs 6.0%; P = .0596).
IN PRACTICE:
These findings “encourage dermatologists to continue treating [psoriasis]/HS confidently despite the ongoing COVID-19 pandemic,” the authors concluded.
SOURCE:
The study led by Bella R. Lee from Ohio State University Wexner Medical Center, Columbus, was published online on March 13, 2024, in the Journal of the American Academy of Dermatology.
LIMITATIONS:
Multivariable adjustments could not be performed in this study due to unavailability of individual-level data, and hospital admissions that occurred outside the Epic system were not captured.
DISCLOSURES:
The study did not receive any funding. All authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
Pro-Inflammatory Diet, Salt Intake Increases T2D Risk
TOPLINE:
The risk for type 2 diabetes (T2D) was higher in individuals who followed a pro-inflammatory diet and had a high habitual salt intake than in those who followed an anti-inflammatory diet and used less salt.
METHODOLOGY:
- High scores on the dietary inflammatory index (DII) — a scoring system that measures the inflammatory potential of an individual’s diet — and high salt intake are associated with increased cardiovascular disease risk; however, studies investigating the association between DII and salt intake with incident T2D risk are scarce.
- Researchers investigated the association between a pro-inflammatory diet, habitual salt intake, and the risk for T2D among 171,094 participants from the UK Biobank (mean age, 55.98 years; 40.7% men).
- Participants were free of diabetes at baseline, had completed at least one dietary recall questionnaire, and were followed up for a median period of 13.5 years.
- The energy-adjusted DII was calculated on the basis of 28 food and nutrient parameter-specific scores, while habitual salt intake was assessed through self-reported frequency of adding salt to foods.
- Any newly diagnosed cases of T2D were considered the first occurrences of health outcomes.
TAKEAWAY:
- Incident cases of T2D were reported in 6216 individuals over the median follow-up period.
- The risk of developing T2D was 18% higher in individuals who followed a pro-inflammatory vs anti-inflammatory diet (adjusted hazard ratio [HR], 1.18; 95% CI, 1.11-1.25); the risk for T2D was elevated by 4% for each one-point increment in the energy-adjusted DII.
- Compared with participants who never or rarely added salt to foods, the risk for T2D increased gradually in those who sometimes (HR, 1.10; 95% CI, 1.04-1.16), usually (HR, 1.14; 95% CI, 1.05-1.24), and always (HR, 1.30; 95% CI, 1.15-1.47) added salt to foods.
- The risk for T2D was the highest in participants who followed a pro-inflammatory diet and always added salt to foods (HR, 1.60; 95% CI, 1.32-1.90) compared with those who followed an anti-inflammatory diet and never or rarely added salt to foods.
IN PRACTICE:
“Our findings indicate that a pro-inflammatory diet and higher habitual salt intake were associated with an increased risk of type 2 diabetes. These results support the public health promotion of an anti-inflammatory diet and reducing salt intake to prevent the onset of type 2 diabetes,” the authors wrote.
SOURCE:
This study was led by Wenqui Shen, MD, from the Department of Endocrinology and Metabolism, Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China, and published online in Diabetes, Obesity and Metabolism.
LIMITATIONS:
Data from a 24-hour dietary recall questionnaire were used to calculate the energy-adjusted DII, which might have led to incidences of incorrect reporting. This study could not measure all components of the DII score. Unmeasured variables and residual confounders might also be present, which were not considered in this analysis.
DISCLOSURES:
This study was supported by grants from the National Natural Science Foundation of China, Science and Technology Commission of Shanghai Municipality, Project of Biobank from the Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine, and other sources. The authors declared no conflicts of interest.
A version of this article first appeared on Medscape.com.
TOPLINE:
The risk for type 2 diabetes (T2D) was higher in individuals who followed a pro-inflammatory diet and had a high habitual salt intake than in those who followed an anti-inflammatory diet and used less salt.
METHODOLOGY:
- High scores on the dietary inflammatory index (DII) — a scoring system that measures the inflammatory potential of an individual’s diet — and high salt intake are associated with increased cardiovascular disease risk; however, studies investigating the association between DII and salt intake with incident T2D risk are scarce.
- Researchers investigated the association between a pro-inflammatory diet, habitual salt intake, and the risk for T2D among 171,094 participants from the UK Biobank (mean age, 55.98 years; 40.7% men).
- Participants were free of diabetes at baseline, had completed at least one dietary recall questionnaire, and were followed up for a median period of 13.5 years.
- The energy-adjusted DII was calculated on the basis of 28 food and nutrient parameter-specific scores, while habitual salt intake was assessed through self-reported frequency of adding salt to foods.
- Any newly diagnosed cases of T2D were considered the first occurrences of health outcomes.
TAKEAWAY:
- Incident cases of T2D were reported in 6216 individuals over the median follow-up period.
- The risk of developing T2D was 18% higher in individuals who followed a pro-inflammatory vs anti-inflammatory diet (adjusted hazard ratio [HR], 1.18; 95% CI, 1.11-1.25); the risk for T2D was elevated by 4% for each one-point increment in the energy-adjusted DII.
- Compared with participants who never or rarely added salt to foods, the risk for T2D increased gradually in those who sometimes (HR, 1.10; 95% CI, 1.04-1.16), usually (HR, 1.14; 95% CI, 1.05-1.24), and always (HR, 1.30; 95% CI, 1.15-1.47) added salt to foods.
- The risk for T2D was the highest in participants who followed a pro-inflammatory diet and always added salt to foods (HR, 1.60; 95% CI, 1.32-1.90) compared with those who followed an anti-inflammatory diet and never or rarely added salt to foods.
IN PRACTICE:
“Our findings indicate that a pro-inflammatory diet and higher habitual salt intake were associated with an increased risk of type 2 diabetes. These results support the public health promotion of an anti-inflammatory diet and reducing salt intake to prevent the onset of type 2 diabetes,” the authors wrote.
SOURCE:
This study was led by Wenqui Shen, MD, from the Department of Endocrinology and Metabolism, Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China, and published online in Diabetes, Obesity and Metabolism.
LIMITATIONS:
Data from a 24-hour dietary recall questionnaire were used to calculate the energy-adjusted DII, which might have led to incidences of incorrect reporting. This study could not measure all components of the DII score. Unmeasured variables and residual confounders might also be present, which were not considered in this analysis.
DISCLOSURES:
This study was supported by grants from the National Natural Science Foundation of China, Science and Technology Commission of Shanghai Municipality, Project of Biobank from the Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine, and other sources. The authors declared no conflicts of interest.
A version of this article first appeared on Medscape.com.
TOPLINE:
The risk for type 2 diabetes (T2D) was higher in individuals who followed a pro-inflammatory diet and had a high habitual salt intake than in those who followed an anti-inflammatory diet and used less salt.
METHODOLOGY:
- High scores on the dietary inflammatory index (DII) — a scoring system that measures the inflammatory potential of an individual’s diet — and high salt intake are associated with increased cardiovascular disease risk; however, studies investigating the association between DII and salt intake with incident T2D risk are scarce.
- Researchers investigated the association between a pro-inflammatory diet, habitual salt intake, and the risk for T2D among 171,094 participants from the UK Biobank (mean age, 55.98 years; 40.7% men).
- Participants were free of diabetes at baseline, had completed at least one dietary recall questionnaire, and were followed up for a median period of 13.5 years.
- The energy-adjusted DII was calculated on the basis of 28 food and nutrient parameter-specific scores, while habitual salt intake was assessed through self-reported frequency of adding salt to foods.
- Any newly diagnosed cases of T2D were considered the first occurrences of health outcomes.
TAKEAWAY:
- Incident cases of T2D were reported in 6216 individuals over the median follow-up period.
- The risk of developing T2D was 18% higher in individuals who followed a pro-inflammatory vs anti-inflammatory diet (adjusted hazard ratio [HR], 1.18; 95% CI, 1.11-1.25); the risk for T2D was elevated by 4% for each one-point increment in the energy-adjusted DII.
- Compared with participants who never or rarely added salt to foods, the risk for T2D increased gradually in those who sometimes (HR, 1.10; 95% CI, 1.04-1.16), usually (HR, 1.14; 95% CI, 1.05-1.24), and always (HR, 1.30; 95% CI, 1.15-1.47) added salt to foods.
- The risk for T2D was the highest in participants who followed a pro-inflammatory diet and always added salt to foods (HR, 1.60; 95% CI, 1.32-1.90) compared with those who followed an anti-inflammatory diet and never or rarely added salt to foods.
IN PRACTICE:
“Our findings indicate that a pro-inflammatory diet and higher habitual salt intake were associated with an increased risk of type 2 diabetes. These results support the public health promotion of an anti-inflammatory diet and reducing salt intake to prevent the onset of type 2 diabetes,” the authors wrote.
SOURCE:
This study was led by Wenqui Shen, MD, from the Department of Endocrinology and Metabolism, Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China, and published online in Diabetes, Obesity and Metabolism.
LIMITATIONS:
Data from a 24-hour dietary recall questionnaire were used to calculate the energy-adjusted DII, which might have led to incidences of incorrect reporting. This study could not measure all components of the DII score. Unmeasured variables and residual confounders might also be present, which were not considered in this analysis.
DISCLOSURES:
This study was supported by grants from the National Natural Science Foundation of China, Science and Technology Commission of Shanghai Municipality, Project of Biobank from the Shanghai Ninth People’s Hospital, Shanghai Jiao Tong University School of Medicine, and other sources. The authors declared no conflicts of interest.
A version of this article first appeared on Medscape.com.
Does Abdominal Fat Location Matter for Brain Health?
TOPLINE:
METHODOLOGY:
- Obesity is a well-known risk factor for poorer cognition and dementia, but the distribution of body fat may influence the risk and underlying mechanisms in the fat-brain-cognition pathway.
- The study examined associations of several abdominal fat depots with cognitive functioning and AD-related brain volumes.
- The study sample included 204 men and women from the Israel Registry for Alzheimer’s Prevention (mean age, 59 years; 60% women) who had a high AD risk due to parental family history.
- Abdominal MRI scans assessed fat stored as subcutaneous adipose tissue (SAT) beneath the skin, visceral adipose tissue (VAT) around abdominal organs, and ectopic, a harmful condition in which lipids accumulate in lean tissues such as the liver and pancreas.
- A structural volumetric brain MRI scan was undertaken by 142 participants to assess specific regions implicated in chosen previous research.
TAKEAWAY:
- High body mass index was associated with high pancreatic fat percentage in both men and women (P < .001) and with high SAT percentage in women (P = .01) but not with VAT percentage in either sex.
- After adjustment for cardiovascular risk factors, a higher pancreatic fat percentage was linked to lower global cognition (beta, −0.33; P = .02) and executive function (beta, −0.32; P = .02) in men, and with lower hippocampal volume in women (beta, −0.25; P = .03).
- In men only, a higher SAT percentage was associated with a lower middle frontal gyrus volume (beta, −0.27; P = .03), while a higher VAT percentage was linked to higher middle frontal gyrus (beta, 0.29; P = .03) and superior frontal gyrus volumes (beta, 0.31; P = .02).
- Hepatic fat was not associated with brain volumes or cognition in either men or women.
IN PRACTICE:
“These results suggest that already in midlife, abdominal fat accumulation may have deleterious effects on brain health, especially in men,” the authors wrote.
SOURCE:
This study was led by Sapir G. Shekhtman, Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel, and published online in Obesity (Silver Spring).
LIMITATIONS:
No causal inferences could be drawn from this study due to its cross-sectional nature. It did not represent the population of middle-aged adults as a whole, but rather those at high risk of developing AD. Factors contributing to fat accumulation, such as menopausal status or treatment, inflammation, insulin resistance, daily exercise, and dietary factors, were not included in this study.
DISCLOSURES:
This work was supported by grants from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
METHODOLOGY:
- Obesity is a well-known risk factor for poorer cognition and dementia, but the distribution of body fat may influence the risk and underlying mechanisms in the fat-brain-cognition pathway.
- The study examined associations of several abdominal fat depots with cognitive functioning and AD-related brain volumes.
- The study sample included 204 men and women from the Israel Registry for Alzheimer’s Prevention (mean age, 59 years; 60% women) who had a high AD risk due to parental family history.
- Abdominal MRI scans assessed fat stored as subcutaneous adipose tissue (SAT) beneath the skin, visceral adipose tissue (VAT) around abdominal organs, and ectopic, a harmful condition in which lipids accumulate in lean tissues such as the liver and pancreas.
- A structural volumetric brain MRI scan was undertaken by 142 participants to assess specific regions implicated in chosen previous research.
TAKEAWAY:
- High body mass index was associated with high pancreatic fat percentage in both men and women (P < .001) and with high SAT percentage in women (P = .01) but not with VAT percentage in either sex.
- After adjustment for cardiovascular risk factors, a higher pancreatic fat percentage was linked to lower global cognition (beta, −0.33; P = .02) and executive function (beta, −0.32; P = .02) in men, and with lower hippocampal volume in women (beta, −0.25; P = .03).
- In men only, a higher SAT percentage was associated with a lower middle frontal gyrus volume (beta, −0.27; P = .03), while a higher VAT percentage was linked to higher middle frontal gyrus (beta, 0.29; P = .03) and superior frontal gyrus volumes (beta, 0.31; P = .02).
- Hepatic fat was not associated with brain volumes or cognition in either men or women.
IN PRACTICE:
“These results suggest that already in midlife, abdominal fat accumulation may have deleterious effects on brain health, especially in men,” the authors wrote.
SOURCE:
This study was led by Sapir G. Shekhtman, Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel, and published online in Obesity (Silver Spring).
LIMITATIONS:
No causal inferences could be drawn from this study due to its cross-sectional nature. It did not represent the population of middle-aged adults as a whole, but rather those at high risk of developing AD. Factors contributing to fat accumulation, such as menopausal status or treatment, inflammation, insulin resistance, daily exercise, and dietary factors, were not included in this study.
DISCLOSURES:
This work was supported by grants from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
METHODOLOGY:
- Obesity is a well-known risk factor for poorer cognition and dementia, but the distribution of body fat may influence the risk and underlying mechanisms in the fat-brain-cognition pathway.
- The study examined associations of several abdominal fat depots with cognitive functioning and AD-related brain volumes.
- The study sample included 204 men and women from the Israel Registry for Alzheimer’s Prevention (mean age, 59 years; 60% women) who had a high AD risk due to parental family history.
- Abdominal MRI scans assessed fat stored as subcutaneous adipose tissue (SAT) beneath the skin, visceral adipose tissue (VAT) around abdominal organs, and ectopic, a harmful condition in which lipids accumulate in lean tissues such as the liver and pancreas.
- A structural volumetric brain MRI scan was undertaken by 142 participants to assess specific regions implicated in chosen previous research.
TAKEAWAY:
- High body mass index was associated with high pancreatic fat percentage in both men and women (P < .001) and with high SAT percentage in women (P = .01) but not with VAT percentage in either sex.
- After adjustment for cardiovascular risk factors, a higher pancreatic fat percentage was linked to lower global cognition (beta, −0.33; P = .02) and executive function (beta, −0.32; P = .02) in men, and with lower hippocampal volume in women (beta, −0.25; P = .03).
- In men only, a higher SAT percentage was associated with a lower middle frontal gyrus volume (beta, −0.27; P = .03), while a higher VAT percentage was linked to higher middle frontal gyrus (beta, 0.29; P = .03) and superior frontal gyrus volumes (beta, 0.31; P = .02).
- Hepatic fat was not associated with brain volumes or cognition in either men or women.
IN PRACTICE:
“These results suggest that already in midlife, abdominal fat accumulation may have deleterious effects on brain health, especially in men,” the authors wrote.
SOURCE:
This study was led by Sapir G. Shekhtman, Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel, and published online in Obesity (Silver Spring).
LIMITATIONS:
No causal inferences could be drawn from this study due to its cross-sectional nature. It did not represent the population of middle-aged adults as a whole, but rather those at high risk of developing AD. Factors contributing to fat accumulation, such as menopausal status or treatment, inflammation, insulin resistance, daily exercise, and dietary factors, were not included in this study.
DISCLOSURES:
This work was supported by grants from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
Consider These Factors in an Academic Radiation Oncology Position
TOPLINE:
— and accept an offer if the practice is “great” in at least two of those areas and “good” in the third, experts say in a recent editorial.
METHODOLOGY:
- Many physicians choose to go into academic medicine because they want to stay involved in research and education while still treating patients.
- However, graduating radiation oncology residents often lack or have limited guidance on what to look for in a prospective job and how to assess their contract.
- This recent editorial provides guidance to radiation oncologists seeking academic positions. The authors advise prospective employees to evaluate three main factors — compensation, daily duties, and location — as well as provide tips for identifying red flags in each category.
TAKEAWAY:
- Compensation: Prospective faculty should assess both direct compensation, that is, salary, and indirect compensation, which typically includes retirement contributions and other perks. For direct compensation, what is the base salary? Is extra work compensated? How does the salary offer measure up to salary data reported by national agencies? Also: Don’t overlook uncompensated duties, such as time in tumor boards or in meetings, which may be time-consuming, and make sure compensation terms are clearly delineated in a contract and equitable among physicians in a specific rank.
- Daily duties: When it comes to daily life on the job, a prospective employee should consider many factors, including the cancer center’s excitement to hire you, the reputation of the faculty and leaders at the organization, employee turnover rates, diversity among faculty, and the time line of career advancement.
- Location: The location of the job encompasses the geography — such as distance from home to work, the number of practices covered, cost of living, and the area itself — as well as the atmosphere for conducting research and publishing.
- Finally, carefully review the job contract. All the key aspects of the job, including compensation and benefits, should be clearly stated in the contract to “improve communication of expectations.”
IN PRACTICE:
“A prospective faculty member can ask 100 questions, but they can’t make 100 demands; consideration of the three domains can help to focus negotiation efforts where the efforts are needed,” the authors noted.
SOURCE:
This editorial, led by Nicholas G. Zaorsky from the Department of Radiation Oncology, University Hospitals Seidman Cancer Center, Case Western Reserve School of Medicine, Cleveland, Ohio, was published online in Practical Radiation Oncology
DISCLOSURES:
The lead author declared being supported by the American Cancer Society and National Institutes of Health. He also reported having ties with many other sources.
A version of this article appeared on Medscape.com.
TOPLINE:
— and accept an offer if the practice is “great” in at least two of those areas and “good” in the third, experts say in a recent editorial.
METHODOLOGY:
- Many physicians choose to go into academic medicine because they want to stay involved in research and education while still treating patients.
- However, graduating radiation oncology residents often lack or have limited guidance on what to look for in a prospective job and how to assess their contract.
- This recent editorial provides guidance to radiation oncologists seeking academic positions. The authors advise prospective employees to evaluate three main factors — compensation, daily duties, and location — as well as provide tips for identifying red flags in each category.
TAKEAWAY:
- Compensation: Prospective faculty should assess both direct compensation, that is, salary, and indirect compensation, which typically includes retirement contributions and other perks. For direct compensation, what is the base salary? Is extra work compensated? How does the salary offer measure up to salary data reported by national agencies? Also: Don’t overlook uncompensated duties, such as time in tumor boards or in meetings, which may be time-consuming, and make sure compensation terms are clearly delineated in a contract and equitable among physicians in a specific rank.
- Daily duties: When it comes to daily life on the job, a prospective employee should consider many factors, including the cancer center’s excitement to hire you, the reputation of the faculty and leaders at the organization, employee turnover rates, diversity among faculty, and the time line of career advancement.
- Location: The location of the job encompasses the geography — such as distance from home to work, the number of practices covered, cost of living, and the area itself — as well as the atmosphere for conducting research and publishing.
- Finally, carefully review the job contract. All the key aspects of the job, including compensation and benefits, should be clearly stated in the contract to “improve communication of expectations.”
IN PRACTICE:
“A prospective faculty member can ask 100 questions, but they can’t make 100 demands; consideration of the three domains can help to focus negotiation efforts where the efforts are needed,” the authors noted.
SOURCE:
This editorial, led by Nicholas G. Zaorsky from the Department of Radiation Oncology, University Hospitals Seidman Cancer Center, Case Western Reserve School of Medicine, Cleveland, Ohio, was published online in Practical Radiation Oncology
DISCLOSURES:
The lead author declared being supported by the American Cancer Society and National Institutes of Health. He also reported having ties with many other sources.
A version of this article appeared on Medscape.com.
TOPLINE:
— and accept an offer if the practice is “great” in at least two of those areas and “good” in the third, experts say in a recent editorial.
METHODOLOGY:
- Many physicians choose to go into academic medicine because they want to stay involved in research and education while still treating patients.
- However, graduating radiation oncology residents often lack or have limited guidance on what to look for in a prospective job and how to assess their contract.
- This recent editorial provides guidance to radiation oncologists seeking academic positions. The authors advise prospective employees to evaluate three main factors — compensation, daily duties, and location — as well as provide tips for identifying red flags in each category.
TAKEAWAY:
- Compensation: Prospective faculty should assess both direct compensation, that is, salary, and indirect compensation, which typically includes retirement contributions and other perks. For direct compensation, what is the base salary? Is extra work compensated? How does the salary offer measure up to salary data reported by national agencies? Also: Don’t overlook uncompensated duties, such as time in tumor boards or in meetings, which may be time-consuming, and make sure compensation terms are clearly delineated in a contract and equitable among physicians in a specific rank.
- Daily duties: When it comes to daily life on the job, a prospective employee should consider many factors, including the cancer center’s excitement to hire you, the reputation of the faculty and leaders at the organization, employee turnover rates, diversity among faculty, and the time line of career advancement.
- Location: The location of the job encompasses the geography — such as distance from home to work, the number of practices covered, cost of living, and the area itself — as well as the atmosphere for conducting research and publishing.
- Finally, carefully review the job contract. All the key aspects of the job, including compensation and benefits, should be clearly stated in the contract to “improve communication of expectations.”
IN PRACTICE:
“A prospective faculty member can ask 100 questions, but they can’t make 100 demands; consideration of the three domains can help to focus negotiation efforts where the efforts are needed,” the authors noted.
SOURCE:
This editorial, led by Nicholas G. Zaorsky from the Department of Radiation Oncology, University Hospitals Seidman Cancer Center, Case Western Reserve School of Medicine, Cleveland, Ohio, was published online in Practical Radiation Oncology
DISCLOSURES:
The lead author declared being supported by the American Cancer Society and National Institutes of Health. He also reported having ties with many other sources.
A version of this article appeared on Medscape.com.
Nurse-Led Strategy Reduces Cholesterol, BP in HIV
TOPLINE:
A multicomponent strategy of nurse-led communication, home blood pressure monitoring, evidence-based treatment algorithms, and electronic health record tools improved systolic blood pressure (SBP) and non–high-density lipoprotein (non-HDL) cholesterol levels in people living with HIV.
METHODOLOGY:
- Investigators assessed if EXTRA-CVD, a nurse-led multicomponent intervention for preventing cardiovascular diseases (CVD), could effectively improve SBP and non-HDL cholesterol levels in people living with HIV whose viral replication has been controlled effectively using antiretroviral therapy.
- They recruited 297 individuals (median age, 59 years; 20.9% women) from three academic HIV clinics in the United States with an HIV-1 viral load < 200 copies/mL who were diagnosed with both hypertension and hypercholesterolemia.
- Participants were randomly assigned to either the EXTRA-CVD intervention group or a control group comprising individuals who received general prevention education.
- SBP (the primary outcome) was calculated as the mean of two SBP measurements obtained 1 minute apart, and non-HDL cholesterol (the secondary outcome) was calculated as total cholesterol minus HDL cholesterol.
TAKEAWAY:
- Participants in the intervention vs control group reported having significantly lower SBP as early as 4 months after the nurse-led strategy (mean difference, −6.4 mm Hg; P = .002), with the improvements sustaining until 12 months (mean difference, −4.2 mm Hg; P = .04).
- At 12 months, participants in the intervention group showed a 16.9-mg/dL (P < .001) reduction in non-HDL cholesterol levels compared with those in the control group.
- The nurse-led strategy led to a greater reduction in SBP in women with HIV vs men living with HIV (5.9 mm Hg greater SBP difference at 12 months), with the difference being clinically meaningful but not statistically significant.
- This nurse-led strategy did not increase the risk for adverse events in people living with HIV.
IN PRACTICE:
“Although the EXTRA-CVD intervention was limited to BP and cholesterol, nurse-led case management might be beneficial for a range of other primary care conditions in HIV clinics. If HIV clinics choose to implement EXTRA-CVD, they might consider adding staff trained in other chronic comorbidities and/or health promotion activities,” the authors noted.
SOURCE:
This study was led by Christopher T. Longenecker, MD, University of Washington School of Medicine, Seattle, and published online on March 5, 2024, in JAMA Network Open.
LIMITATIONS:
Because this trial was conducted at well-resourced, major academic HIV clinics, the results may not be applicable to other populations, such as smaller community-based clinics or HIV care outside the United States. The sensitivity analyses performed in this study may not have fully accounted for the bias introduced by the differential attrition in the intervention group.
DISCLOSURES:
This study was supported by grants from the National Institutes of Health (NIH). The authors declared receiving grants and personal fees from or having other ties with the NIH and other sources.
A version of this article appeared on Medscape.com.
TOPLINE:
A multicomponent strategy of nurse-led communication, home blood pressure monitoring, evidence-based treatment algorithms, and electronic health record tools improved systolic blood pressure (SBP) and non–high-density lipoprotein (non-HDL) cholesterol levels in people living with HIV.
METHODOLOGY:
- Investigators assessed if EXTRA-CVD, a nurse-led multicomponent intervention for preventing cardiovascular diseases (CVD), could effectively improve SBP and non-HDL cholesterol levels in people living with HIV whose viral replication has been controlled effectively using antiretroviral therapy.
- They recruited 297 individuals (median age, 59 years; 20.9% women) from three academic HIV clinics in the United States with an HIV-1 viral load < 200 copies/mL who were diagnosed with both hypertension and hypercholesterolemia.
- Participants were randomly assigned to either the EXTRA-CVD intervention group or a control group comprising individuals who received general prevention education.
- SBP (the primary outcome) was calculated as the mean of two SBP measurements obtained 1 minute apart, and non-HDL cholesterol (the secondary outcome) was calculated as total cholesterol minus HDL cholesterol.
TAKEAWAY:
- Participants in the intervention vs control group reported having significantly lower SBP as early as 4 months after the nurse-led strategy (mean difference, −6.4 mm Hg; P = .002), with the improvements sustaining until 12 months (mean difference, −4.2 mm Hg; P = .04).
- At 12 months, participants in the intervention group showed a 16.9-mg/dL (P < .001) reduction in non-HDL cholesterol levels compared with those in the control group.
- The nurse-led strategy led to a greater reduction in SBP in women with HIV vs men living with HIV (5.9 mm Hg greater SBP difference at 12 months), with the difference being clinically meaningful but not statistically significant.
- This nurse-led strategy did not increase the risk for adverse events in people living with HIV.
IN PRACTICE:
“Although the EXTRA-CVD intervention was limited to BP and cholesterol, nurse-led case management might be beneficial for a range of other primary care conditions in HIV clinics. If HIV clinics choose to implement EXTRA-CVD, they might consider adding staff trained in other chronic comorbidities and/or health promotion activities,” the authors noted.
SOURCE:
This study was led by Christopher T. Longenecker, MD, University of Washington School of Medicine, Seattle, and published online on March 5, 2024, in JAMA Network Open.
LIMITATIONS:
Because this trial was conducted at well-resourced, major academic HIV clinics, the results may not be applicable to other populations, such as smaller community-based clinics or HIV care outside the United States. The sensitivity analyses performed in this study may not have fully accounted for the bias introduced by the differential attrition in the intervention group.
DISCLOSURES:
This study was supported by grants from the National Institutes of Health (NIH). The authors declared receiving grants and personal fees from or having other ties with the NIH and other sources.
A version of this article appeared on Medscape.com.
TOPLINE:
A multicomponent strategy of nurse-led communication, home blood pressure monitoring, evidence-based treatment algorithms, and electronic health record tools improved systolic blood pressure (SBP) and non–high-density lipoprotein (non-HDL) cholesterol levels in people living with HIV.
METHODOLOGY:
- Investigators assessed if EXTRA-CVD, a nurse-led multicomponent intervention for preventing cardiovascular diseases (CVD), could effectively improve SBP and non-HDL cholesterol levels in people living with HIV whose viral replication has been controlled effectively using antiretroviral therapy.
- They recruited 297 individuals (median age, 59 years; 20.9% women) from three academic HIV clinics in the United States with an HIV-1 viral load < 200 copies/mL who were diagnosed with both hypertension and hypercholesterolemia.
- Participants were randomly assigned to either the EXTRA-CVD intervention group or a control group comprising individuals who received general prevention education.
- SBP (the primary outcome) was calculated as the mean of two SBP measurements obtained 1 minute apart, and non-HDL cholesterol (the secondary outcome) was calculated as total cholesterol minus HDL cholesterol.
TAKEAWAY:
- Participants in the intervention vs control group reported having significantly lower SBP as early as 4 months after the nurse-led strategy (mean difference, −6.4 mm Hg; P = .002), with the improvements sustaining until 12 months (mean difference, −4.2 mm Hg; P = .04).
- At 12 months, participants in the intervention group showed a 16.9-mg/dL (P < .001) reduction in non-HDL cholesterol levels compared with those in the control group.
- The nurse-led strategy led to a greater reduction in SBP in women with HIV vs men living with HIV (5.9 mm Hg greater SBP difference at 12 months), with the difference being clinically meaningful but not statistically significant.
- This nurse-led strategy did not increase the risk for adverse events in people living with HIV.
IN PRACTICE:
“Although the EXTRA-CVD intervention was limited to BP and cholesterol, nurse-led case management might be beneficial for a range of other primary care conditions in HIV clinics. If HIV clinics choose to implement EXTRA-CVD, they might consider adding staff trained in other chronic comorbidities and/or health promotion activities,” the authors noted.
SOURCE:
This study was led by Christopher T. Longenecker, MD, University of Washington School of Medicine, Seattle, and published online on March 5, 2024, in JAMA Network Open.
LIMITATIONS:
Because this trial was conducted at well-resourced, major academic HIV clinics, the results may not be applicable to other populations, such as smaller community-based clinics or HIV care outside the United States. The sensitivity analyses performed in this study may not have fully accounted for the bias introduced by the differential attrition in the intervention group.
DISCLOSURES:
This study was supported by grants from the National Institutes of Health (NIH). The authors declared receiving grants and personal fees from or having other ties with the NIH and other sources.
A version of this article appeared on Medscape.com.
After SABR, 100% Local Control at 1 Year in Kidney Cancer
TOPLINE:
METHODOLOGY:
- SABR is a promising treatment strategy for patients with inoperable kidney cancer because it is a noninvasive procedure that does not require general anesthesia and can be used to treat stages TIa and TIb, as well as larger tumors.
- The nonrandomized, phase 2, FASTRACK II trial, conducted in Australia and the Netherlands, investigated the efficacy of SABR in 70 patients with primary RCC who had a single lesion and were considered medically inoperable, were at a high risk for surgical complications, or had declined surgery. Patients also had an Eastern Cooperative Oncology Group performance status of ≤ 2 and an estimated glomerular filtration rate above 30 mL/min.
- The median age of participants was 77 years, median body mass index was 32, and the median Charlson comorbidity index was 7; 30% of the patients were women.
- Patients with tumors ≤ 4 cm (n = 23) received a single fraction of 26 Gy SABR, while those with tumors of 4-10 cm in maximum diameter (n = 47) received 42 Gy SABR in three fractions. The median tumor size was 4.6 cm.
- The primary endpoint was local control, defined as no progression of the primary RCC.
TAKEAWAY:
- At 1 year, no patients experienced local progression of their cancer, for a 100% local control rate.
- Cancer-specific survival was also 100% at 12 months from the start of SABR treatment, while the overall survival rate was 99% at 1 year and 82% at 3 years.
- Treatment-related grade 3 adverse events, such as nausea and vomiting, colonic obstruction, and diarrhea were reported by 10% of patients, with no incidences of grade 4 treatment-related adverse events or treatment-related or cancer-related deaths.
IN PRACTICE:
“Despite a larger average tumor size (4.6 cm) than in many preexisting prospective trials of surgery or SABR in primary renal cell cancer, there were no local treatment failures observed and no patients died from cancer during the study period,” the authors noted. This trial and others “support SABR as a therapeutic option for patients with inoperable or high-risk primary renal cell cancer.”
SOURCE:
This study was led by Shankar Siva, PhD, Department of Radiation Oncology, Peter MacCallum Cancer Centre, Melbourne, Australia, and published online in The Lancet Oncology.
LIMITATIONS:
The study was limited by a small sample size and less mature data at follow-up. The absence of a control group made it impossible to assess if SABR had superior, inferior, or similar efficacy to other treatment options. The definitions of operability or technically high risk might vary between different multidisciplinary teams.
DISCLOSURES:
This study was funded by a grant from the Cancer Australia Priority-driven Collaborative Cancer Research Scheme. The study authors declared receiving grants, contracts, payments, honoraria, and research funding and having other ties with several sources.
A version of this article first appeared on Medscape.com.
TOPLINE:
METHODOLOGY:
- SABR is a promising treatment strategy for patients with inoperable kidney cancer because it is a noninvasive procedure that does not require general anesthesia and can be used to treat stages TIa and TIb, as well as larger tumors.
- The nonrandomized, phase 2, FASTRACK II trial, conducted in Australia and the Netherlands, investigated the efficacy of SABR in 70 patients with primary RCC who had a single lesion and were considered medically inoperable, were at a high risk for surgical complications, or had declined surgery. Patients also had an Eastern Cooperative Oncology Group performance status of ≤ 2 and an estimated glomerular filtration rate above 30 mL/min.
- The median age of participants was 77 years, median body mass index was 32, and the median Charlson comorbidity index was 7; 30% of the patients were women.
- Patients with tumors ≤ 4 cm (n = 23) received a single fraction of 26 Gy SABR, while those with tumors of 4-10 cm in maximum diameter (n = 47) received 42 Gy SABR in three fractions. The median tumor size was 4.6 cm.
- The primary endpoint was local control, defined as no progression of the primary RCC.
TAKEAWAY:
- At 1 year, no patients experienced local progression of their cancer, for a 100% local control rate.
- Cancer-specific survival was also 100% at 12 months from the start of SABR treatment, while the overall survival rate was 99% at 1 year and 82% at 3 years.
- Treatment-related grade 3 adverse events, such as nausea and vomiting, colonic obstruction, and diarrhea were reported by 10% of patients, with no incidences of grade 4 treatment-related adverse events or treatment-related or cancer-related deaths.
IN PRACTICE:
“Despite a larger average tumor size (4.6 cm) than in many preexisting prospective trials of surgery or SABR in primary renal cell cancer, there were no local treatment failures observed and no patients died from cancer during the study period,” the authors noted. This trial and others “support SABR as a therapeutic option for patients with inoperable or high-risk primary renal cell cancer.”
SOURCE:
This study was led by Shankar Siva, PhD, Department of Radiation Oncology, Peter MacCallum Cancer Centre, Melbourne, Australia, and published online in The Lancet Oncology.
LIMITATIONS:
The study was limited by a small sample size and less mature data at follow-up. The absence of a control group made it impossible to assess if SABR had superior, inferior, or similar efficacy to other treatment options. The definitions of operability or technically high risk might vary between different multidisciplinary teams.
DISCLOSURES:
This study was funded by a grant from the Cancer Australia Priority-driven Collaborative Cancer Research Scheme. The study authors declared receiving grants, contracts, payments, honoraria, and research funding and having other ties with several sources.
A version of this article first appeared on Medscape.com.
TOPLINE:
METHODOLOGY:
- SABR is a promising treatment strategy for patients with inoperable kidney cancer because it is a noninvasive procedure that does not require general anesthesia and can be used to treat stages TIa and TIb, as well as larger tumors.
- The nonrandomized, phase 2, FASTRACK II trial, conducted in Australia and the Netherlands, investigated the efficacy of SABR in 70 patients with primary RCC who had a single lesion and were considered medically inoperable, were at a high risk for surgical complications, or had declined surgery. Patients also had an Eastern Cooperative Oncology Group performance status of ≤ 2 and an estimated glomerular filtration rate above 30 mL/min.
- The median age of participants was 77 years, median body mass index was 32, and the median Charlson comorbidity index was 7; 30% of the patients were women.
- Patients with tumors ≤ 4 cm (n = 23) received a single fraction of 26 Gy SABR, while those with tumors of 4-10 cm in maximum diameter (n = 47) received 42 Gy SABR in three fractions. The median tumor size was 4.6 cm.
- The primary endpoint was local control, defined as no progression of the primary RCC.
TAKEAWAY:
- At 1 year, no patients experienced local progression of their cancer, for a 100% local control rate.
- Cancer-specific survival was also 100% at 12 months from the start of SABR treatment, while the overall survival rate was 99% at 1 year and 82% at 3 years.
- Treatment-related grade 3 adverse events, such as nausea and vomiting, colonic obstruction, and diarrhea were reported by 10% of patients, with no incidences of grade 4 treatment-related adverse events or treatment-related or cancer-related deaths.
IN PRACTICE:
“Despite a larger average tumor size (4.6 cm) than in many preexisting prospective trials of surgery or SABR in primary renal cell cancer, there were no local treatment failures observed and no patients died from cancer during the study period,” the authors noted. This trial and others “support SABR as a therapeutic option for patients with inoperable or high-risk primary renal cell cancer.”
SOURCE:
This study was led by Shankar Siva, PhD, Department of Radiation Oncology, Peter MacCallum Cancer Centre, Melbourne, Australia, and published online in The Lancet Oncology.
LIMITATIONS:
The study was limited by a small sample size and less mature data at follow-up. The absence of a control group made it impossible to assess if SABR had superior, inferior, or similar efficacy to other treatment options. The definitions of operability or technically high risk might vary between different multidisciplinary teams.
DISCLOSURES:
This study was funded by a grant from the Cancer Australia Priority-driven Collaborative Cancer Research Scheme. The study authors declared receiving grants, contracts, payments, honoraria, and research funding and having other ties with several sources.
A version of this article first appeared on Medscape.com.
Study Sounds Alert About GLP-1 RA Use and Aspiration Risk
TOPLINE:
Patients on weekly glucagon-like peptide-1 receptor agonists (GLP-1 RAs) have high residual gastric content, a major risk factor for aspiration under anesthesia, despite following fasting guidelines before undergoing elective procedures.
METHODOLOGY:
- The increasing use of GLP-1 RAs to manage weight and hyperglycemia has sparked safety concerns because of the drugs’ association with slow gastric emptying, a major risk factor for aspiration under anesthesia.
- This cross-sectional study used gastric ultrasonography to examine the link between GLP-1 RA use and the prevalence of increased residual gastric content.
- All 124 participants (median age, 56 years; 60% women) — half of whom received once-weekly GLP-1 RAs such as semaglutide, dulaglutide, or tirzepatide — adhered to the guideline-recommended fasting duration before undergoing elective procedures under anesthesia.
- The primary outcome focused on identifying increased residual gastric content, defined by the presence of solids, thick liquids, or > 1.5 mL/kg of clear liquids on ultrasound.
- An exploratory analysis examined the association between the duration of GLP-1 RA discontinuation and increased residual gastric content.
TAKEAWAY:
- The adjusted prevalence of increased residual gastric content was 30.5% (95% CI, 9.9%-51.2%) higher in participants who received GLP-1 RA than those who did not.
- Most patients took their last dose of GLP-1 RA within 5 days before their procedure, but elevated residual gastric content persisted even after 7 days of GLP-1 RA discontinuation.
- There was also no significant association between the type of GLP-1 RA used and the prevalence of increased residual gastric content.
IN PRACTICE:
“We expect healthcare professionals will encounter these classes of drugs with increasing frequency in the perioperative period. Perioperative physicians, including anesthesiologists, surgeons, and primary care physicians, should be well-informed about the safety implications of GLP-1 RA drugs,” the authors wrote.
SOURCE:
The study was led by Sudipta Sen, MD, from the Department of Anesthesiology, Critical Care and Pain Medicine, McGovern Medical School, University of Texas Health Center at Houston, Houston, Texas, and published online in JAMA Surgery.
LIMITATIONS:
Residual gastric content, the primary outcome, served as a proxy for aspiration risk and does not have an exact threshold of volume associated with increased risk. The study did not directly evaluate aspiration events. The authors also acknowledged potential bias from unmeasured confounders owing to the observational nature of this study. A small sample size limited the ability to detect a risk difference for each additional day of drug discontinuation before surgery.
DISCLOSURES:
One of the authors reported receiving a grant from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Patients on weekly glucagon-like peptide-1 receptor agonists (GLP-1 RAs) have high residual gastric content, a major risk factor for aspiration under anesthesia, despite following fasting guidelines before undergoing elective procedures.
METHODOLOGY:
- The increasing use of GLP-1 RAs to manage weight and hyperglycemia has sparked safety concerns because of the drugs’ association with slow gastric emptying, a major risk factor for aspiration under anesthesia.
- This cross-sectional study used gastric ultrasonography to examine the link between GLP-1 RA use and the prevalence of increased residual gastric content.
- All 124 participants (median age, 56 years; 60% women) — half of whom received once-weekly GLP-1 RAs such as semaglutide, dulaglutide, or tirzepatide — adhered to the guideline-recommended fasting duration before undergoing elective procedures under anesthesia.
- The primary outcome focused on identifying increased residual gastric content, defined by the presence of solids, thick liquids, or > 1.5 mL/kg of clear liquids on ultrasound.
- An exploratory analysis examined the association between the duration of GLP-1 RA discontinuation and increased residual gastric content.
TAKEAWAY:
- The adjusted prevalence of increased residual gastric content was 30.5% (95% CI, 9.9%-51.2%) higher in participants who received GLP-1 RA than those who did not.
- Most patients took their last dose of GLP-1 RA within 5 days before their procedure, but elevated residual gastric content persisted even after 7 days of GLP-1 RA discontinuation.
- There was also no significant association between the type of GLP-1 RA used and the prevalence of increased residual gastric content.
IN PRACTICE:
“We expect healthcare professionals will encounter these classes of drugs with increasing frequency in the perioperative period. Perioperative physicians, including anesthesiologists, surgeons, and primary care physicians, should be well-informed about the safety implications of GLP-1 RA drugs,” the authors wrote.
SOURCE:
The study was led by Sudipta Sen, MD, from the Department of Anesthesiology, Critical Care and Pain Medicine, McGovern Medical School, University of Texas Health Center at Houston, Houston, Texas, and published online in JAMA Surgery.
LIMITATIONS:
Residual gastric content, the primary outcome, served as a proxy for aspiration risk and does not have an exact threshold of volume associated with increased risk. The study did not directly evaluate aspiration events. The authors also acknowledged potential bias from unmeasured confounders owing to the observational nature of this study. A small sample size limited the ability to detect a risk difference for each additional day of drug discontinuation before surgery.
DISCLOSURES:
One of the authors reported receiving a grant from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Patients on weekly glucagon-like peptide-1 receptor agonists (GLP-1 RAs) have high residual gastric content, a major risk factor for aspiration under anesthesia, despite following fasting guidelines before undergoing elective procedures.
METHODOLOGY:
- The increasing use of GLP-1 RAs to manage weight and hyperglycemia has sparked safety concerns because of the drugs’ association with slow gastric emptying, a major risk factor for aspiration under anesthesia.
- This cross-sectional study used gastric ultrasonography to examine the link between GLP-1 RA use and the prevalence of increased residual gastric content.
- All 124 participants (median age, 56 years; 60% women) — half of whom received once-weekly GLP-1 RAs such as semaglutide, dulaglutide, or tirzepatide — adhered to the guideline-recommended fasting duration before undergoing elective procedures under anesthesia.
- The primary outcome focused on identifying increased residual gastric content, defined by the presence of solids, thick liquids, or > 1.5 mL/kg of clear liquids on ultrasound.
- An exploratory analysis examined the association between the duration of GLP-1 RA discontinuation and increased residual gastric content.
TAKEAWAY:
- The adjusted prevalence of increased residual gastric content was 30.5% (95% CI, 9.9%-51.2%) higher in participants who received GLP-1 RA than those who did not.
- Most patients took their last dose of GLP-1 RA within 5 days before their procedure, but elevated residual gastric content persisted even after 7 days of GLP-1 RA discontinuation.
- There was also no significant association between the type of GLP-1 RA used and the prevalence of increased residual gastric content.
IN PRACTICE:
“We expect healthcare professionals will encounter these classes of drugs with increasing frequency in the perioperative period. Perioperative physicians, including anesthesiologists, surgeons, and primary care physicians, should be well-informed about the safety implications of GLP-1 RA drugs,” the authors wrote.
SOURCE:
The study was led by Sudipta Sen, MD, from the Department of Anesthesiology, Critical Care and Pain Medicine, McGovern Medical School, University of Texas Health Center at Houston, Houston, Texas, and published online in JAMA Surgery.
LIMITATIONS:
Residual gastric content, the primary outcome, served as a proxy for aspiration risk and does not have an exact threshold of volume associated with increased risk. The study did not directly evaluate aspiration events. The authors also acknowledged potential bias from unmeasured confounders owing to the observational nature of this study. A small sample size limited the ability to detect a risk difference for each additional day of drug discontinuation before surgery.
DISCLOSURES:
One of the authors reported receiving a grant from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.