FDA clears first mobile rapid test for concussion

Article Type
Changed

The U.S. Food and Drug Administration has cleared SyncThink’s Eye-Sync technology to aid in the diagnosis of mild traumatic brain injury, the company has announced.

Eye-Sync is a virtual reality eye-tracking platform that provides objective measurements to aid in the assessment of concussion. It’s the first mobile, rapid test for concussion that has been cleared by the FDA, the company said.

As reported by this news organization, Eye-Sync received breakthrough designation from the FDA for this indication in March 2019.

The FDA initially cleared the Eye-Sync platform for recording, viewing, and analyzing eye movements to help clinicians identify visual tracking impairment.

The Eye-Sync technology uses a series of 60-second eye tracking assessments, neurocognitive batteries, symptom inventories, and standardized patient inventories to identify the type and severity of impairment after concussion.

“The platform generates customizable and interpretive reports that support clinical decision making and offers visual and vestibular therapies to remedy deficits and monitor improvement over time,” the company said.

In support of the application for use in concussion, SyncThink enrolled 1,655 children and adults into a clinical study that collected comprehensive patient and concussion-related data for over 12 months.

The company used these data to develop proprietary algorithms and deep learning models to identify a positive or negative indication of concussion.

The study showed that Eye-Sinc had sensitivity greater than 82% and specificity greater than 93%, “thereby providing clinicians with significant and actionable data when evaluating individuals with concussion,” the company said in a news release.

“The outcome of this study very clearly shows the effectiveness of our technology at detecting concussion and definitively demonstrates the clinical utility of Eye-Sinc,” SyncThink Chief Clinical Officer Scott Anderson said in the release.

“It also shows that the future of concussion diagnosis is no longer purely symptom-based but that of a technology driven multi-modal approach,” Mr. Anderson said.

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Publications
Topics
Sections

The U.S. Food and Drug Administration has cleared SyncThink’s Eye-Sync technology to aid in the diagnosis of mild traumatic brain injury, the company has announced.

Eye-Sync is a virtual reality eye-tracking platform that provides objective measurements to aid in the assessment of concussion. It’s the first mobile, rapid test for concussion that has been cleared by the FDA, the company said.

As reported by this news organization, Eye-Sync received breakthrough designation from the FDA for this indication in March 2019.

The FDA initially cleared the Eye-Sync platform for recording, viewing, and analyzing eye movements to help clinicians identify visual tracking impairment.

The Eye-Sync technology uses a series of 60-second eye tracking assessments, neurocognitive batteries, symptom inventories, and standardized patient inventories to identify the type and severity of impairment after concussion.

“The platform generates customizable and interpretive reports that support clinical decision making and offers visual and vestibular therapies to remedy deficits and monitor improvement over time,” the company said.

In support of the application for use in concussion, SyncThink enrolled 1,655 children and adults into a clinical study that collected comprehensive patient and concussion-related data for over 12 months.

The company used these data to develop proprietary algorithms and deep learning models to identify a positive or negative indication of concussion.

The study showed that Eye-Sinc had sensitivity greater than 82% and specificity greater than 93%, “thereby providing clinicians with significant and actionable data when evaluating individuals with concussion,” the company said in a news release.

“The outcome of this study very clearly shows the effectiveness of our technology at detecting concussion and definitively demonstrates the clinical utility of Eye-Sinc,” SyncThink Chief Clinical Officer Scott Anderson said in the release.

“It also shows that the future of concussion diagnosis is no longer purely symptom-based but that of a technology driven multi-modal approach,” Mr. Anderson said.

A version of this article first appeared on Medscape.com.

The U.S. Food and Drug Administration has cleared SyncThink’s Eye-Sync technology to aid in the diagnosis of mild traumatic brain injury, the company has announced.

Eye-Sync is a virtual reality eye-tracking platform that provides objective measurements to aid in the assessment of concussion. It’s the first mobile, rapid test for concussion that has been cleared by the FDA, the company said.

As reported by this news organization, Eye-Sync received breakthrough designation from the FDA for this indication in March 2019.

The FDA initially cleared the Eye-Sync platform for recording, viewing, and analyzing eye movements to help clinicians identify visual tracking impairment.

The Eye-Sync technology uses a series of 60-second eye tracking assessments, neurocognitive batteries, symptom inventories, and standardized patient inventories to identify the type and severity of impairment after concussion.

“The platform generates customizable and interpretive reports that support clinical decision making and offers visual and vestibular therapies to remedy deficits and monitor improvement over time,” the company said.

In support of the application for use in concussion, SyncThink enrolled 1,655 children and adults into a clinical study that collected comprehensive patient and concussion-related data for over 12 months.

The company used these data to develop proprietary algorithms and deep learning models to identify a positive or negative indication of concussion.

The study showed that Eye-Sinc had sensitivity greater than 82% and specificity greater than 93%, “thereby providing clinicians with significant and actionable data when evaluating individuals with concussion,” the company said in a news release.

“The outcome of this study very clearly shows the effectiveness of our technology at detecting concussion and definitively demonstrates the clinical utility of Eye-Sinc,” SyncThink Chief Clinical Officer Scott Anderson said in the release.

“It also shows that the future of concussion diagnosis is no longer purely symptom-based but that of a technology driven multi-modal approach,” Mr. Anderson said.

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Issue
Neurology Reviews - 29(11)
Publications
Publications
Topics
Article Type
Sections
Citation Override
Publish date: October 6, 2021
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Is AFib a stroke cause or innocent bystander? The debate continues

Article Type
Changed

Discovery of substantial atrial fibrillation (AFib) is usually an indication to start oral anticoagulation (OAC) for stroke prevention, but it’s far from settled whether such AFib is actually a direct cause of thromboembolic stroke. And that has implications for whether patients with occasional bouts of the arrhythmia need to be on continuous OAC.  

It’s possible that some with infrequent paroxysmal AFib can get away with OAC maintained only about as long as the arrhythmia persists, and then go off the drugs, say researchers based on their study, which, they caution, would need the support of prospective trials before such a strategy could be considered.

But importantly, in their patients who had been continuously monitored by their cardiac implantable electronic devices (CIEDs) prior to experiencing a stroke, the 30-day risk of that stroke more than tripled if their AFib burden on 1 day reached at least 5-6 hours. The risk jumped especially high within the first few days after accumulating that amount of AFib in a day, but then fell off sharply over the next few days.

Based on the study, “Your risk of stroke goes up acutely when you have an episode of AFib, and it decreases rapidly, back to baseline – certainly by 30 days and it looked like in our data by 5 days,” Daniel E. Singer, MD, of Massachusetts General Hospital, Boston, said in an interview.

Increasingly, he noted, “there’s a widespread belief that AFib is a risk marker, not a causal risk factor.” In that scenario, most embolic strokes are caused by thrombi formed as a result of an atrial myopathy, characterized by fibrosis and inflammation, that also happens to trigger AFib.

But the current findings are, “from a mechanistic point of view, very much in favor of AFib being a causal risk factor, acutely raising the risk of stroke,” said Dr. Singer, who is lead author on the analysis published online Sept. 29 in JAMA Cardiology.

Some studies have “shown that anticoagulants seem to lower stroke risk even in patients without atrial fib, and even from sources not likely to be coming from the atrium,” Mintu P. Turakhia, MD, of Stanford (Calif.) University, Palo Alto, said in an interview. Collectively they point to “atrial fibrillation as a cause of and a noncausal marker for stroke.”

For example, Dr. Turakhia pointed out in an editorial accompanying the current report that stroke in patients with CIEDs “may occur during prolonged periods of sinus rhythm.”

The current study, he said in an interview, doesn’t preclude atrial myopathy as one direct cause of stroke-associated thrombus, because probably both the myopathy and AFib can be culprits. Still, AFib itself it may bear more responsibility for strokes in patients with fewer competing risks for stroke.

In such patients at lower vascular risk, who may have a CHA2DS2-VASc score of only 1 or 2, for example, “AFib can become a more important cause” of ischemic stroke, Dr. Turakhia said. That’s when AFib is more likely to be temporally related to stroke as the likely culprit, the mechanism addressed by Dr. Singer and associates.

“I think we’re all trying to grapple with what the truth is,” Dr. Singer observed. Still, the current study was unusual for primarily looking at the temporal relationship between AFib and stroke, rather than stroke risk. “And once again, as we found in our earlier study, but now a much larger study, it’s a tight relationship.”

Based on the current results, he said, the risk is “high when you have AFib, and it decreases very rapidly after the AFib is over.” And, “it takes multiple hours of AFib to raise stroke risk.” Inclusion in the analysis required accumulation of at least 5.5 hours of AFib on at least 1 day in a month, the cut point at which stroke risk started to climb significantly in an earlier trial.  

In the current analysis, however, the 30-day odds ratio for stroke was a nonsignificant 2.75 for an AFib burden of 6-23 hours in a day and jumped to a significant 5.0 for a burden in excess of 23 hours in a day. “That’s a lot of AFib” before the risk actually goes up, and supports AFib as causative, Dr. Singer said. If it were the myopathy itself triggering stroke in these particular patients, the risk would be ongoing and not subject to a threshold of AFib burden.
 

 

 

Implications for noncontinuous OAC

“The hope is that there are people who have very little AFib: They may have several hours, and then they have nothing for 6 months. Do they have to be anticoagulated or not?” Dr. Singer asked.

“If you believe the risk-marker story, you might say they have to be anticoagulated. But if you believe our results, you would certainly think there’s a good chance they don’t have to be anticoagulated,” he said.

“So it is logical to think, if you have the right people and continuous monitoring, that you could have time-delimited anticoagulation.” That is, patients might start right away on a direct OAC once reaching the AFib threshold in a day, Dr. Singer said, “going on and off anticoagulants in parallel with their episodes of AFib.”

The strategy wouldn’t be feasible in patients who often experience AFib, Dr. Singer noted, “but it might work for people who have infrequent paroxysmal AFib.” It certainly would first have to be tested in prospective trials, he said. Such trials would be more practical than ever to carry out given the growing availability of continuous AFib monitoring by wearables.

“We need a trial to make the case whether it’s safe or not,” Dr. Turakhia said of such a rhythm-guided approach to OAC for AFib. The population to start with, he said, would be patients with paroxysmal AFib and low CHA2DS2-VASc scores. “If you think CHA2DS2-VASc as an integrated score of vascular risk, such patients would have a lot fewer reasons to have strokes. And if they do have a stroke, it’s more reasonable to assume that it’s likely caused by atrial fib and not just a marker.”

Importantly, such a strategy could well be safer than continuous OAC for some patients – those at the lowest vascular risk and with the most occasional AFib and lowest AFib burden “who are otherwise doing fine,” Dr. Turakhia said. In such patients on continuous OAC, he proposed, the risks of bleeding and intracranial hemorrhage could potentially exceed the expected degree of protection from ischemic events.
 

Discordant periods of AFib burden

Dr. Singer and his colleagues linked a national electronic health record database with Medtronic CareLink records covering 10 years to identify 891 patients who experienced an ischemic stroke preceded by at least 120 days of continuous heart-rhythm monitoring.

The patients were then categorized by their pattern of AFib, if any, within each of two prestroke periods: the most recent 30 days, which was the test period, and the preceding 91-120 days, the control period.

The analysis then excluded any patients who reached an AFib-burden threshold of at least 5.5 hours on any day during both the test and control periods, and those who did not attain that threshold in either period.

“The ones who had AFib in both periods mostly had permanent AFib, and ones that didn’t have AFib in either period mostly were in sinus rhythm,” Dr. Singer said. It was “close to 100%” in both cases.

Those exclusions left 66 patients, 7.4% of the total, who reached the AFib-burden threshold on at least 1 day during either the test or control periods, but not both. They included 52 and 14 patients, respectively, with “discordant” periods, that is, at least that burden of AFib in a day during either the test or control period, but not both.

Comparing AFib burden at test versus control periods among patients for whom the two periods were discordant yielded an OR for stroke of 3.71 (95% confidence interval, 2.06-6.70).

Stroke risk levels were not evenly spread throughout the 24-hour periods that met the AFib-burden threshold or the 30 days preceding the patients’ strokes. The OR for stroke was 5.00 (95% CI, 2.62-9.55) during days 1-5 following the day in which the AFib-burden threshold was met. And it was 5.00 (95% CI, 2.08-12.01) over 30 days if the AFib burden exceeded 23 hours on any day of the test period.

The study’s case-crossover design, in which each patient served as their own control, is one of its advantages, Dr. Singer observed. Most patient features, including CHA2DS2-VASc score and comorbidities, did not change appreciably from earliest to the latest 30-day period, which strengthens the comparison of the two because “you don’t have to worry about long-term confounding.”

Dr. Singer was supported by the Eliot B. and Edith C. Shoolman fund of the Massachusetts General Hospital. He discloses receiving grants from Boehringer Ingelheim and Bristol-Myers Squibb; personal fees from Boehringer Ingelheim, Bristol-Myers Squibb, Fitbit, Johnson & Johnson, Merck, and Pfizer; and royalties from UpToDate.

Dr. Turakhia discloses personal fees from Medtronic, Abbott, Sanofi, Pfizer, Myokardia, Johnson & Johnson, Milestone Pharmaceuticals, InCarda Therapeutics, 100Plus, Forward Pharma, and AliveCor; and grants from Bristol-Myers Squibb, the American Heart Association, Apple, and Bayer.

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Publications
Topics
Sections

Discovery of substantial atrial fibrillation (AFib) is usually an indication to start oral anticoagulation (OAC) for stroke prevention, but it’s far from settled whether such AFib is actually a direct cause of thromboembolic stroke. And that has implications for whether patients with occasional bouts of the arrhythmia need to be on continuous OAC.  

It’s possible that some with infrequent paroxysmal AFib can get away with OAC maintained only about as long as the arrhythmia persists, and then go off the drugs, say researchers based on their study, which, they caution, would need the support of prospective trials before such a strategy could be considered.

But importantly, in their patients who had been continuously monitored by their cardiac implantable electronic devices (CIEDs) prior to experiencing a stroke, the 30-day risk of that stroke more than tripled if their AFib burden on 1 day reached at least 5-6 hours. The risk jumped especially high within the first few days after accumulating that amount of AFib in a day, but then fell off sharply over the next few days.

Based on the study, “Your risk of stroke goes up acutely when you have an episode of AFib, and it decreases rapidly, back to baseline – certainly by 30 days and it looked like in our data by 5 days,” Daniel E. Singer, MD, of Massachusetts General Hospital, Boston, said in an interview.

Increasingly, he noted, “there’s a widespread belief that AFib is a risk marker, not a causal risk factor.” In that scenario, most embolic strokes are caused by thrombi formed as a result of an atrial myopathy, characterized by fibrosis and inflammation, that also happens to trigger AFib.

But the current findings are, “from a mechanistic point of view, very much in favor of AFib being a causal risk factor, acutely raising the risk of stroke,” said Dr. Singer, who is lead author on the analysis published online Sept. 29 in JAMA Cardiology.

Some studies have “shown that anticoagulants seem to lower stroke risk even in patients without atrial fib, and even from sources not likely to be coming from the atrium,” Mintu P. Turakhia, MD, of Stanford (Calif.) University, Palo Alto, said in an interview. Collectively they point to “atrial fibrillation as a cause of and a noncausal marker for stroke.”

For example, Dr. Turakhia pointed out in an editorial accompanying the current report that stroke in patients with CIEDs “may occur during prolonged periods of sinus rhythm.”

The current study, he said in an interview, doesn’t preclude atrial myopathy as one direct cause of stroke-associated thrombus, because probably both the myopathy and AFib can be culprits. Still, AFib itself it may bear more responsibility for strokes in patients with fewer competing risks for stroke.

In such patients at lower vascular risk, who may have a CHA2DS2-VASc score of only 1 or 2, for example, “AFib can become a more important cause” of ischemic stroke, Dr. Turakhia said. That’s when AFib is more likely to be temporally related to stroke as the likely culprit, the mechanism addressed by Dr. Singer and associates.

“I think we’re all trying to grapple with what the truth is,” Dr. Singer observed. Still, the current study was unusual for primarily looking at the temporal relationship between AFib and stroke, rather than stroke risk. “And once again, as we found in our earlier study, but now a much larger study, it’s a tight relationship.”

Based on the current results, he said, the risk is “high when you have AFib, and it decreases very rapidly after the AFib is over.” And, “it takes multiple hours of AFib to raise stroke risk.” Inclusion in the analysis required accumulation of at least 5.5 hours of AFib on at least 1 day in a month, the cut point at which stroke risk started to climb significantly in an earlier trial.  

In the current analysis, however, the 30-day odds ratio for stroke was a nonsignificant 2.75 for an AFib burden of 6-23 hours in a day and jumped to a significant 5.0 for a burden in excess of 23 hours in a day. “That’s a lot of AFib” before the risk actually goes up, and supports AFib as causative, Dr. Singer said. If it were the myopathy itself triggering stroke in these particular patients, the risk would be ongoing and not subject to a threshold of AFib burden.
 

 

 

Implications for noncontinuous OAC

“The hope is that there are people who have very little AFib: They may have several hours, and then they have nothing for 6 months. Do they have to be anticoagulated or not?” Dr. Singer asked.

“If you believe the risk-marker story, you might say they have to be anticoagulated. But if you believe our results, you would certainly think there’s a good chance they don’t have to be anticoagulated,” he said.

“So it is logical to think, if you have the right people and continuous monitoring, that you could have time-delimited anticoagulation.” That is, patients might start right away on a direct OAC once reaching the AFib threshold in a day, Dr. Singer said, “going on and off anticoagulants in parallel with their episodes of AFib.”

The strategy wouldn’t be feasible in patients who often experience AFib, Dr. Singer noted, “but it might work for people who have infrequent paroxysmal AFib.” It certainly would first have to be tested in prospective trials, he said. Such trials would be more practical than ever to carry out given the growing availability of continuous AFib monitoring by wearables.

“We need a trial to make the case whether it’s safe or not,” Dr. Turakhia said of such a rhythm-guided approach to OAC for AFib. The population to start with, he said, would be patients with paroxysmal AFib and low CHA2DS2-VASc scores. “If you think CHA2DS2-VASc as an integrated score of vascular risk, such patients would have a lot fewer reasons to have strokes. And if they do have a stroke, it’s more reasonable to assume that it’s likely caused by atrial fib and not just a marker.”

Importantly, such a strategy could well be safer than continuous OAC for some patients – those at the lowest vascular risk and with the most occasional AFib and lowest AFib burden “who are otherwise doing fine,” Dr. Turakhia said. In such patients on continuous OAC, he proposed, the risks of bleeding and intracranial hemorrhage could potentially exceed the expected degree of protection from ischemic events.
 

Discordant periods of AFib burden

Dr. Singer and his colleagues linked a national electronic health record database with Medtronic CareLink records covering 10 years to identify 891 patients who experienced an ischemic stroke preceded by at least 120 days of continuous heart-rhythm monitoring.

The patients were then categorized by their pattern of AFib, if any, within each of two prestroke periods: the most recent 30 days, which was the test period, and the preceding 91-120 days, the control period.

The analysis then excluded any patients who reached an AFib-burden threshold of at least 5.5 hours on any day during both the test and control periods, and those who did not attain that threshold in either period.

“The ones who had AFib in both periods mostly had permanent AFib, and ones that didn’t have AFib in either period mostly were in sinus rhythm,” Dr. Singer said. It was “close to 100%” in both cases.

Those exclusions left 66 patients, 7.4% of the total, who reached the AFib-burden threshold on at least 1 day during either the test or control periods, but not both. They included 52 and 14 patients, respectively, with “discordant” periods, that is, at least that burden of AFib in a day during either the test or control period, but not both.

Comparing AFib burden at test versus control periods among patients for whom the two periods were discordant yielded an OR for stroke of 3.71 (95% confidence interval, 2.06-6.70).

Stroke risk levels were not evenly spread throughout the 24-hour periods that met the AFib-burden threshold or the 30 days preceding the patients’ strokes. The OR for stroke was 5.00 (95% CI, 2.62-9.55) during days 1-5 following the day in which the AFib-burden threshold was met. And it was 5.00 (95% CI, 2.08-12.01) over 30 days if the AFib burden exceeded 23 hours on any day of the test period.

The study’s case-crossover design, in which each patient served as their own control, is one of its advantages, Dr. Singer observed. Most patient features, including CHA2DS2-VASc score and comorbidities, did not change appreciably from earliest to the latest 30-day period, which strengthens the comparison of the two because “you don’t have to worry about long-term confounding.”

Dr. Singer was supported by the Eliot B. and Edith C. Shoolman fund of the Massachusetts General Hospital. He discloses receiving grants from Boehringer Ingelheim and Bristol-Myers Squibb; personal fees from Boehringer Ingelheim, Bristol-Myers Squibb, Fitbit, Johnson & Johnson, Merck, and Pfizer; and royalties from UpToDate.

Dr. Turakhia discloses personal fees from Medtronic, Abbott, Sanofi, Pfizer, Myokardia, Johnson & Johnson, Milestone Pharmaceuticals, InCarda Therapeutics, 100Plus, Forward Pharma, and AliveCor; and grants from Bristol-Myers Squibb, the American Heart Association, Apple, and Bayer.

A version of this article first appeared on Medscape.com.

Discovery of substantial atrial fibrillation (AFib) is usually an indication to start oral anticoagulation (OAC) for stroke prevention, but it’s far from settled whether such AFib is actually a direct cause of thromboembolic stroke. And that has implications for whether patients with occasional bouts of the arrhythmia need to be on continuous OAC.  

It’s possible that some with infrequent paroxysmal AFib can get away with OAC maintained only about as long as the arrhythmia persists, and then go off the drugs, say researchers based on their study, which, they caution, would need the support of prospective trials before such a strategy could be considered.

But importantly, in their patients who had been continuously monitored by their cardiac implantable electronic devices (CIEDs) prior to experiencing a stroke, the 30-day risk of that stroke more than tripled if their AFib burden on 1 day reached at least 5-6 hours. The risk jumped especially high within the first few days after accumulating that amount of AFib in a day, but then fell off sharply over the next few days.

Based on the study, “Your risk of stroke goes up acutely when you have an episode of AFib, and it decreases rapidly, back to baseline – certainly by 30 days and it looked like in our data by 5 days,” Daniel E. Singer, MD, of Massachusetts General Hospital, Boston, said in an interview.

Increasingly, he noted, “there’s a widespread belief that AFib is a risk marker, not a causal risk factor.” In that scenario, most embolic strokes are caused by thrombi formed as a result of an atrial myopathy, characterized by fibrosis and inflammation, that also happens to trigger AFib.

But the current findings are, “from a mechanistic point of view, very much in favor of AFib being a causal risk factor, acutely raising the risk of stroke,” said Dr. Singer, who is lead author on the analysis published online Sept. 29 in JAMA Cardiology.

Some studies have “shown that anticoagulants seem to lower stroke risk even in patients without atrial fib, and even from sources not likely to be coming from the atrium,” Mintu P. Turakhia, MD, of Stanford (Calif.) University, Palo Alto, said in an interview. Collectively they point to “atrial fibrillation as a cause of and a noncausal marker for stroke.”

For example, Dr. Turakhia pointed out in an editorial accompanying the current report that stroke in patients with CIEDs “may occur during prolonged periods of sinus rhythm.”

The current study, he said in an interview, doesn’t preclude atrial myopathy as one direct cause of stroke-associated thrombus, because probably both the myopathy and AFib can be culprits. Still, AFib itself it may bear more responsibility for strokes in patients with fewer competing risks for stroke.

In such patients at lower vascular risk, who may have a CHA2DS2-VASc score of only 1 or 2, for example, “AFib can become a more important cause” of ischemic stroke, Dr. Turakhia said. That’s when AFib is more likely to be temporally related to stroke as the likely culprit, the mechanism addressed by Dr. Singer and associates.

“I think we’re all trying to grapple with what the truth is,” Dr. Singer observed. Still, the current study was unusual for primarily looking at the temporal relationship between AFib and stroke, rather than stroke risk. “And once again, as we found in our earlier study, but now a much larger study, it’s a tight relationship.”

Based on the current results, he said, the risk is “high when you have AFib, and it decreases very rapidly after the AFib is over.” And, “it takes multiple hours of AFib to raise stroke risk.” Inclusion in the analysis required accumulation of at least 5.5 hours of AFib on at least 1 day in a month, the cut point at which stroke risk started to climb significantly in an earlier trial.  

In the current analysis, however, the 30-day odds ratio for stroke was a nonsignificant 2.75 for an AFib burden of 6-23 hours in a day and jumped to a significant 5.0 for a burden in excess of 23 hours in a day. “That’s a lot of AFib” before the risk actually goes up, and supports AFib as causative, Dr. Singer said. If it were the myopathy itself triggering stroke in these particular patients, the risk would be ongoing and not subject to a threshold of AFib burden.
 

 

 

Implications for noncontinuous OAC

“The hope is that there are people who have very little AFib: They may have several hours, and then they have nothing for 6 months. Do they have to be anticoagulated or not?” Dr. Singer asked.

“If you believe the risk-marker story, you might say they have to be anticoagulated. But if you believe our results, you would certainly think there’s a good chance they don’t have to be anticoagulated,” he said.

“So it is logical to think, if you have the right people and continuous monitoring, that you could have time-delimited anticoagulation.” That is, patients might start right away on a direct OAC once reaching the AFib threshold in a day, Dr. Singer said, “going on and off anticoagulants in parallel with their episodes of AFib.”

The strategy wouldn’t be feasible in patients who often experience AFib, Dr. Singer noted, “but it might work for people who have infrequent paroxysmal AFib.” It certainly would first have to be tested in prospective trials, he said. Such trials would be more practical than ever to carry out given the growing availability of continuous AFib monitoring by wearables.

“We need a trial to make the case whether it’s safe or not,” Dr. Turakhia said of such a rhythm-guided approach to OAC for AFib. The population to start with, he said, would be patients with paroxysmal AFib and low CHA2DS2-VASc scores. “If you think CHA2DS2-VASc as an integrated score of vascular risk, such patients would have a lot fewer reasons to have strokes. And if they do have a stroke, it’s more reasonable to assume that it’s likely caused by atrial fib and not just a marker.”

Importantly, such a strategy could well be safer than continuous OAC for some patients – those at the lowest vascular risk and with the most occasional AFib and lowest AFib burden “who are otherwise doing fine,” Dr. Turakhia said. In such patients on continuous OAC, he proposed, the risks of bleeding and intracranial hemorrhage could potentially exceed the expected degree of protection from ischemic events.
 

Discordant periods of AFib burden

Dr. Singer and his colleagues linked a national electronic health record database with Medtronic CareLink records covering 10 years to identify 891 patients who experienced an ischemic stroke preceded by at least 120 days of continuous heart-rhythm monitoring.

The patients were then categorized by their pattern of AFib, if any, within each of two prestroke periods: the most recent 30 days, which was the test period, and the preceding 91-120 days, the control period.

The analysis then excluded any patients who reached an AFib-burden threshold of at least 5.5 hours on any day during both the test and control periods, and those who did not attain that threshold in either period.

“The ones who had AFib in both periods mostly had permanent AFib, and ones that didn’t have AFib in either period mostly were in sinus rhythm,” Dr. Singer said. It was “close to 100%” in both cases.

Those exclusions left 66 patients, 7.4% of the total, who reached the AFib-burden threshold on at least 1 day during either the test or control periods, but not both. They included 52 and 14 patients, respectively, with “discordant” periods, that is, at least that burden of AFib in a day during either the test or control period, but not both.

Comparing AFib burden at test versus control periods among patients for whom the two periods were discordant yielded an OR for stroke of 3.71 (95% confidence interval, 2.06-6.70).

Stroke risk levels were not evenly spread throughout the 24-hour periods that met the AFib-burden threshold or the 30 days preceding the patients’ strokes. The OR for stroke was 5.00 (95% CI, 2.62-9.55) during days 1-5 following the day in which the AFib-burden threshold was met. And it was 5.00 (95% CI, 2.08-12.01) over 30 days if the AFib burden exceeded 23 hours on any day of the test period.

The study’s case-crossover design, in which each patient served as their own control, is one of its advantages, Dr. Singer observed. Most patient features, including CHA2DS2-VASc score and comorbidities, did not change appreciably from earliest to the latest 30-day period, which strengthens the comparison of the two because “you don’t have to worry about long-term confounding.”

Dr. Singer was supported by the Eliot B. and Edith C. Shoolman fund of the Massachusetts General Hospital. He discloses receiving grants from Boehringer Ingelheim and Bristol-Myers Squibb; personal fees from Boehringer Ingelheim, Bristol-Myers Squibb, Fitbit, Johnson & Johnson, Merck, and Pfizer; and royalties from UpToDate.

Dr. Turakhia discloses personal fees from Medtronic, Abbott, Sanofi, Pfizer, Myokardia, Johnson & Johnson, Milestone Pharmaceuticals, InCarda Therapeutics, 100Plus, Forward Pharma, and AliveCor; and grants from Bristol-Myers Squibb, the American Heart Association, Apple, and Bayer.

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Issue
Neurology Reviews - 29(11)
Publications
Publications
Topics
Article Type
Sections
Citation Override
Publish date: October 5, 2021
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

ADHD med may reduce apathy in Alzheimer’s disease

Article Type
Changed

Methylphenidate is safe and effective for treating apathy in patients with Alzheimer’s disease (AD), new research suggests.

Dundanim/shutterstock.com

Results from a phase 3 randomized trial showed that, after 6 months of treatment, mean score on the Neuropsychiatric Inventory (NPI) apathy subscale decreased by 4.5 points for patients who received methylphenidate vs. a decrease of 3.1 points for those who received placebo.

In addition, the safety profile showed no significant between-group differences.

“Methylphenidate offers a treatment approach providing a modest but potentially clinically significant benefit for patients and caregivers,” said the investigators, led by Jacobo E. Mintzer, MD, MBA, professor of health studies at the Medical University of South Carolina in Charleston.

The findings were published online Sept. 27 in JAMA Neurology.
 

Common problem

Apathy, which is common among patients with AD, is associated with increased risk for mortality, financial burden, and caregiver burden. No treatment has proved effective for apathy in this population.

Two trials of methylphenidate, a catecholaminergic agent, have provided preliminary evidence of efficacy. Findings from the Apathy in Dementia Methylphenidate trial (ADMET) suggested the drug was associated with improved cognition and few adverse events. However, both trials had small patient populations and short durations.

The current investigators conducted ADMET 2, a 6-month, phase 3 trial, to investigate methylphenidate further. They recruited 200 patients (mean age, 76 years; 66% men; 90% White) at nine clinical centers that specialized in dementia care in the United States and one in Canada.

Eligible patients had a diagnosis of possible or probable AD and a Mini-Mental State Examination (MMSE) score between 10 and 28. They also had clinically significant apathy for at least 4 weeks and an available caregiver who spent more than 10 hours a week with the patient.

The researchers randomly assigned patients to receive methylphenidate (n = 99) or placebo (n = 101). For 3 days, participants in the active group received 10 mg/day of methylphenidate. After that point, they received 20 mg/day of methylphenidate for the rest of the study.

Patients in both treatment groups were given the same number of identical-appearing capsules each day.

In-person follow-up visits took place monthly for 6 months. Participants also were contacted by telephone at days 15, 45, and 75 after treatment assignment.

Participants underwent cognitive testing at baseline and at 2, 4, and 6 months. The battery of tests included the MMSE, Hopkins Verbal Learning Test, and Wechsler Adult Intelligence Scale – Revised Digit Span.

The trial’s two primary outcomes were mean change in NPI apathy score from baseline to 6 months and the odds of an improved rating on the Alzheimer’s Disease Cooperative Study Clinical Global Impression of Change (ADCS-CGIC) between baseline and 6 months.

Significant change on either outcome was to be considered a signal of effective treatment.
 

Treatment-specific benefit

Ten patients in the methylphenidate group and seven in the placebo group withdrew during the study.

Mean MMSE score at baseline was 19.2 in the methylphenidate group vs. 18.5 in the placebo group, indicating moderately severe dementia. Mean baseline score on the NPI apathy subscale was 8.0 vs. 7.6, respectively.

In an adjusted, longitudinal model, mean between-group difference in change in NPI apathy score at 6 months was –1.25 (P = .002). The mean NPI apathy score decreased by 4.5 points in the methylphenidate group vs. 3.1 points in the placebo group.

The largest change in apathy score occurred during the first 2 months of treatment. At 6 months, 27% of the methylphenidate group vs. 14% of the placebo group had an NPI apathy score of 0.

In addition, 43.8% of the methylphenidate group had improvement on the ADCS-CGIC compared with 35.2% of the placebo group. The odds ratio (OR) for improvement on ADCS-CGIC for methylphenidate vs. placebo was 1.90 (P = .07).

There was also a strong association between score improvement on the NPI apathy subscale and improvement on the ADCS-CGIC subscale (OR, 2.95; P = .002).

“It is important to note that there were no group differences in any of the cognitive measures, suggesting that the effect of the treatment is specific to the treatment of apathy and not a secondary effect of improvement in cognition,” the researchers wrote.

In all, 17 serious adverse events occurred in the methylphenidate group and 10 occurred in the placebo group. However, all events were found to be hospitalizations for events not related to treatment.
 

 

 

‘Enduring effect’

Commenting on the findings, Jeffrey L. Cummings, MD, ScD, professor of brain sciences at the University of Nevada, Las Vegas, noted that the reduction in NPI apathy subscale score of more than 50% was clinically meaningful.

Dr. Jeffrey L. Cummings

A more robust outcome on the ADCS-CGIC would have been desirable, he added, although that instrument is not designed specifically for apathy.

Methylphenidate’s effect on apathy observed at 2 months and remaining stable throughout the study makes it appear to be “an enduring effect, and not something that the patient accommodates to,” said Dr. Cummings, who was not involved with the research. Such a change may manifest itself in a patient’s greater willingness to help voluntarily with housework or to suggest going for a walk, he noted.

“These are not dramatic changes in cognition, of course, but they are changes in initiative and that is very important,” Dr. Cummings said. Decreased apathy also may improve quality of life for the patient’s caregiver, he added.

Overall, the findings raise the question of whether the Food and Drug Administration should recognize apathy as an indication for which drugs can be approved, said Dr. Cummings.

“For me, that would be the next major step in this line of investigation,” he concluded.

The study was funded by the National Institute on Aging. Dr. Mintzer has served as an adviser to Praxis Bioresearch and Cerevel Therapeutics on matters unrelated to this study. Dr. Cummings is the author of the Neuropsychiatric Inventory but does not receive payments for it from academic trials such as ADMET 2.
 

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Publications
Topics
Sections

Methylphenidate is safe and effective for treating apathy in patients with Alzheimer’s disease (AD), new research suggests.

Dundanim/shutterstock.com

Results from a phase 3 randomized trial showed that, after 6 months of treatment, mean score on the Neuropsychiatric Inventory (NPI) apathy subscale decreased by 4.5 points for patients who received methylphenidate vs. a decrease of 3.1 points for those who received placebo.

In addition, the safety profile showed no significant between-group differences.

“Methylphenidate offers a treatment approach providing a modest but potentially clinically significant benefit for patients and caregivers,” said the investigators, led by Jacobo E. Mintzer, MD, MBA, professor of health studies at the Medical University of South Carolina in Charleston.

The findings were published online Sept. 27 in JAMA Neurology.
 

Common problem

Apathy, which is common among patients with AD, is associated with increased risk for mortality, financial burden, and caregiver burden. No treatment has proved effective for apathy in this population.

Two trials of methylphenidate, a catecholaminergic agent, have provided preliminary evidence of efficacy. Findings from the Apathy in Dementia Methylphenidate trial (ADMET) suggested the drug was associated with improved cognition and few adverse events. However, both trials had small patient populations and short durations.

The current investigators conducted ADMET 2, a 6-month, phase 3 trial, to investigate methylphenidate further. They recruited 200 patients (mean age, 76 years; 66% men; 90% White) at nine clinical centers that specialized in dementia care in the United States and one in Canada.

Eligible patients had a diagnosis of possible or probable AD and a Mini-Mental State Examination (MMSE) score between 10 and 28. They also had clinically significant apathy for at least 4 weeks and an available caregiver who spent more than 10 hours a week with the patient.

The researchers randomly assigned patients to receive methylphenidate (n = 99) or placebo (n = 101). For 3 days, participants in the active group received 10 mg/day of methylphenidate. After that point, they received 20 mg/day of methylphenidate for the rest of the study.

Patients in both treatment groups were given the same number of identical-appearing capsules each day.

In-person follow-up visits took place monthly for 6 months. Participants also were contacted by telephone at days 15, 45, and 75 after treatment assignment.

Participants underwent cognitive testing at baseline and at 2, 4, and 6 months. The battery of tests included the MMSE, Hopkins Verbal Learning Test, and Wechsler Adult Intelligence Scale – Revised Digit Span.

The trial’s two primary outcomes were mean change in NPI apathy score from baseline to 6 months and the odds of an improved rating on the Alzheimer’s Disease Cooperative Study Clinical Global Impression of Change (ADCS-CGIC) between baseline and 6 months.

Significant change on either outcome was to be considered a signal of effective treatment.
 

Treatment-specific benefit

Ten patients in the methylphenidate group and seven in the placebo group withdrew during the study.

Mean MMSE score at baseline was 19.2 in the methylphenidate group vs. 18.5 in the placebo group, indicating moderately severe dementia. Mean baseline score on the NPI apathy subscale was 8.0 vs. 7.6, respectively.

In an adjusted, longitudinal model, mean between-group difference in change in NPI apathy score at 6 months was –1.25 (P = .002). The mean NPI apathy score decreased by 4.5 points in the methylphenidate group vs. 3.1 points in the placebo group.

The largest change in apathy score occurred during the first 2 months of treatment. At 6 months, 27% of the methylphenidate group vs. 14% of the placebo group had an NPI apathy score of 0.

In addition, 43.8% of the methylphenidate group had improvement on the ADCS-CGIC compared with 35.2% of the placebo group. The odds ratio (OR) for improvement on ADCS-CGIC for methylphenidate vs. placebo was 1.90 (P = .07).

There was also a strong association between score improvement on the NPI apathy subscale and improvement on the ADCS-CGIC subscale (OR, 2.95; P = .002).

“It is important to note that there were no group differences in any of the cognitive measures, suggesting that the effect of the treatment is specific to the treatment of apathy and not a secondary effect of improvement in cognition,” the researchers wrote.

In all, 17 serious adverse events occurred in the methylphenidate group and 10 occurred in the placebo group. However, all events were found to be hospitalizations for events not related to treatment.
 

 

 

‘Enduring effect’

Commenting on the findings, Jeffrey L. Cummings, MD, ScD, professor of brain sciences at the University of Nevada, Las Vegas, noted that the reduction in NPI apathy subscale score of more than 50% was clinically meaningful.

Dr. Jeffrey L. Cummings

A more robust outcome on the ADCS-CGIC would have been desirable, he added, although that instrument is not designed specifically for apathy.

Methylphenidate’s effect on apathy observed at 2 months and remaining stable throughout the study makes it appear to be “an enduring effect, and not something that the patient accommodates to,” said Dr. Cummings, who was not involved with the research. Such a change may manifest itself in a patient’s greater willingness to help voluntarily with housework or to suggest going for a walk, he noted.

“These are not dramatic changes in cognition, of course, but they are changes in initiative and that is very important,” Dr. Cummings said. Decreased apathy also may improve quality of life for the patient’s caregiver, he added.

Overall, the findings raise the question of whether the Food and Drug Administration should recognize apathy as an indication for which drugs can be approved, said Dr. Cummings.

“For me, that would be the next major step in this line of investigation,” he concluded.

The study was funded by the National Institute on Aging. Dr. Mintzer has served as an adviser to Praxis Bioresearch and Cerevel Therapeutics on matters unrelated to this study. Dr. Cummings is the author of the Neuropsychiatric Inventory but does not receive payments for it from academic trials such as ADMET 2.
 

A version of this article first appeared on Medscape.com.

Methylphenidate is safe and effective for treating apathy in patients with Alzheimer’s disease (AD), new research suggests.

Dundanim/shutterstock.com

Results from a phase 3 randomized trial showed that, after 6 months of treatment, mean score on the Neuropsychiatric Inventory (NPI) apathy subscale decreased by 4.5 points for patients who received methylphenidate vs. a decrease of 3.1 points for those who received placebo.

In addition, the safety profile showed no significant between-group differences.

“Methylphenidate offers a treatment approach providing a modest but potentially clinically significant benefit for patients and caregivers,” said the investigators, led by Jacobo E. Mintzer, MD, MBA, professor of health studies at the Medical University of South Carolina in Charleston.

The findings were published online Sept. 27 in JAMA Neurology.
 

Common problem

Apathy, which is common among patients with AD, is associated with increased risk for mortality, financial burden, and caregiver burden. No treatment has proved effective for apathy in this population.

Two trials of methylphenidate, a catecholaminergic agent, have provided preliminary evidence of efficacy. Findings from the Apathy in Dementia Methylphenidate trial (ADMET) suggested the drug was associated with improved cognition and few adverse events. However, both trials had small patient populations and short durations.

The current investigators conducted ADMET 2, a 6-month, phase 3 trial, to investigate methylphenidate further. They recruited 200 patients (mean age, 76 years; 66% men; 90% White) at nine clinical centers that specialized in dementia care in the United States and one in Canada.

Eligible patients had a diagnosis of possible or probable AD and a Mini-Mental State Examination (MMSE) score between 10 and 28. They also had clinically significant apathy for at least 4 weeks and an available caregiver who spent more than 10 hours a week with the patient.

The researchers randomly assigned patients to receive methylphenidate (n = 99) or placebo (n = 101). For 3 days, participants in the active group received 10 mg/day of methylphenidate. After that point, they received 20 mg/day of methylphenidate for the rest of the study.

Patients in both treatment groups were given the same number of identical-appearing capsules each day.

In-person follow-up visits took place monthly for 6 months. Participants also were contacted by telephone at days 15, 45, and 75 after treatment assignment.

Participants underwent cognitive testing at baseline and at 2, 4, and 6 months. The battery of tests included the MMSE, Hopkins Verbal Learning Test, and Wechsler Adult Intelligence Scale – Revised Digit Span.

The trial’s two primary outcomes were mean change in NPI apathy score from baseline to 6 months and the odds of an improved rating on the Alzheimer’s Disease Cooperative Study Clinical Global Impression of Change (ADCS-CGIC) between baseline and 6 months.

Significant change on either outcome was to be considered a signal of effective treatment.
 

Treatment-specific benefit

Ten patients in the methylphenidate group and seven in the placebo group withdrew during the study.

Mean MMSE score at baseline was 19.2 in the methylphenidate group vs. 18.5 in the placebo group, indicating moderately severe dementia. Mean baseline score on the NPI apathy subscale was 8.0 vs. 7.6, respectively.

In an adjusted, longitudinal model, mean between-group difference in change in NPI apathy score at 6 months was –1.25 (P = .002). The mean NPI apathy score decreased by 4.5 points in the methylphenidate group vs. 3.1 points in the placebo group.

The largest change in apathy score occurred during the first 2 months of treatment. At 6 months, 27% of the methylphenidate group vs. 14% of the placebo group had an NPI apathy score of 0.

In addition, 43.8% of the methylphenidate group had improvement on the ADCS-CGIC compared with 35.2% of the placebo group. The odds ratio (OR) for improvement on ADCS-CGIC for methylphenidate vs. placebo was 1.90 (P = .07).

There was also a strong association between score improvement on the NPI apathy subscale and improvement on the ADCS-CGIC subscale (OR, 2.95; P = .002).

“It is important to note that there were no group differences in any of the cognitive measures, suggesting that the effect of the treatment is specific to the treatment of apathy and not a secondary effect of improvement in cognition,” the researchers wrote.

In all, 17 serious adverse events occurred in the methylphenidate group and 10 occurred in the placebo group. However, all events were found to be hospitalizations for events not related to treatment.
 

 

 

‘Enduring effect’

Commenting on the findings, Jeffrey L. Cummings, MD, ScD, professor of brain sciences at the University of Nevada, Las Vegas, noted that the reduction in NPI apathy subscale score of more than 50% was clinically meaningful.

Dr. Jeffrey L. Cummings

A more robust outcome on the ADCS-CGIC would have been desirable, he added, although that instrument is not designed specifically for apathy.

Methylphenidate’s effect on apathy observed at 2 months and remaining stable throughout the study makes it appear to be “an enduring effect, and not something that the patient accommodates to,” said Dr. Cummings, who was not involved with the research. Such a change may manifest itself in a patient’s greater willingness to help voluntarily with housework or to suggest going for a walk, he noted.

“These are not dramatic changes in cognition, of course, but they are changes in initiative and that is very important,” Dr. Cummings said. Decreased apathy also may improve quality of life for the patient’s caregiver, he added.

Overall, the findings raise the question of whether the Food and Drug Administration should recognize apathy as an indication for which drugs can be approved, said Dr. Cummings.

“For me, that would be the next major step in this line of investigation,” he concluded.

The study was funded by the National Institute on Aging. Dr. Mintzer has served as an adviser to Praxis Bioresearch and Cerevel Therapeutics on matters unrelated to this study. Dr. Cummings is the author of the Neuropsychiatric Inventory but does not receive payments for it from academic trials such as ADMET 2.
 

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Issue
Neurology Reviews - 29(11)
Publications
Publications
Topics
Article Type
Sections
Citation Override
Publish date: October 5, 2021
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

More than half of U.S. children under 6 years show detectable blood lead levels

Article Type
Changed

Lead poisoning remains a significant threat to the health of young children in the United States, based on data from blood tests of more than 1 million children.

Any level of lead is potentially harmful, although blood lead levels have decreased over the past several decades in part because of the elimination of lead from many consumer products, as well as from gas, paint, and plumbing fixtures, wrote Marissa Hauptman, MD, of Boston Children’s Hospital and colleagues.

However, “numerous environmental sources of legacy lead still exist,” and children living in poverty and in older housing in particular remain at increased risk for lead exposure, they noted.

In a study published in JAMA Pediatrics, the researchers analyzed deidentified results from blood lead tests performed at a single clinical laboratory for 1,141,441 children younger than 6 years between Oct. 1, 2018, and Feb. 29, 2020. The mean age of the children was 2.3 years; approximately half were boys.

Overall, 50.5% of the children tested (576,092 children) had detectable blood lead levels (BLLs), defined as 1.0 mcg/dL or higher, and 1.9% (21,172 children) had elevated BLLs, defined as 5.0 mcg/dL or higher.

In multivariate analysis, both detectable BLLs and elevated BLLs were significantly more common among children with public insurance (adjusted odds ratios, 2.01 and 1.08, respectively).

Children in the highest vs. lowest quintile of pre-1950s housing had significantly greater odds of both detectable and elevated BLLs (aOR, 1.65 and aOR, 3.06); those in the highest vs. lowest quintiles of poverty showed similarly increased risk of detectable and elevated BLLs (aOR, 1.89 and aOR, 1.99, respectively; P < .001 for all).

When the data were broken out by ZIP code, children in predominantly Black non-Hispanic and non-Latino neighborhoods were more likely than those living in other ZIP codes to have detectable BLLs (aOR, 1.13), but less likely to have elevated BLLs (aOR, 0.83). States with the highest overall proportions of children with detectable BLLs were Nebraska (83%), Missouri (82%), and Michigan (78%).

The study findings were limited by several factors, especially the potential for selection bias because of the use of a single reference laboratory (Quest Diagnostics), that does not perform all lead testing in the United States, the researchers noted. Other limitations included variability in testing at the state level, and the use of ZIP code–level data to estimate race, ethnicity, housing, and poverty, they said.

However, the results suggest that lead exposure remains a problem in young children, with significant disparities at the individual and community level, and national efforts must focus on further reductions of lead exposure in areas of highest risk, they concluded.
 

Step up lead elimination efforts

“The removal of lead from gasoline and new paint produced a precipitous decrease in blood lead levels from a population mean of 17 mcg/dL (all ages) in 1976 to 4 mcg/dL in the early 1990s to less than 2 mcg/dL today,” wrote Philip J. Landrigan, MD, of Boston College and David Bellinger, PhD, of Harvard University, Boston, in an accompanying editorial. However, “The findings from this study underscore the urgent need to eliminate all sources of lead exposure from U.S. children’s environments,” and highlight the persistent disparities in children’s lead exposure, they said.

The authors emphasized the need to remove existing lead paint from U.S. homes, as not only the paint itself, but the dust that enters the environment as the pain wears over time, continue to account for most detectable and elevated BLLs in children. A comprehensive lead paint removal effort would be an investment that would protect children now and would protect future generations, they emphasized. They proposed “creating a lead paint removal workforce through federally supported partnerships between city governments and major unions,” that would not only protect children from disease and disability, but could potentially provide jobs and vocational programs that would have a significant impact on communities.
 

Elevated lead levels may be underreported

In fact, the situation of children’s lead exposure in the United States may be more severe than indicated by the study findings, given the variation in testing at the state and local levels, said Karalyn Kinsella, MD, a pediatrician in private practice in Cheshire, Conn.

“There are no available lead test kits in our offices, so I do worry that many elevated lead levels will be missed,” she said.

“The recent case of elevated lead levels in drinking water in Flint, Michigan, was largely detected through pediatric clinic screening and showed that elevated lead levels may remain a major issue in some communities,” said Tim Joos, MD, a clinician in combined internal medicine/pediatrics in Seattle, Wash., in an interview.

“It is important to highlight to what extent baseline and point-source lead contamination still exists, monitor progress towards lowering levels, and identify communities at high risk,” Dr. Joos emphasized. “The exact prevalence of elevated lead levels among the general pediatric populations is hard to estimate from this study because of the methodology, which looked at demographic characteristics of the subset of the pediatric population that had venous samples sent to Quest Lab,” he noted.

“As the authors pointed out, it is hard to know what biases went into deciding whether to screen or not, and whether these were confirmatory tests for elevated point of care testing done earlier in the clinic,” said Dr. Joos. “Nonetheless, it does point to the role of poverty and pre-1950s housing in elevated blood lead levels,” he added. “The study also highlights that, as the CDC considers lowering the level for what is considered an ‘elevated blood lead level’ from 5.0 to perhaps 3.5 mcg/dL, we still have a lot more work to do,” he said.

The study was funded by Quest Diagnostics and the company provided salaries to several coauthors during the study. Dr. Hauptmann disclosed support from the National Institutes of Health/National Institute of Environmental Health Sciences during the current study and support from the Agency for Toxic Substances and Disease Registry and the U.S. Environmental Protection Agency unrelated to the current study. Dr. Landrigan had no financial conflicts to disclose. Dr. Bellinger disclosed fees from attorneys for testimony in cases unrelated to the editorial. Dr. Kinsella had no financial conflicts to disclose, but serves on the Editorial Advisory Board of Pediatric News. Dr. Joos had no financial conflicts to disclose, but serves on the Pediatric News Editorial Advisory Board.

Publications
Topics
Sections

Lead poisoning remains a significant threat to the health of young children in the United States, based on data from blood tests of more than 1 million children.

Any level of lead is potentially harmful, although blood lead levels have decreased over the past several decades in part because of the elimination of lead from many consumer products, as well as from gas, paint, and plumbing fixtures, wrote Marissa Hauptman, MD, of Boston Children’s Hospital and colleagues.

However, “numerous environmental sources of legacy lead still exist,” and children living in poverty and in older housing in particular remain at increased risk for lead exposure, they noted.

In a study published in JAMA Pediatrics, the researchers analyzed deidentified results from blood lead tests performed at a single clinical laboratory for 1,141,441 children younger than 6 years between Oct. 1, 2018, and Feb. 29, 2020. The mean age of the children was 2.3 years; approximately half were boys.

Overall, 50.5% of the children tested (576,092 children) had detectable blood lead levels (BLLs), defined as 1.0 mcg/dL or higher, and 1.9% (21,172 children) had elevated BLLs, defined as 5.0 mcg/dL or higher.

In multivariate analysis, both detectable BLLs and elevated BLLs were significantly more common among children with public insurance (adjusted odds ratios, 2.01 and 1.08, respectively).

Children in the highest vs. lowest quintile of pre-1950s housing had significantly greater odds of both detectable and elevated BLLs (aOR, 1.65 and aOR, 3.06); those in the highest vs. lowest quintiles of poverty showed similarly increased risk of detectable and elevated BLLs (aOR, 1.89 and aOR, 1.99, respectively; P < .001 for all).

When the data were broken out by ZIP code, children in predominantly Black non-Hispanic and non-Latino neighborhoods were more likely than those living in other ZIP codes to have detectable BLLs (aOR, 1.13), but less likely to have elevated BLLs (aOR, 0.83). States with the highest overall proportions of children with detectable BLLs were Nebraska (83%), Missouri (82%), and Michigan (78%).

The study findings were limited by several factors, especially the potential for selection bias because of the use of a single reference laboratory (Quest Diagnostics), that does not perform all lead testing in the United States, the researchers noted. Other limitations included variability in testing at the state level, and the use of ZIP code–level data to estimate race, ethnicity, housing, and poverty, they said.

However, the results suggest that lead exposure remains a problem in young children, with significant disparities at the individual and community level, and national efforts must focus on further reductions of lead exposure in areas of highest risk, they concluded.
 

Step up lead elimination efforts

“The removal of lead from gasoline and new paint produced a precipitous decrease in blood lead levels from a population mean of 17 mcg/dL (all ages) in 1976 to 4 mcg/dL in the early 1990s to less than 2 mcg/dL today,” wrote Philip J. Landrigan, MD, of Boston College and David Bellinger, PhD, of Harvard University, Boston, in an accompanying editorial. However, “The findings from this study underscore the urgent need to eliminate all sources of lead exposure from U.S. children’s environments,” and highlight the persistent disparities in children’s lead exposure, they said.

The authors emphasized the need to remove existing lead paint from U.S. homes, as not only the paint itself, but the dust that enters the environment as the pain wears over time, continue to account for most detectable and elevated BLLs in children. A comprehensive lead paint removal effort would be an investment that would protect children now and would protect future generations, they emphasized. They proposed “creating a lead paint removal workforce through federally supported partnerships between city governments and major unions,” that would not only protect children from disease and disability, but could potentially provide jobs and vocational programs that would have a significant impact on communities.
 

Elevated lead levels may be underreported

In fact, the situation of children’s lead exposure in the United States may be more severe than indicated by the study findings, given the variation in testing at the state and local levels, said Karalyn Kinsella, MD, a pediatrician in private practice in Cheshire, Conn.

“There are no available lead test kits in our offices, so I do worry that many elevated lead levels will be missed,” she said.

“The recent case of elevated lead levels in drinking water in Flint, Michigan, was largely detected through pediatric clinic screening and showed that elevated lead levels may remain a major issue in some communities,” said Tim Joos, MD, a clinician in combined internal medicine/pediatrics in Seattle, Wash., in an interview.

“It is important to highlight to what extent baseline and point-source lead contamination still exists, monitor progress towards lowering levels, and identify communities at high risk,” Dr. Joos emphasized. “The exact prevalence of elevated lead levels among the general pediatric populations is hard to estimate from this study because of the methodology, which looked at demographic characteristics of the subset of the pediatric population that had venous samples sent to Quest Lab,” he noted.

“As the authors pointed out, it is hard to know what biases went into deciding whether to screen or not, and whether these were confirmatory tests for elevated point of care testing done earlier in the clinic,” said Dr. Joos. “Nonetheless, it does point to the role of poverty and pre-1950s housing in elevated blood lead levels,” he added. “The study also highlights that, as the CDC considers lowering the level for what is considered an ‘elevated blood lead level’ from 5.0 to perhaps 3.5 mcg/dL, we still have a lot more work to do,” he said.

The study was funded by Quest Diagnostics and the company provided salaries to several coauthors during the study. Dr. Hauptmann disclosed support from the National Institutes of Health/National Institute of Environmental Health Sciences during the current study and support from the Agency for Toxic Substances and Disease Registry and the U.S. Environmental Protection Agency unrelated to the current study. Dr. Landrigan had no financial conflicts to disclose. Dr. Bellinger disclosed fees from attorneys for testimony in cases unrelated to the editorial. Dr. Kinsella had no financial conflicts to disclose, but serves on the Editorial Advisory Board of Pediatric News. Dr. Joos had no financial conflicts to disclose, but serves on the Pediatric News Editorial Advisory Board.

Lead poisoning remains a significant threat to the health of young children in the United States, based on data from blood tests of more than 1 million children.

Any level of lead is potentially harmful, although blood lead levels have decreased over the past several decades in part because of the elimination of lead from many consumer products, as well as from gas, paint, and plumbing fixtures, wrote Marissa Hauptman, MD, of Boston Children’s Hospital and colleagues.

However, “numerous environmental sources of legacy lead still exist,” and children living in poverty and in older housing in particular remain at increased risk for lead exposure, they noted.

In a study published in JAMA Pediatrics, the researchers analyzed deidentified results from blood lead tests performed at a single clinical laboratory for 1,141,441 children younger than 6 years between Oct. 1, 2018, and Feb. 29, 2020. The mean age of the children was 2.3 years; approximately half were boys.

Overall, 50.5% of the children tested (576,092 children) had detectable blood lead levels (BLLs), defined as 1.0 mcg/dL or higher, and 1.9% (21,172 children) had elevated BLLs, defined as 5.0 mcg/dL or higher.

In multivariate analysis, both detectable BLLs and elevated BLLs were significantly more common among children with public insurance (adjusted odds ratios, 2.01 and 1.08, respectively).

Children in the highest vs. lowest quintile of pre-1950s housing had significantly greater odds of both detectable and elevated BLLs (aOR, 1.65 and aOR, 3.06); those in the highest vs. lowest quintiles of poverty showed similarly increased risk of detectable and elevated BLLs (aOR, 1.89 and aOR, 1.99, respectively; P < .001 for all).

When the data were broken out by ZIP code, children in predominantly Black non-Hispanic and non-Latino neighborhoods were more likely than those living in other ZIP codes to have detectable BLLs (aOR, 1.13), but less likely to have elevated BLLs (aOR, 0.83). States with the highest overall proportions of children with detectable BLLs were Nebraska (83%), Missouri (82%), and Michigan (78%).

The study findings were limited by several factors, especially the potential for selection bias because of the use of a single reference laboratory (Quest Diagnostics), that does not perform all lead testing in the United States, the researchers noted. Other limitations included variability in testing at the state level, and the use of ZIP code–level data to estimate race, ethnicity, housing, and poverty, they said.

However, the results suggest that lead exposure remains a problem in young children, with significant disparities at the individual and community level, and national efforts must focus on further reductions of lead exposure in areas of highest risk, they concluded.
 

Step up lead elimination efforts

“The removal of lead from gasoline and new paint produced a precipitous decrease in blood lead levels from a population mean of 17 mcg/dL (all ages) in 1976 to 4 mcg/dL in the early 1990s to less than 2 mcg/dL today,” wrote Philip J. Landrigan, MD, of Boston College and David Bellinger, PhD, of Harvard University, Boston, in an accompanying editorial. However, “The findings from this study underscore the urgent need to eliminate all sources of lead exposure from U.S. children’s environments,” and highlight the persistent disparities in children’s lead exposure, they said.

The authors emphasized the need to remove existing lead paint from U.S. homes, as not only the paint itself, but the dust that enters the environment as the pain wears over time, continue to account for most detectable and elevated BLLs in children. A comprehensive lead paint removal effort would be an investment that would protect children now and would protect future generations, they emphasized. They proposed “creating a lead paint removal workforce through federally supported partnerships between city governments and major unions,” that would not only protect children from disease and disability, but could potentially provide jobs and vocational programs that would have a significant impact on communities.
 

Elevated lead levels may be underreported

In fact, the situation of children’s lead exposure in the United States may be more severe than indicated by the study findings, given the variation in testing at the state and local levels, said Karalyn Kinsella, MD, a pediatrician in private practice in Cheshire, Conn.

“There are no available lead test kits in our offices, so I do worry that many elevated lead levels will be missed,” she said.

“The recent case of elevated lead levels in drinking water in Flint, Michigan, was largely detected through pediatric clinic screening and showed that elevated lead levels may remain a major issue in some communities,” said Tim Joos, MD, a clinician in combined internal medicine/pediatrics in Seattle, Wash., in an interview.

“It is important to highlight to what extent baseline and point-source lead contamination still exists, monitor progress towards lowering levels, and identify communities at high risk,” Dr. Joos emphasized. “The exact prevalence of elevated lead levels among the general pediatric populations is hard to estimate from this study because of the methodology, which looked at demographic characteristics of the subset of the pediatric population that had venous samples sent to Quest Lab,” he noted.

“As the authors pointed out, it is hard to know what biases went into deciding whether to screen or not, and whether these were confirmatory tests for elevated point of care testing done earlier in the clinic,” said Dr. Joos. “Nonetheless, it does point to the role of poverty and pre-1950s housing in elevated blood lead levels,” he added. “The study also highlights that, as the CDC considers lowering the level for what is considered an ‘elevated blood lead level’ from 5.0 to perhaps 3.5 mcg/dL, we still have a lot more work to do,” he said.

The study was funded by Quest Diagnostics and the company provided salaries to several coauthors during the study. Dr. Hauptmann disclosed support from the National Institutes of Health/National Institute of Environmental Health Sciences during the current study and support from the Agency for Toxic Substances and Disease Registry and the U.S. Environmental Protection Agency unrelated to the current study. Dr. Landrigan had no financial conflicts to disclose. Dr. Bellinger disclosed fees from attorneys for testimony in cases unrelated to the editorial. Dr. Kinsella had no financial conflicts to disclose, but serves on the Editorial Advisory Board of Pediatric News. Dr. Joos had no financial conflicts to disclose, but serves on the Pediatric News Editorial Advisory Board.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM JAMA PEDIATRICS

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Opioid prescribing mapped: Alabama highest, New York lowest

Article Type
Changed

 

Medicare beneficiaries in Alabama were more likely to get a prescription for an opioid than in any other state in 2019, based on newly released data.

That year, opioids represented 6.48% of all drug claims for part D enrollees in the state, just ahead of Utah at 6.41%. Idaho, at 6.07%, was the only other state with an opioid prescribing rate over 6%, while Oklahoma came in at an even 6.0%, according to the latest update of the Centers for Medicare & Medicaid Services’ dataset.

The lowest rate in 2019 belonged to New York, where 2.51% of drug claims, including original prescriptions and refills, involved an opioid. Rhode Island was next at 2.87%, followed by New Jersey (3.23%), Massachusetts (3.26%), and North Dakota (3.39%),

Altogether, Medicare part D processed 1.5 billion drug claims in 2019, of which 66.1 million, or 4.41%, involved opioids. Both of the opioid numbers were down from 2018, when opioids represented 4.68% (70.2 million) of the 1.5 billion total claims, and from 2014, when opioids were involved in 5.73% (81,026,831) of the 1.41 billion drug claims, the CMS data show. That works out to 5.77% fewer opioids in 2019, compared with 2014, despite the increase in total volume.

Among the states, Delaware had the largest 5-year decrease, 2.38 percentage points, as its opioid prescribing rate dropped from 6.61% to 4.23% from 2014 to 2019, with Hawaii showing the smallest decline as it slipped 0.41 percentage points from 3.9% to 3.49%, according to the CMS.

In 2019, part D beneficiaries in Vermont were the most likely to receive a long-acting opioid, which accounted for 20.14% of all opioid prescriptions in the state, while Kentucky had the lowest share of prescriptions written for long-acting forms at 6.41%. The national average was 11.02%, dropping from 11.79% in 2018 and 12.75% in 2014, the CMS reported.

Publications
Topics
Sections

 

Medicare beneficiaries in Alabama were more likely to get a prescription for an opioid than in any other state in 2019, based on newly released data.

That year, opioids represented 6.48% of all drug claims for part D enrollees in the state, just ahead of Utah at 6.41%. Idaho, at 6.07%, was the only other state with an opioid prescribing rate over 6%, while Oklahoma came in at an even 6.0%, according to the latest update of the Centers for Medicare & Medicaid Services’ dataset.

The lowest rate in 2019 belonged to New York, where 2.51% of drug claims, including original prescriptions and refills, involved an opioid. Rhode Island was next at 2.87%, followed by New Jersey (3.23%), Massachusetts (3.26%), and North Dakota (3.39%),

Altogether, Medicare part D processed 1.5 billion drug claims in 2019, of which 66.1 million, or 4.41%, involved opioids. Both of the opioid numbers were down from 2018, when opioids represented 4.68% (70.2 million) of the 1.5 billion total claims, and from 2014, when opioids were involved in 5.73% (81,026,831) of the 1.41 billion drug claims, the CMS data show. That works out to 5.77% fewer opioids in 2019, compared with 2014, despite the increase in total volume.

Among the states, Delaware had the largest 5-year decrease, 2.38 percentage points, as its opioid prescribing rate dropped from 6.61% to 4.23% from 2014 to 2019, with Hawaii showing the smallest decline as it slipped 0.41 percentage points from 3.9% to 3.49%, according to the CMS.

In 2019, part D beneficiaries in Vermont were the most likely to receive a long-acting opioid, which accounted for 20.14% of all opioid prescriptions in the state, while Kentucky had the lowest share of prescriptions written for long-acting forms at 6.41%. The national average was 11.02%, dropping from 11.79% in 2018 and 12.75% in 2014, the CMS reported.

 

Medicare beneficiaries in Alabama were more likely to get a prescription for an opioid than in any other state in 2019, based on newly released data.

That year, opioids represented 6.48% of all drug claims for part D enrollees in the state, just ahead of Utah at 6.41%. Idaho, at 6.07%, was the only other state with an opioid prescribing rate over 6%, while Oklahoma came in at an even 6.0%, according to the latest update of the Centers for Medicare & Medicaid Services’ dataset.

The lowest rate in 2019 belonged to New York, where 2.51% of drug claims, including original prescriptions and refills, involved an opioid. Rhode Island was next at 2.87%, followed by New Jersey (3.23%), Massachusetts (3.26%), and North Dakota (3.39%),

Altogether, Medicare part D processed 1.5 billion drug claims in 2019, of which 66.1 million, or 4.41%, involved opioids. Both of the opioid numbers were down from 2018, when opioids represented 4.68% (70.2 million) of the 1.5 billion total claims, and from 2014, when opioids were involved in 5.73% (81,026,831) of the 1.41 billion drug claims, the CMS data show. That works out to 5.77% fewer opioids in 2019, compared with 2014, despite the increase in total volume.

Among the states, Delaware had the largest 5-year decrease, 2.38 percentage points, as its opioid prescribing rate dropped from 6.61% to 4.23% from 2014 to 2019, with Hawaii showing the smallest decline as it slipped 0.41 percentage points from 3.9% to 3.49%, according to the CMS.

In 2019, part D beneficiaries in Vermont were the most likely to receive a long-acting opioid, which accounted for 20.14% of all opioid prescriptions in the state, while Kentucky had the lowest share of prescriptions written for long-acting forms at 6.41%. The national average was 11.02%, dropping from 11.79% in 2018 and 12.75% in 2014, the CMS reported.

Publications
Publications
Topics
Article Type
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

MIND diet preserves cognition, new data show

Article Type
Changed

 

Adherence to the MIND diet can improve memory and thinking skills of older adults, even in the presence of Alzheimer’s disease pathology, new data from the Rush Memory and Aging Project (MAP) show.

Suprijono Suharjoto/Fotolia

“The MIND diet was associated with better cognitive functions independently of brain pathologies related to Alzheimer’s disease, suggesting that diet may contribute to cognitive resilience, which ultimately indicates that it is never too late for dementia prevention,” lead author Klodian Dhana, MD, PhD, with the Rush Institute of Healthy Aging at Rush University, Chicago, said in an interview.

The study was published online Sept. 14, 2021, in the Journal of Alzheimer’s Disease.
 

Impact on brain pathology

“While previous investigations determined that the MIND diet is associated with a slower cognitive decline, the current study furthered the diet and brain health evidence by assessing the impact of brain pathology in the diet-cognition relationship,” Dr. Dhana said.

The MIND diet was pioneered by the late Martha Clare Morris, ScD, a Rush nutritional epidemiologist, who died in 2020 of cancer at age 64. A hybrid of the Mediterranean and DASH (Dietary Approaches to Stop Hypertension) diets, the MIND diet includes green leafy vegetables, fish, nuts, berries, beans, and whole grains and limits consumption of fried and fast foods, sweets, and pastries.

The current study focused on 569 older adults who died while participating in the MAP study, which began in 1997. Participants in the study were mostly White and were without known dementia. All of the participants agreed to undergo annual clinical evaluations. They also agreed to undergo brain autopsy after death.

Beginning in 2004, participants completed annual food frequency questionnaires, which were used to calculate a MIND diet score based on how often the participants ate specific foods.

The researchers used a series of regression analyses to examine associations of the MIND diet, dementia-related brain pathologies, and global cognition near the time of death. Analyses were adjusted for age, sex, education, apo E4, late-life cognitive activities, and total energy intake.

The results show that a higher MIND diet score was associated with better global cognitive functioning around the time of death (beta, 0.119; P = .003).

Notably, the researchers said, neither the strength nor the significance of association changed markedly when AD pathology and other brain pathologies were included in the model (beta, 0.111; P = .003).

The relationship between better adherence to the MIND diet and better cognition remained significant when the analysis was restricted to individuals without mild cognitive impairment at baseline (beta, 0.121; P = .005) as well as to persons in whom a postmortem diagnosis of AD was made on the basis of NIA-Reagan consensus recommendations (beta, 0.114; P = .023).

The limitations of the study include the reliance on self-reported diet information and a sample made up of mostly White volunteers who agreed to annual evaluations and postmortem organ donation, thus limiting generalizability.

Strengths of the study include the prospective design with annual assessment of cognitive function using standardized tests and collection of the dietary information using validated questionnaires. Also, the neuropathologic evaluations were performed by examiners blinded to clinical data.

“Diet changes can impact cognitive functioning and risk of dementia, for better or worse. There are fairly simple diet and lifestyle changes a person could make that may help to slow cognitive decline with aging and contribute to brain health,” Dr. Dhana said in a news release.
 

 

 

Builds resilience

Weighing in on the study, Heather Snyder, PhD, vice president of medical and scientific relations for the Alzheimer’s Association, said this “interesting study sheds light on the impact of nutrition on cognitive function.

“The findings add to the growing literature that lifestyle factors – like access to a heart-healthy diet – may help the brain be more resilient to disease-specific changes,” Snyder said in an interview.

“The Alzheimer’s Association’s US POINTER study is investigating how lifestyle interventions, including nutrition guidance, like the MIND diet, may impact a person’s risk of cognitive decline. An ancillary study of the US POINTER will include brain imaging to investigate how these lifestyle interventions impact the biology of the brain,” Dr. Snyder noted.

The research was supported by the National Institute on Aging of the National Institutes of Health. Dr. Dhana and Dr. Snyder disclosed no relevant financial relationships.

A version of this article first appeared on Medscape.com.

Publications
Topics
Sections

 

Adherence to the MIND diet can improve memory and thinking skills of older adults, even in the presence of Alzheimer’s disease pathology, new data from the Rush Memory and Aging Project (MAP) show.

Suprijono Suharjoto/Fotolia

“The MIND diet was associated with better cognitive functions independently of brain pathologies related to Alzheimer’s disease, suggesting that diet may contribute to cognitive resilience, which ultimately indicates that it is never too late for dementia prevention,” lead author Klodian Dhana, MD, PhD, with the Rush Institute of Healthy Aging at Rush University, Chicago, said in an interview.

The study was published online Sept. 14, 2021, in the Journal of Alzheimer’s Disease.
 

Impact on brain pathology

“While previous investigations determined that the MIND diet is associated with a slower cognitive decline, the current study furthered the diet and brain health evidence by assessing the impact of brain pathology in the diet-cognition relationship,” Dr. Dhana said.

The MIND diet was pioneered by the late Martha Clare Morris, ScD, a Rush nutritional epidemiologist, who died in 2020 of cancer at age 64. A hybrid of the Mediterranean and DASH (Dietary Approaches to Stop Hypertension) diets, the MIND diet includes green leafy vegetables, fish, nuts, berries, beans, and whole grains and limits consumption of fried and fast foods, sweets, and pastries.

The current study focused on 569 older adults who died while participating in the MAP study, which began in 1997. Participants in the study were mostly White and were without known dementia. All of the participants agreed to undergo annual clinical evaluations. They also agreed to undergo brain autopsy after death.

Beginning in 2004, participants completed annual food frequency questionnaires, which were used to calculate a MIND diet score based on how often the participants ate specific foods.

The researchers used a series of regression analyses to examine associations of the MIND diet, dementia-related brain pathologies, and global cognition near the time of death. Analyses were adjusted for age, sex, education, apo E4, late-life cognitive activities, and total energy intake.

The results show that a higher MIND diet score was associated with better global cognitive functioning around the time of death (beta, 0.119; P = .003).

Notably, the researchers said, neither the strength nor the significance of association changed markedly when AD pathology and other brain pathologies were included in the model (beta, 0.111; P = .003).

The relationship between better adherence to the MIND diet and better cognition remained significant when the analysis was restricted to individuals without mild cognitive impairment at baseline (beta, 0.121; P = .005) as well as to persons in whom a postmortem diagnosis of AD was made on the basis of NIA-Reagan consensus recommendations (beta, 0.114; P = .023).

The limitations of the study include the reliance on self-reported diet information and a sample made up of mostly White volunteers who agreed to annual evaluations and postmortem organ donation, thus limiting generalizability.

Strengths of the study include the prospective design with annual assessment of cognitive function using standardized tests and collection of the dietary information using validated questionnaires. Also, the neuropathologic evaluations were performed by examiners blinded to clinical data.

“Diet changes can impact cognitive functioning and risk of dementia, for better or worse. There are fairly simple diet and lifestyle changes a person could make that may help to slow cognitive decline with aging and contribute to brain health,” Dr. Dhana said in a news release.
 

 

 

Builds resilience

Weighing in on the study, Heather Snyder, PhD, vice president of medical and scientific relations for the Alzheimer’s Association, said this “interesting study sheds light on the impact of nutrition on cognitive function.

“The findings add to the growing literature that lifestyle factors – like access to a heart-healthy diet – may help the brain be more resilient to disease-specific changes,” Snyder said in an interview.

“The Alzheimer’s Association’s US POINTER study is investigating how lifestyle interventions, including nutrition guidance, like the MIND diet, may impact a person’s risk of cognitive decline. An ancillary study of the US POINTER will include brain imaging to investigate how these lifestyle interventions impact the biology of the brain,” Dr. Snyder noted.

The research was supported by the National Institute on Aging of the National Institutes of Health. Dr. Dhana and Dr. Snyder disclosed no relevant financial relationships.

A version of this article first appeared on Medscape.com.

 

Adherence to the MIND diet can improve memory and thinking skills of older adults, even in the presence of Alzheimer’s disease pathology, new data from the Rush Memory and Aging Project (MAP) show.

Suprijono Suharjoto/Fotolia

“The MIND diet was associated with better cognitive functions independently of brain pathologies related to Alzheimer’s disease, suggesting that diet may contribute to cognitive resilience, which ultimately indicates that it is never too late for dementia prevention,” lead author Klodian Dhana, MD, PhD, with the Rush Institute of Healthy Aging at Rush University, Chicago, said in an interview.

The study was published online Sept. 14, 2021, in the Journal of Alzheimer’s Disease.
 

Impact on brain pathology

“While previous investigations determined that the MIND diet is associated with a slower cognitive decline, the current study furthered the diet and brain health evidence by assessing the impact of brain pathology in the diet-cognition relationship,” Dr. Dhana said.

The MIND diet was pioneered by the late Martha Clare Morris, ScD, a Rush nutritional epidemiologist, who died in 2020 of cancer at age 64. A hybrid of the Mediterranean and DASH (Dietary Approaches to Stop Hypertension) diets, the MIND diet includes green leafy vegetables, fish, nuts, berries, beans, and whole grains and limits consumption of fried and fast foods, sweets, and pastries.

The current study focused on 569 older adults who died while participating in the MAP study, which began in 1997. Participants in the study were mostly White and were without known dementia. All of the participants agreed to undergo annual clinical evaluations. They also agreed to undergo brain autopsy after death.

Beginning in 2004, participants completed annual food frequency questionnaires, which were used to calculate a MIND diet score based on how often the participants ate specific foods.

The researchers used a series of regression analyses to examine associations of the MIND diet, dementia-related brain pathologies, and global cognition near the time of death. Analyses were adjusted for age, sex, education, apo E4, late-life cognitive activities, and total energy intake.

The results show that a higher MIND diet score was associated with better global cognitive functioning around the time of death (beta, 0.119; P = .003).

Notably, the researchers said, neither the strength nor the significance of association changed markedly when AD pathology and other brain pathologies were included in the model (beta, 0.111; P = .003).

The relationship between better adherence to the MIND diet and better cognition remained significant when the analysis was restricted to individuals without mild cognitive impairment at baseline (beta, 0.121; P = .005) as well as to persons in whom a postmortem diagnosis of AD was made on the basis of NIA-Reagan consensus recommendations (beta, 0.114; P = .023).

The limitations of the study include the reliance on self-reported diet information and a sample made up of mostly White volunteers who agreed to annual evaluations and postmortem organ donation, thus limiting generalizability.

Strengths of the study include the prospective design with annual assessment of cognitive function using standardized tests and collection of the dietary information using validated questionnaires. Also, the neuropathologic evaluations were performed by examiners blinded to clinical data.

“Diet changes can impact cognitive functioning and risk of dementia, for better or worse. There are fairly simple diet and lifestyle changes a person could make that may help to slow cognitive decline with aging and contribute to brain health,” Dr. Dhana said in a news release.
 

 

 

Builds resilience

Weighing in on the study, Heather Snyder, PhD, vice president of medical and scientific relations for the Alzheimer’s Association, said this “interesting study sheds light on the impact of nutrition on cognitive function.

“The findings add to the growing literature that lifestyle factors – like access to a heart-healthy diet – may help the brain be more resilient to disease-specific changes,” Snyder said in an interview.

“The Alzheimer’s Association’s US POINTER study is investigating how lifestyle interventions, including nutrition guidance, like the MIND diet, may impact a person’s risk of cognitive decline. An ancillary study of the US POINTER will include brain imaging to investigate how these lifestyle interventions impact the biology of the brain,” Dr. Snyder noted.

The research was supported by the National Institute on Aging of the National Institutes of Health. Dr. Dhana and Dr. Snyder disclosed no relevant financial relationships.

A version of this article first appeared on Medscape.com.

Publications
Publications
Topics
Article Type
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Migraine history linked to more severe hot flashes in postmenopausal women

Article Type
Changed

Women with a history of migraine are more likely to experience severe or very severe hot flashes than women without migraines, according to research presented Sept. 24 at the hybrid annual meeting of the North American Menopause Society. An estimated one in five women experience migraine, and women tend to have greater migraine symptoms and disability, the authors note in their background information. Since migraines are also linked to a higher risk of cardiovascular disease, the authors sought to learn whether migraines were associated with vasomotor symptoms, another cardiovascular risk factor.

“The question in my mind is, can we do better at predicting cardiovascular risk in women because the risk prediction models that we have really don’t work all that well in women because they were designed for use in men,” Stephanie S. Faubion, MD, MBA, Penny and Bill George Director for Mayo Clinic’s Center for Women’s Health said in an interview. “My ultimate goal is to see if we can somehow use big data, artificial intelligence to figure out how to weight some of these female-specific or female-predominant factors to come up with a better model for cardiovascular risk prediction.”

The researchers analyzed cross-sectional data from 3,308 women who participated in the Data Registry on the Experiences of Aging, Menopause and Sexuality (DREAMS) study through Mayo Clinic sites in Rochester, Minn.; Scottsdale, Ariz.; and Jacksonville, Fla.. The women ranged in age from 45 to 60 years old, with an average age of 53, and the vast majority of them were white (95%) and had at least some college (93%). Most were also in a long-term relationship (85%), and a majority were employed (69%) and postmenopausal (67%).

The data, collected between May 2015 and December 2019, included a self-reported history of migraine and questionnaires that included the Menopause Rating Scale of menopause-related symptoms.

The researchers adjusted their findings to account for body mass index (BMI), menopause status, smoking status, depression, anxiety, current use of hormone therapy, and presence of low back pain within the past year. ”The diagnosis of low back pain, another pain disorder, was used to test the specificity of the association of migraine and vasomotor symptoms,” the authors write.

Just over a quarter of the women (27%) reported a history of migraine, and these women’s Menopause Rating Scale scores were an average 1.36 points greater than women without a history of migraines (P < .001). Women with self-reported migraine were also 40% more likely than women without migraines to report severe or very severe flashes versus reporting no hot flashes at all (odds ratio, 1.4; P = .02).

“The odds of reporting more severe hot flashes increased monotonically in women with a history of migraine,” the authors report. “In addition, women with low back pain had higher Menopause Rating Scale scores, but were no more likely to have severe/very severe hot flashes than those without back pain, confirming the specificity of the link between vasomotor symptoms and migraine.”

It’s not clear if migraine or hot flashes are risk factors that add to a woman’s existing cardiovascular risk profile or whether they are simply biomarkers of a shared pathway, Dr Faubion said in an interview. She speculates that the common link between migraine and vasomotor symptoms could be neurovascular dysregulation.

Rachael B. Smith, DO, of the department of ob.gyn. at the University of Arizona, Phoenix, was not involved in the research but found that hypothesis plausible as well.

“Our neurologic and vascular systems are coordinated physiologic processes working together for basic brain and body function,” Dr. Smith said in an interview. Some of the symptoms of migraines and menopause are similar and both are often explained by the dysfunction of these systems. The association between history of migraines and severity of vasomotor symptoms is very likely to be explained by this dysregulation between the neurologic and vascular systems.”

Dr. Smith also pointed out, however, that the largely homogeneous study population, all from the same national clinic system, makes it difficult to know how generalizable these findings are.

The primary clinical implications of these findings are that women’s providers need to be sure they’re asking their patients about migraine history and symptoms.

“The counseling we provide on menopausal symptoms should be better tailored to our patients’ medical history, specifically inquiring about history of migraines and how this may impact their symptoms,” Dr. Smith said.

The research was funded by the National Institutes of Health. Dr. Faubion and Dr. Smith had no disclosures.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

Women with a history of migraine are more likely to experience severe or very severe hot flashes than women without migraines, according to research presented Sept. 24 at the hybrid annual meeting of the North American Menopause Society. An estimated one in five women experience migraine, and women tend to have greater migraine symptoms and disability, the authors note in their background information. Since migraines are also linked to a higher risk of cardiovascular disease, the authors sought to learn whether migraines were associated with vasomotor symptoms, another cardiovascular risk factor.

“The question in my mind is, can we do better at predicting cardiovascular risk in women because the risk prediction models that we have really don’t work all that well in women because they were designed for use in men,” Stephanie S. Faubion, MD, MBA, Penny and Bill George Director for Mayo Clinic’s Center for Women’s Health said in an interview. “My ultimate goal is to see if we can somehow use big data, artificial intelligence to figure out how to weight some of these female-specific or female-predominant factors to come up with a better model for cardiovascular risk prediction.”

The researchers analyzed cross-sectional data from 3,308 women who participated in the Data Registry on the Experiences of Aging, Menopause and Sexuality (DREAMS) study through Mayo Clinic sites in Rochester, Minn.; Scottsdale, Ariz.; and Jacksonville, Fla.. The women ranged in age from 45 to 60 years old, with an average age of 53, and the vast majority of them were white (95%) and had at least some college (93%). Most were also in a long-term relationship (85%), and a majority were employed (69%) and postmenopausal (67%).

The data, collected between May 2015 and December 2019, included a self-reported history of migraine and questionnaires that included the Menopause Rating Scale of menopause-related symptoms.

The researchers adjusted their findings to account for body mass index (BMI), menopause status, smoking status, depression, anxiety, current use of hormone therapy, and presence of low back pain within the past year. ”The diagnosis of low back pain, another pain disorder, was used to test the specificity of the association of migraine and vasomotor symptoms,” the authors write.

Just over a quarter of the women (27%) reported a history of migraine, and these women’s Menopause Rating Scale scores were an average 1.36 points greater than women without a history of migraines (P < .001). Women with self-reported migraine were also 40% more likely than women without migraines to report severe or very severe flashes versus reporting no hot flashes at all (odds ratio, 1.4; P = .02).

“The odds of reporting more severe hot flashes increased monotonically in women with a history of migraine,” the authors report. “In addition, women with low back pain had higher Menopause Rating Scale scores, but were no more likely to have severe/very severe hot flashes than those without back pain, confirming the specificity of the link between vasomotor symptoms and migraine.”

It’s not clear if migraine or hot flashes are risk factors that add to a woman’s existing cardiovascular risk profile or whether they are simply biomarkers of a shared pathway, Dr Faubion said in an interview. She speculates that the common link between migraine and vasomotor symptoms could be neurovascular dysregulation.

Rachael B. Smith, DO, of the department of ob.gyn. at the University of Arizona, Phoenix, was not involved in the research but found that hypothesis plausible as well.

“Our neurologic and vascular systems are coordinated physiologic processes working together for basic brain and body function,” Dr. Smith said in an interview. Some of the symptoms of migraines and menopause are similar and both are often explained by the dysfunction of these systems. The association between history of migraines and severity of vasomotor symptoms is very likely to be explained by this dysregulation between the neurologic and vascular systems.”

Dr. Smith also pointed out, however, that the largely homogeneous study population, all from the same national clinic system, makes it difficult to know how generalizable these findings are.

The primary clinical implications of these findings are that women’s providers need to be sure they’re asking their patients about migraine history and symptoms.

“The counseling we provide on menopausal symptoms should be better tailored to our patients’ medical history, specifically inquiring about history of migraines and how this may impact their symptoms,” Dr. Smith said.

The research was funded by the National Institutes of Health. Dr. Faubion and Dr. Smith had no disclosures.

Women with a history of migraine are more likely to experience severe or very severe hot flashes than women without migraines, according to research presented Sept. 24 at the hybrid annual meeting of the North American Menopause Society. An estimated one in five women experience migraine, and women tend to have greater migraine symptoms and disability, the authors note in their background information. Since migraines are also linked to a higher risk of cardiovascular disease, the authors sought to learn whether migraines were associated with vasomotor symptoms, another cardiovascular risk factor.

“The question in my mind is, can we do better at predicting cardiovascular risk in women because the risk prediction models that we have really don’t work all that well in women because they were designed for use in men,” Stephanie S. Faubion, MD, MBA, Penny and Bill George Director for Mayo Clinic’s Center for Women’s Health said in an interview. “My ultimate goal is to see if we can somehow use big data, artificial intelligence to figure out how to weight some of these female-specific or female-predominant factors to come up with a better model for cardiovascular risk prediction.”

The researchers analyzed cross-sectional data from 3,308 women who participated in the Data Registry on the Experiences of Aging, Menopause and Sexuality (DREAMS) study through Mayo Clinic sites in Rochester, Minn.; Scottsdale, Ariz.; and Jacksonville, Fla.. The women ranged in age from 45 to 60 years old, with an average age of 53, and the vast majority of them were white (95%) and had at least some college (93%). Most were also in a long-term relationship (85%), and a majority were employed (69%) and postmenopausal (67%).

The data, collected between May 2015 and December 2019, included a self-reported history of migraine and questionnaires that included the Menopause Rating Scale of menopause-related symptoms.

The researchers adjusted their findings to account for body mass index (BMI), menopause status, smoking status, depression, anxiety, current use of hormone therapy, and presence of low back pain within the past year. ”The diagnosis of low back pain, another pain disorder, was used to test the specificity of the association of migraine and vasomotor symptoms,” the authors write.

Just over a quarter of the women (27%) reported a history of migraine, and these women’s Menopause Rating Scale scores were an average 1.36 points greater than women without a history of migraines (P < .001). Women with self-reported migraine were also 40% more likely than women without migraines to report severe or very severe flashes versus reporting no hot flashes at all (odds ratio, 1.4; P = .02).

“The odds of reporting more severe hot flashes increased monotonically in women with a history of migraine,” the authors report. “In addition, women with low back pain had higher Menopause Rating Scale scores, but were no more likely to have severe/very severe hot flashes than those without back pain, confirming the specificity of the link between vasomotor symptoms and migraine.”

It’s not clear if migraine or hot flashes are risk factors that add to a woman’s existing cardiovascular risk profile or whether they are simply biomarkers of a shared pathway, Dr Faubion said in an interview. She speculates that the common link between migraine and vasomotor symptoms could be neurovascular dysregulation.

Rachael B. Smith, DO, of the department of ob.gyn. at the University of Arizona, Phoenix, was not involved in the research but found that hypothesis plausible as well.

“Our neurologic and vascular systems are coordinated physiologic processes working together for basic brain and body function,” Dr. Smith said in an interview. Some of the symptoms of migraines and menopause are similar and both are often explained by the dysfunction of these systems. The association between history of migraines and severity of vasomotor symptoms is very likely to be explained by this dysregulation between the neurologic and vascular systems.”

Dr. Smith also pointed out, however, that the largely homogeneous study population, all from the same national clinic system, makes it difficult to know how generalizable these findings are.

The primary clinical implications of these findings are that women’s providers need to be sure they’re asking their patients about migraine history and symptoms.

“The counseling we provide on menopausal symptoms should be better tailored to our patients’ medical history, specifically inquiring about history of migraines and how this may impact their symptoms,” Dr. Smith said.

The research was funded by the National Institutes of Health. Dr. Faubion and Dr. Smith had no disclosures.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM NAMS 2021

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

FDA okays new oral CGRP antagonist for migraine prevention

Article Type
Changed

The Food and Drug Administration has approved atogepant (Qulipta), a novel calcitonin gene-related peptide (CGRP) receptor antagonist, for the prevention of episodic migraine, the manufacturer announced in a release.

The once-daily medication will be available in doses of 10 mg, 30 mg, and 60 mg.

“Qulipta provides a simple oral treatment option specifically developed to prevent migraine attacks and target CGRP, which is believed to be crucially involved in migraine in many patients,” coinvestigator Peter J. Goadsby, MD, PhD, DSc, neurologist and professor at the University of California, Los Angeles, and King’s College London, said in the release.

Approval was based partly on the findings from the phase 3 ADVANCE trial, in which patients with episodic migraine were randomly assigned to receive placebo or a 10-mg, 30-mg, or 60-mg daily dose of atogepant for 12 weeks.

As reported by this news organization, all three doses of atogepant reduced the number of mean monthly migraine days.

With this approval, neurologists will be able to choose from four monoclonal antibodies and two gepants for the preventive treatment of migraine.

“Having another gepant that can also be given preventively is a good idea, because one may be better than the other for a patient,” Alan M. Rapoport, MD, past president of the International Headache Society and founder and director emeritus of the New England Center for Headache, Stamford, Conn., told this news organization.

“Once we have a year or so of experience with atogepant, we’ll have a pretty good idea of which one works better preventively,” said Dr. Rapoport, who was not involved with the research.
 

Practice changing?

In the ADVANCE trial, there was a reduction of 3.69 migraine days with the 10-mg dose, 3.86 days with the 30-mg dose, and 4.2 days with the 60-mg dose. Placebo was associated with a reduction of 2.48 migraine days.

In addition, more than half of patients in each atogepant arm achieved a reduction in mean monthly migraine days of 50% or greater. This outcome occurred in 55.6% of the 10-mg atogepant group, 58.7% of the 30-mg group, and 60.8% of the 60-mg group. Approximately 29% patients who received placebo achieved this outcome.

The data indicated that atogepant has a favorable safety profile. The most common adverse events associated with treatment were constipation, nausea, and upper respiratory tract infection.

Dr. Rapoport, who is also a clinical professor of neurology at UCLA, noted that he was impressed with the efficacy.

“I’m not as impressed with the adverse events, but they’re not serious, and they don’t necessarily last,” he said.

Although being able to prescribe a single drug for acute and preventive treatment may be an advantage, it remains to be seen whether the tolerability and price of atogepant will be barriers for patients, Dr. Rapoport added.

How the approval will affect clinical practice is also unclear, he noted.

“If you’re going to start someone on a preventive, especially if it’s a woman of childbearing potential, you might just consider one of the two gepants. Doctors will decide once they see how they work,” said Dr. Rapoport.
 

 

 

Not a ‘breakthrough’ treatment

Also commenting ahead of the approval, Elizabeth W. Loder, MD, vice chair for academic affairs in the department of neurology at Brigham and Women’s Hospital, Boston, noted that the “safety of these CGRP medications in pregnancy is uncertain, and there are theoretical reasons to be concerned about it.”

Unlike injectable CGRP medications, atogepant is eliminated from the body relatively quickly after the patient stops taking it, said Dr. Loder, who is also professor of neurology at Harvard Medical School, Boston. However, atogepant may not otherwise differ greatly from other medications of its type.

“I don’t see a reason to think that one of these oral CGRP medicines is much more effective than another one,” said Dr. Loder.

“In my mind, as a clinician who will be prescribing these for patients, it will be cost and the ease of getting it covered that makes the difference,” she added.

These questions may raise concerns. “Those of us who treat patients who do not have private insurance find it very difficult to get these medications for them, even in situations where they have exhausted other alternatives,” said Dr. Loder.

Patients insured by Medicare or Medicaid “usually have no avenue to get some of these new, expensive treatments,” she said.

The approval of atogepant for acute and preventive treatment shows that the distinction between these indications may be artificial, Dr. Loder noted. The approval “will, I hope, help people think more flexibly about the way in which we use medications.”

It is a positive that atogepant has emerged as another option for preventive therapy, but the treatment cannot be considered a breakthrough, Dr. Loder added. The efficacy of atogepant, like that of other preventive treatments for migraine, is modest.

“It would be so nice if we could find things that were more effective than the treatments we currently have,” said Dr. Loder.

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Publications
Topics
Sections

The Food and Drug Administration has approved atogepant (Qulipta), a novel calcitonin gene-related peptide (CGRP) receptor antagonist, for the prevention of episodic migraine, the manufacturer announced in a release.

The once-daily medication will be available in doses of 10 mg, 30 mg, and 60 mg.

“Qulipta provides a simple oral treatment option specifically developed to prevent migraine attacks and target CGRP, which is believed to be crucially involved in migraine in many patients,” coinvestigator Peter J. Goadsby, MD, PhD, DSc, neurologist and professor at the University of California, Los Angeles, and King’s College London, said in the release.

Approval was based partly on the findings from the phase 3 ADVANCE trial, in which patients with episodic migraine were randomly assigned to receive placebo or a 10-mg, 30-mg, or 60-mg daily dose of atogepant for 12 weeks.

As reported by this news organization, all three doses of atogepant reduced the number of mean monthly migraine days.

With this approval, neurologists will be able to choose from four monoclonal antibodies and two gepants for the preventive treatment of migraine.

“Having another gepant that can also be given preventively is a good idea, because one may be better than the other for a patient,” Alan M. Rapoport, MD, past president of the International Headache Society and founder and director emeritus of the New England Center for Headache, Stamford, Conn., told this news organization.

“Once we have a year or so of experience with atogepant, we’ll have a pretty good idea of which one works better preventively,” said Dr. Rapoport, who was not involved with the research.
 

Practice changing?

In the ADVANCE trial, there was a reduction of 3.69 migraine days with the 10-mg dose, 3.86 days with the 30-mg dose, and 4.2 days with the 60-mg dose. Placebo was associated with a reduction of 2.48 migraine days.

In addition, more than half of patients in each atogepant arm achieved a reduction in mean monthly migraine days of 50% or greater. This outcome occurred in 55.6% of the 10-mg atogepant group, 58.7% of the 30-mg group, and 60.8% of the 60-mg group. Approximately 29% patients who received placebo achieved this outcome.

The data indicated that atogepant has a favorable safety profile. The most common adverse events associated with treatment were constipation, nausea, and upper respiratory tract infection.

Dr. Rapoport, who is also a clinical professor of neurology at UCLA, noted that he was impressed with the efficacy.

“I’m not as impressed with the adverse events, but they’re not serious, and they don’t necessarily last,” he said.

Although being able to prescribe a single drug for acute and preventive treatment may be an advantage, it remains to be seen whether the tolerability and price of atogepant will be barriers for patients, Dr. Rapoport added.

How the approval will affect clinical practice is also unclear, he noted.

“If you’re going to start someone on a preventive, especially if it’s a woman of childbearing potential, you might just consider one of the two gepants. Doctors will decide once they see how they work,” said Dr. Rapoport.
 

 

 

Not a ‘breakthrough’ treatment

Also commenting ahead of the approval, Elizabeth W. Loder, MD, vice chair for academic affairs in the department of neurology at Brigham and Women’s Hospital, Boston, noted that the “safety of these CGRP medications in pregnancy is uncertain, and there are theoretical reasons to be concerned about it.”

Unlike injectable CGRP medications, atogepant is eliminated from the body relatively quickly after the patient stops taking it, said Dr. Loder, who is also professor of neurology at Harvard Medical School, Boston. However, atogepant may not otherwise differ greatly from other medications of its type.

“I don’t see a reason to think that one of these oral CGRP medicines is much more effective than another one,” said Dr. Loder.

“In my mind, as a clinician who will be prescribing these for patients, it will be cost and the ease of getting it covered that makes the difference,” she added.

These questions may raise concerns. “Those of us who treat patients who do not have private insurance find it very difficult to get these medications for them, even in situations where they have exhausted other alternatives,” said Dr. Loder.

Patients insured by Medicare or Medicaid “usually have no avenue to get some of these new, expensive treatments,” she said.

The approval of atogepant for acute and preventive treatment shows that the distinction between these indications may be artificial, Dr. Loder noted. The approval “will, I hope, help people think more flexibly about the way in which we use medications.”

It is a positive that atogepant has emerged as another option for preventive therapy, but the treatment cannot be considered a breakthrough, Dr. Loder added. The efficacy of atogepant, like that of other preventive treatments for migraine, is modest.

“It would be so nice if we could find things that were more effective than the treatments we currently have,” said Dr. Loder.

A version of this article first appeared on Medscape.com.

The Food and Drug Administration has approved atogepant (Qulipta), a novel calcitonin gene-related peptide (CGRP) receptor antagonist, for the prevention of episodic migraine, the manufacturer announced in a release.

The once-daily medication will be available in doses of 10 mg, 30 mg, and 60 mg.

“Qulipta provides a simple oral treatment option specifically developed to prevent migraine attacks and target CGRP, which is believed to be crucially involved in migraine in many patients,” coinvestigator Peter J. Goadsby, MD, PhD, DSc, neurologist and professor at the University of California, Los Angeles, and King’s College London, said in the release.

Approval was based partly on the findings from the phase 3 ADVANCE trial, in which patients with episodic migraine were randomly assigned to receive placebo or a 10-mg, 30-mg, or 60-mg daily dose of atogepant for 12 weeks.

As reported by this news organization, all three doses of atogepant reduced the number of mean monthly migraine days.

With this approval, neurologists will be able to choose from four monoclonal antibodies and two gepants for the preventive treatment of migraine.

“Having another gepant that can also be given preventively is a good idea, because one may be better than the other for a patient,” Alan M. Rapoport, MD, past president of the International Headache Society and founder and director emeritus of the New England Center for Headache, Stamford, Conn., told this news organization.

“Once we have a year or so of experience with atogepant, we’ll have a pretty good idea of which one works better preventively,” said Dr. Rapoport, who was not involved with the research.
 

Practice changing?

In the ADVANCE trial, there was a reduction of 3.69 migraine days with the 10-mg dose, 3.86 days with the 30-mg dose, and 4.2 days with the 60-mg dose. Placebo was associated with a reduction of 2.48 migraine days.

In addition, more than half of patients in each atogepant arm achieved a reduction in mean monthly migraine days of 50% or greater. This outcome occurred in 55.6% of the 10-mg atogepant group, 58.7% of the 30-mg group, and 60.8% of the 60-mg group. Approximately 29% patients who received placebo achieved this outcome.

The data indicated that atogepant has a favorable safety profile. The most common adverse events associated with treatment were constipation, nausea, and upper respiratory tract infection.

Dr. Rapoport, who is also a clinical professor of neurology at UCLA, noted that he was impressed with the efficacy.

“I’m not as impressed with the adverse events, but they’re not serious, and they don’t necessarily last,” he said.

Although being able to prescribe a single drug for acute and preventive treatment may be an advantage, it remains to be seen whether the tolerability and price of atogepant will be barriers for patients, Dr. Rapoport added.

How the approval will affect clinical practice is also unclear, he noted.

“If you’re going to start someone on a preventive, especially if it’s a woman of childbearing potential, you might just consider one of the two gepants. Doctors will decide once they see how they work,” said Dr. Rapoport.
 

 

 

Not a ‘breakthrough’ treatment

Also commenting ahead of the approval, Elizabeth W. Loder, MD, vice chair for academic affairs in the department of neurology at Brigham and Women’s Hospital, Boston, noted that the “safety of these CGRP medications in pregnancy is uncertain, and there are theoretical reasons to be concerned about it.”

Unlike injectable CGRP medications, atogepant is eliminated from the body relatively quickly after the patient stops taking it, said Dr. Loder, who is also professor of neurology at Harvard Medical School, Boston. However, atogepant may not otherwise differ greatly from other medications of its type.

“I don’t see a reason to think that one of these oral CGRP medicines is much more effective than another one,” said Dr. Loder.

“In my mind, as a clinician who will be prescribing these for patients, it will be cost and the ease of getting it covered that makes the difference,” she added.

These questions may raise concerns. “Those of us who treat patients who do not have private insurance find it very difficult to get these medications for them, even in situations where they have exhausted other alternatives,” said Dr. Loder.

Patients insured by Medicare or Medicaid “usually have no avenue to get some of these new, expensive treatments,” she said.

The approval of atogepant for acute and preventive treatment shows that the distinction between these indications may be artificial, Dr. Loder noted. The approval “will, I hope, help people think more flexibly about the way in which we use medications.”

It is a positive that atogepant has emerged as another option for preventive therapy, but the treatment cannot be considered a breakthrough, Dr. Loder added. The efficacy of atogepant, like that of other preventive treatments for migraine, is modest.

“It would be so nice if we could find things that were more effective than the treatments we currently have,” said Dr. Loder.

A version of this article first appeared on Medscape.com.

Issue
Neurology Reviews - 29(11)
Issue
Neurology Reviews - 29(11)
Publications
Publications
Topics
Article Type
Sections
Citation Override
Publish date: September 29, 2021
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Apple devices identify early Parkinson’s disease

Article Type
Changed

Apple Watches and iPhones can differentiate between individuals with early, untreated Parkinson’s disease and healthy controls, new research shows. Results from the WATCH-PD study show clear differences in a finger-tapping task in the Parkinson’s disease versus control group. The finger-tapping task also correlated with “traditional measures,” such as the Movement Disorder Society–Unified Parkinson’s Disease Rating Scale (MDS-UPDRS), investigators reported.

“And then the smartphone and smartwatch also showed differences in gait between groups,” said lead investigator Jamie Adams, MD, University of Rochester, New York.

The findings were presented at the International Congress of Parkinson’s Disease and Movement Disorders.
 

WATCH-PD

The 12-month WATCH-PD study included 132 individuals at 17 Parkinson’s Study Group sites, 82 with Parkinson’s disease and 50 controls.

Participants with Parkinson’s disease were untreated, were no more than 2 years out from diagnosis (mean disease duration, 10.0 ±7.3 months), and were in Hoehn and Yahr stage 1 or 2.

Apple Watches and iPhones were provided to participants, all of whom underwent in-clinic assessments at baseline and at months 1, 3, 6, 9, and 12. The assessments included motor and cognitive tasks using the devices, which contained motion sensors.

The phone also contained an app that could assess verbal, cognitive, and other abilities. Participants wore a set of inertial sensors (APDM Mobility Lab) while performing the MDS-UPDRS Part III motor examination.

In addition, there were biweekly at-home tasks. Questions and tests on the watch assessed symptoms of mood, fatigue, cognition, and falls as well as cognitive performance involving perceptual, verbal, visual spatial, and fine motor abilities. Both the watch and iPhone were used to gauge gait, balance, and tremor.

Ages of the participants were approximately the same in the Parkinson’s disease and control groups (63.3 years vs. 60.2 years, respectively), but male to female ratios differed between the groups. There were more men in the Parkinson’s disease cohort (56% men vs. 44% women) and more women in the control cohort (36% vs. 64%; P =.03).
 

Between-group differences

Results showed that MDS-UPDRS total scores and on all individual parts of the rating scale were significantly better for the control group (lower scores are better), as shown in the following table.



Similarly, the control group performed better than the Parkinson’s disease group on the Montreal Cognitive Assessment (MoCA), with higher scores showing better performance on the 0 to 30 scale (28.1 vs. 27.6, respectively).

Touchscreen assessments on the phone also showed group differences in a finger-tapping task, with more taps by the control group than by the Parkinson’s disease group. The difference was more pronounced when the dominant hand was used.

The median numbers of taps in 20 seconds for the dominant hand were 103.7 for the Parkinson’s disease cohort versus 131.9 for control cohort (P < .005); and for the nondominant hand the numbers of taps were 106.6 versus 122.1 (P < .05), respectively. The control group also scored better on tests of hand fine-motor control (P < .01) and on the mobile digit symbols modalities test (P < .05)

Measures of gait in a 1-minute walk test also showed group differences.

“The five gait measures that differed most were cadence, which is steps per minute, double support, arm swing amplitude, arm swing variation, and turn duration,” Dr. Adams said.


 

‘Tremendous interest’

Commenting on the findings, Ludy Shih, MD, MMSc, of Boston University, noted that in the future, devices such as the ones used in this study may help clinicians remotely monitor their patients’ Parkinson’s disease conditions and response to therapy.

That would “eliminate some of the transportation barrier for people with Parkinson’s disease,” said Dr. Shih, who was not involved with the research.

The devices can give objective measurements, reducing inter-rater variability in assessment of movements, she noted.

“I think there’s tremendous interest in using digital measures to pick up on subtle disease phenotypes earlier than a clinical diagnosis can be made,” Dr. Shih said.

She also referred to literature “going back a few decades” showing that finger tapping can be used as a pharmacodynamic measure of how well a patient’s dopaminergic medications are working, so the devices may be a way to remotely assess treatment efficacy and decide when it is time to make adjustments.

Dr. Shih said she thinks regulatory agencies are now open “to consider these as part of the totality of evidence that a therapeutic [device] might be working.”

Whether these would need to be professional grade and approved as medical devices or if patients could just buy smartwatches and smartphones to generate useful data is still a question, she said. Already, there are several Parkinson’s apps that the public can download to track symptoms, improve voice, provide exercises, find support groups or research studies, and more.

Dr. Shih predicted that the biweekly at-home tasks, as in the current protocol, could be a burden to some people. If only a segment of the population were willing to comply, it could call into question how generalizable the results were, she added.

“There’s even a prior publication showing that compliance rate really dropped like a rock,” she noted. However, for those people willing to perform the tasks on a regular schedule, the results could be valuable, Dr. Shih said.

Dr. Adams concurred, saying that she had received feedback from some of her study participants that the biweekly tasks were a bit much.

The study was supported by Biogen and Takeda Pharmaceuticals. Dr. Adams receives research support from Biogen. Dr. Shih has disclosed no relevant financial relationships.

A version of this article first appeared on Medscape.com.

Meeting/Event
Publications
Topics
Sections
Meeting/Event
Meeting/Event

Apple Watches and iPhones can differentiate between individuals with early, untreated Parkinson’s disease and healthy controls, new research shows. Results from the WATCH-PD study show clear differences in a finger-tapping task in the Parkinson’s disease versus control group. The finger-tapping task also correlated with “traditional measures,” such as the Movement Disorder Society–Unified Parkinson’s Disease Rating Scale (MDS-UPDRS), investigators reported.

“And then the smartphone and smartwatch also showed differences in gait between groups,” said lead investigator Jamie Adams, MD, University of Rochester, New York.

The findings were presented at the International Congress of Parkinson’s Disease and Movement Disorders.
 

WATCH-PD

The 12-month WATCH-PD study included 132 individuals at 17 Parkinson’s Study Group sites, 82 with Parkinson’s disease and 50 controls.

Participants with Parkinson’s disease were untreated, were no more than 2 years out from diagnosis (mean disease duration, 10.0 ±7.3 months), and were in Hoehn and Yahr stage 1 or 2.

Apple Watches and iPhones were provided to participants, all of whom underwent in-clinic assessments at baseline and at months 1, 3, 6, 9, and 12. The assessments included motor and cognitive tasks using the devices, which contained motion sensors.

The phone also contained an app that could assess verbal, cognitive, and other abilities. Participants wore a set of inertial sensors (APDM Mobility Lab) while performing the MDS-UPDRS Part III motor examination.

In addition, there were biweekly at-home tasks. Questions and tests on the watch assessed symptoms of mood, fatigue, cognition, and falls as well as cognitive performance involving perceptual, verbal, visual spatial, and fine motor abilities. Both the watch and iPhone were used to gauge gait, balance, and tremor.

Ages of the participants were approximately the same in the Parkinson’s disease and control groups (63.3 years vs. 60.2 years, respectively), but male to female ratios differed between the groups. There were more men in the Parkinson’s disease cohort (56% men vs. 44% women) and more women in the control cohort (36% vs. 64%; P =.03).
 

Between-group differences

Results showed that MDS-UPDRS total scores and on all individual parts of the rating scale were significantly better for the control group (lower scores are better), as shown in the following table.



Similarly, the control group performed better than the Parkinson’s disease group on the Montreal Cognitive Assessment (MoCA), with higher scores showing better performance on the 0 to 30 scale (28.1 vs. 27.6, respectively).

Touchscreen assessments on the phone also showed group differences in a finger-tapping task, with more taps by the control group than by the Parkinson’s disease group. The difference was more pronounced when the dominant hand was used.

The median numbers of taps in 20 seconds for the dominant hand were 103.7 for the Parkinson’s disease cohort versus 131.9 for control cohort (P < .005); and for the nondominant hand the numbers of taps were 106.6 versus 122.1 (P < .05), respectively. The control group also scored better on tests of hand fine-motor control (P < .01) and on the mobile digit symbols modalities test (P < .05)

Measures of gait in a 1-minute walk test also showed group differences.

“The five gait measures that differed most were cadence, which is steps per minute, double support, arm swing amplitude, arm swing variation, and turn duration,” Dr. Adams said.


 

‘Tremendous interest’

Commenting on the findings, Ludy Shih, MD, MMSc, of Boston University, noted that in the future, devices such as the ones used in this study may help clinicians remotely monitor their patients’ Parkinson’s disease conditions and response to therapy.

That would “eliminate some of the transportation barrier for people with Parkinson’s disease,” said Dr. Shih, who was not involved with the research.

The devices can give objective measurements, reducing inter-rater variability in assessment of movements, she noted.

“I think there’s tremendous interest in using digital measures to pick up on subtle disease phenotypes earlier than a clinical diagnosis can be made,” Dr. Shih said.

She also referred to literature “going back a few decades” showing that finger tapping can be used as a pharmacodynamic measure of how well a patient’s dopaminergic medications are working, so the devices may be a way to remotely assess treatment efficacy and decide when it is time to make adjustments.

Dr. Shih said she thinks regulatory agencies are now open “to consider these as part of the totality of evidence that a therapeutic [device] might be working.”

Whether these would need to be professional grade and approved as medical devices or if patients could just buy smartwatches and smartphones to generate useful data is still a question, she said. Already, there are several Parkinson’s apps that the public can download to track symptoms, improve voice, provide exercises, find support groups or research studies, and more.

Dr. Shih predicted that the biweekly at-home tasks, as in the current protocol, could be a burden to some people. If only a segment of the population were willing to comply, it could call into question how generalizable the results were, she added.

“There’s even a prior publication showing that compliance rate really dropped like a rock,” she noted. However, for those people willing to perform the tasks on a regular schedule, the results could be valuable, Dr. Shih said.

Dr. Adams concurred, saying that she had received feedback from some of her study participants that the biweekly tasks were a bit much.

The study was supported by Biogen and Takeda Pharmaceuticals. Dr. Adams receives research support from Biogen. Dr. Shih has disclosed no relevant financial relationships.

A version of this article first appeared on Medscape.com.

Apple Watches and iPhones can differentiate between individuals with early, untreated Parkinson’s disease and healthy controls, new research shows. Results from the WATCH-PD study show clear differences in a finger-tapping task in the Parkinson’s disease versus control group. The finger-tapping task also correlated with “traditional measures,” such as the Movement Disorder Society–Unified Parkinson’s Disease Rating Scale (MDS-UPDRS), investigators reported.

“And then the smartphone and smartwatch also showed differences in gait between groups,” said lead investigator Jamie Adams, MD, University of Rochester, New York.

The findings were presented at the International Congress of Parkinson’s Disease and Movement Disorders.
 

WATCH-PD

The 12-month WATCH-PD study included 132 individuals at 17 Parkinson’s Study Group sites, 82 with Parkinson’s disease and 50 controls.

Participants with Parkinson’s disease were untreated, were no more than 2 years out from diagnosis (mean disease duration, 10.0 ±7.3 months), and were in Hoehn and Yahr stage 1 or 2.

Apple Watches and iPhones were provided to participants, all of whom underwent in-clinic assessments at baseline and at months 1, 3, 6, 9, and 12. The assessments included motor and cognitive tasks using the devices, which contained motion sensors.

The phone also contained an app that could assess verbal, cognitive, and other abilities. Participants wore a set of inertial sensors (APDM Mobility Lab) while performing the MDS-UPDRS Part III motor examination.

In addition, there were biweekly at-home tasks. Questions and tests on the watch assessed symptoms of mood, fatigue, cognition, and falls as well as cognitive performance involving perceptual, verbal, visual spatial, and fine motor abilities. Both the watch and iPhone were used to gauge gait, balance, and tremor.

Ages of the participants were approximately the same in the Parkinson’s disease and control groups (63.3 years vs. 60.2 years, respectively), but male to female ratios differed between the groups. There were more men in the Parkinson’s disease cohort (56% men vs. 44% women) and more women in the control cohort (36% vs. 64%; P =.03).
 

Between-group differences

Results showed that MDS-UPDRS total scores and on all individual parts of the rating scale were significantly better for the control group (lower scores are better), as shown in the following table.



Similarly, the control group performed better than the Parkinson’s disease group on the Montreal Cognitive Assessment (MoCA), with higher scores showing better performance on the 0 to 30 scale (28.1 vs. 27.6, respectively).

Touchscreen assessments on the phone also showed group differences in a finger-tapping task, with more taps by the control group than by the Parkinson’s disease group. The difference was more pronounced when the dominant hand was used.

The median numbers of taps in 20 seconds for the dominant hand were 103.7 for the Parkinson’s disease cohort versus 131.9 for control cohort (P < .005); and for the nondominant hand the numbers of taps were 106.6 versus 122.1 (P < .05), respectively. The control group also scored better on tests of hand fine-motor control (P < .01) and on the mobile digit symbols modalities test (P < .05)

Measures of gait in a 1-minute walk test also showed group differences.

“The five gait measures that differed most were cadence, which is steps per minute, double support, arm swing amplitude, arm swing variation, and turn duration,” Dr. Adams said.


 

‘Tremendous interest’

Commenting on the findings, Ludy Shih, MD, MMSc, of Boston University, noted that in the future, devices such as the ones used in this study may help clinicians remotely monitor their patients’ Parkinson’s disease conditions and response to therapy.

That would “eliminate some of the transportation barrier for people with Parkinson’s disease,” said Dr. Shih, who was not involved with the research.

The devices can give objective measurements, reducing inter-rater variability in assessment of movements, she noted.

“I think there’s tremendous interest in using digital measures to pick up on subtle disease phenotypes earlier than a clinical diagnosis can be made,” Dr. Shih said.

She also referred to literature “going back a few decades” showing that finger tapping can be used as a pharmacodynamic measure of how well a patient’s dopaminergic medications are working, so the devices may be a way to remotely assess treatment efficacy and decide when it is time to make adjustments.

Dr. Shih said she thinks regulatory agencies are now open “to consider these as part of the totality of evidence that a therapeutic [device] might be working.”

Whether these would need to be professional grade and approved as medical devices or if patients could just buy smartwatches and smartphones to generate useful data is still a question, she said. Already, there are several Parkinson’s apps that the public can download to track symptoms, improve voice, provide exercises, find support groups or research studies, and more.

Dr. Shih predicted that the biweekly at-home tasks, as in the current protocol, could be a burden to some people. If only a segment of the population were willing to comply, it could call into question how generalizable the results were, she added.

“There’s even a prior publication showing that compliance rate really dropped like a rock,” she noted. However, for those people willing to perform the tasks on a regular schedule, the results could be valuable, Dr. Shih said.

Dr. Adams concurred, saying that she had received feedback from some of her study participants that the biweekly tasks were a bit much.

The study was supported by Biogen and Takeda Pharmaceuticals. Dr. Adams receives research support from Biogen. Dr. Shih has disclosed no relevant financial relationships.

A version of this article first appeared on Medscape.com.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM MDS VIRTUAL CONGRESS 2021

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article

Differences in Care by Race in Older Nursing Home Residents With Dementia

Article Type
Changed
Display Headline
Differences in Care by Race in Older Nursing Home Residents With Dementia

Study Overview

Objective. To examine differences in care, specifically hospitalization towards the end of life, among nursing home residents with dementia who were Black compared with those who were White.

Design. Population based cohort study in the US. The study included all decedents with Alzheimer’s disease or related dementia (ADRD) who resided in a nursing home from 2014 to 2017. Decedents from nursing homes were identified by death within 1 day of an identified nursing home stay or within 8 days of a hospital transfer from nursing home. Data were obtained from Minimum Data Set 3.0 (MDS) which contains clinical data from all Medicaid or Medicare certified nursing homes, and from the Medicare Beneficiary Summary File (MBSF) and Medicare Provider and Analysis and Review (MedPAR) which contains hospitalization events for all Medicare Beneficiaries. These files were linked to identify nursing home residents with ADRD who were hospitalized at the end of life. ADRD diagnosis was identified from the chronic condition list from the MBSF and from MDS diagnosis list.

Setting and participants. The study included 665 033 residents from 14 595 nursing homes who died during the study period. Resident race was categorized as White or Black based on the MBSF. Severe cognitive impairment was identified using the MDS that categorized residents as severe or not using the Brief Interview for Mental Status and the Cognitive Performance Scale. The mean (SD) age of the study population was 86.7 (9.2) years for White residents and 82.6 (11.1) years for Black residents. Of the participants, 68.8% and 61.2% were female for Black and White residents, respectively. Approximately 23.4% of White and 32.5% of Black residents had severe cognitive impairment. For nursing home characteristics, 71.5% of the 14 595 nursing homes represented were for profit; average bedside was 109.5 (57.0) and occupancy rate was on average 81.2% (14.3%).

Main outcome measures. The study outcome measure was any hospitalization within 30 days prior to death. The outcome was selected as an indicator of quality of care because as older adults living with ADRD experience progressive worsening of cognitive symptoms, at the end of life when dementia is severe, advance care planning and communication with health care proxies and surrogates often result in coordinated care that avoids acute hospitalizations, which are often burdensome to both patient and family and may yield poorer quality of life.

Main results. The study found that approximately 29.5% of White decedents and 40.7% of Black decedents were hospitalized towards the end of life. Nursing homes with a higher proportion of Black residents were more likely to have residents hospitalized towards the end of life with 35% of residents hospitalized in the highest quartile (27% Black) compared with 17% hospitalized for nursing homes in the lowest quartile (0% Black).After adjusting for covariates, Black residents were 7.9% more likely to be hospitalized in the last 30 days of life compared with White residents. Blacks with severe cognitive impairment has elevated risk of hospitalization by 4.9% when compared with White residents. After accounting for nursing home facility–level characteristics, nursing homes with a low proportion of Black residents had a 5.2% higher risk of hospitalizations compared with nursing homes with no Black residents, and nursing homes with a higher percentage of Black residents had a 13.3% higher risk of hospitalization compared with nursing homes with no Black residents.

Conclusion. Race is associated with care disparities in older nursing home residents with dementia. This study suggests that hospitalization towards the end of life as a quality of care marker differs across nursing homes, and nursing homes with a higher proportion of Black residents were more likely to be hospitalized. This suggests that these nursing homes may have fewer resources and delivered poorer quality of care, and that disparities in health systems or institutions contribute to differences in quality of care for this vulnerable group.

Commentary

Disparities of health status, health care, and affordability across race and ethnicity have persisted throughout the past 20 years.1 There is further evidence to support systemic differences that can contribute to differences in health outcomes.2 Although changes in health care policy such as the Affordable Care Act have expanded health care coverage, and instituted changes that aims to improve health care quality and reduce disparities, it is clear that factors contributing to disparities in care are structural and perhaps systemic. The latest evidence comes in this study that examines racial disparities in health care quality in one of the most vulnerable populations—older adults with Alzheimer’s disease and dementia. The finding that Black nursing home residents, when compared with White residents, often has higher risk of hospitalization at the end of life, even among those with severe dementia where better coordinated care, clear goals of care and perhaps instituting palliative care would result in lower rate of hospitalization. The disparities were observed across nursing homes as well, where nursing homes with higher proportion of Black residents appear to have lower quality of care.

 

 

These findings are consistent with prior work that has examined differences in Black and White population on uptake of palliative care, discussion, and the documentation of advance care planning.3 Factors that may contribute to these differences include mistrust of the health care system among minorities, and not being connected to adequate health care resources. Family members and surrogate health care decision makers may consider receiving more aggressive care as advocating for better health care for their family members.4 These differences may contribute to the differences in hospitalization rates among residents within the same nursing home; however, the differences between nursing homes even after accounting for individual differences may indicate more widespread systemic differences that is associated with race. Policy changes that will address these differences are needed to level these differences so that quality care can be delivered regardless of race.5 For this vulnerable population with a terminal illness, approaches to enhance uptake of palliative approaches and care delivery for dementia patients at terminal stage are needed and understanding and targeting factors that contribute to low uptake of these approaches will enhance end of life care. Understanding the differences in resources and systems of care in nursing homes and perhaps how palliative care is integrated in these settings will be important to address care disparities that occurs across nursing homes.

Applications for Clinical Practice

Clinicians who take care of this population of older adults with advanced dementia should be aware of the potential for racial disparities that may lead to differences in the quality of care. The underlying reasons for these differences could be targeted so that older adults in all racial groups may have equal access to quality care including palliative approaches that avoid aggressive care for terminal illnesses across settings that may yield better care and quality of life. Policy makers and health systems leaders need to consider the current realities with racial disparities that policies need to address these differences so that they may not continue to persist in our systems of care.

Financial disclosures: None.

References

1. Mahajan S, Caraballo C, Lu Y, et al. Trends in Differences in Health Status and Health Care Access and Affordability by Race and Ethnicity in the United States, 1999-2018. JAMA. 2021;326(7):637-648. doi:10.1001/jama.2021.9907

2. Gill TM, Zang EX, Murphy TE, et al. Association Between Neighborhood Disadvantage and Functional Well-being in Community-Living Older Persons. [published online ahead of print, 2021 Aug 23]. JAMA Intern Med. doi:10.1001/jamainternmed.2021.4260

3. Bazargan M, Bazargan-Hejazi S. Disparities in Palliative and Hospice Care and Completion of Advance Care Planning and Directives Among Non-Hispanic Blacks: A Scoping Review of Recent Literature. Am J Hosp Palliat Care. 2021;38(6):688-718. doi:10.1177/1049909120966585

4. Siler S, Arora K, Doyon K, Fischer SM. Spirituality and the Illness Experience: Perspectives of African American Older Adults. Am J Hosp Palliat Care. 2021;38(6):618-625. doi:10.1177/1049909120988280

5. Council on Ethical and Judicial Affairs. Black-white disparities in health care. JAMA. 1990;263(17):2344-2346. doi:10.1001/jama.1990.03440170066038

Article PDF
Issue
Journal of Clinical Outcomes Management - 28(5)
Publications
Topics
Page Number
199-201
Sections
Article PDF
Article PDF

Study Overview

Objective. To examine differences in care, specifically hospitalization towards the end of life, among nursing home residents with dementia who were Black compared with those who were White.

Design. Population based cohort study in the US. The study included all decedents with Alzheimer’s disease or related dementia (ADRD) who resided in a nursing home from 2014 to 2017. Decedents from nursing homes were identified by death within 1 day of an identified nursing home stay or within 8 days of a hospital transfer from nursing home. Data were obtained from Minimum Data Set 3.0 (MDS) which contains clinical data from all Medicaid or Medicare certified nursing homes, and from the Medicare Beneficiary Summary File (MBSF) and Medicare Provider and Analysis and Review (MedPAR) which contains hospitalization events for all Medicare Beneficiaries. These files were linked to identify nursing home residents with ADRD who were hospitalized at the end of life. ADRD diagnosis was identified from the chronic condition list from the MBSF and from MDS diagnosis list.

Setting and participants. The study included 665 033 residents from 14 595 nursing homes who died during the study period. Resident race was categorized as White or Black based on the MBSF. Severe cognitive impairment was identified using the MDS that categorized residents as severe or not using the Brief Interview for Mental Status and the Cognitive Performance Scale. The mean (SD) age of the study population was 86.7 (9.2) years for White residents and 82.6 (11.1) years for Black residents. Of the participants, 68.8% and 61.2% were female for Black and White residents, respectively. Approximately 23.4% of White and 32.5% of Black residents had severe cognitive impairment. For nursing home characteristics, 71.5% of the 14 595 nursing homes represented were for profit; average bedside was 109.5 (57.0) and occupancy rate was on average 81.2% (14.3%).

Main outcome measures. The study outcome measure was any hospitalization within 30 days prior to death. The outcome was selected as an indicator of quality of care because as older adults living with ADRD experience progressive worsening of cognitive symptoms, at the end of life when dementia is severe, advance care planning and communication with health care proxies and surrogates often result in coordinated care that avoids acute hospitalizations, which are often burdensome to both patient and family and may yield poorer quality of life.

Main results. The study found that approximately 29.5% of White decedents and 40.7% of Black decedents were hospitalized towards the end of life. Nursing homes with a higher proportion of Black residents were more likely to have residents hospitalized towards the end of life with 35% of residents hospitalized in the highest quartile (27% Black) compared with 17% hospitalized for nursing homes in the lowest quartile (0% Black).After adjusting for covariates, Black residents were 7.9% more likely to be hospitalized in the last 30 days of life compared with White residents. Blacks with severe cognitive impairment has elevated risk of hospitalization by 4.9% when compared with White residents. After accounting for nursing home facility–level characteristics, nursing homes with a low proportion of Black residents had a 5.2% higher risk of hospitalizations compared with nursing homes with no Black residents, and nursing homes with a higher percentage of Black residents had a 13.3% higher risk of hospitalization compared with nursing homes with no Black residents.

Conclusion. Race is associated with care disparities in older nursing home residents with dementia. This study suggests that hospitalization towards the end of life as a quality of care marker differs across nursing homes, and nursing homes with a higher proportion of Black residents were more likely to be hospitalized. This suggests that these nursing homes may have fewer resources and delivered poorer quality of care, and that disparities in health systems or institutions contribute to differences in quality of care for this vulnerable group.

Commentary

Disparities of health status, health care, and affordability across race and ethnicity have persisted throughout the past 20 years.1 There is further evidence to support systemic differences that can contribute to differences in health outcomes.2 Although changes in health care policy such as the Affordable Care Act have expanded health care coverage, and instituted changes that aims to improve health care quality and reduce disparities, it is clear that factors contributing to disparities in care are structural and perhaps systemic. The latest evidence comes in this study that examines racial disparities in health care quality in one of the most vulnerable populations—older adults with Alzheimer’s disease and dementia. The finding that Black nursing home residents, when compared with White residents, often has higher risk of hospitalization at the end of life, even among those with severe dementia where better coordinated care, clear goals of care and perhaps instituting palliative care would result in lower rate of hospitalization. The disparities were observed across nursing homes as well, where nursing homes with higher proportion of Black residents appear to have lower quality of care.

 

 

These findings are consistent with prior work that has examined differences in Black and White population on uptake of palliative care, discussion, and the documentation of advance care planning.3 Factors that may contribute to these differences include mistrust of the health care system among minorities, and not being connected to adequate health care resources. Family members and surrogate health care decision makers may consider receiving more aggressive care as advocating for better health care for their family members.4 These differences may contribute to the differences in hospitalization rates among residents within the same nursing home; however, the differences between nursing homes even after accounting for individual differences may indicate more widespread systemic differences that is associated with race. Policy changes that will address these differences are needed to level these differences so that quality care can be delivered regardless of race.5 For this vulnerable population with a terminal illness, approaches to enhance uptake of palliative approaches and care delivery for dementia patients at terminal stage are needed and understanding and targeting factors that contribute to low uptake of these approaches will enhance end of life care. Understanding the differences in resources and systems of care in nursing homes and perhaps how palliative care is integrated in these settings will be important to address care disparities that occurs across nursing homes.

Applications for Clinical Practice

Clinicians who take care of this population of older adults with advanced dementia should be aware of the potential for racial disparities that may lead to differences in the quality of care. The underlying reasons for these differences could be targeted so that older adults in all racial groups may have equal access to quality care including palliative approaches that avoid aggressive care for terminal illnesses across settings that may yield better care and quality of life. Policy makers and health systems leaders need to consider the current realities with racial disparities that policies need to address these differences so that they may not continue to persist in our systems of care.

Financial disclosures: None.

Study Overview

Objective. To examine differences in care, specifically hospitalization towards the end of life, among nursing home residents with dementia who were Black compared with those who were White.

Design. Population based cohort study in the US. The study included all decedents with Alzheimer’s disease or related dementia (ADRD) who resided in a nursing home from 2014 to 2017. Decedents from nursing homes were identified by death within 1 day of an identified nursing home stay or within 8 days of a hospital transfer from nursing home. Data were obtained from Minimum Data Set 3.0 (MDS) which contains clinical data from all Medicaid or Medicare certified nursing homes, and from the Medicare Beneficiary Summary File (MBSF) and Medicare Provider and Analysis and Review (MedPAR) which contains hospitalization events for all Medicare Beneficiaries. These files were linked to identify nursing home residents with ADRD who were hospitalized at the end of life. ADRD diagnosis was identified from the chronic condition list from the MBSF and from MDS diagnosis list.

Setting and participants. The study included 665 033 residents from 14 595 nursing homes who died during the study period. Resident race was categorized as White or Black based on the MBSF. Severe cognitive impairment was identified using the MDS that categorized residents as severe or not using the Brief Interview for Mental Status and the Cognitive Performance Scale. The mean (SD) age of the study population was 86.7 (9.2) years for White residents and 82.6 (11.1) years for Black residents. Of the participants, 68.8% and 61.2% were female for Black and White residents, respectively. Approximately 23.4% of White and 32.5% of Black residents had severe cognitive impairment. For nursing home characteristics, 71.5% of the 14 595 nursing homes represented were for profit; average bedside was 109.5 (57.0) and occupancy rate was on average 81.2% (14.3%).

Main outcome measures. The study outcome measure was any hospitalization within 30 days prior to death. The outcome was selected as an indicator of quality of care because as older adults living with ADRD experience progressive worsening of cognitive symptoms, at the end of life when dementia is severe, advance care planning and communication with health care proxies and surrogates often result in coordinated care that avoids acute hospitalizations, which are often burdensome to both patient and family and may yield poorer quality of life.

Main results. The study found that approximately 29.5% of White decedents and 40.7% of Black decedents were hospitalized towards the end of life. Nursing homes with a higher proportion of Black residents were more likely to have residents hospitalized towards the end of life with 35% of residents hospitalized in the highest quartile (27% Black) compared with 17% hospitalized for nursing homes in the lowest quartile (0% Black).After adjusting for covariates, Black residents were 7.9% more likely to be hospitalized in the last 30 days of life compared with White residents. Blacks with severe cognitive impairment has elevated risk of hospitalization by 4.9% when compared with White residents. After accounting for nursing home facility–level characteristics, nursing homes with a low proportion of Black residents had a 5.2% higher risk of hospitalizations compared with nursing homes with no Black residents, and nursing homes with a higher percentage of Black residents had a 13.3% higher risk of hospitalization compared with nursing homes with no Black residents.

Conclusion. Race is associated with care disparities in older nursing home residents with dementia. This study suggests that hospitalization towards the end of life as a quality of care marker differs across nursing homes, and nursing homes with a higher proportion of Black residents were more likely to be hospitalized. This suggests that these nursing homes may have fewer resources and delivered poorer quality of care, and that disparities in health systems or institutions contribute to differences in quality of care for this vulnerable group.

Commentary

Disparities of health status, health care, and affordability across race and ethnicity have persisted throughout the past 20 years.1 There is further evidence to support systemic differences that can contribute to differences in health outcomes.2 Although changes in health care policy such as the Affordable Care Act have expanded health care coverage, and instituted changes that aims to improve health care quality and reduce disparities, it is clear that factors contributing to disparities in care are structural and perhaps systemic. The latest evidence comes in this study that examines racial disparities in health care quality in one of the most vulnerable populations—older adults with Alzheimer’s disease and dementia. The finding that Black nursing home residents, when compared with White residents, often has higher risk of hospitalization at the end of life, even among those with severe dementia where better coordinated care, clear goals of care and perhaps instituting palliative care would result in lower rate of hospitalization. The disparities were observed across nursing homes as well, where nursing homes with higher proportion of Black residents appear to have lower quality of care.

 

 

These findings are consistent with prior work that has examined differences in Black and White population on uptake of palliative care, discussion, and the documentation of advance care planning.3 Factors that may contribute to these differences include mistrust of the health care system among minorities, and not being connected to adequate health care resources. Family members and surrogate health care decision makers may consider receiving more aggressive care as advocating for better health care for their family members.4 These differences may contribute to the differences in hospitalization rates among residents within the same nursing home; however, the differences between nursing homes even after accounting for individual differences may indicate more widespread systemic differences that is associated with race. Policy changes that will address these differences are needed to level these differences so that quality care can be delivered regardless of race.5 For this vulnerable population with a terminal illness, approaches to enhance uptake of palliative approaches and care delivery for dementia patients at terminal stage are needed and understanding and targeting factors that contribute to low uptake of these approaches will enhance end of life care. Understanding the differences in resources and systems of care in nursing homes and perhaps how palliative care is integrated in these settings will be important to address care disparities that occurs across nursing homes.

Applications for Clinical Practice

Clinicians who take care of this population of older adults with advanced dementia should be aware of the potential for racial disparities that may lead to differences in the quality of care. The underlying reasons for these differences could be targeted so that older adults in all racial groups may have equal access to quality care including palliative approaches that avoid aggressive care for terminal illnesses across settings that may yield better care and quality of life. Policy makers and health systems leaders need to consider the current realities with racial disparities that policies need to address these differences so that they may not continue to persist in our systems of care.

Financial disclosures: None.

References

1. Mahajan S, Caraballo C, Lu Y, et al. Trends in Differences in Health Status and Health Care Access and Affordability by Race and Ethnicity in the United States, 1999-2018. JAMA. 2021;326(7):637-648. doi:10.1001/jama.2021.9907

2. Gill TM, Zang EX, Murphy TE, et al. Association Between Neighborhood Disadvantage and Functional Well-being in Community-Living Older Persons. [published online ahead of print, 2021 Aug 23]. JAMA Intern Med. doi:10.1001/jamainternmed.2021.4260

3. Bazargan M, Bazargan-Hejazi S. Disparities in Palliative and Hospice Care and Completion of Advance Care Planning and Directives Among Non-Hispanic Blacks: A Scoping Review of Recent Literature. Am J Hosp Palliat Care. 2021;38(6):688-718. doi:10.1177/1049909120966585

4. Siler S, Arora K, Doyon K, Fischer SM. Spirituality and the Illness Experience: Perspectives of African American Older Adults. Am J Hosp Palliat Care. 2021;38(6):618-625. doi:10.1177/1049909120988280

5. Council on Ethical and Judicial Affairs. Black-white disparities in health care. JAMA. 1990;263(17):2344-2346. doi:10.1001/jama.1990.03440170066038

References

1. Mahajan S, Caraballo C, Lu Y, et al. Trends in Differences in Health Status and Health Care Access and Affordability by Race and Ethnicity in the United States, 1999-2018. JAMA. 2021;326(7):637-648. doi:10.1001/jama.2021.9907

2. Gill TM, Zang EX, Murphy TE, et al. Association Between Neighborhood Disadvantage and Functional Well-being in Community-Living Older Persons. [published online ahead of print, 2021 Aug 23]. JAMA Intern Med. doi:10.1001/jamainternmed.2021.4260

3. Bazargan M, Bazargan-Hejazi S. Disparities in Palliative and Hospice Care and Completion of Advance Care Planning and Directives Among Non-Hispanic Blacks: A Scoping Review of Recent Literature. Am J Hosp Palliat Care. 2021;38(6):688-718. doi:10.1177/1049909120966585

4. Siler S, Arora K, Doyon K, Fischer SM. Spirituality and the Illness Experience: Perspectives of African American Older Adults. Am J Hosp Palliat Care. 2021;38(6):618-625. doi:10.1177/1049909120988280

5. Council on Ethical and Judicial Affairs. Black-white disparities in health care. JAMA. 1990;263(17):2344-2346. doi:10.1001/jama.1990.03440170066038

Issue
Journal of Clinical Outcomes Management - 28(5)
Issue
Journal of Clinical Outcomes Management - 28(5)
Page Number
199-201
Page Number
199-201
Publications
Publications
Topics
Article Type
Display Headline
Differences in Care by Race in Older Nursing Home Residents With Dementia
Display Headline
Differences in Care by Race in Older Nursing Home Residents With Dementia
Sections
Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads
WebMD Article
Article PDF Media