Allowed Publications
Slot System
Featured Buckets
Featured Buckets Admin

Lasting norovirus immunity may depend on T cells

Norovirus-specific cell immunity is durable
Article Type
Changed
Fri, 04/09/2021 - 10:18

 

Protection against norovirus gastroenteritis is supported in part by norovirus-specific CD8+ T cells that reside in peripheral, intestinal, and lymphoid tissues, according to investigators.

These findings, and the molecular tools used to discover them, could guide development of a norovirus vaccine and novel cellular therapies, according to lead author Ajinkya Pattekar, MD, of the University of Pennsylvania, Philadelphia, and colleagues.

“Currently, there are no approved pharmacologic therapies against norovirus, and despite several promising clinical trials, an effective vaccine is not available,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology, which may stem from an incomplete understanding of norovirus immunity, according to Dr. Pattekar and colleagues.

They noted that most previous research has focused on humoral immunity, which appears variable between individuals, with some people exhibiting a strong humoral response, while others mount only partial humoral protection. The investigators also noted that, depending on which studies were examined, this type of defense could last years or fade within weeks to months and that “immune mechanisms other than antibodies may be important for protection against noroviruses.”

Specifically, cellular immunity may be at work. A 2020 study involving volunteers showed that T cells were cross-reactive to a type of norovirus the participants had never been exposed to.

“These findings suggest that T cells may target conserved epitopes and could offer cross-protection against a broad range of noroviruses,” Dr. Pattekar and colleagues wrote.

To test this hypothesis, they first collected peripheral blood mononuclear cells (PBMCs) from three healthy volunteers with unknown norovirus exposure history. Then serum samples were screened for norovirus functional antibodies via the binding between virus-like particles (VLPs) and histo–blood group antigens (HBGAs). This revealed disparate profiles of blocking antibodies against various norovirus strains. While donor 1 and donor 2 had antibodies against multiple strains, donor 3 lacked norovirus antibodies. Further testing showed that this latter individual was a nonsecretor with limited exposure history.

Next, the investigators tested donor PBMCs for norovirus-specific T-cell responses with use of overlapping libraries of peptides for each of the three norovirus open reading frames (ORF1, ORF2, and ORF3). T-cell responses, predominantly involving CD8+ T cells, were observed in all donors. While donor 1 had the greatest response to ORF1, donors 2 and 3 had responses that focused on ORF2.

“Thus, norovirus-specific T cells targeting ORF1 and ORF2 epitopes are present in peripheral blood from healthy donors regardless of secretor status,” the investigators wrote.

To better characterize T-cell epitopes, the investigators subdivided the overlapping peptide libraries into groups of shorter peptides, then exposed serum to these smaller component pools. This revealed eight HLA class I restricted epitopes that were derived from a genogroup II.4 pandemic norovirus strain; this group of variants has been responsible for all six of the norovirus pandemics since 1996.

Closer examination of the epitopes showed that they were “broadly conserved beyond GII.4.” Only one epitope exhibited variation in the C-terminal aromatic anchor, and it was nondominant. The investigators therefore identified seven immunodominant CD8+ epitopes, which they considered “valuable targets for vaccine and cell-based therapies.

“These data further confirm that epitope-specific CD8+ T cells are a universal feature of the overall norovirus immune response and could be an attractive target for future vaccines,” the investigators wrote.

Additional testing involving samples of spleen, mesenteric lymph nodes, and duodenum from deceased individuals showed presence of norovirus-specific CD8+ T cells, with particular abundance in intestinal tissue, and distinct phenotypes and functional properties in different tissue types.

“Future studies using tetramers and intestinal samples should build on these observations and fully define the location and microenvironment of norovirus-specific T cells,” the investigators wrote. “If carried out in the context of a vaccine trial, such studies could be highly valuable in elucidating tissue-resident memory correlates of norovirus immunity.”

The study was funded by the National Institutes of Health, the Wellcome Trust, and Deutsche Forschungsgemeinschaft. The investigators reported no conflicts of interest.

Body

 

Understanding the immune correlates of protection for norovirus is important for the development and evaluation of candidate vaccines and to better clarify the variation in host susceptibility to infection.

Dr. Craig B. Wilen

Prior research on the human immune response to norovirus infection has largely focused on the antibody response. There is less known about the antinorovirus T cell response, which can target and clear virus-infected cells. Notably, anti-viral CD8+ T cells are critical for control of norovirus infection in mouse models, which suggests a similarly important role in humans. In this study by Dr. Pattekar and colleagues, the authors generated human norovirus-specific peptides covering the entire viral proteome, and then they used these peptides to identify and characterize norovirus-specific CD8+ T cells from the blood, spleen, lymph nodes, and intestinal lamina propria of human donors who were not actively infected by norovirus. The authors identified virus-specific memory T cells in the blood and intestines. Further, they found several HLA class I restricted virus epitopes that are highly conserved amongst the most commonly circulating GII.4 noroviruses. These norovirus-specific T cells represented about 0.5% of all cells and reveal that norovirus induces a durable population of memory T cells.

Further research is needed to determine whether norovirus-specific CD8+ T cells are necessary or sufficient for preventing norovirus infection and disease in people. This important study provides novel tools and increases our understanding of cell-mediated immunity to human norovirus infection that will influence future vaccine design and evaluation for this important human pathogen.

Craig B. Wilen, MD, PhD, is assistant professor of laboratory medicine and immunobiology at Yale University, New Haven, Conn. He does not have any conflicts to disclose.

Publications
Topics
Sections
Body

 

Understanding the immune correlates of protection for norovirus is important for the development and evaluation of candidate vaccines and to better clarify the variation in host susceptibility to infection.

Dr. Craig B. Wilen

Prior research on the human immune response to norovirus infection has largely focused on the antibody response. There is less known about the antinorovirus T cell response, which can target and clear virus-infected cells. Notably, anti-viral CD8+ T cells are critical for control of norovirus infection in mouse models, which suggests a similarly important role in humans. In this study by Dr. Pattekar and colleagues, the authors generated human norovirus-specific peptides covering the entire viral proteome, and then they used these peptides to identify and characterize norovirus-specific CD8+ T cells from the blood, spleen, lymph nodes, and intestinal lamina propria of human donors who were not actively infected by norovirus. The authors identified virus-specific memory T cells in the blood and intestines. Further, they found several HLA class I restricted virus epitopes that are highly conserved amongst the most commonly circulating GII.4 noroviruses. These norovirus-specific T cells represented about 0.5% of all cells and reveal that norovirus induces a durable population of memory T cells.

Further research is needed to determine whether norovirus-specific CD8+ T cells are necessary or sufficient for preventing norovirus infection and disease in people. This important study provides novel tools and increases our understanding of cell-mediated immunity to human norovirus infection that will influence future vaccine design and evaluation for this important human pathogen.

Craig B. Wilen, MD, PhD, is assistant professor of laboratory medicine and immunobiology at Yale University, New Haven, Conn. He does not have any conflicts to disclose.

Body

 

Understanding the immune correlates of protection for norovirus is important for the development and evaluation of candidate vaccines and to better clarify the variation in host susceptibility to infection.

Dr. Craig B. Wilen

Prior research on the human immune response to norovirus infection has largely focused on the antibody response. There is less known about the antinorovirus T cell response, which can target and clear virus-infected cells. Notably, anti-viral CD8+ T cells are critical for control of norovirus infection in mouse models, which suggests a similarly important role in humans. In this study by Dr. Pattekar and colleagues, the authors generated human norovirus-specific peptides covering the entire viral proteome, and then they used these peptides to identify and characterize norovirus-specific CD8+ T cells from the blood, spleen, lymph nodes, and intestinal lamina propria of human donors who were not actively infected by norovirus. The authors identified virus-specific memory T cells in the blood and intestines. Further, they found several HLA class I restricted virus epitopes that are highly conserved amongst the most commonly circulating GII.4 noroviruses. These norovirus-specific T cells represented about 0.5% of all cells and reveal that norovirus induces a durable population of memory T cells.

Further research is needed to determine whether norovirus-specific CD8+ T cells are necessary or sufficient for preventing norovirus infection and disease in people. This important study provides novel tools and increases our understanding of cell-mediated immunity to human norovirus infection that will influence future vaccine design and evaluation for this important human pathogen.

Craig B. Wilen, MD, PhD, is assistant professor of laboratory medicine and immunobiology at Yale University, New Haven, Conn. He does not have any conflicts to disclose.

Title
Norovirus-specific cell immunity is durable
Norovirus-specific cell immunity is durable

 

Protection against norovirus gastroenteritis is supported in part by norovirus-specific CD8+ T cells that reside in peripheral, intestinal, and lymphoid tissues, according to investigators.

These findings, and the molecular tools used to discover them, could guide development of a norovirus vaccine and novel cellular therapies, according to lead author Ajinkya Pattekar, MD, of the University of Pennsylvania, Philadelphia, and colleagues.

“Currently, there are no approved pharmacologic therapies against norovirus, and despite several promising clinical trials, an effective vaccine is not available,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology, which may stem from an incomplete understanding of norovirus immunity, according to Dr. Pattekar and colleagues.

They noted that most previous research has focused on humoral immunity, which appears variable between individuals, with some people exhibiting a strong humoral response, while others mount only partial humoral protection. The investigators also noted that, depending on which studies were examined, this type of defense could last years or fade within weeks to months and that “immune mechanisms other than antibodies may be important for protection against noroviruses.”

Specifically, cellular immunity may be at work. A 2020 study involving volunteers showed that T cells were cross-reactive to a type of norovirus the participants had never been exposed to.

“These findings suggest that T cells may target conserved epitopes and could offer cross-protection against a broad range of noroviruses,” Dr. Pattekar and colleagues wrote.

To test this hypothesis, they first collected peripheral blood mononuclear cells (PBMCs) from three healthy volunteers with unknown norovirus exposure history. Then serum samples were screened for norovirus functional antibodies via the binding between virus-like particles (VLPs) and histo–blood group antigens (HBGAs). This revealed disparate profiles of blocking antibodies against various norovirus strains. While donor 1 and donor 2 had antibodies against multiple strains, donor 3 lacked norovirus antibodies. Further testing showed that this latter individual was a nonsecretor with limited exposure history.

Next, the investigators tested donor PBMCs for norovirus-specific T-cell responses with use of overlapping libraries of peptides for each of the three norovirus open reading frames (ORF1, ORF2, and ORF3). T-cell responses, predominantly involving CD8+ T cells, were observed in all donors. While donor 1 had the greatest response to ORF1, donors 2 and 3 had responses that focused on ORF2.

“Thus, norovirus-specific T cells targeting ORF1 and ORF2 epitopes are present in peripheral blood from healthy donors regardless of secretor status,” the investigators wrote.

To better characterize T-cell epitopes, the investigators subdivided the overlapping peptide libraries into groups of shorter peptides, then exposed serum to these smaller component pools. This revealed eight HLA class I restricted epitopes that were derived from a genogroup II.4 pandemic norovirus strain; this group of variants has been responsible for all six of the norovirus pandemics since 1996.

Closer examination of the epitopes showed that they were “broadly conserved beyond GII.4.” Only one epitope exhibited variation in the C-terminal aromatic anchor, and it was nondominant. The investigators therefore identified seven immunodominant CD8+ epitopes, which they considered “valuable targets for vaccine and cell-based therapies.

“These data further confirm that epitope-specific CD8+ T cells are a universal feature of the overall norovirus immune response and could be an attractive target for future vaccines,” the investigators wrote.

Additional testing involving samples of spleen, mesenteric lymph nodes, and duodenum from deceased individuals showed presence of norovirus-specific CD8+ T cells, with particular abundance in intestinal tissue, and distinct phenotypes and functional properties in different tissue types.

“Future studies using tetramers and intestinal samples should build on these observations and fully define the location and microenvironment of norovirus-specific T cells,” the investigators wrote. “If carried out in the context of a vaccine trial, such studies could be highly valuable in elucidating tissue-resident memory correlates of norovirus immunity.”

The study was funded by the National Institutes of Health, the Wellcome Trust, and Deutsche Forschungsgemeinschaft. The investigators reported no conflicts of interest.

 

Protection against norovirus gastroenteritis is supported in part by norovirus-specific CD8+ T cells that reside in peripheral, intestinal, and lymphoid tissues, according to investigators.

These findings, and the molecular tools used to discover them, could guide development of a norovirus vaccine and novel cellular therapies, according to lead author Ajinkya Pattekar, MD, of the University of Pennsylvania, Philadelphia, and colleagues.

“Currently, there are no approved pharmacologic therapies against norovirus, and despite several promising clinical trials, an effective vaccine is not available,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology, which may stem from an incomplete understanding of norovirus immunity, according to Dr. Pattekar and colleagues.

They noted that most previous research has focused on humoral immunity, which appears variable between individuals, with some people exhibiting a strong humoral response, while others mount only partial humoral protection. The investigators also noted that, depending on which studies were examined, this type of defense could last years or fade within weeks to months and that “immune mechanisms other than antibodies may be important for protection against noroviruses.”

Specifically, cellular immunity may be at work. A 2020 study involving volunteers showed that T cells were cross-reactive to a type of norovirus the participants had never been exposed to.

“These findings suggest that T cells may target conserved epitopes and could offer cross-protection against a broad range of noroviruses,” Dr. Pattekar and colleagues wrote.

To test this hypothesis, they first collected peripheral blood mononuclear cells (PBMCs) from three healthy volunteers with unknown norovirus exposure history. Then serum samples were screened for norovirus functional antibodies via the binding between virus-like particles (VLPs) and histo–blood group antigens (HBGAs). This revealed disparate profiles of blocking antibodies against various norovirus strains. While donor 1 and donor 2 had antibodies against multiple strains, donor 3 lacked norovirus antibodies. Further testing showed that this latter individual was a nonsecretor with limited exposure history.

Next, the investigators tested donor PBMCs for norovirus-specific T-cell responses with use of overlapping libraries of peptides for each of the three norovirus open reading frames (ORF1, ORF2, and ORF3). T-cell responses, predominantly involving CD8+ T cells, were observed in all donors. While donor 1 had the greatest response to ORF1, donors 2 and 3 had responses that focused on ORF2.

“Thus, norovirus-specific T cells targeting ORF1 and ORF2 epitopes are present in peripheral blood from healthy donors regardless of secretor status,” the investigators wrote.

To better characterize T-cell epitopes, the investigators subdivided the overlapping peptide libraries into groups of shorter peptides, then exposed serum to these smaller component pools. This revealed eight HLA class I restricted epitopes that were derived from a genogroup II.4 pandemic norovirus strain; this group of variants has been responsible for all six of the norovirus pandemics since 1996.

Closer examination of the epitopes showed that they were “broadly conserved beyond GII.4.” Only one epitope exhibited variation in the C-terminal aromatic anchor, and it was nondominant. The investigators therefore identified seven immunodominant CD8+ epitopes, which they considered “valuable targets for vaccine and cell-based therapies.

“These data further confirm that epitope-specific CD8+ T cells are a universal feature of the overall norovirus immune response and could be an attractive target for future vaccines,” the investigators wrote.

Additional testing involving samples of spleen, mesenteric lymph nodes, and duodenum from deceased individuals showed presence of norovirus-specific CD8+ T cells, with particular abundance in intestinal tissue, and distinct phenotypes and functional properties in different tissue types.

“Future studies using tetramers and intestinal samples should build on these observations and fully define the location and microenvironment of norovirus-specific T cells,” the investigators wrote. “If carried out in the context of a vaccine trial, such studies could be highly valuable in elucidating tissue-resident memory correlates of norovirus immunity.”

The study was funded by the National Institutes of Health, the Wellcome Trust, and Deutsche Forschungsgemeinschaft. The investigators reported no conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CELLULAR AND MOLECULAR GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads

Low-risk adenomas may not elevate risk of CRC-related death

What’s the best timing for CRC surveillance?
Article Type
Changed
Fri, 04/09/2021 - 09:08

Unlike high-risk adenomas (HRAs), low-risk adenomas (LRAs) have a minimal association with risk of metachronous colorectal cancer (CRC), and no relationship with odds of metachronous CRC-related mortality, according to a meta-analysis of more than 500,000 individuals.

Dr. Abhiram Duvvuri

These findings should impact surveillance guidelines and make follow-up the same for individuals with LRAs or no adenomas, reported lead author Abhiram Duvvuri, MD, of the division of gastroenterology and hepatology at the University of Kansas, Kansas City, and colleagues. Currently, the United States Multi-Society Task Force on Colorectal Cancer advises colonoscopy intervals of 3 years for individuals with HRAs, 7-10 years for those with LRAs, and 10 years for those without adenomas.

“The evidence supporting these surveillance recommendations for clinically relevant endpoints such as cancer and cancer-related deaths among patients who undergo adenoma removal, particularly LRA, is minimal, because most of the evidence was based on the surrogate risk of metachronous advanced neoplasia,” the investigators wrote in Gastroenterology.

To provide more solid evidence, the investigators performed a systematic review and meta-analysis, ultimately analyzing 12 studies with data from 510,019 individuals at a mean age of 59.2 years. All studies reported rates of LRA, HRA, or no adenoma at baseline colonoscopy, plus incidence of metachronous CRC and/or CRC-related mortality. With these data, the investigators determined incidence of metachronous CRC and CRC-related mortality for each of the adenoma groups and also compared these incidences per 10,000 person-years of follow-up across groups.

After a mean follow-up of 8.5 years, patients with HRAs had a significantly higher rate of CRC compared with patients who had LRAs (13.81 vs. 4.5; odds ratio, 2.35; 95% confidence interval, 1.72-3.20) or no adenomas (13.81 vs. 3.4; OR, 2.92; 95% CI, 2.31-3.69). Similarly, but to a lesser degree, LRAs were associated with significantly greater risk of CRC than that of no adenomas (4.5 vs. 3.4; OR, 1.26; 95% CI, 1.06-1.51).

Data on CRC- related mortality further supported these minimal risk profiles because LRAs did not significantly increase the risk of CRC-related mortality compared with no adenomas (OR, 1.15; 95% CI, 0.76-1.74). In contrast, HRAs were associated with significantly greater risk of CRC-related death than that of both LRAs (OR, 2.48; 95% CI, 1.30-4.75) and no adenomas (OR, 2.69; 95% CI, 1.87-3.87).

The investigators acknowledged certain limitations of their study. For one, there were no randomized controlled trials in the meta-analysis, which can introduce bias. Loss of patients to follow-up is also possible; however, the investigators noted that there was a robust sample of patients available for study outcomes all the same. There is also risk of comparability bias in that HRA and LRA groups underwent more colonoscopies; however, the duration of follow-up and timing of last colonoscopy were similar among groups. Lastly, it’s possible the patient sample wasn’t representative because of healthy screenee bias, but the investigators compared groups against general population to minimize that bias.

The investigators also highlighted several strengths of their study that make their findings more reliable than those of past meta-analyses. For one, their study is the largest of its kind to date, and involved a significantly higher number of patients with LRA and no adenomas. Also, in contrast with previous studies, CRC and CRC-related mortality were evaluated rather than advanced adenomas, they noted.

“Furthermore, we also analyzed CRC incidence and mortality in the LRA group compared with the general population, with the [standardized incidence ratio] being lower and [standardized mortality ratio] being comparable, confirming that it is indeed a low-risk group,” they wrote.

Considering these strengths and the nature of their findings, Dr. Duvvuri and colleagues called for a more conservative approach to CRC surveillance among individuals with LRAs, and more research to investigate extending colonoscopy intervals even further.

“We recommend that the interval for follow-up colonoscopy should be the same in patients with LRAs or no adenomas but that the HRA group should have a more frequent surveillance interval for CRC surveillance compared with these groups,” they concluded. “Future studies should evaluate whether surveillance intervals could be lengthened beyond 10 years in the no-adenoma and LRA groups after an initial high-quality index colonoscopy.”

One author disclosed affiliations with Erbe, Cdx Labs, Aries, and others. Dr. Duvvuri and the remaining authors disclosed no conflicts.

Body

 

Despite evidence suggesting that colorectal cancer (CRC) incidence and mortality can be decreased through the endoscopic removal of adenomatous polyps, the question remains as to whether further endoscopic surveillance is necessary after polypectomy and, if so, how often. The most recent iteration of the United States Multi-Society Task Force guidelines endorsed a lengthening of the surveillance interval following the removal of low-risk adenomas (LRAs), defined as 1-2 tubular adenomas <10 mm with low-grade dysplasia, while maintaining a shorter interval for high-risk adenomas (HRAs), defined as advanced adenomas (villous histology, high-grade dysplasia, or >10 mm) or >3 adenomas.

Dr. Reid M. Ness

Dr. Duvvuri and colleagues present the results of a systematic review and meta-analysis of studies examining metachronous CRC incidence and mortality following index colonoscopy. They found a small but statistically significant increase in the incidence of CRC but no significant difference in CRC mortality when comparing patients with LRAs to those with no adenomas. In contrast, they found both a statistically and clinically significant difference in CRC incidence/mortality when comparing patients with HRAs to both those with no adenomas and those with LRAs. They concluded that these results support a recommendation for no difference in follow-up surveillance between patients with LRAs and no adenomas but do support more frequent surveillance for patients with HRAs at index colonoscopy.

Future studies should better examine the timing of neoplasm incidence/recurrence following adenoma removal and also examine metachronous CRC incidence/mortality in patients with sessile serrated lesions at index colonoscopy.

Reid M. Ness, MD, MPH, AGAF, is an associate professor in the division of gastroenterology, hepatology, and nutrition at Vanderbilt University Medical Center and at the VA Tennessee Valley Healthcare System, Nashville, campus. He is an investigator in the Vanderbilt-Ingram Cancer Center. Dr. Ness has no financial relationships to disclose.

Publications
Topics
Sections
Body

 

Despite evidence suggesting that colorectal cancer (CRC) incidence and mortality can be decreased through the endoscopic removal of adenomatous polyps, the question remains as to whether further endoscopic surveillance is necessary after polypectomy and, if so, how often. The most recent iteration of the United States Multi-Society Task Force guidelines endorsed a lengthening of the surveillance interval following the removal of low-risk adenomas (LRAs), defined as 1-2 tubular adenomas <10 mm with low-grade dysplasia, while maintaining a shorter interval for high-risk adenomas (HRAs), defined as advanced adenomas (villous histology, high-grade dysplasia, or >10 mm) or >3 adenomas.

Dr. Reid M. Ness

Dr. Duvvuri and colleagues present the results of a systematic review and meta-analysis of studies examining metachronous CRC incidence and mortality following index colonoscopy. They found a small but statistically significant increase in the incidence of CRC but no significant difference in CRC mortality when comparing patients with LRAs to those with no adenomas. In contrast, they found both a statistically and clinically significant difference in CRC incidence/mortality when comparing patients with HRAs to both those with no adenomas and those with LRAs. They concluded that these results support a recommendation for no difference in follow-up surveillance between patients with LRAs and no adenomas but do support more frequent surveillance for patients with HRAs at index colonoscopy.

Future studies should better examine the timing of neoplasm incidence/recurrence following adenoma removal and also examine metachronous CRC incidence/mortality in patients with sessile serrated lesions at index colonoscopy.

Reid M. Ness, MD, MPH, AGAF, is an associate professor in the division of gastroenterology, hepatology, and nutrition at Vanderbilt University Medical Center and at the VA Tennessee Valley Healthcare System, Nashville, campus. He is an investigator in the Vanderbilt-Ingram Cancer Center. Dr. Ness has no financial relationships to disclose.

Body

 

Despite evidence suggesting that colorectal cancer (CRC) incidence and mortality can be decreased through the endoscopic removal of adenomatous polyps, the question remains as to whether further endoscopic surveillance is necessary after polypectomy and, if so, how often. The most recent iteration of the United States Multi-Society Task Force guidelines endorsed a lengthening of the surveillance interval following the removal of low-risk adenomas (LRAs), defined as 1-2 tubular adenomas <10 mm with low-grade dysplasia, while maintaining a shorter interval for high-risk adenomas (HRAs), defined as advanced adenomas (villous histology, high-grade dysplasia, or >10 mm) or >3 adenomas.

Dr. Reid M. Ness

Dr. Duvvuri and colleagues present the results of a systematic review and meta-analysis of studies examining metachronous CRC incidence and mortality following index colonoscopy. They found a small but statistically significant increase in the incidence of CRC but no significant difference in CRC mortality when comparing patients with LRAs to those with no adenomas. In contrast, they found both a statistically and clinically significant difference in CRC incidence/mortality when comparing patients with HRAs to both those with no adenomas and those with LRAs. They concluded that these results support a recommendation for no difference in follow-up surveillance between patients with LRAs and no adenomas but do support more frequent surveillance for patients with HRAs at index colonoscopy.

Future studies should better examine the timing of neoplasm incidence/recurrence following adenoma removal and also examine metachronous CRC incidence/mortality in patients with sessile serrated lesions at index colonoscopy.

Reid M. Ness, MD, MPH, AGAF, is an associate professor in the division of gastroenterology, hepatology, and nutrition at Vanderbilt University Medical Center and at the VA Tennessee Valley Healthcare System, Nashville, campus. He is an investigator in the Vanderbilt-Ingram Cancer Center. Dr. Ness has no financial relationships to disclose.

Title
What’s the best timing for CRC surveillance?
What’s the best timing for CRC surveillance?

Unlike high-risk adenomas (HRAs), low-risk adenomas (LRAs) have a minimal association with risk of metachronous colorectal cancer (CRC), and no relationship with odds of metachronous CRC-related mortality, according to a meta-analysis of more than 500,000 individuals.

Dr. Abhiram Duvvuri

These findings should impact surveillance guidelines and make follow-up the same for individuals with LRAs or no adenomas, reported lead author Abhiram Duvvuri, MD, of the division of gastroenterology and hepatology at the University of Kansas, Kansas City, and colleagues. Currently, the United States Multi-Society Task Force on Colorectal Cancer advises colonoscopy intervals of 3 years for individuals with HRAs, 7-10 years for those with LRAs, and 10 years for those without adenomas.

“The evidence supporting these surveillance recommendations for clinically relevant endpoints such as cancer and cancer-related deaths among patients who undergo adenoma removal, particularly LRA, is minimal, because most of the evidence was based on the surrogate risk of metachronous advanced neoplasia,” the investigators wrote in Gastroenterology.

To provide more solid evidence, the investigators performed a systematic review and meta-analysis, ultimately analyzing 12 studies with data from 510,019 individuals at a mean age of 59.2 years. All studies reported rates of LRA, HRA, or no adenoma at baseline colonoscopy, plus incidence of metachronous CRC and/or CRC-related mortality. With these data, the investigators determined incidence of metachronous CRC and CRC-related mortality for each of the adenoma groups and also compared these incidences per 10,000 person-years of follow-up across groups.

After a mean follow-up of 8.5 years, patients with HRAs had a significantly higher rate of CRC compared with patients who had LRAs (13.81 vs. 4.5; odds ratio, 2.35; 95% confidence interval, 1.72-3.20) or no adenomas (13.81 vs. 3.4; OR, 2.92; 95% CI, 2.31-3.69). Similarly, but to a lesser degree, LRAs were associated with significantly greater risk of CRC than that of no adenomas (4.5 vs. 3.4; OR, 1.26; 95% CI, 1.06-1.51).

Data on CRC- related mortality further supported these minimal risk profiles because LRAs did not significantly increase the risk of CRC-related mortality compared with no adenomas (OR, 1.15; 95% CI, 0.76-1.74). In contrast, HRAs were associated with significantly greater risk of CRC-related death than that of both LRAs (OR, 2.48; 95% CI, 1.30-4.75) and no adenomas (OR, 2.69; 95% CI, 1.87-3.87).

The investigators acknowledged certain limitations of their study. For one, there were no randomized controlled trials in the meta-analysis, which can introduce bias. Loss of patients to follow-up is also possible; however, the investigators noted that there was a robust sample of patients available for study outcomes all the same. There is also risk of comparability bias in that HRA and LRA groups underwent more colonoscopies; however, the duration of follow-up and timing of last colonoscopy were similar among groups. Lastly, it’s possible the patient sample wasn’t representative because of healthy screenee bias, but the investigators compared groups against general population to minimize that bias.

The investigators also highlighted several strengths of their study that make their findings more reliable than those of past meta-analyses. For one, their study is the largest of its kind to date, and involved a significantly higher number of patients with LRA and no adenomas. Also, in contrast with previous studies, CRC and CRC-related mortality were evaluated rather than advanced adenomas, they noted.

“Furthermore, we also analyzed CRC incidence and mortality in the LRA group compared with the general population, with the [standardized incidence ratio] being lower and [standardized mortality ratio] being comparable, confirming that it is indeed a low-risk group,” they wrote.

Considering these strengths and the nature of their findings, Dr. Duvvuri and colleagues called for a more conservative approach to CRC surveillance among individuals with LRAs, and more research to investigate extending colonoscopy intervals even further.

“We recommend that the interval for follow-up colonoscopy should be the same in patients with LRAs or no adenomas but that the HRA group should have a more frequent surveillance interval for CRC surveillance compared with these groups,” they concluded. “Future studies should evaluate whether surveillance intervals could be lengthened beyond 10 years in the no-adenoma and LRA groups after an initial high-quality index colonoscopy.”

One author disclosed affiliations with Erbe, Cdx Labs, Aries, and others. Dr. Duvvuri and the remaining authors disclosed no conflicts.

Unlike high-risk adenomas (HRAs), low-risk adenomas (LRAs) have a minimal association with risk of metachronous colorectal cancer (CRC), and no relationship with odds of metachronous CRC-related mortality, according to a meta-analysis of more than 500,000 individuals.

Dr. Abhiram Duvvuri

These findings should impact surveillance guidelines and make follow-up the same for individuals with LRAs or no adenomas, reported lead author Abhiram Duvvuri, MD, of the division of gastroenterology and hepatology at the University of Kansas, Kansas City, and colleagues. Currently, the United States Multi-Society Task Force on Colorectal Cancer advises colonoscopy intervals of 3 years for individuals with HRAs, 7-10 years for those with LRAs, and 10 years for those without adenomas.

“The evidence supporting these surveillance recommendations for clinically relevant endpoints such as cancer and cancer-related deaths among patients who undergo adenoma removal, particularly LRA, is minimal, because most of the evidence was based on the surrogate risk of metachronous advanced neoplasia,” the investigators wrote in Gastroenterology.

To provide more solid evidence, the investigators performed a systematic review and meta-analysis, ultimately analyzing 12 studies with data from 510,019 individuals at a mean age of 59.2 years. All studies reported rates of LRA, HRA, or no adenoma at baseline colonoscopy, plus incidence of metachronous CRC and/or CRC-related mortality. With these data, the investigators determined incidence of metachronous CRC and CRC-related mortality for each of the adenoma groups and also compared these incidences per 10,000 person-years of follow-up across groups.

After a mean follow-up of 8.5 years, patients with HRAs had a significantly higher rate of CRC compared with patients who had LRAs (13.81 vs. 4.5; odds ratio, 2.35; 95% confidence interval, 1.72-3.20) or no adenomas (13.81 vs. 3.4; OR, 2.92; 95% CI, 2.31-3.69). Similarly, but to a lesser degree, LRAs were associated with significantly greater risk of CRC than that of no adenomas (4.5 vs. 3.4; OR, 1.26; 95% CI, 1.06-1.51).

Data on CRC- related mortality further supported these minimal risk profiles because LRAs did not significantly increase the risk of CRC-related mortality compared with no adenomas (OR, 1.15; 95% CI, 0.76-1.74). In contrast, HRAs were associated with significantly greater risk of CRC-related death than that of both LRAs (OR, 2.48; 95% CI, 1.30-4.75) and no adenomas (OR, 2.69; 95% CI, 1.87-3.87).

The investigators acknowledged certain limitations of their study. For one, there were no randomized controlled trials in the meta-analysis, which can introduce bias. Loss of patients to follow-up is also possible; however, the investigators noted that there was a robust sample of patients available for study outcomes all the same. There is also risk of comparability bias in that HRA and LRA groups underwent more colonoscopies; however, the duration of follow-up and timing of last colonoscopy were similar among groups. Lastly, it’s possible the patient sample wasn’t representative because of healthy screenee bias, but the investigators compared groups against general population to minimize that bias.

The investigators also highlighted several strengths of their study that make their findings more reliable than those of past meta-analyses. For one, their study is the largest of its kind to date, and involved a significantly higher number of patients with LRA and no adenomas. Also, in contrast with previous studies, CRC and CRC-related mortality were evaluated rather than advanced adenomas, they noted.

“Furthermore, we also analyzed CRC incidence and mortality in the LRA group compared with the general population, with the [standardized incidence ratio] being lower and [standardized mortality ratio] being comparable, confirming that it is indeed a low-risk group,” they wrote.

Considering these strengths and the nature of their findings, Dr. Duvvuri and colleagues called for a more conservative approach to CRC surveillance among individuals with LRAs, and more research to investigate extending colonoscopy intervals even further.

“We recommend that the interval for follow-up colonoscopy should be the same in patients with LRAs or no adenomas but that the HRA group should have a more frequent surveillance interval for CRC surveillance compared with these groups,” they concluded. “Future studies should evaluate whether surveillance intervals could be lengthened beyond 10 years in the no-adenoma and LRA groups after an initial high-quality index colonoscopy.”

One author disclosed affiliations with Erbe, Cdx Labs, Aries, and others. Dr. Duvvuri and the remaining authors disclosed no conflicts.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM GASTROENTEROLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads

Surveillance endoscopy in Barrett’s may perform better than expected

There’s still room for improvement
Article Type
Changed
Thu, 04/08/2021 - 14:41

For patients with Barrett’s esophagus, surveillance endoscopy detects high-grade dysplasia (HGD) and esophageal adenocarcinoma (EAC) more often than previously reported, according to a retrospective analysis of more than 1,000 patients.

Dr. Lovekirat Dhaliwal

Neoplasia detection rate, defined as findings on initial surveillance endoscopy, was also lower than that observed in past studies, according to lead author Lovekirat Dhaliwal, MBBS, of Mayo Clinic, Rochester, Minn., and colleagues.

This study’s findings may help define quality control benchmarks for endoscopic surveillance of Barrett’s esophagus, the investigators wrote in Clinical Gastroenterology and Hepatology. Accurate metrics are needed, they noted, because almost 9 out of 10 patients with Barrett’s esophagus present with EAC outside of a surveillance program, which “may represent missed opportunities at screening.” At the same time, a previous study by the investigators and one from another group, have suggested that 25%-33% of HGD/EAC cases may go undetected by initial surveillance endoscopy.

“Dysplasia detection in [Barrett’s esophagus] is challenging because of its patchy distribution and often subtle appearance,” the investigators noted. “Lack of compliance with recommended biopsy guidelines is also well-documented.”

On the other hand, Dr. Dhaliwal and colleagues suggested that previous studies may not accurately portray community practice and, therefore, have limited value in determining quality control metrics. A 2019 review, for instance, reported a neoplasia detection rate of 7% among patients with Barrett’s esophagus, but this finding “is composed of data from largely referral center cohorts with endoscopy performed by experienced academic gastroenterologists,” they wrote, which may lead to overestimation of such detection.

To better characterize this landscape, the investigators conducted a retrospective analysis involving 1,066 patients with Barrett’s esophagus who underwent initial surveillance endoscopy between 1991 and 2019. Approximately three out of four surveillance endoscopies (77%) were performed by gastroenterologists, while the remaining were performed by nongastroenterologists, such as family practitioners or surgeons. About 60% of patients were adequately biopsied according to the Seattle protocol.

Analysis revealed that the neoplasia detection rate was 4.9% (95% confidence interval, 3.8%-6.4%), which is less than the previously reported rate of 7%. HGD was more common than EAC (33 cases vs. 20 cases). Out of 1,066 patients, 391 without neoplasia on initial endoscopy underwent repeat endoscopy within a year. Among these individuals, HGD or EAC was detected in eight patients, which suggests that 13% of diagnoses were missed on initial endoscopy, a rate well below the previously reported range of 25%-33%.
 

Technology challenged by technique

The neoplasia detection rate “appeared to increase significantly from 1991 to 2019 on univariate analysis (particularly after 2000), but this was not observed on multivariate analysis,” the investigators wrote. “This was despite the introduction of high definition monitors and high resolution endoscopes in subsequent years.

“This may suggest that in a low dysplasia prevalence setting, basic techniques such as careful white light inspection of the [Barrett’s esophagus] mucosa along with targeted and Seattle protocol biopsies may be more important,” they noted.

The importance of technique may be further supported by another finding: Gastroenterologists detected neoplasia almost four times as often as did nongastroenterologists (odds ratio, 3.6; P = .0154).

“This finding is novel and may be due to additional training in endoscopy, lesion recognition, and familiarity with surveillance guidelines in gastroenterologists,” the investigators wrote. “If this finding is replicated in other cohorts, it may support recommendations for the performance of surveillance by endoscopists trained in gastrointestinal endoscopy and well-versed in surveillance guidelines.

“[U]sing neoplasia detection as a quality metric coupled with outcome measures such as missed dysplasia rates could improve adherence to established biopsy protocols and improve the quality of care to patients,” they wrote. “Ultimately, this can be an opportunity to develop a high-value, evidence-based quality metric in [Barrett’s esophagus] surveillance.”

The authors acknowledged some limitations to their study. Its retrospective design meant no one biopsy protocol could be adopted across the entire study period; however, the results were “unchanged” when restricted to the period after introduction of the Seattle protocol in 2000. The study’s long period could have left results susceptible to changing guidelines, but the neoplasia detection rates remained relatively stable over time.

“Because prior reports consisted largely of tertiary care center cohorts, our findings may reflect the absence of referral bias and be more generalizable,” the investigators wrote.

The study was funded by the National Institute of Aging and the National Cancer Institute. The investigators disclosed relationships with Celgene, Nine Point Medical, Takeda, and others.

Body

 

The current study by Dr. Dhaliwal and colleagues evaluates the neoplasia detection rate (NDR) for high-grade dysplasia (HGD) or esophageal adenocarcinoma (EAC) during surveillance endoscopy, which is a proposed novel quality metric for BE. Within a population cohort, the investigators found the NDR was 4.9%, and this did not increase significantly during the study period from 1991 to 2019. Gastroenterologists were more likely to report visible abnormalities during endoscopy and this was a significant predictor of neoplasia detection in a multivariable model. However, the overall rate of missed HGD or EAC was 13%, and this was not associated with procedural specialty. Interestingly, even with only 57% adherence to Seattle protocol in this study, there was no association with missed lesions.

Dr. David A. Leiman

Despite advances in endoscopic imaging and measures establishing quality for biopsy technique, there remains substantial room for improvement in the endoscopic management of patients with BE. While unable to evaluate all factors associated with neoplasia detection, the authors have provided an important real-world benchmark for NDR. Further study is needed to establish the connection between NDR and missed dysplasia, as well as its impact on outcomes such as EAC staging and mortality. Critically, understanding the role of specialized training and other factors such as inspection time to improve NDR is needed.

David A. Leiman, MD, MSHP, is the chair of the AGA Quality Committee. He is an assistant professor of medicine at Duke University, Durham, N.C., where he serves as director of esophageal research and quality. He has no conflicts.

Publications
Topics
Sections
Body

 

The current study by Dr. Dhaliwal and colleagues evaluates the neoplasia detection rate (NDR) for high-grade dysplasia (HGD) or esophageal adenocarcinoma (EAC) during surveillance endoscopy, which is a proposed novel quality metric for BE. Within a population cohort, the investigators found the NDR was 4.9%, and this did not increase significantly during the study period from 1991 to 2019. Gastroenterologists were more likely to report visible abnormalities during endoscopy and this was a significant predictor of neoplasia detection in a multivariable model. However, the overall rate of missed HGD or EAC was 13%, and this was not associated with procedural specialty. Interestingly, even with only 57% adherence to Seattle protocol in this study, there was no association with missed lesions.

Dr. David A. Leiman

Despite advances in endoscopic imaging and measures establishing quality for biopsy technique, there remains substantial room for improvement in the endoscopic management of patients with BE. While unable to evaluate all factors associated with neoplasia detection, the authors have provided an important real-world benchmark for NDR. Further study is needed to establish the connection between NDR and missed dysplasia, as well as its impact on outcomes such as EAC staging and mortality. Critically, understanding the role of specialized training and other factors such as inspection time to improve NDR is needed.

David A. Leiman, MD, MSHP, is the chair of the AGA Quality Committee. He is an assistant professor of medicine at Duke University, Durham, N.C., where he serves as director of esophageal research and quality. He has no conflicts.

Body

 

The current study by Dr. Dhaliwal and colleagues evaluates the neoplasia detection rate (NDR) for high-grade dysplasia (HGD) or esophageal adenocarcinoma (EAC) during surveillance endoscopy, which is a proposed novel quality metric for BE. Within a population cohort, the investigators found the NDR was 4.9%, and this did not increase significantly during the study period from 1991 to 2019. Gastroenterologists were more likely to report visible abnormalities during endoscopy and this was a significant predictor of neoplasia detection in a multivariable model. However, the overall rate of missed HGD or EAC was 13%, and this was not associated with procedural specialty. Interestingly, even with only 57% adherence to Seattle protocol in this study, there was no association with missed lesions.

Dr. David A. Leiman

Despite advances in endoscopic imaging and measures establishing quality for biopsy technique, there remains substantial room for improvement in the endoscopic management of patients with BE. While unable to evaluate all factors associated with neoplasia detection, the authors have provided an important real-world benchmark for NDR. Further study is needed to establish the connection between NDR and missed dysplasia, as well as its impact on outcomes such as EAC staging and mortality. Critically, understanding the role of specialized training and other factors such as inspection time to improve NDR is needed.

David A. Leiman, MD, MSHP, is the chair of the AGA Quality Committee. He is an assistant professor of medicine at Duke University, Durham, N.C., where he serves as director of esophageal research and quality. He has no conflicts.

Title
There’s still room for improvement
There’s still room for improvement

For patients with Barrett’s esophagus, surveillance endoscopy detects high-grade dysplasia (HGD) and esophageal adenocarcinoma (EAC) more often than previously reported, according to a retrospective analysis of more than 1,000 patients.

Dr. Lovekirat Dhaliwal

Neoplasia detection rate, defined as findings on initial surveillance endoscopy, was also lower than that observed in past studies, according to lead author Lovekirat Dhaliwal, MBBS, of Mayo Clinic, Rochester, Minn., and colleagues.

This study’s findings may help define quality control benchmarks for endoscopic surveillance of Barrett’s esophagus, the investigators wrote in Clinical Gastroenterology and Hepatology. Accurate metrics are needed, they noted, because almost 9 out of 10 patients with Barrett’s esophagus present with EAC outside of a surveillance program, which “may represent missed opportunities at screening.” At the same time, a previous study by the investigators and one from another group, have suggested that 25%-33% of HGD/EAC cases may go undetected by initial surveillance endoscopy.

“Dysplasia detection in [Barrett’s esophagus] is challenging because of its patchy distribution and often subtle appearance,” the investigators noted. “Lack of compliance with recommended biopsy guidelines is also well-documented.”

On the other hand, Dr. Dhaliwal and colleagues suggested that previous studies may not accurately portray community practice and, therefore, have limited value in determining quality control metrics. A 2019 review, for instance, reported a neoplasia detection rate of 7% among patients with Barrett’s esophagus, but this finding “is composed of data from largely referral center cohorts with endoscopy performed by experienced academic gastroenterologists,” they wrote, which may lead to overestimation of such detection.

To better characterize this landscape, the investigators conducted a retrospective analysis involving 1,066 patients with Barrett’s esophagus who underwent initial surveillance endoscopy between 1991 and 2019. Approximately three out of four surveillance endoscopies (77%) were performed by gastroenterologists, while the remaining were performed by nongastroenterologists, such as family practitioners or surgeons. About 60% of patients were adequately biopsied according to the Seattle protocol.

Analysis revealed that the neoplasia detection rate was 4.9% (95% confidence interval, 3.8%-6.4%), which is less than the previously reported rate of 7%. HGD was more common than EAC (33 cases vs. 20 cases). Out of 1,066 patients, 391 without neoplasia on initial endoscopy underwent repeat endoscopy within a year. Among these individuals, HGD or EAC was detected in eight patients, which suggests that 13% of diagnoses were missed on initial endoscopy, a rate well below the previously reported range of 25%-33%.
 

Technology challenged by technique

The neoplasia detection rate “appeared to increase significantly from 1991 to 2019 on univariate analysis (particularly after 2000), but this was not observed on multivariate analysis,” the investigators wrote. “This was despite the introduction of high definition monitors and high resolution endoscopes in subsequent years.

“This may suggest that in a low dysplasia prevalence setting, basic techniques such as careful white light inspection of the [Barrett’s esophagus] mucosa along with targeted and Seattle protocol biopsies may be more important,” they noted.

The importance of technique may be further supported by another finding: Gastroenterologists detected neoplasia almost four times as often as did nongastroenterologists (odds ratio, 3.6; P = .0154).

“This finding is novel and may be due to additional training in endoscopy, lesion recognition, and familiarity with surveillance guidelines in gastroenterologists,” the investigators wrote. “If this finding is replicated in other cohorts, it may support recommendations for the performance of surveillance by endoscopists trained in gastrointestinal endoscopy and well-versed in surveillance guidelines.

“[U]sing neoplasia detection as a quality metric coupled with outcome measures such as missed dysplasia rates could improve adherence to established biopsy protocols and improve the quality of care to patients,” they wrote. “Ultimately, this can be an opportunity to develop a high-value, evidence-based quality metric in [Barrett’s esophagus] surveillance.”

The authors acknowledged some limitations to their study. Its retrospective design meant no one biopsy protocol could be adopted across the entire study period; however, the results were “unchanged” when restricted to the period after introduction of the Seattle protocol in 2000. The study’s long period could have left results susceptible to changing guidelines, but the neoplasia detection rates remained relatively stable over time.

“Because prior reports consisted largely of tertiary care center cohorts, our findings may reflect the absence of referral bias and be more generalizable,” the investigators wrote.

The study was funded by the National Institute of Aging and the National Cancer Institute. The investigators disclosed relationships with Celgene, Nine Point Medical, Takeda, and others.

For patients with Barrett’s esophagus, surveillance endoscopy detects high-grade dysplasia (HGD) and esophageal adenocarcinoma (EAC) more often than previously reported, according to a retrospective analysis of more than 1,000 patients.

Dr. Lovekirat Dhaliwal

Neoplasia detection rate, defined as findings on initial surveillance endoscopy, was also lower than that observed in past studies, according to lead author Lovekirat Dhaliwal, MBBS, of Mayo Clinic, Rochester, Minn., and colleagues.

This study’s findings may help define quality control benchmarks for endoscopic surveillance of Barrett’s esophagus, the investigators wrote in Clinical Gastroenterology and Hepatology. Accurate metrics are needed, they noted, because almost 9 out of 10 patients with Barrett’s esophagus present with EAC outside of a surveillance program, which “may represent missed opportunities at screening.” At the same time, a previous study by the investigators and one from another group, have suggested that 25%-33% of HGD/EAC cases may go undetected by initial surveillance endoscopy.

“Dysplasia detection in [Barrett’s esophagus] is challenging because of its patchy distribution and often subtle appearance,” the investigators noted. “Lack of compliance with recommended biopsy guidelines is also well-documented.”

On the other hand, Dr. Dhaliwal and colleagues suggested that previous studies may not accurately portray community practice and, therefore, have limited value in determining quality control metrics. A 2019 review, for instance, reported a neoplasia detection rate of 7% among patients with Barrett’s esophagus, but this finding “is composed of data from largely referral center cohorts with endoscopy performed by experienced academic gastroenterologists,” they wrote, which may lead to overestimation of such detection.

To better characterize this landscape, the investigators conducted a retrospective analysis involving 1,066 patients with Barrett’s esophagus who underwent initial surveillance endoscopy between 1991 and 2019. Approximately three out of four surveillance endoscopies (77%) were performed by gastroenterologists, while the remaining were performed by nongastroenterologists, such as family practitioners or surgeons. About 60% of patients were adequately biopsied according to the Seattle protocol.

Analysis revealed that the neoplasia detection rate was 4.9% (95% confidence interval, 3.8%-6.4%), which is less than the previously reported rate of 7%. HGD was more common than EAC (33 cases vs. 20 cases). Out of 1,066 patients, 391 without neoplasia on initial endoscopy underwent repeat endoscopy within a year. Among these individuals, HGD or EAC was detected in eight patients, which suggests that 13% of diagnoses were missed on initial endoscopy, a rate well below the previously reported range of 25%-33%.
 

Technology challenged by technique

The neoplasia detection rate “appeared to increase significantly from 1991 to 2019 on univariate analysis (particularly after 2000), but this was not observed on multivariate analysis,” the investigators wrote. “This was despite the introduction of high definition monitors and high resolution endoscopes in subsequent years.

“This may suggest that in a low dysplasia prevalence setting, basic techniques such as careful white light inspection of the [Barrett’s esophagus] mucosa along with targeted and Seattle protocol biopsies may be more important,” they noted.

The importance of technique may be further supported by another finding: Gastroenterologists detected neoplasia almost four times as often as did nongastroenterologists (odds ratio, 3.6; P = .0154).

“This finding is novel and may be due to additional training in endoscopy, lesion recognition, and familiarity with surveillance guidelines in gastroenterologists,” the investigators wrote. “If this finding is replicated in other cohorts, it may support recommendations for the performance of surveillance by endoscopists trained in gastrointestinal endoscopy and well-versed in surveillance guidelines.

“[U]sing neoplasia detection as a quality metric coupled with outcome measures such as missed dysplasia rates could improve adherence to established biopsy protocols and improve the quality of care to patients,” they wrote. “Ultimately, this can be an opportunity to develop a high-value, evidence-based quality metric in [Barrett’s esophagus] surveillance.”

The authors acknowledged some limitations to their study. Its retrospective design meant no one biopsy protocol could be adopted across the entire study period; however, the results were “unchanged” when restricted to the period after introduction of the Seattle protocol in 2000. The study’s long period could have left results susceptible to changing guidelines, but the neoplasia detection rates remained relatively stable over time.

“Because prior reports consisted largely of tertiary care center cohorts, our findings may reflect the absence of referral bias and be more generalizable,” the investigators wrote.

The study was funded by the National Institute of Aging and the National Cancer Institute. The investigators disclosed relationships with Celgene, Nine Point Medical, Takeda, and others.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CLINICAL GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content
Disable Inline Native ads

Liver stiffness predicts hepatic events in NAFLD

Study highlights potential for noninvasive LSMs
Article Type
Changed
Tue, 03/23/2021 - 14:39

Among patients with nonalcoholic fatty liver disease (NAFLD) and compensated advanced chronic liver disease, liver stiffness measurements (LSMs) are associated with risks of hepatic events, according to a retrospective analysis of more than 1,000 patients.

“[N]oninvasive markers that can predict liver disease severity and outcomes in patients with NAFLD and advanced fibrosis are a major unmet need,” wrote lead author Salvatore Petta, MD, of the University of Palermo, Italy, and colleagues. Their report is in Clinical Gastroenterology and Hepatology. “Data about the accuracy of LSM in the prediction of events in NAFLD, and especially in patients with NAFLD and F3-F4 fibrosis, are scarce.”

To address this knowledge gap, the investigators retrospectively analyzed data from 1,039 consecutive patients with NAFLD who had baseline LSMs of more than 10 kPa and/or histologically diagnosed F3-F4 fibrosis. Patients were prospectively recruited at 10 centers in 6 countries, then followed for a median of 35 months, ranging from 19 to 63 months.

All patients had their liver stiffness measured with an M or XL probe at baseline. In addition, approximately half of the patients (n = 533) had a follow-up measurement using the same method, generating a subgroup with changes in liver stiffness. “Improved” liver stiffness was defined as a decrease in LSM greater than 20% from baseline, “impaired” liver stiffness was defined as an increase in LSM greater than 20% from baseline, and “stable” liver stiffness was defined as a change falling between 20% lower and 20% higher than baseline.

At baseline, mean LSM was 17.6 kPa. Cox regression analysis revealed that baseline LSM was independently associated with HCC (hazard ratio, 1.03; 95% confidence interval, 1.00-1.04; P = .003), liver decompensation (HR, 1.03; 95% CI, 1.02-1.04; P < .001), and liver-related death (HR, 1.02; 95% CI, 1.00-1.03; P = .005), but not extrahepatic events.

According to the investigators, the association between LSM at baseline and risk of liver decompensation was maintained after adjustment for the severity of liver disease and for surrogate markers of portal hypertension, they noted. Furthermore, patients with a baseline LSM of at least 21 kPa – which indicates high risk of clinically significant portal hypertension (CSPH) – were at greater risk of liver decompensation than were those with an LSM less than 21 kPa (HR, 3.71; 95% CI, 1.89-6.78; P = .04).

In the subgroup with follow-up measurements, approximately half of the patients had an improved LSM (53.3%), while 27.2% had a stable LSM, and 19.5% had an impaired LSM, a pattern that was significantly associated with diabetes at baseline (P = .01).

“These data agree with the available literature identifying diabetes as a risk factor for liver disease progression and liver-related complications,” the investigators wrote.

Cox regression showed that, among those with follow-up LSM, changes in LSM were independently associated with HCC (HR, 1.72; 95% CI, 1.01-3.02; P = .04), liver decompensation (HR, 1.56; 95% CI, 1.05-2.51; P = . 04), liver-related mortality (HR, 1.96; 95% CI, 1.10-3.38; P = .02), and mortality of any cause (HR, 1.73; 95% CI, 1.11-2.69; P = .01).

These risks could be further stratified by level of change in liver stiffness, with greater impairment predicting greater risk: The crude rate of liver decompensation was 14.4% among those with impaired LSM, compared with 6.2% among those with stable LSM and 3.8% among those with LSM improvement. That said, the categories of changes in LSM were not predictive of decompensation among patients with high risk of CSPH at baseline; however, they remained predictive among those with low risk of CSPH at baseline.

“[T]his study … showed that an integrated assessment of baseline LSM or [changes in LSM] can help in stratifying the risk of development of liver-related complications and of both hepatic and overall mortality,” the investigators concluded. “These data, if further validated, could help personalize prognosis and follow-up in NAFLD with [compensated advanced chronic liver disease].”

The investigators disclosed relationships with AbbVie, Novo Nordisk, Gilead, and others.

Body

 

As the prevalence of nonalcoholic fatty liver disease (NAFLD) continues to rise, risk-stratifying those who will develop liver-related complications remains a major challenge. Although progression of liver fibrosis is a key risk factor for developing liver-related complications, the clinical application of noninvasive fibrosis markers for prognostication has been largely unexplored in NAFLD.

Dr. George Cholankeril

This study by Dr. Petta and colleagues highlights the potential for liver stiffness measurements (LSMs) as a noninvasive method. Increased LSM that was suggestive of clinically significant portal hypertension (kPa >21) had a nearly fourfold risk of hepatic decompensation. Furthermore, a longitudinal increase in LSM by greater than 20% was associated with a greater than 50% increased risk for hepatic decompensation, hepatocellular carcinoma, and death.

Transient elastography is a widely available and accurate tool for the noninvasive assessment of liver fibrosis for NAFLD in routine clinical practice. Routine serial measurements of LSM with transient elastography during clinic visits can provide clinicians with important information in the management of NAFLD, which can aid in treatment decisions, response to therapy, and monitoring of disease progression.

Further research is needed to validate these findings and to evaluate how longitudinal changes in LSM and other noninvasive fibrosis markers can prognosticate outcomes in NAFLD.

George Cholankeril MD, MS, is an assistant professor in the section of gastroenterology & hepatology of the department of medicine and in the division of abdominal transplantation of the department of surgery at Baylor College of Medicine in Houston. He reported having no conflicts of interest.

Publications
Topics
Sections
Body

 

As the prevalence of nonalcoholic fatty liver disease (NAFLD) continues to rise, risk-stratifying those who will develop liver-related complications remains a major challenge. Although progression of liver fibrosis is a key risk factor for developing liver-related complications, the clinical application of noninvasive fibrosis markers for prognostication has been largely unexplored in NAFLD.

Dr. George Cholankeril

This study by Dr. Petta and colleagues highlights the potential for liver stiffness measurements (LSMs) as a noninvasive method. Increased LSM that was suggestive of clinically significant portal hypertension (kPa >21) had a nearly fourfold risk of hepatic decompensation. Furthermore, a longitudinal increase in LSM by greater than 20% was associated with a greater than 50% increased risk for hepatic decompensation, hepatocellular carcinoma, and death.

Transient elastography is a widely available and accurate tool for the noninvasive assessment of liver fibrosis for NAFLD in routine clinical practice. Routine serial measurements of LSM with transient elastography during clinic visits can provide clinicians with important information in the management of NAFLD, which can aid in treatment decisions, response to therapy, and monitoring of disease progression.

Further research is needed to validate these findings and to evaluate how longitudinal changes in LSM and other noninvasive fibrosis markers can prognosticate outcomes in NAFLD.

George Cholankeril MD, MS, is an assistant professor in the section of gastroenterology & hepatology of the department of medicine and in the division of abdominal transplantation of the department of surgery at Baylor College of Medicine in Houston. He reported having no conflicts of interest.

Body

 

As the prevalence of nonalcoholic fatty liver disease (NAFLD) continues to rise, risk-stratifying those who will develop liver-related complications remains a major challenge. Although progression of liver fibrosis is a key risk factor for developing liver-related complications, the clinical application of noninvasive fibrosis markers for prognostication has been largely unexplored in NAFLD.

Dr. George Cholankeril

This study by Dr. Petta and colleagues highlights the potential for liver stiffness measurements (LSMs) as a noninvasive method. Increased LSM that was suggestive of clinically significant portal hypertension (kPa >21) had a nearly fourfold risk of hepatic decompensation. Furthermore, a longitudinal increase in LSM by greater than 20% was associated with a greater than 50% increased risk for hepatic decompensation, hepatocellular carcinoma, and death.

Transient elastography is a widely available and accurate tool for the noninvasive assessment of liver fibrosis for NAFLD in routine clinical practice. Routine serial measurements of LSM with transient elastography during clinic visits can provide clinicians with important information in the management of NAFLD, which can aid in treatment decisions, response to therapy, and monitoring of disease progression.

Further research is needed to validate these findings and to evaluate how longitudinal changes in LSM and other noninvasive fibrosis markers can prognosticate outcomes in NAFLD.

George Cholankeril MD, MS, is an assistant professor in the section of gastroenterology & hepatology of the department of medicine and in the division of abdominal transplantation of the department of surgery at Baylor College of Medicine in Houston. He reported having no conflicts of interest.

Title
Study highlights potential for noninvasive LSMs
Study highlights potential for noninvasive LSMs

Among patients with nonalcoholic fatty liver disease (NAFLD) and compensated advanced chronic liver disease, liver stiffness measurements (LSMs) are associated with risks of hepatic events, according to a retrospective analysis of more than 1,000 patients.

“[N]oninvasive markers that can predict liver disease severity and outcomes in patients with NAFLD and advanced fibrosis are a major unmet need,” wrote lead author Salvatore Petta, MD, of the University of Palermo, Italy, and colleagues. Their report is in Clinical Gastroenterology and Hepatology. “Data about the accuracy of LSM in the prediction of events in NAFLD, and especially in patients with NAFLD and F3-F4 fibrosis, are scarce.”

To address this knowledge gap, the investigators retrospectively analyzed data from 1,039 consecutive patients with NAFLD who had baseline LSMs of more than 10 kPa and/or histologically diagnosed F3-F4 fibrosis. Patients were prospectively recruited at 10 centers in 6 countries, then followed for a median of 35 months, ranging from 19 to 63 months.

All patients had their liver stiffness measured with an M or XL probe at baseline. In addition, approximately half of the patients (n = 533) had a follow-up measurement using the same method, generating a subgroup with changes in liver stiffness. “Improved” liver stiffness was defined as a decrease in LSM greater than 20% from baseline, “impaired” liver stiffness was defined as an increase in LSM greater than 20% from baseline, and “stable” liver stiffness was defined as a change falling between 20% lower and 20% higher than baseline.

At baseline, mean LSM was 17.6 kPa. Cox regression analysis revealed that baseline LSM was independently associated with HCC (hazard ratio, 1.03; 95% confidence interval, 1.00-1.04; P = .003), liver decompensation (HR, 1.03; 95% CI, 1.02-1.04; P < .001), and liver-related death (HR, 1.02; 95% CI, 1.00-1.03; P = .005), but not extrahepatic events.

According to the investigators, the association between LSM at baseline and risk of liver decompensation was maintained after adjustment for the severity of liver disease and for surrogate markers of portal hypertension, they noted. Furthermore, patients with a baseline LSM of at least 21 kPa – which indicates high risk of clinically significant portal hypertension (CSPH) – were at greater risk of liver decompensation than were those with an LSM less than 21 kPa (HR, 3.71; 95% CI, 1.89-6.78; P = .04).

In the subgroup with follow-up measurements, approximately half of the patients had an improved LSM (53.3%), while 27.2% had a stable LSM, and 19.5% had an impaired LSM, a pattern that was significantly associated with diabetes at baseline (P = .01).

“These data agree with the available literature identifying diabetes as a risk factor for liver disease progression and liver-related complications,” the investigators wrote.

Cox regression showed that, among those with follow-up LSM, changes in LSM were independently associated with HCC (HR, 1.72; 95% CI, 1.01-3.02; P = .04), liver decompensation (HR, 1.56; 95% CI, 1.05-2.51; P = . 04), liver-related mortality (HR, 1.96; 95% CI, 1.10-3.38; P = .02), and mortality of any cause (HR, 1.73; 95% CI, 1.11-2.69; P = .01).

These risks could be further stratified by level of change in liver stiffness, with greater impairment predicting greater risk: The crude rate of liver decompensation was 14.4% among those with impaired LSM, compared with 6.2% among those with stable LSM and 3.8% among those with LSM improvement. That said, the categories of changes in LSM were not predictive of decompensation among patients with high risk of CSPH at baseline; however, they remained predictive among those with low risk of CSPH at baseline.

“[T]his study … showed that an integrated assessment of baseline LSM or [changes in LSM] can help in stratifying the risk of development of liver-related complications and of both hepatic and overall mortality,” the investigators concluded. “These data, if further validated, could help personalize prognosis and follow-up in NAFLD with [compensated advanced chronic liver disease].”

The investigators disclosed relationships with AbbVie, Novo Nordisk, Gilead, and others.

Among patients with nonalcoholic fatty liver disease (NAFLD) and compensated advanced chronic liver disease, liver stiffness measurements (LSMs) are associated with risks of hepatic events, according to a retrospective analysis of more than 1,000 patients.

“[N]oninvasive markers that can predict liver disease severity and outcomes in patients with NAFLD and advanced fibrosis are a major unmet need,” wrote lead author Salvatore Petta, MD, of the University of Palermo, Italy, and colleagues. Their report is in Clinical Gastroenterology and Hepatology. “Data about the accuracy of LSM in the prediction of events in NAFLD, and especially in patients with NAFLD and F3-F4 fibrosis, are scarce.”

To address this knowledge gap, the investigators retrospectively analyzed data from 1,039 consecutive patients with NAFLD who had baseline LSMs of more than 10 kPa and/or histologically diagnosed F3-F4 fibrosis. Patients were prospectively recruited at 10 centers in 6 countries, then followed for a median of 35 months, ranging from 19 to 63 months.

All patients had their liver stiffness measured with an M or XL probe at baseline. In addition, approximately half of the patients (n = 533) had a follow-up measurement using the same method, generating a subgroup with changes in liver stiffness. “Improved” liver stiffness was defined as a decrease in LSM greater than 20% from baseline, “impaired” liver stiffness was defined as an increase in LSM greater than 20% from baseline, and “stable” liver stiffness was defined as a change falling between 20% lower and 20% higher than baseline.

At baseline, mean LSM was 17.6 kPa. Cox regression analysis revealed that baseline LSM was independently associated with HCC (hazard ratio, 1.03; 95% confidence interval, 1.00-1.04; P = .003), liver decompensation (HR, 1.03; 95% CI, 1.02-1.04; P < .001), and liver-related death (HR, 1.02; 95% CI, 1.00-1.03; P = .005), but not extrahepatic events.

According to the investigators, the association between LSM at baseline and risk of liver decompensation was maintained after adjustment for the severity of liver disease and for surrogate markers of portal hypertension, they noted. Furthermore, patients with a baseline LSM of at least 21 kPa – which indicates high risk of clinically significant portal hypertension (CSPH) – were at greater risk of liver decompensation than were those with an LSM less than 21 kPa (HR, 3.71; 95% CI, 1.89-6.78; P = .04).

In the subgroup with follow-up measurements, approximately half of the patients had an improved LSM (53.3%), while 27.2% had a stable LSM, and 19.5% had an impaired LSM, a pattern that was significantly associated with diabetes at baseline (P = .01).

“These data agree with the available literature identifying diabetes as a risk factor for liver disease progression and liver-related complications,” the investigators wrote.

Cox regression showed that, among those with follow-up LSM, changes in LSM were independently associated with HCC (HR, 1.72; 95% CI, 1.01-3.02; P = .04), liver decompensation (HR, 1.56; 95% CI, 1.05-2.51; P = . 04), liver-related mortality (HR, 1.96; 95% CI, 1.10-3.38; P = .02), and mortality of any cause (HR, 1.73; 95% CI, 1.11-2.69; P = .01).

These risks could be further stratified by level of change in liver stiffness, with greater impairment predicting greater risk: The crude rate of liver decompensation was 14.4% among those with impaired LSM, compared with 6.2% among those with stable LSM and 3.8% among those with LSM improvement. That said, the categories of changes in LSM were not predictive of decompensation among patients with high risk of CSPH at baseline; however, they remained predictive among those with low risk of CSPH at baseline.

“[T]his study … showed that an integrated assessment of baseline LSM or [changes in LSM] can help in stratifying the risk of development of liver-related complications and of both hepatic and overall mortality,” the investigators concluded. “These data, if further validated, could help personalize prognosis and follow-up in NAFLD with [compensated advanced chronic liver disease].”

The investigators disclosed relationships with AbbVie, Novo Nordisk, Gilead, and others.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CLINICAL GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content

Mitochondrial DNA variant increases gallstone risk

Gallstone disease is multifactorial
Article Type
Changed
Tue, 03/23/2021 - 14:41

A mitochondrial DNA variant may increase the risk of gallstone disease more than fourfold, according to investigators.

wir0man/GettyImages

Mitochondrial DNA 827A>G disrupts mitochondrial function and leads to abnormal cholesterol transport, which increases gallstone development, reported Dayan Sun, of Fudan University, Shanghai, China, and colleagues.

The investigators noted that the findings add support to a genetic role in gallstone development, which could allow for identification of at-risk individuals and implementation of preventive measures.

“The etiology of gallstone disease is multifactorial; age, sex, pregnancy, diet (macronutrients, alcohol, and coffee), and other factors are involved,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology. “Moreover, the significant familial predisposition and ethnic differences in prevalence of this disease indicate the potential influences of genetic factors.”

In 2002, Nakeeb and colleagues reported that at least 30% of gallstone disease cases stemmed from genetic factors. And genetics may play an even greater role in certain populations, such as Native Americans, among whom more than 70% of women have gallstone disease, based on a study by Everhart and colleagues.

According to Ms. Sun and colleagues, a variety of genetic drivers of gallstone disease have been identified, such as ABCG8, identified as the most common genetic risk factor by at least one study, along with a list of other rare mutations, such as one affecting CFTR that leads to altered bile composition.

Based on previous research that linked mitochondrial DNA variants with metabolic defects and, more specifically, aberrations in lipid metabolism, as well as an observed “maternal bias in the maternal transmission of gallstone disease” that suggest mitochondrial influence, the investigators looked for patterns specifically in mitochondrial DNA variants among patients with gallstones.

The study enrolled 104 probands with confirmed gallstone disease and 300 unrelated controls. After collecting DNA samples from all participants, the investigators sequenced mitochondrial DNA HVS1 regions. A comparison of haplogroups showed that B4b’d’e’j was more common among patients with gallstone disease than among controls (odds ratio, 4.428; P = .00012), and further analysis pinpointed 827A>G, a variant in 12S ribosomal RNA.

“During the evolutionary history of modern humans, haplogroup B4 might have originated in East Asia approximately 40,000 years ago,” the investigators wrote, noting that B2, a subhaplogroup of B4, “was a founder haplogroup and expanded in the Americas after the Last Glacial Maximum (approximately 20,000 years ago).”

According to the investigators, this may explain why Native Americans have a higher prevalence of gallstones than East Asians (14%-35% vs. 3%-12%) because they are more often carriers of B4 (14%-44% vs. 2%-8%).

The investigators sought to characterize the impact that the 827A>G variant has on mitochondrial function and found effects ranging from lower respiratory chain complex activity, diminished mitochondrial function, activated mitochondrial protein quality control and retrograde signaling pathways, abnormal lipid metabolism, and abnormal cholesterol transport processes.

For example, the investigators investigated respiratory chain complex activity by creating two sister branch haplogroup cell models, including six cybrids for 827A and six more for 827G, which is how they detected the lower activity. Another step the investigators took was corroborating this finding by detecting OXPHOS function in the 827A and 827G cybrids to determine mitochondrial function.

“In summary, our study demonstrates a potential link between mitochondrial DNA 827A>G and gallstone disease,” the investigators wrote. “Our findings provide a significant biological basis for the clinical diagnosis and prevention of gallstone disease in the future.”

The study was funded by the National Natural Science Foundation of China, the 111 Project, the Shanghai Municipal Science and Technology Major Project, the Scientific and Technology Committee of Shanghai Municipality, and the CAMS Innovation Fund for Medical Sciences. The investigators reported no conflicts of interest.

Body

 

Cholesterol gallstone disease results from imbalances in cholesterol metabolism. Other than the well-known lifestyle risk factors, there is also a strong genetic predisposition to gallstone formation. This study by Sun and colleagues examined the possible association between mitochondrial DNA (mtDNA) variants and cholesterol gallstone development because of the importance of the mitochondria in cellular metabolism and the increased maternal transmission of gallstone disease.

Dr. Xiao Zhao
The investigators compared sequencing data obtained from 104 patients with gallstones versus 300 controls in the Chinese population and identified 827A>G in the mitochondrial 12S ribosomal RNA as the most likely disease conferring variant. In order to functionally validate this polymorphism, the investigators generated cybrid cell lines. They found that, compared with the 827A cybrids, the 827G cybrids exhibited diminished mitochondrial function and increased production of reactive oxygen species. Moreover, there was activation of mitochondrial-nuclear signaling pathways in the 827G cybrids that increased the expression of the lithogenic genes ABCG5/8, which mediate hepatobiliary cholesterol export, especially in gallstone promoting conditions.

This study highlighted gallstone disease as a multifactorial condition that results from complex interaction between genetic and environmental factors. Interestingly, the allele frequency of the 827A>G mtDNA variant was noted to be higher in Native Americans, which may partially explain the high prevalence of gallstones in this population. Further studies are needed to identify additional genetic risk factors in ethnic groups that also have a significant burden of cholelithiasis.

Xiao Zhao, MD, is an assistant professor of medicine of division of digestive diseases in the department of medicine at Columbia University, New York. She reported having no conflicts of interest.

Publications
Topics
Sections
Body

 

Cholesterol gallstone disease results from imbalances in cholesterol metabolism. Other than the well-known lifestyle risk factors, there is also a strong genetic predisposition to gallstone formation. This study by Sun and colleagues examined the possible association between mitochondrial DNA (mtDNA) variants and cholesterol gallstone development because of the importance of the mitochondria in cellular metabolism and the increased maternal transmission of gallstone disease.

Dr. Xiao Zhao
The investigators compared sequencing data obtained from 104 patients with gallstones versus 300 controls in the Chinese population and identified 827A>G in the mitochondrial 12S ribosomal RNA as the most likely disease conferring variant. In order to functionally validate this polymorphism, the investigators generated cybrid cell lines. They found that, compared with the 827A cybrids, the 827G cybrids exhibited diminished mitochondrial function and increased production of reactive oxygen species. Moreover, there was activation of mitochondrial-nuclear signaling pathways in the 827G cybrids that increased the expression of the lithogenic genes ABCG5/8, which mediate hepatobiliary cholesterol export, especially in gallstone promoting conditions.

This study highlighted gallstone disease as a multifactorial condition that results from complex interaction between genetic and environmental factors. Interestingly, the allele frequency of the 827A>G mtDNA variant was noted to be higher in Native Americans, which may partially explain the high prevalence of gallstones in this population. Further studies are needed to identify additional genetic risk factors in ethnic groups that also have a significant burden of cholelithiasis.

Xiao Zhao, MD, is an assistant professor of medicine of division of digestive diseases in the department of medicine at Columbia University, New York. She reported having no conflicts of interest.

Body

 

Cholesterol gallstone disease results from imbalances in cholesterol metabolism. Other than the well-known lifestyle risk factors, there is also a strong genetic predisposition to gallstone formation. This study by Sun and colleagues examined the possible association between mitochondrial DNA (mtDNA) variants and cholesterol gallstone development because of the importance of the mitochondria in cellular metabolism and the increased maternal transmission of gallstone disease.

Dr. Xiao Zhao
The investigators compared sequencing data obtained from 104 patients with gallstones versus 300 controls in the Chinese population and identified 827A>G in the mitochondrial 12S ribosomal RNA as the most likely disease conferring variant. In order to functionally validate this polymorphism, the investigators generated cybrid cell lines. They found that, compared with the 827A cybrids, the 827G cybrids exhibited diminished mitochondrial function and increased production of reactive oxygen species. Moreover, there was activation of mitochondrial-nuclear signaling pathways in the 827G cybrids that increased the expression of the lithogenic genes ABCG5/8, which mediate hepatobiliary cholesterol export, especially in gallstone promoting conditions.

This study highlighted gallstone disease as a multifactorial condition that results from complex interaction between genetic and environmental factors. Interestingly, the allele frequency of the 827A>G mtDNA variant was noted to be higher in Native Americans, which may partially explain the high prevalence of gallstones in this population. Further studies are needed to identify additional genetic risk factors in ethnic groups that also have a significant burden of cholelithiasis.

Xiao Zhao, MD, is an assistant professor of medicine of division of digestive diseases in the department of medicine at Columbia University, New York. She reported having no conflicts of interest.

Title
Gallstone disease is multifactorial
Gallstone disease is multifactorial

A mitochondrial DNA variant may increase the risk of gallstone disease more than fourfold, according to investigators.

wir0man/GettyImages

Mitochondrial DNA 827A>G disrupts mitochondrial function and leads to abnormal cholesterol transport, which increases gallstone development, reported Dayan Sun, of Fudan University, Shanghai, China, and colleagues.

The investigators noted that the findings add support to a genetic role in gallstone development, which could allow for identification of at-risk individuals and implementation of preventive measures.

“The etiology of gallstone disease is multifactorial; age, sex, pregnancy, diet (macronutrients, alcohol, and coffee), and other factors are involved,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology. “Moreover, the significant familial predisposition and ethnic differences in prevalence of this disease indicate the potential influences of genetic factors.”

In 2002, Nakeeb and colleagues reported that at least 30% of gallstone disease cases stemmed from genetic factors. And genetics may play an even greater role in certain populations, such as Native Americans, among whom more than 70% of women have gallstone disease, based on a study by Everhart and colleagues.

According to Ms. Sun and colleagues, a variety of genetic drivers of gallstone disease have been identified, such as ABCG8, identified as the most common genetic risk factor by at least one study, along with a list of other rare mutations, such as one affecting CFTR that leads to altered bile composition.

Based on previous research that linked mitochondrial DNA variants with metabolic defects and, more specifically, aberrations in lipid metabolism, as well as an observed “maternal bias in the maternal transmission of gallstone disease” that suggest mitochondrial influence, the investigators looked for patterns specifically in mitochondrial DNA variants among patients with gallstones.

The study enrolled 104 probands with confirmed gallstone disease and 300 unrelated controls. After collecting DNA samples from all participants, the investigators sequenced mitochondrial DNA HVS1 regions. A comparison of haplogroups showed that B4b’d’e’j was more common among patients with gallstone disease than among controls (odds ratio, 4.428; P = .00012), and further analysis pinpointed 827A>G, a variant in 12S ribosomal RNA.

“During the evolutionary history of modern humans, haplogroup B4 might have originated in East Asia approximately 40,000 years ago,” the investigators wrote, noting that B2, a subhaplogroup of B4, “was a founder haplogroup and expanded in the Americas after the Last Glacial Maximum (approximately 20,000 years ago).”

According to the investigators, this may explain why Native Americans have a higher prevalence of gallstones than East Asians (14%-35% vs. 3%-12%) because they are more often carriers of B4 (14%-44% vs. 2%-8%).

The investigators sought to characterize the impact that the 827A>G variant has on mitochondrial function and found effects ranging from lower respiratory chain complex activity, diminished mitochondrial function, activated mitochondrial protein quality control and retrograde signaling pathways, abnormal lipid metabolism, and abnormal cholesterol transport processes.

For example, the investigators investigated respiratory chain complex activity by creating two sister branch haplogroup cell models, including six cybrids for 827A and six more for 827G, which is how they detected the lower activity. Another step the investigators took was corroborating this finding by detecting OXPHOS function in the 827A and 827G cybrids to determine mitochondrial function.

“In summary, our study demonstrates a potential link between mitochondrial DNA 827A>G and gallstone disease,” the investigators wrote. “Our findings provide a significant biological basis for the clinical diagnosis and prevention of gallstone disease in the future.”

The study was funded by the National Natural Science Foundation of China, the 111 Project, the Shanghai Municipal Science and Technology Major Project, the Scientific and Technology Committee of Shanghai Municipality, and the CAMS Innovation Fund for Medical Sciences. The investigators reported no conflicts of interest.

A mitochondrial DNA variant may increase the risk of gallstone disease more than fourfold, according to investigators.

wir0man/GettyImages

Mitochondrial DNA 827A>G disrupts mitochondrial function and leads to abnormal cholesterol transport, which increases gallstone development, reported Dayan Sun, of Fudan University, Shanghai, China, and colleagues.

The investigators noted that the findings add support to a genetic role in gallstone development, which could allow for identification of at-risk individuals and implementation of preventive measures.

“The etiology of gallstone disease is multifactorial; age, sex, pregnancy, diet (macronutrients, alcohol, and coffee), and other factors are involved,” the investigators wrote in Cellular and Molecular Gastroenterology and Hepatology. “Moreover, the significant familial predisposition and ethnic differences in prevalence of this disease indicate the potential influences of genetic factors.”

In 2002, Nakeeb and colleagues reported that at least 30% of gallstone disease cases stemmed from genetic factors. And genetics may play an even greater role in certain populations, such as Native Americans, among whom more than 70% of women have gallstone disease, based on a study by Everhart and colleagues.

According to Ms. Sun and colleagues, a variety of genetic drivers of gallstone disease have been identified, such as ABCG8, identified as the most common genetic risk factor by at least one study, along with a list of other rare mutations, such as one affecting CFTR that leads to altered bile composition.

Based on previous research that linked mitochondrial DNA variants with metabolic defects and, more specifically, aberrations in lipid metabolism, as well as an observed “maternal bias in the maternal transmission of gallstone disease” that suggest mitochondrial influence, the investigators looked for patterns specifically in mitochondrial DNA variants among patients with gallstones.

The study enrolled 104 probands with confirmed gallstone disease and 300 unrelated controls. After collecting DNA samples from all participants, the investigators sequenced mitochondrial DNA HVS1 regions. A comparison of haplogroups showed that B4b’d’e’j was more common among patients with gallstone disease than among controls (odds ratio, 4.428; P = .00012), and further analysis pinpointed 827A>G, a variant in 12S ribosomal RNA.

“During the evolutionary history of modern humans, haplogroup B4 might have originated in East Asia approximately 40,000 years ago,” the investigators wrote, noting that B2, a subhaplogroup of B4, “was a founder haplogroup and expanded in the Americas after the Last Glacial Maximum (approximately 20,000 years ago).”

According to the investigators, this may explain why Native Americans have a higher prevalence of gallstones than East Asians (14%-35% vs. 3%-12%) because they are more often carriers of B4 (14%-44% vs. 2%-8%).

The investigators sought to characterize the impact that the 827A>G variant has on mitochondrial function and found effects ranging from lower respiratory chain complex activity, diminished mitochondrial function, activated mitochondrial protein quality control and retrograde signaling pathways, abnormal lipid metabolism, and abnormal cholesterol transport processes.

For example, the investigators investigated respiratory chain complex activity by creating two sister branch haplogroup cell models, including six cybrids for 827A and six more for 827G, which is how they detected the lower activity. Another step the investigators took was corroborating this finding by detecting OXPHOS function in the 827A and 827G cybrids to determine mitochondrial function.

“In summary, our study demonstrates a potential link between mitochondrial DNA 827A>G and gallstone disease,” the investigators wrote. “Our findings provide a significant biological basis for the clinical diagnosis and prevention of gallstone disease in the future.”

The study was funded by the National Natural Science Foundation of China, the 111 Project, the Shanghai Municipal Science and Technology Major Project, the Scientific and Technology Committee of Shanghai Municipality, and the CAMS Innovation Fund for Medical Sciences. The investigators reported no conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CELLULAR AND MOLECULAR GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content

PPIs improve functional dyspepsia via anti-inflammatory effects

‘Perplexing’ findings provide future direction
Article Type
Changed
Fri, 03/12/2021 - 09:54

Proton pump inhibitors (PPIs) improve functional dyspepsia (FD) by reducing duodenal eosinophils and mast cells, according to a prospective study.

Dr. Lucas Wauters

This suggests that the anti-inflammatory effects of PPIs are responsible for symptom improvement, and not barrier-protective or acid-suppressive effects, a finding that may guide future therapies and biomarkers, reported lead author Lucas Wauters, PhD, of University Hospitals Leuven (Belgium), and colleagues reported in Gastroenterology.

“FD is a common and unexplained disorder with unknown pathophysiology, hampering a conclusive diagnosis and the development of effective drugs,” the investigators wrote.

Although PPIs are currently used as first-line FD therapy, ostensibly for acid suppression, “the exact mechanism of action of PPIs in FD is unknown,” the investigators noted.

According to Dr. Wauters and colleagues, previous FD studies, such as a 2020 study published in Gut, have reported a variety of pathophysiological findings in the duodenum, including increased eosinophils and mast cells, as well as activation of duodenogastric reflexes, which suggests “a primary role for duodenal pathology in FD symptom generation.” Several drivers of this pathology have been proposed. Some, such as aberrations in bile salts and acidity, point to local, luminal changes, whereas others, such as dysregulated hypothalamic-pituitary-adrenal axis responsiveness and psychosocial factors, implicate a broader set of drivers, the investigators wrote.

The present study explored this landscape through a prospective trial that enrolled 30 healthy volunteers and 47 patients with FD (2 patients with FD did not complete the study).

Patients with FD were subgrouped into “FD-starters” who had not taken PPIs and/or acid suppression for at least 3 months leading up to the trial (n = 28) and “FD-stoppers” who had refractory symptoms after at least 1 month of daily PPI usage (n = 19). Among participants with FD, 25 had postprandial distress syndrome (PDS), 9 had epigastric pain syndrome (EPS), and 13 had subtype overlap.

For the trial, FD-starters and healthy volunteers took 4 weeks of pantoprazole 40 mg once daily, whereas FD-stoppers ceased PPI therapy for 8 weeks. Before and after these respective periods, certain study procedures were conducted, including duodenal biopsy collection, duodenal fluid aspiration, and questionnaires for symptoms and stress. The study also included use of Ussing chambers for biopsies, immunohistochemistry, and bile salt measurements.

FD-starters were significantly more symptomatic than healthy volunteers were at baseline. After starting PPIs, those with FD had symptom improvements, confirming “clinical efficacy of a standard course of PPIs in all FD subtypes,” whereas healthy volunteers showed no significant change in symptoms.

Similarly, baseline duodenal eosinophil counts were higher in FD-starters than in healthy volunteers. On starting PPIs, however, eosinophil counts in these two groups moved in opposite directions: FD-starters’ counts dropped from a mean of 331 to 183 eosinophils/mm2, whereas healthy volunteers’ counts rose from a mean of 115 to 229 eosinophils/mm2 (P < .0001). Changes in mast cells and paracellular passage followed the same pattern, falling in FD-starters and rising in healthy volunteers. On the other hand, symptoms actually improved in the FD-stoppers after they went off PPIs, although they did not reach symptom levels of the healthy volunteers.

“Differential effects of PPIs in healthy volunteers point to the role of luminal changes in determining low-grade mucosal immune activation in the duodenum, which can also occur in FD after long-term use and provide arguments against continued use in refractory patients,” the investigators wrote.

Dr. Wauters and colleagues suggested that their findings could guide future approaches to FD management.

“Our results suggest that quantification of duodenal eosinophils has the potential to become part of diagnostic workup and guide therapeutic decisions in FD,” they wrote. “Additional study of the underlying mediators might lead to the discovery of new potential biomarkers or novel therapeutic targets, potentially allowing the identification of subgroups responding to biologically targeted rather than symptom-based treatments.”

The study was supported by the clinical research fund of the University Hospitals Leuven. The investigators reported no conflicts of interest.

Body

Functional dyspepsia (FD) is a commonly encountered diagnosis among primary care and gastroenterology clinics with estimated prevalence of 5%-11% worldwide. However, the pathophysiology of this entity is not well understood, and most of the patients who undergo upper endoscopy for dyspepsia tend to have normal findings.

Dr. Dyanesh A. Patel

The differential effects of PPIs on duodenal inflammation among the groups in this study are perplexing and the findings are limited by the short duration of follow-up testing and lack of a placebo group. Duodenal eosinophils, mast cells, and permeability were higher in treatment-naive FD, and a 4-week course of PPIs reduced this inflammation in FD but apparently increased it in healthy volunteers. Furthermore, when patients with PPI-refractory FD were studied, withdrawal of the PPI lead to improvement in symptoms and trends toward less duodenal inflammation.

The discordant impact of PPIs among these groups on duodenal inflammation raises questions regarding the true effect from PPIs or changes driven by other systemic factors (brain-gut axis). This is hard to differentiate without a placebo group. Focus in future studies should be placed on how to clinically phenotype and predict PPI responders versus nonresponders along with use of longer durations to see if even PPI responders tend to lose response over time, regardless of changes in the duodenal inflammation and permeability.

Dhyanesh A. Patel, MD, is an assistant professor of medicine at the center for esophageal disorders, Vanderbilt University Medical Center, Nashville, Tenn. He has no conflicts.

Publications
Topics
Sections
Body

Functional dyspepsia (FD) is a commonly encountered diagnosis among primary care and gastroenterology clinics with estimated prevalence of 5%-11% worldwide. However, the pathophysiology of this entity is not well understood, and most of the patients who undergo upper endoscopy for dyspepsia tend to have normal findings.

Dr. Dyanesh A. Patel

The differential effects of PPIs on duodenal inflammation among the groups in this study are perplexing and the findings are limited by the short duration of follow-up testing and lack of a placebo group. Duodenal eosinophils, mast cells, and permeability were higher in treatment-naive FD, and a 4-week course of PPIs reduced this inflammation in FD but apparently increased it in healthy volunteers. Furthermore, when patients with PPI-refractory FD were studied, withdrawal of the PPI lead to improvement in symptoms and trends toward less duodenal inflammation.

The discordant impact of PPIs among these groups on duodenal inflammation raises questions regarding the true effect from PPIs or changes driven by other systemic factors (brain-gut axis). This is hard to differentiate without a placebo group. Focus in future studies should be placed on how to clinically phenotype and predict PPI responders versus nonresponders along with use of longer durations to see if even PPI responders tend to lose response over time, regardless of changes in the duodenal inflammation and permeability.

Dhyanesh A. Patel, MD, is an assistant professor of medicine at the center for esophageal disorders, Vanderbilt University Medical Center, Nashville, Tenn. He has no conflicts.

Body

Functional dyspepsia (FD) is a commonly encountered diagnosis among primary care and gastroenterology clinics with estimated prevalence of 5%-11% worldwide. However, the pathophysiology of this entity is not well understood, and most of the patients who undergo upper endoscopy for dyspepsia tend to have normal findings.

Dr. Dyanesh A. Patel

The differential effects of PPIs on duodenal inflammation among the groups in this study are perplexing and the findings are limited by the short duration of follow-up testing and lack of a placebo group. Duodenal eosinophils, mast cells, and permeability were higher in treatment-naive FD, and a 4-week course of PPIs reduced this inflammation in FD but apparently increased it in healthy volunteers. Furthermore, when patients with PPI-refractory FD were studied, withdrawal of the PPI lead to improvement in symptoms and trends toward less duodenal inflammation.

The discordant impact of PPIs among these groups on duodenal inflammation raises questions regarding the true effect from PPIs or changes driven by other systemic factors (brain-gut axis). This is hard to differentiate without a placebo group. Focus in future studies should be placed on how to clinically phenotype and predict PPI responders versus nonresponders along with use of longer durations to see if even PPI responders tend to lose response over time, regardless of changes in the duodenal inflammation and permeability.

Dhyanesh A. Patel, MD, is an assistant professor of medicine at the center for esophageal disorders, Vanderbilt University Medical Center, Nashville, Tenn. He has no conflicts.

Title
‘Perplexing’ findings provide future direction
‘Perplexing’ findings provide future direction

Proton pump inhibitors (PPIs) improve functional dyspepsia (FD) by reducing duodenal eosinophils and mast cells, according to a prospective study.

Dr. Lucas Wauters

This suggests that the anti-inflammatory effects of PPIs are responsible for symptom improvement, and not barrier-protective or acid-suppressive effects, a finding that may guide future therapies and biomarkers, reported lead author Lucas Wauters, PhD, of University Hospitals Leuven (Belgium), and colleagues reported in Gastroenterology.

“FD is a common and unexplained disorder with unknown pathophysiology, hampering a conclusive diagnosis and the development of effective drugs,” the investigators wrote.

Although PPIs are currently used as first-line FD therapy, ostensibly for acid suppression, “the exact mechanism of action of PPIs in FD is unknown,” the investigators noted.

According to Dr. Wauters and colleagues, previous FD studies, such as a 2020 study published in Gut, have reported a variety of pathophysiological findings in the duodenum, including increased eosinophils and mast cells, as well as activation of duodenogastric reflexes, which suggests “a primary role for duodenal pathology in FD symptom generation.” Several drivers of this pathology have been proposed. Some, such as aberrations in bile salts and acidity, point to local, luminal changes, whereas others, such as dysregulated hypothalamic-pituitary-adrenal axis responsiveness and psychosocial factors, implicate a broader set of drivers, the investigators wrote.

The present study explored this landscape through a prospective trial that enrolled 30 healthy volunteers and 47 patients with FD (2 patients with FD did not complete the study).

Patients with FD were subgrouped into “FD-starters” who had not taken PPIs and/or acid suppression for at least 3 months leading up to the trial (n = 28) and “FD-stoppers” who had refractory symptoms after at least 1 month of daily PPI usage (n = 19). Among participants with FD, 25 had postprandial distress syndrome (PDS), 9 had epigastric pain syndrome (EPS), and 13 had subtype overlap.

For the trial, FD-starters and healthy volunteers took 4 weeks of pantoprazole 40 mg once daily, whereas FD-stoppers ceased PPI therapy for 8 weeks. Before and after these respective periods, certain study procedures were conducted, including duodenal biopsy collection, duodenal fluid aspiration, and questionnaires for symptoms and stress. The study also included use of Ussing chambers for biopsies, immunohistochemistry, and bile salt measurements.

FD-starters were significantly more symptomatic than healthy volunteers were at baseline. After starting PPIs, those with FD had symptom improvements, confirming “clinical efficacy of a standard course of PPIs in all FD subtypes,” whereas healthy volunteers showed no significant change in symptoms.

Similarly, baseline duodenal eosinophil counts were higher in FD-starters than in healthy volunteers. On starting PPIs, however, eosinophil counts in these two groups moved in opposite directions: FD-starters’ counts dropped from a mean of 331 to 183 eosinophils/mm2, whereas healthy volunteers’ counts rose from a mean of 115 to 229 eosinophils/mm2 (P < .0001). Changes in mast cells and paracellular passage followed the same pattern, falling in FD-starters and rising in healthy volunteers. On the other hand, symptoms actually improved in the FD-stoppers after they went off PPIs, although they did not reach symptom levels of the healthy volunteers.

“Differential effects of PPIs in healthy volunteers point to the role of luminal changes in determining low-grade mucosal immune activation in the duodenum, which can also occur in FD after long-term use and provide arguments against continued use in refractory patients,” the investigators wrote.

Dr. Wauters and colleagues suggested that their findings could guide future approaches to FD management.

“Our results suggest that quantification of duodenal eosinophils has the potential to become part of diagnostic workup and guide therapeutic decisions in FD,” they wrote. “Additional study of the underlying mediators might lead to the discovery of new potential biomarkers or novel therapeutic targets, potentially allowing the identification of subgroups responding to biologically targeted rather than symptom-based treatments.”

The study was supported by the clinical research fund of the University Hospitals Leuven. The investigators reported no conflicts of interest.

Proton pump inhibitors (PPIs) improve functional dyspepsia (FD) by reducing duodenal eosinophils and mast cells, according to a prospective study.

Dr. Lucas Wauters

This suggests that the anti-inflammatory effects of PPIs are responsible for symptom improvement, and not barrier-protective or acid-suppressive effects, a finding that may guide future therapies and biomarkers, reported lead author Lucas Wauters, PhD, of University Hospitals Leuven (Belgium), and colleagues reported in Gastroenterology.

“FD is a common and unexplained disorder with unknown pathophysiology, hampering a conclusive diagnosis and the development of effective drugs,” the investigators wrote.

Although PPIs are currently used as first-line FD therapy, ostensibly for acid suppression, “the exact mechanism of action of PPIs in FD is unknown,” the investigators noted.

According to Dr. Wauters and colleagues, previous FD studies, such as a 2020 study published in Gut, have reported a variety of pathophysiological findings in the duodenum, including increased eosinophils and mast cells, as well as activation of duodenogastric reflexes, which suggests “a primary role for duodenal pathology in FD symptom generation.” Several drivers of this pathology have been proposed. Some, such as aberrations in bile salts and acidity, point to local, luminal changes, whereas others, such as dysregulated hypothalamic-pituitary-adrenal axis responsiveness and psychosocial factors, implicate a broader set of drivers, the investigators wrote.

The present study explored this landscape through a prospective trial that enrolled 30 healthy volunteers and 47 patients with FD (2 patients with FD did not complete the study).

Patients with FD were subgrouped into “FD-starters” who had not taken PPIs and/or acid suppression for at least 3 months leading up to the trial (n = 28) and “FD-stoppers” who had refractory symptoms after at least 1 month of daily PPI usage (n = 19). Among participants with FD, 25 had postprandial distress syndrome (PDS), 9 had epigastric pain syndrome (EPS), and 13 had subtype overlap.

For the trial, FD-starters and healthy volunteers took 4 weeks of pantoprazole 40 mg once daily, whereas FD-stoppers ceased PPI therapy for 8 weeks. Before and after these respective periods, certain study procedures were conducted, including duodenal biopsy collection, duodenal fluid aspiration, and questionnaires for symptoms and stress. The study also included use of Ussing chambers for biopsies, immunohistochemistry, and bile salt measurements.

FD-starters were significantly more symptomatic than healthy volunteers were at baseline. After starting PPIs, those with FD had symptom improvements, confirming “clinical efficacy of a standard course of PPIs in all FD subtypes,” whereas healthy volunteers showed no significant change in symptoms.

Similarly, baseline duodenal eosinophil counts were higher in FD-starters than in healthy volunteers. On starting PPIs, however, eosinophil counts in these two groups moved in opposite directions: FD-starters’ counts dropped from a mean of 331 to 183 eosinophils/mm2, whereas healthy volunteers’ counts rose from a mean of 115 to 229 eosinophils/mm2 (P < .0001). Changes in mast cells and paracellular passage followed the same pattern, falling in FD-starters and rising in healthy volunteers. On the other hand, symptoms actually improved in the FD-stoppers after they went off PPIs, although they did not reach symptom levels of the healthy volunteers.

“Differential effects of PPIs in healthy volunteers point to the role of luminal changes in determining low-grade mucosal immune activation in the duodenum, which can also occur in FD after long-term use and provide arguments against continued use in refractory patients,” the investigators wrote.

Dr. Wauters and colleagues suggested that their findings could guide future approaches to FD management.

“Our results suggest that quantification of duodenal eosinophils has the potential to become part of diagnostic workup and guide therapeutic decisions in FD,” they wrote. “Additional study of the underlying mediators might lead to the discovery of new potential biomarkers or novel therapeutic targets, potentially allowing the identification of subgroups responding to biologically targeted rather than symptom-based treatments.”

The study was supported by the clinical research fund of the University Hospitals Leuven. The investigators reported no conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM GASTROENTEROLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content

Automated software accurately generates ERCP quality reports

Possibilities abound
Article Type
Changed
Tue, 03/23/2021 - 14:40

 

Modified endoscopy documentation software can automatically generate endoscopic retrograde cholangiopancreatography (ERCP) quality metrics, based on a trial at two referral centers.

Providers were prompted during procedures, and inputting any missed data took providers less than 30 additional seconds per patient. The approach led to highly accurate quality reports, lead author Gregory A. Coté, MD, MS, of the Medical University of South Carolina, Charleston, and colleagues wrote in Techniques and Innovations in Gastrointestinal Endoscopy.

The investigators suggested that these findings may lead to the kind of quality reports already used for colonoscopy, which are easier to produce. Such reports are important, they wrote, as the U.S. health care system shifts to value-based reimbursement models, which in turn puts greater scrutiny on the quality of endoscopic procedures. However, doing so with ERCP isn’t entirely straightforward.

“Measuring adherence to ERCP quality indicators is especially challenging given: variance in indications, intraprocedural maneuvers, potential outcomes of a complex procedure, and variability in physician report documentation,” Dr. Coté and colleagues wrote. “In order to operationalize robust tracking of clinically relevant adherence to ERCP quality indicators in clinical practice – that is, to provide real-time feedback to providers, health systems, payors, and patients – an automated system of measurement must be developed.”

The quality indicators used in the study were largely drawn from an American Society for Gastrointestinal Endoscopy/American College of Gastroenterology task force document, with exclusion of those that were subjective or required systematic follow-up. The investigators modified existing endoscopy documentation software at two referral centers to include mandatory, structured data fields, principally with inclusion of quality improvements deemed high priority by the society consensus document, study authors, or both. For instance, providers were obligated to select a specific indication instead of various, synonymous terms (for example, “biliary stricture” vs. “common bile duct stricture”). Examples of quality indicators included successful cannulation of the desired duct, successful retrieval of stone less than 10 mm, or successful placement of a bile duct stent when indicated. Endoscopists were also required to note the presence of postoperative foregut anatomy or presence of existing sphincterotomy, variables which serve to stratefy the quality indicator outcome for degree of difficulty and allow appropriate comparisons of data. In addition, the study authors included inquiries about use of rectal indomethacin, use of prophylactic pancreatic duct stent, and documentation of need for repeat ERCP, follow-up x-ray, or both.

After 9 months, the system recorded 1,376 ERCP procedures conducted by eight providers, with a median annualized volume of 237 procedures (range, 37-336). Almost one-third (29%) of the patients had not had prior sphincterotomy.

Automated reporting of ERCP was compared with manual record review, which confirmed high (98%-100%) accuracy. This high level of accuracy “obviates the need for manual adjudication of medical records,” the investigators wrote.

They used data from one provider to create a template report card, and while exact comparisons across providers and institutions were not published, an example report card that was published with the study showed how such comparisons could be generated in the real world.

“The tool presented in this study allows for an objective assessment of ERCP performance which can provide explicit feedback to providers and allow transparent assessment of quality outcomes; it has the potential to improve the quality of ERCP akin to what has been demonstrated using colonoscopy report cards,” the investigators wrote. “Importantly, this can be achieved with minimal alteration to providers’ routine procedure documentation.”

Dr. Coté and colleagues also noted that the software modifications “can be implemented in other endoscopy units using the same or similar software.”

Taking the project to the next level would require widespread collaboration, according to the investigators.

“A key next step is to operationalize the transfer of data across multiple institutions, allowing for the creation of interim, standard-quality indicator reports that could be disseminated to providers, health systems, and payors,” they wrote. “If applied to a national cohort, this tool could accurately assess the current landscape of ERCP quality and provide tremendous opportunities for systematic improvement.”

One author disclosed a relationship with Provation Medical, but the remaining authors declared no relevant conflicts.

Body

 

Quality indicators have been proposed to improve the outcome of patients undergoing endoscopic procedures. The path toward quality improvement begins with selection of parameters, which matters a great deal and have wide performance variation. Endoscopists then track their own performance, compare it with targets based on community standards, and improve their patients’ outcomes using this feedback. Great progress has been made in the area of tracking and improving adenoma detection rate, an indicator closely tied to reduction in colorectal cancer mortality.

Endoscopic retrograde cholangiopancreatography (ERCP) is a high-stakes procedure with great potential therapeutic benefit and with a small but significant risk of life-threatening complications such as pancreatitis. This study by Coté and colleagues illuminates an effective and straightforward step to making ERCP quality improvement feasible. The report card concept is not new, but the novel innovation is to leverage the use of required fields in the electronic report generator. Seamlessly, this produces nuanced reports that link provider performance to patient characteristics and indication. The authors have shown extremely high accuracy of automatic electronic ERCP quality indicator recording, compared with manual data collection. Such data have clear and immediate utility in the credentialing process and quality improvement arena. With this means of recording outcomes, deidentified ERCP quality data might soon join colonoscopy data in national data repositories such as the GI Quality Improvement Consortium, and government quality reporting on ERCP outcomes would become much more feasible. Fellow self-assessment and logging of progress could also be facilitated if report generators were further amended to require recording of fellow participation.

Jonathan Cohen, MD, FASGE, is a clinical professor of medicine at New York University Langone Health. He reported having no relevant conflicts of interest.

Publications
Topics
Sections
Body

 

Quality indicators have been proposed to improve the outcome of patients undergoing endoscopic procedures. The path toward quality improvement begins with selection of parameters, which matters a great deal and have wide performance variation. Endoscopists then track their own performance, compare it with targets based on community standards, and improve their patients’ outcomes using this feedback. Great progress has been made in the area of tracking and improving adenoma detection rate, an indicator closely tied to reduction in colorectal cancer mortality.

Endoscopic retrograde cholangiopancreatography (ERCP) is a high-stakes procedure with great potential therapeutic benefit and with a small but significant risk of life-threatening complications such as pancreatitis. This study by Coté and colleagues illuminates an effective and straightforward step to making ERCP quality improvement feasible. The report card concept is not new, but the novel innovation is to leverage the use of required fields in the electronic report generator. Seamlessly, this produces nuanced reports that link provider performance to patient characteristics and indication. The authors have shown extremely high accuracy of automatic electronic ERCP quality indicator recording, compared with manual data collection. Such data have clear and immediate utility in the credentialing process and quality improvement arena. With this means of recording outcomes, deidentified ERCP quality data might soon join colonoscopy data in national data repositories such as the GI Quality Improvement Consortium, and government quality reporting on ERCP outcomes would become much more feasible. Fellow self-assessment and logging of progress could also be facilitated if report generators were further amended to require recording of fellow participation.

Jonathan Cohen, MD, FASGE, is a clinical professor of medicine at New York University Langone Health. He reported having no relevant conflicts of interest.

Body

 

Quality indicators have been proposed to improve the outcome of patients undergoing endoscopic procedures. The path toward quality improvement begins with selection of parameters, which matters a great deal and have wide performance variation. Endoscopists then track their own performance, compare it with targets based on community standards, and improve their patients’ outcomes using this feedback. Great progress has been made in the area of tracking and improving adenoma detection rate, an indicator closely tied to reduction in colorectal cancer mortality.

Endoscopic retrograde cholangiopancreatography (ERCP) is a high-stakes procedure with great potential therapeutic benefit and with a small but significant risk of life-threatening complications such as pancreatitis. This study by Coté and colleagues illuminates an effective and straightforward step to making ERCP quality improvement feasible. The report card concept is not new, but the novel innovation is to leverage the use of required fields in the electronic report generator. Seamlessly, this produces nuanced reports that link provider performance to patient characteristics and indication. The authors have shown extremely high accuracy of automatic electronic ERCP quality indicator recording, compared with manual data collection. Such data have clear and immediate utility in the credentialing process and quality improvement arena. With this means of recording outcomes, deidentified ERCP quality data might soon join colonoscopy data in national data repositories such as the GI Quality Improvement Consortium, and government quality reporting on ERCP outcomes would become much more feasible. Fellow self-assessment and logging of progress could also be facilitated if report generators were further amended to require recording of fellow participation.

Jonathan Cohen, MD, FASGE, is a clinical professor of medicine at New York University Langone Health. He reported having no relevant conflicts of interest.

Title
Possibilities abound
Possibilities abound

 

Modified endoscopy documentation software can automatically generate endoscopic retrograde cholangiopancreatography (ERCP) quality metrics, based on a trial at two referral centers.

Providers were prompted during procedures, and inputting any missed data took providers less than 30 additional seconds per patient. The approach led to highly accurate quality reports, lead author Gregory A. Coté, MD, MS, of the Medical University of South Carolina, Charleston, and colleagues wrote in Techniques and Innovations in Gastrointestinal Endoscopy.

The investigators suggested that these findings may lead to the kind of quality reports already used for colonoscopy, which are easier to produce. Such reports are important, they wrote, as the U.S. health care system shifts to value-based reimbursement models, which in turn puts greater scrutiny on the quality of endoscopic procedures. However, doing so with ERCP isn’t entirely straightforward.

“Measuring adherence to ERCP quality indicators is especially challenging given: variance in indications, intraprocedural maneuvers, potential outcomes of a complex procedure, and variability in physician report documentation,” Dr. Coté and colleagues wrote. “In order to operationalize robust tracking of clinically relevant adherence to ERCP quality indicators in clinical practice – that is, to provide real-time feedback to providers, health systems, payors, and patients – an automated system of measurement must be developed.”

The quality indicators used in the study were largely drawn from an American Society for Gastrointestinal Endoscopy/American College of Gastroenterology task force document, with exclusion of those that were subjective or required systematic follow-up. The investigators modified existing endoscopy documentation software at two referral centers to include mandatory, structured data fields, principally with inclusion of quality improvements deemed high priority by the society consensus document, study authors, or both. For instance, providers were obligated to select a specific indication instead of various, synonymous terms (for example, “biliary stricture” vs. “common bile duct stricture”). Examples of quality indicators included successful cannulation of the desired duct, successful retrieval of stone less than 10 mm, or successful placement of a bile duct stent when indicated. Endoscopists were also required to note the presence of postoperative foregut anatomy or presence of existing sphincterotomy, variables which serve to stratefy the quality indicator outcome for degree of difficulty and allow appropriate comparisons of data. In addition, the study authors included inquiries about use of rectal indomethacin, use of prophylactic pancreatic duct stent, and documentation of need for repeat ERCP, follow-up x-ray, or both.

After 9 months, the system recorded 1,376 ERCP procedures conducted by eight providers, with a median annualized volume of 237 procedures (range, 37-336). Almost one-third (29%) of the patients had not had prior sphincterotomy.

Automated reporting of ERCP was compared with manual record review, which confirmed high (98%-100%) accuracy. This high level of accuracy “obviates the need for manual adjudication of medical records,” the investigators wrote.

They used data from one provider to create a template report card, and while exact comparisons across providers and institutions were not published, an example report card that was published with the study showed how such comparisons could be generated in the real world.

“The tool presented in this study allows for an objective assessment of ERCP performance which can provide explicit feedback to providers and allow transparent assessment of quality outcomes; it has the potential to improve the quality of ERCP akin to what has been demonstrated using colonoscopy report cards,” the investigators wrote. “Importantly, this can be achieved with minimal alteration to providers’ routine procedure documentation.”

Dr. Coté and colleagues also noted that the software modifications “can be implemented in other endoscopy units using the same or similar software.”

Taking the project to the next level would require widespread collaboration, according to the investigators.

“A key next step is to operationalize the transfer of data across multiple institutions, allowing for the creation of interim, standard-quality indicator reports that could be disseminated to providers, health systems, and payors,” they wrote. “If applied to a national cohort, this tool could accurately assess the current landscape of ERCP quality and provide tremendous opportunities for systematic improvement.”

One author disclosed a relationship with Provation Medical, but the remaining authors declared no relevant conflicts.

 

Modified endoscopy documentation software can automatically generate endoscopic retrograde cholangiopancreatography (ERCP) quality metrics, based on a trial at two referral centers.

Providers were prompted during procedures, and inputting any missed data took providers less than 30 additional seconds per patient. The approach led to highly accurate quality reports, lead author Gregory A. Coté, MD, MS, of the Medical University of South Carolina, Charleston, and colleagues wrote in Techniques and Innovations in Gastrointestinal Endoscopy.

The investigators suggested that these findings may lead to the kind of quality reports already used for colonoscopy, which are easier to produce. Such reports are important, they wrote, as the U.S. health care system shifts to value-based reimbursement models, which in turn puts greater scrutiny on the quality of endoscopic procedures. However, doing so with ERCP isn’t entirely straightforward.

“Measuring adherence to ERCP quality indicators is especially challenging given: variance in indications, intraprocedural maneuvers, potential outcomes of a complex procedure, and variability in physician report documentation,” Dr. Coté and colleagues wrote. “In order to operationalize robust tracking of clinically relevant adherence to ERCP quality indicators in clinical practice – that is, to provide real-time feedback to providers, health systems, payors, and patients – an automated system of measurement must be developed.”

The quality indicators used in the study were largely drawn from an American Society for Gastrointestinal Endoscopy/American College of Gastroenterology task force document, with exclusion of those that were subjective or required systematic follow-up. The investigators modified existing endoscopy documentation software at two referral centers to include mandatory, structured data fields, principally with inclusion of quality improvements deemed high priority by the society consensus document, study authors, or both. For instance, providers were obligated to select a specific indication instead of various, synonymous terms (for example, “biliary stricture” vs. “common bile duct stricture”). Examples of quality indicators included successful cannulation of the desired duct, successful retrieval of stone less than 10 mm, or successful placement of a bile duct stent when indicated. Endoscopists were also required to note the presence of postoperative foregut anatomy or presence of existing sphincterotomy, variables which serve to stratefy the quality indicator outcome for degree of difficulty and allow appropriate comparisons of data. In addition, the study authors included inquiries about use of rectal indomethacin, use of prophylactic pancreatic duct stent, and documentation of need for repeat ERCP, follow-up x-ray, or both.

After 9 months, the system recorded 1,376 ERCP procedures conducted by eight providers, with a median annualized volume of 237 procedures (range, 37-336). Almost one-third (29%) of the patients had not had prior sphincterotomy.

Automated reporting of ERCP was compared with manual record review, which confirmed high (98%-100%) accuracy. This high level of accuracy “obviates the need for manual adjudication of medical records,” the investigators wrote.

They used data from one provider to create a template report card, and while exact comparisons across providers and institutions were not published, an example report card that was published with the study showed how such comparisons could be generated in the real world.

“The tool presented in this study allows for an objective assessment of ERCP performance which can provide explicit feedback to providers and allow transparent assessment of quality outcomes; it has the potential to improve the quality of ERCP akin to what has been demonstrated using colonoscopy report cards,” the investigators wrote. “Importantly, this can be achieved with minimal alteration to providers’ routine procedure documentation.”

Dr. Coté and colleagues also noted that the software modifications “can be implemented in other endoscopy units using the same or similar software.”

Taking the project to the next level would require widespread collaboration, according to the investigators.

“A key next step is to operationalize the transfer of data across multiple institutions, allowing for the creation of interim, standard-quality indicator reports that could be disseminated to providers, health systems, and payors,” they wrote. “If applied to a national cohort, this tool could accurately assess the current landscape of ERCP quality and provide tremendous opportunities for systematic improvement.”

One author disclosed a relationship with Provation Medical, but the remaining authors declared no relevant conflicts.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM TECHNIQUES AND INNOVATIONS IN GASTROINTESTINAL ENDOSCOPY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content

AGA Clinical Practice Update: Palliative care management in cirrhosis

Article Type
Changed
Tue, 03/02/2021 - 16:19

 

Clinicians who manage patients with cirrhosis should incorporate palliative care “irrespective of transplant candidacy,” according to a clinical practice update from the American Gastroenterological Association.

“[T]his care should be based on needs assessment instead of prognosis alone, delivered concurrently with curative or life-prolonging treatments, and tailored to the stage of disease,” wrote Puneeta Tandon, MD, of University of Alberta in Edmonton, Alta., and associates. Their report is in Clinical Gastroenterology and Hepatology.

Cirrhosis has a median survival ranging from 2 years for decompensated disease to 12 years for compensated disease, according to one systemic review. Moreover, even compensated cirrhosis incurs “a high burden of physical and psychological symptoms,” which increases as cirrhosis progresses, the update authors noted.

According to another review, there is established evidence outside cirrhosis that palliative care – including comprehensive symptom management, advance care planning, and timely referrals to specialty palliative care and hospice support – has the potential to significantly improve quality of life, end-of-life care, health care costs, coordination among providers, and caregiver outcomes.

However, the update authors noted that there remain few guidelines or guidance statements regarding palliative care in cirrhosis. Hence, the clinical practice update reviews 10 best practices to help clinicians fill this gap.

Providers “from any specialty, within any healthcare setting” can help provide palliative care for patients with cirrhosis, the experts emphasized. This is, in part, because of the growing population with cirrhosis being met with a limited number of palliative care specialists; dealing with this reality can be helped by inviting other providers to learn about and engage in palliative care.

Another best practice statement addressed assessing symptoms “within physical, psychological, social, and spiritual domains related to [patients’] liver disease, its treatment, and prognosis.” This approach is needed because of the complex effects that a life-threatening illness and its symptoms can have on many variables, including loss of independence/identity, financial stress, and impact on personal relationships. A systematic review of symptom prevalence in end-stage liver disease revealed a complex milieu, including pain, muscle cramps, sexual dysfunction, insomnia, and anxiety.

High-quality communication is important in palliative care, including discussion of prognosis and goals of care. Providers specializing in gastroenterology/hepatology should reevaluate prognosis and clarify prognosis and goals of care with patients and caregivers during routine visits and sentinel events, such as new complications, a hospital or intensive care admission, and when transplant eligibility is determined. However, prognostication in cirrhosis can be challenging, the experts noted. The update authors also acknowledged that, while more research is needed to inform practice regarding communicating with patients with serious illness about palliative care and goals of care, there are courses and resources meant to help improve those skills, including those provided by Vital Talk, Respecting Choices, and the Serious Illness Conversation Guide.

Cirrhosis “has physical, mental, and financial consequences” for caregivers, especially when patients have decompensated disease. To support caregivers, clinicians can routinely evaluate their burdens and needs. Tools such as the Caregiver Strain Index are useful and can be administered by ancillary staff. Clinicians also can reach out to primary care and palliative care providers to identify local resources for caregiver support.

“Because lack of time is one of the major barriers to administering palliative care, healthcare providers should consider how they can optimize efficiencies in palliative care delivery,” the experts wrote. Examples include identifying local billing codes, arranging for ancillary staff to screen patients on their palliative care needs, and setting up multidisciplinary teams that work together to deliver palliative care. If access to specialty palliative care is limited, providers can collaborate with local specialist teams to set “clear triggers and pathways for referral.”

Finally, hospice referrals are often delayed for patients with cirrhosis. “Find out your local referral criteria for hospice and what would be required to refer a cirrhosis patient there,” the experts advised. “Healthcare providers caring for patients with cirrhosis should provide timely referral to hospice for patients who have comfort-oriented goals and prognosis of 6 months or less.”

The authors of the clinical practice update received no funding support. They reported having no relevant conflicts of interest.

Publications
Topics
Sections

 

Clinicians who manage patients with cirrhosis should incorporate palliative care “irrespective of transplant candidacy,” according to a clinical practice update from the American Gastroenterological Association.

“[T]his care should be based on needs assessment instead of prognosis alone, delivered concurrently with curative or life-prolonging treatments, and tailored to the stage of disease,” wrote Puneeta Tandon, MD, of University of Alberta in Edmonton, Alta., and associates. Their report is in Clinical Gastroenterology and Hepatology.

Cirrhosis has a median survival ranging from 2 years for decompensated disease to 12 years for compensated disease, according to one systemic review. Moreover, even compensated cirrhosis incurs “a high burden of physical and psychological symptoms,” which increases as cirrhosis progresses, the update authors noted.

According to another review, there is established evidence outside cirrhosis that palliative care – including comprehensive symptom management, advance care planning, and timely referrals to specialty palliative care and hospice support – has the potential to significantly improve quality of life, end-of-life care, health care costs, coordination among providers, and caregiver outcomes.

However, the update authors noted that there remain few guidelines or guidance statements regarding palliative care in cirrhosis. Hence, the clinical practice update reviews 10 best practices to help clinicians fill this gap.

Providers “from any specialty, within any healthcare setting” can help provide palliative care for patients with cirrhosis, the experts emphasized. This is, in part, because of the growing population with cirrhosis being met with a limited number of palliative care specialists; dealing with this reality can be helped by inviting other providers to learn about and engage in palliative care.

Another best practice statement addressed assessing symptoms “within physical, psychological, social, and spiritual domains related to [patients’] liver disease, its treatment, and prognosis.” This approach is needed because of the complex effects that a life-threatening illness and its symptoms can have on many variables, including loss of independence/identity, financial stress, and impact on personal relationships. A systematic review of symptom prevalence in end-stage liver disease revealed a complex milieu, including pain, muscle cramps, sexual dysfunction, insomnia, and anxiety.

High-quality communication is important in palliative care, including discussion of prognosis and goals of care. Providers specializing in gastroenterology/hepatology should reevaluate prognosis and clarify prognosis and goals of care with patients and caregivers during routine visits and sentinel events, such as new complications, a hospital or intensive care admission, and when transplant eligibility is determined. However, prognostication in cirrhosis can be challenging, the experts noted. The update authors also acknowledged that, while more research is needed to inform practice regarding communicating with patients with serious illness about palliative care and goals of care, there are courses and resources meant to help improve those skills, including those provided by Vital Talk, Respecting Choices, and the Serious Illness Conversation Guide.

Cirrhosis “has physical, mental, and financial consequences” for caregivers, especially when patients have decompensated disease. To support caregivers, clinicians can routinely evaluate their burdens and needs. Tools such as the Caregiver Strain Index are useful and can be administered by ancillary staff. Clinicians also can reach out to primary care and palliative care providers to identify local resources for caregiver support.

“Because lack of time is one of the major barriers to administering palliative care, healthcare providers should consider how they can optimize efficiencies in palliative care delivery,” the experts wrote. Examples include identifying local billing codes, arranging for ancillary staff to screen patients on their palliative care needs, and setting up multidisciplinary teams that work together to deliver palliative care. If access to specialty palliative care is limited, providers can collaborate with local specialist teams to set “clear triggers and pathways for referral.”

Finally, hospice referrals are often delayed for patients with cirrhosis. “Find out your local referral criteria for hospice and what would be required to refer a cirrhosis patient there,” the experts advised. “Healthcare providers caring for patients with cirrhosis should provide timely referral to hospice for patients who have comfort-oriented goals and prognosis of 6 months or less.”

The authors of the clinical practice update received no funding support. They reported having no relevant conflicts of interest.

 

Clinicians who manage patients with cirrhosis should incorporate palliative care “irrespective of transplant candidacy,” according to a clinical practice update from the American Gastroenterological Association.

“[T]his care should be based on needs assessment instead of prognosis alone, delivered concurrently with curative or life-prolonging treatments, and tailored to the stage of disease,” wrote Puneeta Tandon, MD, of University of Alberta in Edmonton, Alta., and associates. Their report is in Clinical Gastroenterology and Hepatology.

Cirrhosis has a median survival ranging from 2 years for decompensated disease to 12 years for compensated disease, according to one systemic review. Moreover, even compensated cirrhosis incurs “a high burden of physical and psychological symptoms,” which increases as cirrhosis progresses, the update authors noted.

According to another review, there is established evidence outside cirrhosis that palliative care – including comprehensive symptom management, advance care planning, and timely referrals to specialty palliative care and hospice support – has the potential to significantly improve quality of life, end-of-life care, health care costs, coordination among providers, and caregiver outcomes.

However, the update authors noted that there remain few guidelines or guidance statements regarding palliative care in cirrhosis. Hence, the clinical practice update reviews 10 best practices to help clinicians fill this gap.

Providers “from any specialty, within any healthcare setting” can help provide palliative care for patients with cirrhosis, the experts emphasized. This is, in part, because of the growing population with cirrhosis being met with a limited number of palliative care specialists; dealing with this reality can be helped by inviting other providers to learn about and engage in palliative care.

Another best practice statement addressed assessing symptoms “within physical, psychological, social, and spiritual domains related to [patients’] liver disease, its treatment, and prognosis.” This approach is needed because of the complex effects that a life-threatening illness and its symptoms can have on many variables, including loss of independence/identity, financial stress, and impact on personal relationships. A systematic review of symptom prevalence in end-stage liver disease revealed a complex milieu, including pain, muscle cramps, sexual dysfunction, insomnia, and anxiety.

High-quality communication is important in palliative care, including discussion of prognosis and goals of care. Providers specializing in gastroenterology/hepatology should reevaluate prognosis and clarify prognosis and goals of care with patients and caregivers during routine visits and sentinel events, such as new complications, a hospital or intensive care admission, and when transplant eligibility is determined. However, prognostication in cirrhosis can be challenging, the experts noted. The update authors also acknowledged that, while more research is needed to inform practice regarding communicating with patients with serious illness about palliative care and goals of care, there are courses and resources meant to help improve those skills, including those provided by Vital Talk, Respecting Choices, and the Serious Illness Conversation Guide.

Cirrhosis “has physical, mental, and financial consequences” for caregivers, especially when patients have decompensated disease. To support caregivers, clinicians can routinely evaluate their burdens and needs. Tools such as the Caregiver Strain Index are useful and can be administered by ancillary staff. Clinicians also can reach out to primary care and palliative care providers to identify local resources for caregiver support.

“Because lack of time is one of the major barriers to administering palliative care, healthcare providers should consider how they can optimize efficiencies in palliative care delivery,” the experts wrote. Examples include identifying local billing codes, arranging for ancillary staff to screen patients on their palliative care needs, and setting up multidisciplinary teams that work together to deliver palliative care. If access to specialty palliative care is limited, providers can collaborate with local specialist teams to set “clear triggers and pathways for referral.”

Finally, hospice referrals are often delayed for patients with cirrhosis. “Find out your local referral criteria for hospice and what would be required to refer a cirrhosis patient there,” the experts advised. “Healthcare providers caring for patients with cirrhosis should provide timely referral to hospice for patients who have comfort-oriented goals and prognosis of 6 months or less.”

The authors of the clinical practice update received no funding support. They reported having no relevant conflicts of interest.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM CLINICAL GASTROENTEROLOGY AND HEPATOLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer

AGA Clinical Practice Update: Eradication strategies for H. pylori infection

Article Type
Changed
Sat, 03/27/2021 - 16:40
Display Headline
AGA Clinical Practice Update: Eradication strategies for H. pylori infection

Antimicrobial resistance is the most common cause of treatment-refractory Helicobacter pylori infection, but before switching antibiotics, clinicians should screen for factors such as treatment nonadherence or inadequate suppression of gastric acid, according to a clinical practice update from the American Gastroenterological Association.

Dr. Shailja C. Shah

“Inadequate acid suppression is associated with H. pylori eradication failure. The use of high-dose and more potent PPIs, PPIs not metabolized by CYP2C19, or potassium-competitive acid blockers, if available, should be considered in cases of refractory H. pylori infection,” wrote Shailja C. Shah, MD, MPH, of Vanderbilt University Medical Center in Nashville, Tenn., and coauthors Prasad G. Iyer, MD, and Steven F. Moss, MD. . Their report is in Gastroenterology.

H. pylori infection is the most common cause of gastric cancer. Although eradication is widely recommended, it can be challenging because of strain diversity, rising antimicrobial resistance, a dearth of recent head-to-head clinical trials, and sparse epidemiologic and sensitivity data, the experts noted. For this reason, before selecting an eradication regimen, it is vital to thoroughly review a patient’s history of antibiotics – for example, any prior macrolide or fluoroquinolone exposure should preclude the use of clarithromycin- or levofloxacin-based regimens “given the high likelihood of resistance,” the experts wrote. They also advised that clinicians should avoid levofloxacin unless the H. pylori strain is known to be sensitive to it or if population rates of levofloxacin resistance rates are known to be less than 15%. However, amoxicillin, tetracycline, and rifabutin resistance are rare, and these agents “can be considered for subsequent therapies in refractory H. pylori infection.”

A longer antimicrobial regimen (such as 14 vs. 7 days) is more likely to eradicate H. pylori. If first-line bismuth quadruple therapy (such as a PPI plus bismuth, metronidazole, and tetracycline) fails, then second-line options include another bismuth-containing quadruple-agent regimen, or triple therapy with rifabutin or levofloxacin plus high-dose dual PPI therapy and amoxicillin. If patient history contains “penicillin allergy” but does not list anaphylaxis, then penicillin allergy testing can help determine if amoxicillin-based regimens are an option. The authors also note that, when used, amoxicillin should be dosed at 2 g/day in divided doses three to four times per day in order to avoid low trough levels because this might be associated with H. pylori eradication failure. For metronidazole, regardless of in vitro resistance, eradication is more likely if patients receive 1.5-2 g/day, in divided doses, with concomitant bismuth.



Treatment nonadherence contributes to refractory H. pylori infection and may be caused by the complexity of the treatment regimen, high pill burden, and side effects. To improve adherence, the experts advised counseling patients on the rationale for the treatment regimen, the dosing instructions, the importance of completing the full course of therapy, and providing anticipatory guidance regarding common side effects. If a patient adheres to second-line treatment and it still fails, then susceptibility testing is advised before starting another regimen. Depending on the results, options may include levofloxacin-based quadruple therapy, another round of bismuth-based quadruple therapy, a PPI plus amoxicillin and rifabutin, or high-dose PPI therapy plus high-dose amoxicillin (2-3 g/day divided across three to four doses).

Other considerations include how to approach patients and caregivers, particularly the elderly and other vulnerable patients, with shared decision-making to help them weigh the potential benefits of continuing to try to eradicate H. pylori against the risk of possible adverse effects and the “inconvenience of repeated exposure to antibiotics and high-dose acid suppression,” the experts wrote. They also advised tracking rates of eradication success and relevant demographic and clinical data, including patients’ antibiotic history. Publicly sharing aggregated, deidentified results can help other local clinicians select eradication regimens. Finally, the use of probiotics and other adjunctive therapies “should be considered experimental” since these have no clear benefit for treating refractory H. pylori infection.

Dr. Shah was funded by an AGA Research Scholar Award and a Veterans Affairs Career Development Award. She reported having no conflicts of interest. Dr. Iyer and Dr. Moss disclosed ties to Exact Sciences, Pentax Medical, Redhill Biopharma, Phathom, American Molecular Laboratories, and Takeda.

Publications
Topics
Sections

Antimicrobial resistance is the most common cause of treatment-refractory Helicobacter pylori infection, but before switching antibiotics, clinicians should screen for factors such as treatment nonadherence or inadequate suppression of gastric acid, according to a clinical practice update from the American Gastroenterological Association.

Dr. Shailja C. Shah

“Inadequate acid suppression is associated with H. pylori eradication failure. The use of high-dose and more potent PPIs, PPIs not metabolized by CYP2C19, or potassium-competitive acid blockers, if available, should be considered in cases of refractory H. pylori infection,” wrote Shailja C. Shah, MD, MPH, of Vanderbilt University Medical Center in Nashville, Tenn., and coauthors Prasad G. Iyer, MD, and Steven F. Moss, MD. . Their report is in Gastroenterology.

H. pylori infection is the most common cause of gastric cancer. Although eradication is widely recommended, it can be challenging because of strain diversity, rising antimicrobial resistance, a dearth of recent head-to-head clinical trials, and sparse epidemiologic and sensitivity data, the experts noted. For this reason, before selecting an eradication regimen, it is vital to thoroughly review a patient’s history of antibiotics – for example, any prior macrolide or fluoroquinolone exposure should preclude the use of clarithromycin- or levofloxacin-based regimens “given the high likelihood of resistance,” the experts wrote. They also advised that clinicians should avoid levofloxacin unless the H. pylori strain is known to be sensitive to it or if population rates of levofloxacin resistance rates are known to be less than 15%. However, amoxicillin, tetracycline, and rifabutin resistance are rare, and these agents “can be considered for subsequent therapies in refractory H. pylori infection.”

A longer antimicrobial regimen (such as 14 vs. 7 days) is more likely to eradicate H. pylori. If first-line bismuth quadruple therapy (such as a PPI plus bismuth, metronidazole, and tetracycline) fails, then second-line options include another bismuth-containing quadruple-agent regimen, or triple therapy with rifabutin or levofloxacin plus high-dose dual PPI therapy and amoxicillin. If patient history contains “penicillin allergy” but does not list anaphylaxis, then penicillin allergy testing can help determine if amoxicillin-based regimens are an option. The authors also note that, when used, amoxicillin should be dosed at 2 g/day in divided doses three to four times per day in order to avoid low trough levels because this might be associated with H. pylori eradication failure. For metronidazole, regardless of in vitro resistance, eradication is more likely if patients receive 1.5-2 g/day, in divided doses, with concomitant bismuth.



Treatment nonadherence contributes to refractory H. pylori infection and may be caused by the complexity of the treatment regimen, high pill burden, and side effects. To improve adherence, the experts advised counseling patients on the rationale for the treatment regimen, the dosing instructions, the importance of completing the full course of therapy, and providing anticipatory guidance regarding common side effects. If a patient adheres to second-line treatment and it still fails, then susceptibility testing is advised before starting another regimen. Depending on the results, options may include levofloxacin-based quadruple therapy, another round of bismuth-based quadruple therapy, a PPI plus amoxicillin and rifabutin, or high-dose PPI therapy plus high-dose amoxicillin (2-3 g/day divided across three to four doses).

Other considerations include how to approach patients and caregivers, particularly the elderly and other vulnerable patients, with shared decision-making to help them weigh the potential benefits of continuing to try to eradicate H. pylori against the risk of possible adverse effects and the “inconvenience of repeated exposure to antibiotics and high-dose acid suppression,” the experts wrote. They also advised tracking rates of eradication success and relevant demographic and clinical data, including patients’ antibiotic history. Publicly sharing aggregated, deidentified results can help other local clinicians select eradication regimens. Finally, the use of probiotics and other adjunctive therapies “should be considered experimental” since these have no clear benefit for treating refractory H. pylori infection.

Dr. Shah was funded by an AGA Research Scholar Award and a Veterans Affairs Career Development Award. She reported having no conflicts of interest. Dr. Iyer and Dr. Moss disclosed ties to Exact Sciences, Pentax Medical, Redhill Biopharma, Phathom, American Molecular Laboratories, and Takeda.

Antimicrobial resistance is the most common cause of treatment-refractory Helicobacter pylori infection, but before switching antibiotics, clinicians should screen for factors such as treatment nonadherence or inadequate suppression of gastric acid, according to a clinical practice update from the American Gastroenterological Association.

Dr. Shailja C. Shah

“Inadequate acid suppression is associated with H. pylori eradication failure. The use of high-dose and more potent PPIs, PPIs not metabolized by CYP2C19, or potassium-competitive acid blockers, if available, should be considered in cases of refractory H. pylori infection,” wrote Shailja C. Shah, MD, MPH, of Vanderbilt University Medical Center in Nashville, Tenn., and coauthors Prasad G. Iyer, MD, and Steven F. Moss, MD. . Their report is in Gastroenterology.

H. pylori infection is the most common cause of gastric cancer. Although eradication is widely recommended, it can be challenging because of strain diversity, rising antimicrobial resistance, a dearth of recent head-to-head clinical trials, and sparse epidemiologic and sensitivity data, the experts noted. For this reason, before selecting an eradication regimen, it is vital to thoroughly review a patient’s history of antibiotics – for example, any prior macrolide or fluoroquinolone exposure should preclude the use of clarithromycin- or levofloxacin-based regimens “given the high likelihood of resistance,” the experts wrote. They also advised that clinicians should avoid levofloxacin unless the H. pylori strain is known to be sensitive to it or if population rates of levofloxacin resistance rates are known to be less than 15%. However, amoxicillin, tetracycline, and rifabutin resistance are rare, and these agents “can be considered for subsequent therapies in refractory H. pylori infection.”

A longer antimicrobial regimen (such as 14 vs. 7 days) is more likely to eradicate H. pylori. If first-line bismuth quadruple therapy (such as a PPI plus bismuth, metronidazole, and tetracycline) fails, then second-line options include another bismuth-containing quadruple-agent regimen, or triple therapy with rifabutin or levofloxacin plus high-dose dual PPI therapy and amoxicillin. If patient history contains “penicillin allergy” but does not list anaphylaxis, then penicillin allergy testing can help determine if amoxicillin-based regimens are an option. The authors also note that, when used, amoxicillin should be dosed at 2 g/day in divided doses three to four times per day in order to avoid low trough levels because this might be associated with H. pylori eradication failure. For metronidazole, regardless of in vitro resistance, eradication is more likely if patients receive 1.5-2 g/day, in divided doses, with concomitant bismuth.



Treatment nonadherence contributes to refractory H. pylori infection and may be caused by the complexity of the treatment regimen, high pill burden, and side effects. To improve adherence, the experts advised counseling patients on the rationale for the treatment regimen, the dosing instructions, the importance of completing the full course of therapy, and providing anticipatory guidance regarding common side effects. If a patient adheres to second-line treatment and it still fails, then susceptibility testing is advised before starting another regimen. Depending on the results, options may include levofloxacin-based quadruple therapy, another round of bismuth-based quadruple therapy, a PPI plus amoxicillin and rifabutin, or high-dose PPI therapy plus high-dose amoxicillin (2-3 g/day divided across three to four doses).

Other considerations include how to approach patients and caregivers, particularly the elderly and other vulnerable patients, with shared decision-making to help them weigh the potential benefits of continuing to try to eradicate H. pylori against the risk of possible adverse effects and the “inconvenience of repeated exposure to antibiotics and high-dose acid suppression,” the experts wrote. They also advised tracking rates of eradication success and relevant demographic and clinical data, including patients’ antibiotic history. Publicly sharing aggregated, deidentified results can help other local clinicians select eradication regimens. Finally, the use of probiotics and other adjunctive therapies “should be considered experimental” since these have no clear benefit for treating refractory H. pylori infection.

Dr. Shah was funded by an AGA Research Scholar Award and a Veterans Affairs Career Development Award. She reported having no conflicts of interest. Dr. Iyer and Dr. Moss disclosed ties to Exact Sciences, Pentax Medical, Redhill Biopharma, Phathom, American Molecular Laboratories, and Takeda.

Publications
Publications
Topics
Article Type
Display Headline
AGA Clinical Practice Update: Eradication strategies for H. pylori infection
Display Headline
AGA Clinical Practice Update: Eradication strategies for H. pylori infection
Sections
Article Source

FROM GASTROENTEROLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article
Display survey writer
Reuters content

Combo testing improves CRC screening participation, but not advanced disease detection

Strategies of choice
Article Type
Changed
Fri, 02/05/2021 - 15:14

 

Offering a combination of colonoscopy and fecal immunochemical testing (FIT), either in sequence or by choice, may significantly increase participation in colorectal cancer (CRC) screening, according to a prospective study involving more than 12,000 individuals in Poland.

Still, greater participation did not lead to significantly higher rates of advanced disease detection, reported lead author Nastazja Dagny Pilonis, MD, of the Maria Sklodowska-Curie National Research Institute of Oncology, Warsaw, and colleagues in Gastroenterology.

According to the investigators, screening programs that offer colonoscopy and FIT are more effective than those that offer colonoscopy alone, but an optimal combination protocol has yet to be established, and some parts of the world still rely upon a single diagnostic method.

“In Europe, CRC screening programs often implement only one screening modality: colonoscopy, sigmoidoscopy, or stool testing, depending on the health care provider,” the investigators wrote in Gastroenterology. They noted, however, that national guidelines in the United States recommend strategies that include more than one screening method. “‘One-size-fits-all’ approaches to CRC screening do not result in satisfactory participation” because of behavioral, cultural, and socioeconomic variation among individuals.

To improve understanding of the best ways to improve participation, the investigators conducted a prospective randomized trial, PICCOLINO, via the Polish Colonoscopy Screening Program. In total, 12,485 eligible individuals aged between 55 and 64 years received postal invitations to participate in CRC screening. Individuals were randomized in a 1:1:1 ratio into one of three mailing protocols, each of which involved an initial invitation, and, if needed, a second invitation that offered the following:

  • Control group: colonoscopy, with nonresponders receiving the same invitation again
  • Sequential group: colonoscopy, with nonresponders or refusers receiving a second invitation that offered FIT
  • Choice group: choice between colonoscopy or FIT, with nonresponders receiving the same invitation again

The primary outcome was participation in screening within 18 weeks of enrollment. The secondary outcome was diagnostic yield for either advanced adenoma or CRC.

Out of the three groups, the control group had the lowest participation rate, at 17.5%, compared with 25.8% for the sequential group and 26.5% for the choice group. Multivariable logistic regression showed that individuals in the sequential and choice groups had 64% and 70% higher rates of participation, respectively. Across all groups, age of 60 years or older predicted 12% higher likelihood of participation; in contrast, location more than 40 kilometers from a testing center was associated with an 18% decrease in participation, compared with individuals who lived less than 20 kilometers away.

While the control and sequential groups had similar rates of colonoscopy participation, at 17.5% and 15.9%, respectively (P = .788), this rate was significantly lower, at 8.5%, in the choice group (P = .001). Conversely, the sequential group had a significantly lower rate of FITs than the choice group, at 9.9% versus 17.9%, respectively (P = .001). Among participants with a positive FIT, diagnostic work-up colonoscopies were performed in 70.0% of those in the sequential group and 73.3% in the choice group, “despite active call-recall efforts.”

Across all invited individuals, advanced disease detection rates were similar across groups, at 1.1% for both the control and the sequential group and 1.2% for the choice group. Among those who were actually screened, the control group had a slightly higher diagnostic yield for advanced neoplasia, at 6.5%, compared with 4.2% in the sequential group and 4.4% in the choice group; however, these differences were not statistically significant. In contrast, significantly more adenomas of any kind were detected in the control and sequential groups (5.6% for both) than the choice group (3.9%) (P < .001).

“Although the strategies which included FIT showed higher participation rates than the strategy of offering colonoscopy alone, these strategies did not result in increased detection rates of advanced neoplasia in the intention to screen analysis,” the investigators wrote. “An absolute increase in participation rates of 8%-10% seems insufficient to translate into higher advanced neoplasia detection at the population level.”

Dr. Pilonis and colleagues also suggested that the relatively low rate of diagnostic colonoscopy after positive FIT contributed to the suboptimal diagnostic yield.

“These rates are unsatisfactory taking into account significant call-recall efforts, but are within the range reported in other studies,” they wrote.

They also wrote that their study compared participation and detection between one-time colonoscopy and one-time screening strategies combining colonoscopy and FIT. In acknowledging this, they noted that these approaches have different screening intervals and uptake over time: “FIT has been shown to achieve higher participation rates than colonoscopy for one time screening, but its uptake over several rounds may not be superior to one time colonoscopy.” Furthermore, detection rates of the sequential or choice strategies for advanced disease may rise over time with further implementation, so the one-time screening may not be sufficient to reveal what could become significant differences.

The study was funded by the Polish Ministry of Health, the Polish Foundation of Gastroenterology, and the Centre of Postgraduate Medical Education in Warsaw. FITs, materials, and reagents were provided by Eiken Chemical. The investigators disclosed relationships with Boston Scientific, AbbVie, Olympus, and others.

Body

 

Multiple strategies have been validated for CRC screening, showing different characteristics that may affect their acceptability. Indeed, dislike of specific tests has been reported as a barrier to screening for some patients. While adopting more than one method to account for subjects’ preferences would then seem a potentially effective approach to increase uptake, most population-based programs are offering only one screening modality.

The PICCOLINO study, conducted within the Polish CRC screening program, showed that offering fecal immunochemical tests (FIT) together with colonoscopy, either as an active choice or in sequence, may substantially improve participation as compared with the offer of colonoscopy alone.

The combination approaches offered the opportunity to respond to the screening invitation also to those subjects who prefer a noninvasive test, which may have limited the impact of organizational barriers on participation. Making the test immediately available with the invitation letter likely helped enhance the response rate in the choice group, which may explain the high proportion of subjects opting for FIT. Offering FIT might also reduce disparities related to distance from the endoscopy center seen when using primary colonoscopy screening. A longer follow-up is needed to assess the neoplasia yield of the combination strategies, accounting for the cumulative detection rate of FIT over several rounds.

This study shows that implementing combination approaches within population-based programs represents a feasible option, although the low compliance with referral for colonoscopy assessment would suggest the need to implement communication efforts specifically addressing negative attitudes to colonoscopy among subjects opting for FIT.

Carlo Senore, MD, MSc, is an epidemiologist at the epidemiology and screening unit–CPO at the University Hospital Città della Salute e della Scienza in Turin, Italy; he is the director of Piedmont Region Screening Committee. He has no conflicts.

Publications
Topics
Sections
Body

 

Multiple strategies have been validated for CRC screening, showing different characteristics that may affect their acceptability. Indeed, dislike of specific tests has been reported as a barrier to screening for some patients. While adopting more than one method to account for subjects’ preferences would then seem a potentially effective approach to increase uptake, most population-based programs are offering only one screening modality.

The PICCOLINO study, conducted within the Polish CRC screening program, showed that offering fecal immunochemical tests (FIT) together with colonoscopy, either as an active choice or in sequence, may substantially improve participation as compared with the offer of colonoscopy alone.

The combination approaches offered the opportunity to respond to the screening invitation also to those subjects who prefer a noninvasive test, which may have limited the impact of organizational barriers on participation. Making the test immediately available with the invitation letter likely helped enhance the response rate in the choice group, which may explain the high proportion of subjects opting for FIT. Offering FIT might also reduce disparities related to distance from the endoscopy center seen when using primary colonoscopy screening. A longer follow-up is needed to assess the neoplasia yield of the combination strategies, accounting for the cumulative detection rate of FIT over several rounds.

This study shows that implementing combination approaches within population-based programs represents a feasible option, although the low compliance with referral for colonoscopy assessment would suggest the need to implement communication efforts specifically addressing negative attitudes to colonoscopy among subjects opting for FIT.

Carlo Senore, MD, MSc, is an epidemiologist at the epidemiology and screening unit–CPO at the University Hospital Città della Salute e della Scienza in Turin, Italy; he is the director of Piedmont Region Screening Committee. He has no conflicts.

Body

 

Multiple strategies have been validated for CRC screening, showing different characteristics that may affect their acceptability. Indeed, dislike of specific tests has been reported as a barrier to screening for some patients. While adopting more than one method to account for subjects’ preferences would then seem a potentially effective approach to increase uptake, most population-based programs are offering only one screening modality.

The PICCOLINO study, conducted within the Polish CRC screening program, showed that offering fecal immunochemical tests (FIT) together with colonoscopy, either as an active choice or in sequence, may substantially improve participation as compared with the offer of colonoscopy alone.

The combination approaches offered the opportunity to respond to the screening invitation also to those subjects who prefer a noninvasive test, which may have limited the impact of organizational barriers on participation. Making the test immediately available with the invitation letter likely helped enhance the response rate in the choice group, which may explain the high proportion of subjects opting for FIT. Offering FIT might also reduce disparities related to distance from the endoscopy center seen when using primary colonoscopy screening. A longer follow-up is needed to assess the neoplasia yield of the combination strategies, accounting for the cumulative detection rate of FIT over several rounds.

This study shows that implementing combination approaches within population-based programs represents a feasible option, although the low compliance with referral for colonoscopy assessment would suggest the need to implement communication efforts specifically addressing negative attitudes to colonoscopy among subjects opting for FIT.

Carlo Senore, MD, MSc, is an epidemiologist at the epidemiology and screening unit–CPO at the University Hospital Città della Salute e della Scienza in Turin, Italy; he is the director of Piedmont Region Screening Committee. He has no conflicts.

Title
Strategies of choice
Strategies of choice

 

Offering a combination of colonoscopy and fecal immunochemical testing (FIT), either in sequence or by choice, may significantly increase participation in colorectal cancer (CRC) screening, according to a prospective study involving more than 12,000 individuals in Poland.

Still, greater participation did not lead to significantly higher rates of advanced disease detection, reported lead author Nastazja Dagny Pilonis, MD, of the Maria Sklodowska-Curie National Research Institute of Oncology, Warsaw, and colleagues in Gastroenterology.

According to the investigators, screening programs that offer colonoscopy and FIT are more effective than those that offer colonoscopy alone, but an optimal combination protocol has yet to be established, and some parts of the world still rely upon a single diagnostic method.

“In Europe, CRC screening programs often implement only one screening modality: colonoscopy, sigmoidoscopy, or stool testing, depending on the health care provider,” the investigators wrote in Gastroenterology. They noted, however, that national guidelines in the United States recommend strategies that include more than one screening method. “‘One-size-fits-all’ approaches to CRC screening do not result in satisfactory participation” because of behavioral, cultural, and socioeconomic variation among individuals.

To improve understanding of the best ways to improve participation, the investigators conducted a prospective randomized trial, PICCOLINO, via the Polish Colonoscopy Screening Program. In total, 12,485 eligible individuals aged between 55 and 64 years received postal invitations to participate in CRC screening. Individuals were randomized in a 1:1:1 ratio into one of three mailing protocols, each of which involved an initial invitation, and, if needed, a second invitation that offered the following:

  • Control group: colonoscopy, with nonresponders receiving the same invitation again
  • Sequential group: colonoscopy, with nonresponders or refusers receiving a second invitation that offered FIT
  • Choice group: choice between colonoscopy or FIT, with nonresponders receiving the same invitation again

The primary outcome was participation in screening within 18 weeks of enrollment. The secondary outcome was diagnostic yield for either advanced adenoma or CRC.

Out of the three groups, the control group had the lowest participation rate, at 17.5%, compared with 25.8% for the sequential group and 26.5% for the choice group. Multivariable logistic regression showed that individuals in the sequential and choice groups had 64% and 70% higher rates of participation, respectively. Across all groups, age of 60 years or older predicted 12% higher likelihood of participation; in contrast, location more than 40 kilometers from a testing center was associated with an 18% decrease in participation, compared with individuals who lived less than 20 kilometers away.

While the control and sequential groups had similar rates of colonoscopy participation, at 17.5% and 15.9%, respectively (P = .788), this rate was significantly lower, at 8.5%, in the choice group (P = .001). Conversely, the sequential group had a significantly lower rate of FITs than the choice group, at 9.9% versus 17.9%, respectively (P = .001). Among participants with a positive FIT, diagnostic work-up colonoscopies were performed in 70.0% of those in the sequential group and 73.3% in the choice group, “despite active call-recall efforts.”

Across all invited individuals, advanced disease detection rates were similar across groups, at 1.1% for both the control and the sequential group and 1.2% for the choice group. Among those who were actually screened, the control group had a slightly higher diagnostic yield for advanced neoplasia, at 6.5%, compared with 4.2% in the sequential group and 4.4% in the choice group; however, these differences were not statistically significant. In contrast, significantly more adenomas of any kind were detected in the control and sequential groups (5.6% for both) than the choice group (3.9%) (P < .001).

“Although the strategies which included FIT showed higher participation rates than the strategy of offering colonoscopy alone, these strategies did not result in increased detection rates of advanced neoplasia in the intention to screen analysis,” the investigators wrote. “An absolute increase in participation rates of 8%-10% seems insufficient to translate into higher advanced neoplasia detection at the population level.”

Dr. Pilonis and colleagues also suggested that the relatively low rate of diagnostic colonoscopy after positive FIT contributed to the suboptimal diagnostic yield.

“These rates are unsatisfactory taking into account significant call-recall efforts, but are within the range reported in other studies,” they wrote.

They also wrote that their study compared participation and detection between one-time colonoscopy and one-time screening strategies combining colonoscopy and FIT. In acknowledging this, they noted that these approaches have different screening intervals and uptake over time: “FIT has been shown to achieve higher participation rates than colonoscopy for one time screening, but its uptake over several rounds may not be superior to one time colonoscopy.” Furthermore, detection rates of the sequential or choice strategies for advanced disease may rise over time with further implementation, so the one-time screening may not be sufficient to reveal what could become significant differences.

The study was funded by the Polish Ministry of Health, the Polish Foundation of Gastroenterology, and the Centre of Postgraduate Medical Education in Warsaw. FITs, materials, and reagents were provided by Eiken Chemical. The investigators disclosed relationships with Boston Scientific, AbbVie, Olympus, and others.

 

Offering a combination of colonoscopy and fecal immunochemical testing (FIT), either in sequence or by choice, may significantly increase participation in colorectal cancer (CRC) screening, according to a prospective study involving more than 12,000 individuals in Poland.

Still, greater participation did not lead to significantly higher rates of advanced disease detection, reported lead author Nastazja Dagny Pilonis, MD, of the Maria Sklodowska-Curie National Research Institute of Oncology, Warsaw, and colleagues in Gastroenterology.

According to the investigators, screening programs that offer colonoscopy and FIT are more effective than those that offer colonoscopy alone, but an optimal combination protocol has yet to be established, and some parts of the world still rely upon a single diagnostic method.

“In Europe, CRC screening programs often implement only one screening modality: colonoscopy, sigmoidoscopy, or stool testing, depending on the health care provider,” the investigators wrote in Gastroenterology. They noted, however, that national guidelines in the United States recommend strategies that include more than one screening method. “‘One-size-fits-all’ approaches to CRC screening do not result in satisfactory participation” because of behavioral, cultural, and socioeconomic variation among individuals.

To improve understanding of the best ways to improve participation, the investigators conducted a prospective randomized trial, PICCOLINO, via the Polish Colonoscopy Screening Program. In total, 12,485 eligible individuals aged between 55 and 64 years received postal invitations to participate in CRC screening. Individuals were randomized in a 1:1:1 ratio into one of three mailing protocols, each of which involved an initial invitation, and, if needed, a second invitation that offered the following:

  • Control group: colonoscopy, with nonresponders receiving the same invitation again
  • Sequential group: colonoscopy, with nonresponders or refusers receiving a second invitation that offered FIT
  • Choice group: choice between colonoscopy or FIT, with nonresponders receiving the same invitation again

The primary outcome was participation in screening within 18 weeks of enrollment. The secondary outcome was diagnostic yield for either advanced adenoma or CRC.

Out of the three groups, the control group had the lowest participation rate, at 17.5%, compared with 25.8% for the sequential group and 26.5% for the choice group. Multivariable logistic regression showed that individuals in the sequential and choice groups had 64% and 70% higher rates of participation, respectively. Across all groups, age of 60 years or older predicted 12% higher likelihood of participation; in contrast, location more than 40 kilometers from a testing center was associated with an 18% decrease in participation, compared with individuals who lived less than 20 kilometers away.

While the control and sequential groups had similar rates of colonoscopy participation, at 17.5% and 15.9%, respectively (P = .788), this rate was significantly lower, at 8.5%, in the choice group (P = .001). Conversely, the sequential group had a significantly lower rate of FITs than the choice group, at 9.9% versus 17.9%, respectively (P = .001). Among participants with a positive FIT, diagnostic work-up colonoscopies were performed in 70.0% of those in the sequential group and 73.3% in the choice group, “despite active call-recall efforts.”

Across all invited individuals, advanced disease detection rates were similar across groups, at 1.1% for both the control and the sequential group and 1.2% for the choice group. Among those who were actually screened, the control group had a slightly higher diagnostic yield for advanced neoplasia, at 6.5%, compared with 4.2% in the sequential group and 4.4% in the choice group; however, these differences were not statistically significant. In contrast, significantly more adenomas of any kind were detected in the control and sequential groups (5.6% for both) than the choice group (3.9%) (P < .001).

“Although the strategies which included FIT showed higher participation rates than the strategy of offering colonoscopy alone, these strategies did not result in increased detection rates of advanced neoplasia in the intention to screen analysis,” the investigators wrote. “An absolute increase in participation rates of 8%-10% seems insufficient to translate into higher advanced neoplasia detection at the population level.”

Dr. Pilonis and colleagues also suggested that the relatively low rate of diagnostic colonoscopy after positive FIT contributed to the suboptimal diagnostic yield.

“These rates are unsatisfactory taking into account significant call-recall efforts, but are within the range reported in other studies,” they wrote.

They also wrote that their study compared participation and detection between one-time colonoscopy and one-time screening strategies combining colonoscopy and FIT. In acknowledging this, they noted that these approaches have different screening intervals and uptake over time: “FIT has been shown to achieve higher participation rates than colonoscopy for one time screening, but its uptake over several rounds may not be superior to one time colonoscopy.” Furthermore, detection rates of the sequential or choice strategies for advanced disease may rise over time with further implementation, so the one-time screening may not be sufficient to reveal what could become significant differences.

The study was funded by the Polish Ministry of Health, the Polish Foundation of Gastroenterology, and the Centre of Postgraduate Medical Education in Warsaw. FITs, materials, and reagents were provided by Eiken Chemical. The investigators disclosed relationships with Boston Scientific, AbbVie, Olympus, and others.

Publications
Publications
Topics
Article Type
Sections
Article Source

FROM GASTROENTEROLOGY

Disallow All Ads
Content Gating
No Gating (article Unlocked/Free)
Alternative CME
Disqus Comments
Default
Use ProPublica
Hide sidebar & use full width
render the right sidebar.
Conference Recap Checkbox
Not Conference Recap
Clinical Edge
Display the Slideshow in this Article
Medscape Article