User login
High Marks for New CAR T Toxicity Grading Tool
“Hematotoxicity after CAR T is common and clinically relevant, but it also remains poorly understood [with] a high degree of heterogeneity in terms of grading its clinical management,” said first author Kai Rejeski, MD, in presenting on the findings at the 6th European CAR T-cell Meeting, held in Spain and jointly sponsored by the European Society for Blood and Marrow Transplantation (EBMT) and the European Hematology Association (EHA).
“We hope that this novel grading system helps with this by enabling harmonized reporting using the same nomenclature and allowing the comparison of the expected incidence rates of grade 3 or higher [hematological toxicities] across several disease entities and CAR T products,” said Dr. Rejeski, of the Adult BMT (Blood Marrow Transplant) and Cellular Therapy Service, Memorial Sloan Kettering Cancer Center, New York City.
ICAHT Grading System
In a recent meta-analysis, Dr. Rejeski and his team found that infections are the cause of as many as 49% of non–relapse related deaths after CAR T-cell therapy, representing the most common cause of death and numbering significantly more than the more prominent causes of cytokine release syndrome (CRS) or immune effector cell–associated neurotoxicity (ICANS), which paradoxically have been the focus of significantly more research. In addition, the authors have reported substantial inconsistency among CAR T centers in the grading and management of the post–CAR T cytopenias that can cause those infections, underscoring the need for better guidelines.
“The narrative around CAR T toxicity has long centered on CRS and ICANS as novel and prototypical side effects with distinct management protocols,” Dr. Rejeski said in an interview. “However, it is cytopenias and the associated infections that drive nonrelapse mortality after CAR T.”
To address the need, the EHA and EBMT established the grading system for Immune Effector Cell–Associated HematoToxicity (ICAHT) that is applicable across disease types, indications, and treatment settings.
The details of the grading system were published in September 2023 in the journal Blood. The new system, which specifically focuses on neutrophil count and timing, importantly addresses the biphasic nature of ICAHT by distinguishing “early” ICAHT, occurring within 30 days of the CAR T administration, and “late” ICAHT, occurring more that 30 days following the treatment.
By contrast, conventional grading scales for CAR T–related cytopenias, such as the Common Terminology Criteria for Adverse Events (CTCAE) scale, “neither reflect the unique quality of post–CAR T neutrophil recovery, nor do they reflect the inherent risk of infections due to protracted neutropenia,” the authors report in the study.
Real-World Evaluation
To assess the ICAHT grading system’s relevance in a real-world clinical setting of CAR T-cell therapy recipients, Dr. Rejeski and colleagues conducted a multicenter observational study, published in January 2024 in Blood Advances.
The study involved 549 patients at 12 international CAR T centers treated with BCMA- or CD19- directed CAR T therapy for relapsed/refractory B-cell malignancies.
Of the patients, 112 were treated for multiple myeloma (MM), 334 for large B cell lymphoma (LBCL), and 103 for mantle cell lymphoma (MCL).
Using the grading system, grade 3 (severe) or 4 (life-threatening) ICAHT (n = 125), was found to be strongly associated with key factors including a cumulative duration of severe neutropenia (P < .0001), the presence of multilineage cytopenias, such as severe thrombocytopenia (90%, compared with 46% in nonsevere ICAHT) and severe anemia (92% vs 49%; both P < .001), as well as the use of platelet and red blood cell transfusions.
Grade 3 or higher ICAHT was more common in patients with MCL (28%), compared with LBCL (23%) and MM (15%).
Key factors at baseline that were independently associated with severe ICAHT after multivariate adjustment included the presence of bone marrow infiltration, increased serum LDH levels, elevated CAR-HEMATOTOX scores (all P < .001), and receipt of CD28z costimulatory domain products, including axi-cel or brexu-cel (P = .01).
Those with grade 3 or higher ICAHT scores had a significantly higher rate of severe infections, compared with lower ICAHT scores (49% vs 13%; P < .0001), as well as increased nonrelapse mortality (14% vs 4.5%; P < .0001), primarily attributable to fatal infections.
Survival outcomes were also worse with grade 3 or higher ICAHT, including significantly lower rates of 1-year progression-free survival (35% vs 51%) and 1-year overall survival (52% vs 73%; both P < .0001).
Grade 3 or higher ICAHT was also significantly associated with prolonged hospital stays (median 21 vs 16 days; P < .0001).
However, contrary to findings from some previous studies, the current study showed no association between ICAHT severity and the prior administration of autologous stem cell transplant.
The number of prior treatment lines was not associated with grade 3 or higher ICAHT. However, grade 3 or higher CRS was more common as a cotoxicity (15% vs 5% without severe ICAHT), as was severe ICANS (26% vs 13%; both P < .001).
Notably, ICAHT grading showed superiority in the prediction of severe infections, compared with CTCAE grading (c-index 0.73 vs 0.55, P < .0001 vs nonsignificant).
While mild to moderate toxicity after CAR T-cell therapy has been associated with more favorable outcomes, the poor survival rates associated with severe ICAHT “underscore that high-grade toxicity and inferior treatment outcomes often go hand-in-hand,” the authors write.
Conversely, “the patients with grade 1 or 2 ICAHT exhibited excellent treatment outcomes in our study,” they point out.
Recommendations in Clinical Practice
For clinical guidance, the ICAHT grading system provides best practice recommendations based on severity for diagnostic work-up and management, such as measures including use of granulocyte-colony stimulating factor (G-CSF), anti-infective prophylaxis and stem cell boosts.
The authors add that preinfusion scoring systems, including the CAR-HEMATOTOX prognostic score, may be optimized by ICAHT grading in terms of modeling for severe or life-threatening ICAHT as an important endpoint.
“We have had an absence of the standardized severity-based guidelines that we know very well for CRS and ICANS, both in terms of the diagnostic work-up and the grading but also the management,” Dr. Rejeski said at the meeting.
“We hope that the new ICAHT grading focuses future research efforts to not only understand this important side effect better, but also develop specific management strategies that mitigate the risk of infections in high-risk patients,” Dr. Rejeski added.
“The multiply validated CAR-HEMATOTOX score, assessed at time of lymphodepletion, may be helpful in this regard,” he added.
An accompanying editorial published with the guidelines underscored that “this is the first such guideline by a major organization and is a much-needed development for the management of this important CAR T-cell–associated toxicity.”
The improved standardized reporting of ICAHT “could also inform hematotoxicity management protocols,” said the editorial authors, David Qualls, MD, of the Memorial Sloan Kettering Cancer Center in New York City and Caron Jacobson, MD, of the Dana-Farber Cancer Institute, in Boston, Massachusetts.
“While providing comprehensive recommendations for ICAHT, the EHA/EBMT guidelines also highlight important gaps in our current knowledge of ICAHT, which are significant,” the editorial authors add.
Further commenting, Ulrich Jaeger, MD, a professor of hematology at the Medical University of Vienna, Vienna, Austria, agreed that the research fills an important need in post–CAR T-cell therapy management.
“Dr. Rejeski´s work is really seminal in the field and confirmed by validation cohorts in other centers,” he said in an interview. “I think the story is absolutely clear. It will be of increasing importance, with more patients surviving. [The system] will have to be adapted to novel indications as well.”
Dr. Rejeski disclosed ties with Kite/Gilead, Novartis, GMS/Celgene, and Pierre-Fabre. Jaeger reports relationships with Novartis, Gilead Sciences, Celgene/BMS, Janssen, Roche, Miltenyi Biotec, and Innovative Medicines Initiative.
“Hematotoxicity after CAR T is common and clinically relevant, but it also remains poorly understood [with] a high degree of heterogeneity in terms of grading its clinical management,” said first author Kai Rejeski, MD, in presenting on the findings at the 6th European CAR T-cell Meeting, held in Spain and jointly sponsored by the European Society for Blood and Marrow Transplantation (EBMT) and the European Hematology Association (EHA).
“We hope that this novel grading system helps with this by enabling harmonized reporting using the same nomenclature and allowing the comparison of the expected incidence rates of grade 3 or higher [hematological toxicities] across several disease entities and CAR T products,” said Dr. Rejeski, of the Adult BMT (Blood Marrow Transplant) and Cellular Therapy Service, Memorial Sloan Kettering Cancer Center, New York City.
ICAHT Grading System
In a recent meta-analysis, Dr. Rejeski and his team found that infections are the cause of as many as 49% of non–relapse related deaths after CAR T-cell therapy, representing the most common cause of death and numbering significantly more than the more prominent causes of cytokine release syndrome (CRS) or immune effector cell–associated neurotoxicity (ICANS), which paradoxically have been the focus of significantly more research. In addition, the authors have reported substantial inconsistency among CAR T centers in the grading and management of the post–CAR T cytopenias that can cause those infections, underscoring the need for better guidelines.
“The narrative around CAR T toxicity has long centered on CRS and ICANS as novel and prototypical side effects with distinct management protocols,” Dr. Rejeski said in an interview. “However, it is cytopenias and the associated infections that drive nonrelapse mortality after CAR T.”
To address the need, the EHA and EBMT established the grading system for Immune Effector Cell–Associated HematoToxicity (ICAHT) that is applicable across disease types, indications, and treatment settings.
The details of the grading system were published in September 2023 in the journal Blood. The new system, which specifically focuses on neutrophil count and timing, importantly addresses the biphasic nature of ICAHT by distinguishing “early” ICAHT, occurring within 30 days of the CAR T administration, and “late” ICAHT, occurring more that 30 days following the treatment.
By contrast, conventional grading scales for CAR T–related cytopenias, such as the Common Terminology Criteria for Adverse Events (CTCAE) scale, “neither reflect the unique quality of post–CAR T neutrophil recovery, nor do they reflect the inherent risk of infections due to protracted neutropenia,” the authors report in the study.
Real-World Evaluation
To assess the ICAHT grading system’s relevance in a real-world clinical setting of CAR T-cell therapy recipients, Dr. Rejeski and colleagues conducted a multicenter observational study, published in January 2024 in Blood Advances.
The study involved 549 patients at 12 international CAR T centers treated with BCMA- or CD19- directed CAR T therapy for relapsed/refractory B-cell malignancies.
Of the patients, 112 were treated for multiple myeloma (MM), 334 for large B cell lymphoma (LBCL), and 103 for mantle cell lymphoma (MCL).
Using the grading system, grade 3 (severe) or 4 (life-threatening) ICAHT (n = 125), was found to be strongly associated with key factors including a cumulative duration of severe neutropenia (P < .0001), the presence of multilineage cytopenias, such as severe thrombocytopenia (90%, compared with 46% in nonsevere ICAHT) and severe anemia (92% vs 49%; both P < .001), as well as the use of platelet and red blood cell transfusions.
Grade 3 or higher ICAHT was more common in patients with MCL (28%), compared with LBCL (23%) and MM (15%).
Key factors at baseline that were independently associated with severe ICAHT after multivariate adjustment included the presence of bone marrow infiltration, increased serum LDH levels, elevated CAR-HEMATOTOX scores (all P < .001), and receipt of CD28z costimulatory domain products, including axi-cel or brexu-cel (P = .01).
Those with grade 3 or higher ICAHT scores had a significantly higher rate of severe infections, compared with lower ICAHT scores (49% vs 13%; P < .0001), as well as increased nonrelapse mortality (14% vs 4.5%; P < .0001), primarily attributable to fatal infections.
Survival outcomes were also worse with grade 3 or higher ICAHT, including significantly lower rates of 1-year progression-free survival (35% vs 51%) and 1-year overall survival (52% vs 73%; both P < .0001).
Grade 3 or higher ICAHT was also significantly associated with prolonged hospital stays (median 21 vs 16 days; P < .0001).
However, contrary to findings from some previous studies, the current study showed no association between ICAHT severity and the prior administration of autologous stem cell transplant.
The number of prior treatment lines was not associated with grade 3 or higher ICAHT. However, grade 3 or higher CRS was more common as a cotoxicity (15% vs 5% without severe ICAHT), as was severe ICANS (26% vs 13%; both P < .001).
Notably, ICAHT grading showed superiority in the prediction of severe infections, compared with CTCAE grading (c-index 0.73 vs 0.55, P < .0001 vs nonsignificant).
While mild to moderate toxicity after CAR T-cell therapy has been associated with more favorable outcomes, the poor survival rates associated with severe ICAHT “underscore that high-grade toxicity and inferior treatment outcomes often go hand-in-hand,” the authors write.
Conversely, “the patients with grade 1 or 2 ICAHT exhibited excellent treatment outcomes in our study,” they point out.
Recommendations in Clinical Practice
For clinical guidance, the ICAHT grading system provides best practice recommendations based on severity for diagnostic work-up and management, such as measures including use of granulocyte-colony stimulating factor (G-CSF), anti-infective prophylaxis and stem cell boosts.
The authors add that preinfusion scoring systems, including the CAR-HEMATOTOX prognostic score, may be optimized by ICAHT grading in terms of modeling for severe or life-threatening ICAHT as an important endpoint.
“We have had an absence of the standardized severity-based guidelines that we know very well for CRS and ICANS, both in terms of the diagnostic work-up and the grading but also the management,” Dr. Rejeski said at the meeting.
“We hope that the new ICAHT grading focuses future research efforts to not only understand this important side effect better, but also develop specific management strategies that mitigate the risk of infections in high-risk patients,” Dr. Rejeski added.
“The multiply validated CAR-HEMATOTOX score, assessed at time of lymphodepletion, may be helpful in this regard,” he added.
An accompanying editorial published with the guidelines underscored that “this is the first such guideline by a major organization and is a much-needed development for the management of this important CAR T-cell–associated toxicity.”
The improved standardized reporting of ICAHT “could also inform hematotoxicity management protocols,” said the editorial authors, David Qualls, MD, of the Memorial Sloan Kettering Cancer Center in New York City and Caron Jacobson, MD, of the Dana-Farber Cancer Institute, in Boston, Massachusetts.
“While providing comprehensive recommendations for ICAHT, the EHA/EBMT guidelines also highlight important gaps in our current knowledge of ICAHT, which are significant,” the editorial authors add.
Further commenting, Ulrich Jaeger, MD, a professor of hematology at the Medical University of Vienna, Vienna, Austria, agreed that the research fills an important need in post–CAR T-cell therapy management.
“Dr. Rejeski´s work is really seminal in the field and confirmed by validation cohorts in other centers,” he said in an interview. “I think the story is absolutely clear. It will be of increasing importance, with more patients surviving. [The system] will have to be adapted to novel indications as well.”
Dr. Rejeski disclosed ties with Kite/Gilead, Novartis, GMS/Celgene, and Pierre-Fabre. Jaeger reports relationships with Novartis, Gilead Sciences, Celgene/BMS, Janssen, Roche, Miltenyi Biotec, and Innovative Medicines Initiative.
“Hematotoxicity after CAR T is common and clinically relevant, but it also remains poorly understood [with] a high degree of heterogeneity in terms of grading its clinical management,” said first author Kai Rejeski, MD, in presenting on the findings at the 6th European CAR T-cell Meeting, held in Spain and jointly sponsored by the European Society for Blood and Marrow Transplantation (EBMT) and the European Hematology Association (EHA).
“We hope that this novel grading system helps with this by enabling harmonized reporting using the same nomenclature and allowing the comparison of the expected incidence rates of grade 3 or higher [hematological toxicities] across several disease entities and CAR T products,” said Dr. Rejeski, of the Adult BMT (Blood Marrow Transplant) and Cellular Therapy Service, Memorial Sloan Kettering Cancer Center, New York City.
ICAHT Grading System
In a recent meta-analysis, Dr. Rejeski and his team found that infections are the cause of as many as 49% of non–relapse related deaths after CAR T-cell therapy, representing the most common cause of death and numbering significantly more than the more prominent causes of cytokine release syndrome (CRS) or immune effector cell–associated neurotoxicity (ICANS), which paradoxically have been the focus of significantly more research. In addition, the authors have reported substantial inconsistency among CAR T centers in the grading and management of the post–CAR T cytopenias that can cause those infections, underscoring the need for better guidelines.
“The narrative around CAR T toxicity has long centered on CRS and ICANS as novel and prototypical side effects with distinct management protocols,” Dr. Rejeski said in an interview. “However, it is cytopenias and the associated infections that drive nonrelapse mortality after CAR T.”
To address the need, the EHA and EBMT established the grading system for Immune Effector Cell–Associated HematoToxicity (ICAHT) that is applicable across disease types, indications, and treatment settings.
The details of the grading system were published in September 2023 in the journal Blood. The new system, which specifically focuses on neutrophil count and timing, importantly addresses the biphasic nature of ICAHT by distinguishing “early” ICAHT, occurring within 30 days of the CAR T administration, and “late” ICAHT, occurring more that 30 days following the treatment.
By contrast, conventional grading scales for CAR T–related cytopenias, such as the Common Terminology Criteria for Adverse Events (CTCAE) scale, “neither reflect the unique quality of post–CAR T neutrophil recovery, nor do they reflect the inherent risk of infections due to protracted neutropenia,” the authors report in the study.
Real-World Evaluation
To assess the ICAHT grading system’s relevance in a real-world clinical setting of CAR T-cell therapy recipients, Dr. Rejeski and colleagues conducted a multicenter observational study, published in January 2024 in Blood Advances.
The study involved 549 patients at 12 international CAR T centers treated with BCMA- or CD19- directed CAR T therapy for relapsed/refractory B-cell malignancies.
Of the patients, 112 were treated for multiple myeloma (MM), 334 for large B cell lymphoma (LBCL), and 103 for mantle cell lymphoma (MCL).
Using the grading system, grade 3 (severe) or 4 (life-threatening) ICAHT (n = 125), was found to be strongly associated with key factors including a cumulative duration of severe neutropenia (P < .0001), the presence of multilineage cytopenias, such as severe thrombocytopenia (90%, compared with 46% in nonsevere ICAHT) and severe anemia (92% vs 49%; both P < .001), as well as the use of platelet and red blood cell transfusions.
Grade 3 or higher ICAHT was more common in patients with MCL (28%), compared with LBCL (23%) and MM (15%).
Key factors at baseline that were independently associated with severe ICAHT after multivariate adjustment included the presence of bone marrow infiltration, increased serum LDH levels, elevated CAR-HEMATOTOX scores (all P < .001), and receipt of CD28z costimulatory domain products, including axi-cel or brexu-cel (P = .01).
Those with grade 3 or higher ICAHT scores had a significantly higher rate of severe infections, compared with lower ICAHT scores (49% vs 13%; P < .0001), as well as increased nonrelapse mortality (14% vs 4.5%; P < .0001), primarily attributable to fatal infections.
Survival outcomes were also worse with grade 3 or higher ICAHT, including significantly lower rates of 1-year progression-free survival (35% vs 51%) and 1-year overall survival (52% vs 73%; both P < .0001).
Grade 3 or higher ICAHT was also significantly associated with prolonged hospital stays (median 21 vs 16 days; P < .0001).
However, contrary to findings from some previous studies, the current study showed no association between ICAHT severity and the prior administration of autologous stem cell transplant.
The number of prior treatment lines was not associated with grade 3 or higher ICAHT. However, grade 3 or higher CRS was more common as a cotoxicity (15% vs 5% without severe ICAHT), as was severe ICANS (26% vs 13%; both P < .001).
Notably, ICAHT grading showed superiority in the prediction of severe infections, compared with CTCAE grading (c-index 0.73 vs 0.55, P < .0001 vs nonsignificant).
While mild to moderate toxicity after CAR T-cell therapy has been associated with more favorable outcomes, the poor survival rates associated with severe ICAHT “underscore that high-grade toxicity and inferior treatment outcomes often go hand-in-hand,” the authors write.
Conversely, “the patients with grade 1 or 2 ICAHT exhibited excellent treatment outcomes in our study,” they point out.
Recommendations in Clinical Practice
For clinical guidance, the ICAHT grading system provides best practice recommendations based on severity for diagnostic work-up and management, such as measures including use of granulocyte-colony stimulating factor (G-CSF), anti-infective prophylaxis and stem cell boosts.
The authors add that preinfusion scoring systems, including the CAR-HEMATOTOX prognostic score, may be optimized by ICAHT grading in terms of modeling for severe or life-threatening ICAHT as an important endpoint.
“We have had an absence of the standardized severity-based guidelines that we know very well for CRS and ICANS, both in terms of the diagnostic work-up and the grading but also the management,” Dr. Rejeski said at the meeting.
“We hope that the new ICAHT grading focuses future research efforts to not only understand this important side effect better, but also develop specific management strategies that mitigate the risk of infections in high-risk patients,” Dr. Rejeski added.
“The multiply validated CAR-HEMATOTOX score, assessed at time of lymphodepletion, may be helpful in this regard,” he added.
An accompanying editorial published with the guidelines underscored that “this is the first such guideline by a major organization and is a much-needed development for the management of this important CAR T-cell–associated toxicity.”
The improved standardized reporting of ICAHT “could also inform hematotoxicity management protocols,” said the editorial authors, David Qualls, MD, of the Memorial Sloan Kettering Cancer Center in New York City and Caron Jacobson, MD, of the Dana-Farber Cancer Institute, in Boston, Massachusetts.
“While providing comprehensive recommendations for ICAHT, the EHA/EBMT guidelines also highlight important gaps in our current knowledge of ICAHT, which are significant,” the editorial authors add.
Further commenting, Ulrich Jaeger, MD, a professor of hematology at the Medical University of Vienna, Vienna, Austria, agreed that the research fills an important need in post–CAR T-cell therapy management.
“Dr. Rejeski´s work is really seminal in the field and confirmed by validation cohorts in other centers,” he said in an interview. “I think the story is absolutely clear. It will be of increasing importance, with more patients surviving. [The system] will have to be adapted to novel indications as well.”
Dr. Rejeski disclosed ties with Kite/Gilead, Novartis, GMS/Celgene, and Pierre-Fabre. Jaeger reports relationships with Novartis, Gilead Sciences, Celgene/BMS, Janssen, Roche, Miltenyi Biotec, and Innovative Medicines Initiative.
FROM THE 6TH EUROPEAN CAR T-CELL MEETING
Hypertension Stable in US, Antihypertensive Med Use Rises
TOPLINE:
Hypertension prevalence remained stable in the United States at 30% after guidelines updated in 2017 lowered the threshold for the condition, while antihypertensive medication use rose about 3%, new research from the Centers for Disease Control and Prevention (CDC) shows.
METHODOLOGY:
- Researchers analyzed data from the Behavioral Risk Factor Surveillance System, a telephone survey of US adults aged 18 years and older.
- Self-reported diagnosed hypertension was defined as an affirmative response to the question, “Have you ever been told by a doctor, nurse, or other health professional that you have high blood pressure?”
- To determine treatment, respondents who answered the first question affirmatively were then asked, “Are you currently taking medicine for your high blood pressure?”
- Hypertension and treatment were assessed by age group (18-44, 45-64, and > 65 years), sex, race, ethnicity, level of education, and state of residence.
TAKEAWAY:
- The final analytic samples for 2017, 2019, and 2021 included 425,417, 392,100, and 410,318 participants, respectively.
- From 2017 to 2021, the overall age-standardized prevalence of hypertension did not change, remaining at almost exactly 30%.
- The age-standardized prevalence of antihypertensive medication use among individuals with hypertension increased by 3.1 percentage points, from 59.8% to 62.9%.
- Increases in medication use were seen in most sociodemographic groups; for example, in 2021, the prevalence was higher among women than among men (68.5% vs 59.4%), among adults aged ≥ 65 years than among those aged 18-44 years (92.5% vs 42.5%), and among Black patients than among White patients (71.3% vs 62%).
- Increases in medication use were also seen by state; use increased in 11 states, ranging from 52.2% in Utah to 72.8% in Mississippi in 2021, and did not decrease significantly in any state.
IN PRACTICE:
“These findings can be used to increase awareness of hypertension and promote lifestyle modifications and antihypertensive medication use to optimize blood pressure control and reduce disparities in prevalence and control,” the authors wrote.
SOURCE:
The study was led by Ahlia Sekkarie, PhD, of CDC’s Division for Heart Disease and Stroke Prevention, and published online in Morbidity and Mortality Weekly Report.
LIMITATIONS:
The study had several limitations. The findings were based on self-report. Median response rates of less than 50% could lead to under- or overestimates of prevalence. Parts of the population, such as those in long-term care facilities or without a telephone, were not included in the analysis. Some demographic categories had small sample sizes; therefore, prevalence changes might not be detectable.
DISCLOSURES:
No specific funding was reported. The authors reported no potential conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Hypertension prevalence remained stable in the United States at 30% after guidelines updated in 2017 lowered the threshold for the condition, while antihypertensive medication use rose about 3%, new research from the Centers for Disease Control and Prevention (CDC) shows.
METHODOLOGY:
- Researchers analyzed data from the Behavioral Risk Factor Surveillance System, a telephone survey of US adults aged 18 years and older.
- Self-reported diagnosed hypertension was defined as an affirmative response to the question, “Have you ever been told by a doctor, nurse, or other health professional that you have high blood pressure?”
- To determine treatment, respondents who answered the first question affirmatively were then asked, “Are you currently taking medicine for your high blood pressure?”
- Hypertension and treatment were assessed by age group (18-44, 45-64, and > 65 years), sex, race, ethnicity, level of education, and state of residence.
TAKEAWAY:
- The final analytic samples for 2017, 2019, and 2021 included 425,417, 392,100, and 410,318 participants, respectively.
- From 2017 to 2021, the overall age-standardized prevalence of hypertension did not change, remaining at almost exactly 30%.
- The age-standardized prevalence of antihypertensive medication use among individuals with hypertension increased by 3.1 percentage points, from 59.8% to 62.9%.
- Increases in medication use were seen in most sociodemographic groups; for example, in 2021, the prevalence was higher among women than among men (68.5% vs 59.4%), among adults aged ≥ 65 years than among those aged 18-44 years (92.5% vs 42.5%), and among Black patients than among White patients (71.3% vs 62%).
- Increases in medication use were also seen by state; use increased in 11 states, ranging from 52.2% in Utah to 72.8% in Mississippi in 2021, and did not decrease significantly in any state.
IN PRACTICE:
“These findings can be used to increase awareness of hypertension and promote lifestyle modifications and antihypertensive medication use to optimize blood pressure control and reduce disparities in prevalence and control,” the authors wrote.
SOURCE:
The study was led by Ahlia Sekkarie, PhD, of CDC’s Division for Heart Disease and Stroke Prevention, and published online in Morbidity and Mortality Weekly Report.
LIMITATIONS:
The study had several limitations. The findings were based on self-report. Median response rates of less than 50% could lead to under- or overestimates of prevalence. Parts of the population, such as those in long-term care facilities or without a telephone, were not included in the analysis. Some demographic categories had small sample sizes; therefore, prevalence changes might not be detectable.
DISCLOSURES:
No specific funding was reported. The authors reported no potential conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Hypertension prevalence remained stable in the United States at 30% after guidelines updated in 2017 lowered the threshold for the condition, while antihypertensive medication use rose about 3%, new research from the Centers for Disease Control and Prevention (CDC) shows.
METHODOLOGY:
- Researchers analyzed data from the Behavioral Risk Factor Surveillance System, a telephone survey of US adults aged 18 years and older.
- Self-reported diagnosed hypertension was defined as an affirmative response to the question, “Have you ever been told by a doctor, nurse, or other health professional that you have high blood pressure?”
- To determine treatment, respondents who answered the first question affirmatively were then asked, “Are you currently taking medicine for your high blood pressure?”
- Hypertension and treatment were assessed by age group (18-44, 45-64, and > 65 years), sex, race, ethnicity, level of education, and state of residence.
TAKEAWAY:
- The final analytic samples for 2017, 2019, and 2021 included 425,417, 392,100, and 410,318 participants, respectively.
- From 2017 to 2021, the overall age-standardized prevalence of hypertension did not change, remaining at almost exactly 30%.
- The age-standardized prevalence of antihypertensive medication use among individuals with hypertension increased by 3.1 percentage points, from 59.8% to 62.9%.
- Increases in medication use were seen in most sociodemographic groups; for example, in 2021, the prevalence was higher among women than among men (68.5% vs 59.4%), among adults aged ≥ 65 years than among those aged 18-44 years (92.5% vs 42.5%), and among Black patients than among White patients (71.3% vs 62%).
- Increases in medication use were also seen by state; use increased in 11 states, ranging from 52.2% in Utah to 72.8% in Mississippi in 2021, and did not decrease significantly in any state.
IN PRACTICE:
“These findings can be used to increase awareness of hypertension and promote lifestyle modifications and antihypertensive medication use to optimize blood pressure control and reduce disparities in prevalence and control,” the authors wrote.
SOURCE:
The study was led by Ahlia Sekkarie, PhD, of CDC’s Division for Heart Disease and Stroke Prevention, and published online in Morbidity and Mortality Weekly Report.
LIMITATIONS:
The study had several limitations. The findings were based on self-report. Median response rates of less than 50% could lead to under- or overestimates of prevalence. Parts of the population, such as those in long-term care facilities or without a telephone, were not included in the analysis. Some demographic categories had small sample sizes; therefore, prevalence changes might not be detectable.
DISCLOSURES:
No specific funding was reported. The authors reported no potential conflicts of interest.
A version of this article appeared on Medscape.com.
Healthy Lifestyle Linked to Lower Risk for IBS
TOPLINE:
Adherence to a higher number of the five key lifestyle behaviors — not smoking, vigorous physical activity, optimal sleep, high-quality diet, and moderate alcohol consumption — is associated with a lower risk for irritable bowel syndrome (IBS), new research suggested.
METHODOLOGY:
- Researchers assessed the association between healthy lifestyle behaviors and IBS incidence using UK Biobank data from 64,268 adults (mean age, 56 years; 55%, women) with no IBS diagnosis at baseline.
- Participants were enrolled during 2006-2010 and followed up to 2022.
- Self-reported healthy lifestyle behaviors were never smoking, optimal sleep, high level of vigorous physical activity, high dietary quality, and moderate alcohol intake.
TAKEAWAY:
- At baseline, 11.8% of participants reported none of the five healthy lifestyle behaviors, 32.1% reported one behavior, 34.1% reported two behaviors, and 21.9% reported three to five behaviors.
- During a mean follow-up of 12.6 years, 961 (1.5%) incident IBS cases occurred.
- Adjusted hazard ratios associated with IBS incidence and having one, two, and three to five behaviors were 0.79, 0.64, and 0.58, respectively.
- Significant independent inverse associations with IBS incidence were seen for never smoking (0.86), high level of vigorous physical activity (0.83), and optimal sleep (0.73).
- After adjustment for age, sex, employment status, geographic location, gastrointestinal infection, endometriosis, and family history of IBS, adherence to a higher number of healthy lifestyle behaviors remained significantly associated with a lower risk of incident IBS.
IN PRACTICE:
“This study provides evidence that adherence to a higher number of healthy lifestyle behaviors — never smoking, optimal sleep, high level of physical activity, high dietary quality and moderate alcohol intake — is significantly associated with a lower risk of subsequent IBS incidence. These findings suggest that lifestyle modifications should be considered as key primary prevention strategies for IBS,” the authors wrote.
SOURCE:
The study, led by Fai Fai Ho of The Chinese University of Hong Kong, was published online in Gut.
LIMITATIONS:
The study was observational, so it could not show cause and effect. It relied on self-report, which is not always accurate, and the findings may not be applicable to younger age groups. Lifestyle changes made during the follow-up period could not be considered.
DISCLOSURES:
The study was funded by the National Key R&D Program of China and the National Natural Science Foundation of China. The authors declared no competing interests.
A version of this article appeared on Medscape.com.
TOPLINE:
Adherence to a higher number of the five key lifestyle behaviors — not smoking, vigorous physical activity, optimal sleep, high-quality diet, and moderate alcohol consumption — is associated with a lower risk for irritable bowel syndrome (IBS), new research suggested.
METHODOLOGY:
- Researchers assessed the association between healthy lifestyle behaviors and IBS incidence using UK Biobank data from 64,268 adults (mean age, 56 years; 55%, women) with no IBS diagnosis at baseline.
- Participants were enrolled during 2006-2010 and followed up to 2022.
- Self-reported healthy lifestyle behaviors were never smoking, optimal sleep, high level of vigorous physical activity, high dietary quality, and moderate alcohol intake.
TAKEAWAY:
- At baseline, 11.8% of participants reported none of the five healthy lifestyle behaviors, 32.1% reported one behavior, 34.1% reported two behaviors, and 21.9% reported three to five behaviors.
- During a mean follow-up of 12.6 years, 961 (1.5%) incident IBS cases occurred.
- Adjusted hazard ratios associated with IBS incidence and having one, two, and three to five behaviors were 0.79, 0.64, and 0.58, respectively.
- Significant independent inverse associations with IBS incidence were seen for never smoking (0.86), high level of vigorous physical activity (0.83), and optimal sleep (0.73).
- After adjustment for age, sex, employment status, geographic location, gastrointestinal infection, endometriosis, and family history of IBS, adherence to a higher number of healthy lifestyle behaviors remained significantly associated with a lower risk of incident IBS.
IN PRACTICE:
“This study provides evidence that adherence to a higher number of healthy lifestyle behaviors — never smoking, optimal sleep, high level of physical activity, high dietary quality and moderate alcohol intake — is significantly associated with a lower risk of subsequent IBS incidence. These findings suggest that lifestyle modifications should be considered as key primary prevention strategies for IBS,” the authors wrote.
SOURCE:
The study, led by Fai Fai Ho of The Chinese University of Hong Kong, was published online in Gut.
LIMITATIONS:
The study was observational, so it could not show cause and effect. It relied on self-report, which is not always accurate, and the findings may not be applicable to younger age groups. Lifestyle changes made during the follow-up period could not be considered.
DISCLOSURES:
The study was funded by the National Key R&D Program of China and the National Natural Science Foundation of China. The authors declared no competing interests.
A version of this article appeared on Medscape.com.
TOPLINE:
Adherence to a higher number of the five key lifestyle behaviors — not smoking, vigorous physical activity, optimal sleep, high-quality diet, and moderate alcohol consumption — is associated with a lower risk for irritable bowel syndrome (IBS), new research suggested.
METHODOLOGY:
- Researchers assessed the association between healthy lifestyle behaviors and IBS incidence using UK Biobank data from 64,268 adults (mean age, 56 years; 55%, women) with no IBS diagnosis at baseline.
- Participants were enrolled during 2006-2010 and followed up to 2022.
- Self-reported healthy lifestyle behaviors were never smoking, optimal sleep, high level of vigorous physical activity, high dietary quality, and moderate alcohol intake.
TAKEAWAY:
- At baseline, 11.8% of participants reported none of the five healthy lifestyle behaviors, 32.1% reported one behavior, 34.1% reported two behaviors, and 21.9% reported three to five behaviors.
- During a mean follow-up of 12.6 years, 961 (1.5%) incident IBS cases occurred.
- Adjusted hazard ratios associated with IBS incidence and having one, two, and three to five behaviors were 0.79, 0.64, and 0.58, respectively.
- Significant independent inverse associations with IBS incidence were seen for never smoking (0.86), high level of vigorous physical activity (0.83), and optimal sleep (0.73).
- After adjustment for age, sex, employment status, geographic location, gastrointestinal infection, endometriosis, and family history of IBS, adherence to a higher number of healthy lifestyle behaviors remained significantly associated with a lower risk of incident IBS.
IN PRACTICE:
“This study provides evidence that adherence to a higher number of healthy lifestyle behaviors — never smoking, optimal sleep, high level of physical activity, high dietary quality and moderate alcohol intake — is significantly associated with a lower risk of subsequent IBS incidence. These findings suggest that lifestyle modifications should be considered as key primary prevention strategies for IBS,” the authors wrote.
SOURCE:
The study, led by Fai Fai Ho of The Chinese University of Hong Kong, was published online in Gut.
LIMITATIONS:
The study was observational, so it could not show cause and effect. It relied on self-report, which is not always accurate, and the findings may not be applicable to younger age groups. Lifestyle changes made during the follow-up period could not be considered.
DISCLOSURES:
The study was funded by the National Key R&D Program of China and the National Natural Science Foundation of China. The authors declared no competing interests.
A version of this article appeared on Medscape.com.
IBS Placebo Responses Predicted By Patient Beliefs, Relationship with Provider
, according to investigators.
These findings may improve prediction of placebo responses in IBS, and may help avoid patient-provider “mismatch,” both of which can alter treatment outcomes and confound clinical trial findings, reported lead author Jeffrey M. Lackner, PsyD, chief of the division of behavioral medicine at the University of Buffalo, New York, and colleagues.
“A relatively large (40%) placebo response in IBS trials obscures potentially useful, mechanistic, and pharmacodynamically induced symptom changes among agents that do reach market,” the investigators wrote in Gastro Hep Advances. “This begs the question of what individual difference factors distinguish placebo responders.”
While previous studies have explored placebo patient predictors in IBS, most focused on study design and baseline personal characteristics such as age and sex, with none yielding prognostically reliable findings, according to Dr. Lackner and colleagues. Mid-treatment factors such as patient-provider dynamics have not been featured in published meta-analyses, they noted, despite their potential importance.
“This limitation partly reflects the demands of efficacy trials that prioritize pre- and posttreatment data over that collected during acute phase, when the putative mechanisms underpinning placebo effects play out,” the investigators wrote. “The expectation that one can benefit from a treatment, for example, is optimally assessed after its rationale is delivered but before a clinically thorough regimen is provided, meaning that it cannot be fruitfully assessed at baseline along with other personal characteristics when treatment rationale is not fully disclosed. The same applies to relational factors such as patient-physician interactions that define the context where treatment is delivered, and placebo response presumably incubates.”
To explore the above factors, Dr. Lackner and colleagues conducted a secondary analysis of 145 patients with Rome III-diagnosed IBS from the Irritable Bowel Syndrome Outcome Study.
During the study, patients were randomized to receive either 10 sessions of clinic-based cognitive behavioral therapy (CBT), 4 sessions of minimal-contact CBT, or 4 sessions of supportive counseling and education without any prescribed behavior changes. Responses were measured by the IBS version of the Clinical Global Improvement Scale, with evaluations conducted at the treatment midpoint and 2 weeks after treatment.
Candidate predictors at baseline included pain catastrophizing, somatization, emotion regulation, neuroticism, stress, and others, while clinical factors included treatment expectancy/credibility and patient-provider relationship.
Responses during treatment were significantly associated with lower somatization and stress level at baseline, as well as greater patient-provider agreement on treatment tasks (P less than .001).
Posttreatment responses were significantly associated with baseline gastroenterologist-rated IBS severity, anxiety, agreement that the patient and the provider shared goals from a provider perspective, and ability to reframe stressful events in a positive light (P less than .001). That ability to reconsider emotions was also associated with a faster placebo response (P = .011).
“The strength of placebo responsiveness is subject to the influence of patient factors that precede treatment delivery (rethinking or reinterpreting stressful situations in everyday life in a way that reduces their subsequent impact) and specific elements of provider-patient interactions that occur while treatment is delivered, particularly practitioners’ estimation that patients agree on their goals and tasks to achieve them,” Dr. Lackner and colleagues concluded. “We believe this line of research can help identify factors that drive placebo response and narrow the patient-provider ‘mismatch’ that undermines the quality, satisfaction, and efficiency of IBS care regardless of what treatment is delivered.”
The study was supported by the NIH. The investigators disclosed no conflicts of interest.
Irritable bowel syndrome (IBS) is associated with impaired functioning and work or school absenteeism. Current treatments are suboptimal and there is a need for improved management strategies. A challenge in designing trials can be placebo response. Placebo can also be a treatment modality with approximately 40% response in adults and children with IBS. The study by Lackner et al. provides predictors of the magnitude, and timing of placebo response. Accordingly, certain behaviors and strategies adopted by patients and clinicians in addition to pharmacotherapy can harness greater clinical improvements.
While patient factors such as stress levels, somatization, and anxiety played a role in predicting rapid and delayed placebo response, an interesting domain was “cognitive reappraisal,” the ability to alter the impact of stressful events by reframing unpleasantness toward them. This was associated with greater global improvement post treatment and differed between rapid and delayed responders. Cognitive reappraisal has shown changes in the limbic system such as activation of the prefrontal cortex like placebo analgesia. Thus, optimal introduction of treatments to patients may be important to maximize the cognitive appraisal abilities, enhance expectation effects, and improve treatment outcomes. Similarly, minimizing nocebo effects may be equally important to decrease side effects.
The agreement between patients and clinicians on treatment goals and tasks also predicted response. Thus, developing thorough treatment goals beforehand could be crucial to sustain treatment responses. For example, improved functioning may be a goal to agree upon rather than symptom reduction alone before commencement of treatment. Similarly, shared decision-making during treatment may have a tremendous influence on favorable outcomes.
Neha Santucci, MD, MBBS, is director of the Disorders of Gut-Brain Interaction Program at the Neurogastroenterology and Motility Center, Pediatric Gastroenterology, Hepatology and Nutrition, Cincinnati Children’s Hospital Medical Center, and associate professor of pediatrics, University of Cincinnati College of Medicine.
Irritable bowel syndrome (IBS) is associated with impaired functioning and work or school absenteeism. Current treatments are suboptimal and there is a need for improved management strategies. A challenge in designing trials can be placebo response. Placebo can also be a treatment modality with approximately 40% response in adults and children with IBS. The study by Lackner et al. provides predictors of the magnitude, and timing of placebo response. Accordingly, certain behaviors and strategies adopted by patients and clinicians in addition to pharmacotherapy can harness greater clinical improvements.
While patient factors such as stress levels, somatization, and anxiety played a role in predicting rapid and delayed placebo response, an interesting domain was “cognitive reappraisal,” the ability to alter the impact of stressful events by reframing unpleasantness toward them. This was associated with greater global improvement post treatment and differed between rapid and delayed responders. Cognitive reappraisal has shown changes in the limbic system such as activation of the prefrontal cortex like placebo analgesia. Thus, optimal introduction of treatments to patients may be important to maximize the cognitive appraisal abilities, enhance expectation effects, and improve treatment outcomes. Similarly, minimizing nocebo effects may be equally important to decrease side effects.
The agreement between patients and clinicians on treatment goals and tasks also predicted response. Thus, developing thorough treatment goals beforehand could be crucial to sustain treatment responses. For example, improved functioning may be a goal to agree upon rather than symptom reduction alone before commencement of treatment. Similarly, shared decision-making during treatment may have a tremendous influence on favorable outcomes.
Neha Santucci, MD, MBBS, is director of the Disorders of Gut-Brain Interaction Program at the Neurogastroenterology and Motility Center, Pediatric Gastroenterology, Hepatology and Nutrition, Cincinnati Children’s Hospital Medical Center, and associate professor of pediatrics, University of Cincinnati College of Medicine.
Irritable bowel syndrome (IBS) is associated with impaired functioning and work or school absenteeism. Current treatments are suboptimal and there is a need for improved management strategies. A challenge in designing trials can be placebo response. Placebo can also be a treatment modality with approximately 40% response in adults and children with IBS. The study by Lackner et al. provides predictors of the magnitude, and timing of placebo response. Accordingly, certain behaviors and strategies adopted by patients and clinicians in addition to pharmacotherapy can harness greater clinical improvements.
While patient factors such as stress levels, somatization, and anxiety played a role in predicting rapid and delayed placebo response, an interesting domain was “cognitive reappraisal,” the ability to alter the impact of stressful events by reframing unpleasantness toward them. This was associated with greater global improvement post treatment and differed between rapid and delayed responders. Cognitive reappraisal has shown changes in the limbic system such as activation of the prefrontal cortex like placebo analgesia. Thus, optimal introduction of treatments to patients may be important to maximize the cognitive appraisal abilities, enhance expectation effects, and improve treatment outcomes. Similarly, minimizing nocebo effects may be equally important to decrease side effects.
The agreement between patients and clinicians on treatment goals and tasks also predicted response. Thus, developing thorough treatment goals beforehand could be crucial to sustain treatment responses. For example, improved functioning may be a goal to agree upon rather than symptom reduction alone before commencement of treatment. Similarly, shared decision-making during treatment may have a tremendous influence on favorable outcomes.
Neha Santucci, MD, MBBS, is director of the Disorders of Gut-Brain Interaction Program at the Neurogastroenterology and Motility Center, Pediatric Gastroenterology, Hepatology and Nutrition, Cincinnati Children’s Hospital Medical Center, and associate professor of pediatrics, University of Cincinnati College of Medicine.
, according to investigators.
These findings may improve prediction of placebo responses in IBS, and may help avoid patient-provider “mismatch,” both of which can alter treatment outcomes and confound clinical trial findings, reported lead author Jeffrey M. Lackner, PsyD, chief of the division of behavioral medicine at the University of Buffalo, New York, and colleagues.
“A relatively large (40%) placebo response in IBS trials obscures potentially useful, mechanistic, and pharmacodynamically induced symptom changes among agents that do reach market,” the investigators wrote in Gastro Hep Advances. “This begs the question of what individual difference factors distinguish placebo responders.”
While previous studies have explored placebo patient predictors in IBS, most focused on study design and baseline personal characteristics such as age and sex, with none yielding prognostically reliable findings, according to Dr. Lackner and colleagues. Mid-treatment factors such as patient-provider dynamics have not been featured in published meta-analyses, they noted, despite their potential importance.
“This limitation partly reflects the demands of efficacy trials that prioritize pre- and posttreatment data over that collected during acute phase, when the putative mechanisms underpinning placebo effects play out,” the investigators wrote. “The expectation that one can benefit from a treatment, for example, is optimally assessed after its rationale is delivered but before a clinically thorough regimen is provided, meaning that it cannot be fruitfully assessed at baseline along with other personal characteristics when treatment rationale is not fully disclosed. The same applies to relational factors such as patient-physician interactions that define the context where treatment is delivered, and placebo response presumably incubates.”
To explore the above factors, Dr. Lackner and colleagues conducted a secondary analysis of 145 patients with Rome III-diagnosed IBS from the Irritable Bowel Syndrome Outcome Study.
During the study, patients were randomized to receive either 10 sessions of clinic-based cognitive behavioral therapy (CBT), 4 sessions of minimal-contact CBT, or 4 sessions of supportive counseling and education without any prescribed behavior changes. Responses were measured by the IBS version of the Clinical Global Improvement Scale, with evaluations conducted at the treatment midpoint and 2 weeks after treatment.
Candidate predictors at baseline included pain catastrophizing, somatization, emotion regulation, neuroticism, stress, and others, while clinical factors included treatment expectancy/credibility and patient-provider relationship.
Responses during treatment were significantly associated with lower somatization and stress level at baseline, as well as greater patient-provider agreement on treatment tasks (P less than .001).
Posttreatment responses were significantly associated with baseline gastroenterologist-rated IBS severity, anxiety, agreement that the patient and the provider shared goals from a provider perspective, and ability to reframe stressful events in a positive light (P less than .001). That ability to reconsider emotions was also associated with a faster placebo response (P = .011).
“The strength of placebo responsiveness is subject to the influence of patient factors that precede treatment delivery (rethinking or reinterpreting stressful situations in everyday life in a way that reduces their subsequent impact) and specific elements of provider-patient interactions that occur while treatment is delivered, particularly practitioners’ estimation that patients agree on their goals and tasks to achieve them,” Dr. Lackner and colleagues concluded. “We believe this line of research can help identify factors that drive placebo response and narrow the patient-provider ‘mismatch’ that undermines the quality, satisfaction, and efficiency of IBS care regardless of what treatment is delivered.”
The study was supported by the NIH. The investigators disclosed no conflicts of interest.
, according to investigators.
These findings may improve prediction of placebo responses in IBS, and may help avoid patient-provider “mismatch,” both of which can alter treatment outcomes and confound clinical trial findings, reported lead author Jeffrey M. Lackner, PsyD, chief of the division of behavioral medicine at the University of Buffalo, New York, and colleagues.
“A relatively large (40%) placebo response in IBS trials obscures potentially useful, mechanistic, and pharmacodynamically induced symptom changes among agents that do reach market,” the investigators wrote in Gastro Hep Advances. “This begs the question of what individual difference factors distinguish placebo responders.”
While previous studies have explored placebo patient predictors in IBS, most focused on study design and baseline personal characteristics such as age and sex, with none yielding prognostically reliable findings, according to Dr. Lackner and colleagues. Mid-treatment factors such as patient-provider dynamics have not been featured in published meta-analyses, they noted, despite their potential importance.
“This limitation partly reflects the demands of efficacy trials that prioritize pre- and posttreatment data over that collected during acute phase, when the putative mechanisms underpinning placebo effects play out,” the investigators wrote. “The expectation that one can benefit from a treatment, for example, is optimally assessed after its rationale is delivered but before a clinically thorough regimen is provided, meaning that it cannot be fruitfully assessed at baseline along with other personal characteristics when treatment rationale is not fully disclosed. The same applies to relational factors such as patient-physician interactions that define the context where treatment is delivered, and placebo response presumably incubates.”
To explore the above factors, Dr. Lackner and colleagues conducted a secondary analysis of 145 patients with Rome III-diagnosed IBS from the Irritable Bowel Syndrome Outcome Study.
During the study, patients were randomized to receive either 10 sessions of clinic-based cognitive behavioral therapy (CBT), 4 sessions of minimal-contact CBT, or 4 sessions of supportive counseling and education without any prescribed behavior changes. Responses were measured by the IBS version of the Clinical Global Improvement Scale, with evaluations conducted at the treatment midpoint and 2 weeks after treatment.
Candidate predictors at baseline included pain catastrophizing, somatization, emotion regulation, neuroticism, stress, and others, while clinical factors included treatment expectancy/credibility and patient-provider relationship.
Responses during treatment were significantly associated with lower somatization and stress level at baseline, as well as greater patient-provider agreement on treatment tasks (P less than .001).
Posttreatment responses were significantly associated with baseline gastroenterologist-rated IBS severity, anxiety, agreement that the patient and the provider shared goals from a provider perspective, and ability to reframe stressful events in a positive light (P less than .001). That ability to reconsider emotions was also associated with a faster placebo response (P = .011).
“The strength of placebo responsiveness is subject to the influence of patient factors that precede treatment delivery (rethinking or reinterpreting stressful situations in everyday life in a way that reduces their subsequent impact) and specific elements of provider-patient interactions that occur while treatment is delivered, particularly practitioners’ estimation that patients agree on their goals and tasks to achieve them,” Dr. Lackner and colleagues concluded. “We believe this line of research can help identify factors that drive placebo response and narrow the patient-provider ‘mismatch’ that undermines the quality, satisfaction, and efficiency of IBS care regardless of what treatment is delivered.”
The study was supported by the NIH. The investigators disclosed no conflicts of interest.
FROM GASTRO HEP ADVANCES
Is a 1-Hour Glucose Test Better at Predicting T2D Risk?
A new position statement from the International Diabetes Federation advises using a 1-hour 75-g oral glucose tolerance test (OGTT) to improve identification of people at risk of developing type 2 diabetes.
“There are many, many people who may appear ‘normal’ if you use A1c or fasting glucose, but if you do a glucose tolerance test, they may have an abnormality after a glucose load. …The 1-hour plasma glucose has been found to be a more sensitive biomarker for the earlier identification of these high-risk individuals,” lead author Michael Bergman, MD, professor of medicine and population health at New York University Grossman School of Medicine, New York City, told this news organization in an interview.
Dr. Bergman presented the document, written by a 22-member international expert panel, on March 6, 2024, at the annual Advanced Technologies & Treatments for Diabetes meeting. It was simultaneously published in Diabetes Research and Clinical Practice.
This is the International Diabetes Federation’s (IDF’s) proposed screening algorithm for “intermediate hyperglycemia” and type 2 diabetes:
- At-risk high-risk individuals are first screened with a validated questionnaire such as the FINDRISK or the American Diabetes Association’s (ADA’s) risk screening tool.
- People identified as high risk should undergo laboratory screening with a 1-hour 75-g OGTT (although a 2-hour OGTT, fasting glucose, or A1c, as currently recommended by several organizations, is still considered acceptable).
- People with a 1-hour plasma glucose value at or above 155 mg (8.6 mmol/L) are considered to have intermediate hyperglycemia and should be prescribed lifestyle intervention and referred to a diabetes prevention program.
- Those with a 1-hour value greater than or equal to 209 mg/dL (11.6 mmol/L) are considered to have type 2 diabetes and should have a repeat test to confirm the diagnosis, with referral for further evaluation and treatment.
The new guidance is based on increasing evidence that the 1-hour test is a better predictor than other tests, including the 2-hour OGTT, of progression to type 2 diabetes and its associated complications, in a variety of populations. The document cites data showing that a plasma glucose of 155 mg/dl or greater on the 1-hour post-75-g test can identify people with undiagnosed type 2 diabetes or who are at increased risk but who have “normal” glucose tolerance as defined by an A1c < 5.7% (38.8 mmol/mol), a fasting plasma glucose < 100 mg/dL (5.6 mmol/L), or a 2-hour value below 140 mg/dL (7.8 mmol/L).
However, even though a 1-hour test may be more convenient than the traditional 2-hour test, incorporating OGTT into busy clinical practice may still pose logistical problems and may not improve ultimate outcomes, Elizabeth Selvin, PhD, MPH, of the Johns Hopkins University Bloomberg School of Public Health, Baltimore, Maryland, told this news organization. “I worry that emphasizing a burdensome test is not going to improve diabetes screening or diabetes prevention. Doing more 1-hour glucose screening is not going to get more people into diabetes prevention programs.”
When an audience member raised the logistics concern during the session Q&A, Dr. Bergman replied, “It’s no different than recommending colonoscopy or a mammogram. … I tell patients that we want to learn more about what is going on before we make a firm diagnosis. … I’ve done more than a hundred 1-hour glucose tolerance tests and have not seen one patient who refused because of inconvenience. Everything depends on the way we discuss things with patients.”
Recommendation Based on Emerging Evidence
“Intermediate hyperglycemia” is the World Health Organization’s term to refer to either impaired fasting glucose (IFG) or impaired glucose tolerance (IGT). These conditions are often called “prediabetes,” especially in the United States, but that term has been controversial because not everyone with IFG and IGT will go on to develop type 2 diabetes, said Dr. Bergman, who is also director of the NYU Langone Diabetes Prevention Program.
“With ‘prediabetes’ you’re labeling someone with a disease they may not develop. It’s not normal, but it’s not diabetes, so it’s an intermediate state,” he explained.
The statement provides a detailed summary of the data from 19 studies supporting use of a 1-hour plasma glucose of ≥ 155 mg/dL (8.6 mmol/L) to diagnose intermediate hyperglycemia and type 2 diabetes, including links between that level of glycemia and worsened metabolic and atherogenic profiles, risk for microvascular and macrovascular complications and mortality, and identification of risks for obstructive sleep apnea, cystic fibrosis-related diabetes mellitus, fatty liver disease, and premature mortality.
One major problem with current testing, Dr. Bergman said, is that “there is a huge disconnect between A1c and glucose values. … Only about 30% of individuals with an abnormal A1c will have an abnormal fasting glucose, and the inverse is also true. There’s a big mismatch between the two.”
Current guidelines suggest using both A1c and fasting glucose, but Dr. Bergman said that even then “you’re still missing about 20-30% who have IGT. … Part of the problem is that the criteria we use for defining abnormal fasting and 2-hour levels are too high. … 140 [mg/dL] for the 2-hour is too high and 100 [mg/dL] fasting is too high. … And that’s one of the reasons why many people progress to type 2 diabetes, because we’re using screening thresholds that are nonphysiologic.”
But Dr. Selvin disagrees, pointing to her own work showing that “using a combination of fasting glucose and A1c for screening does an excellent job at identifying high-risk individuals.”
She’s also unconvinced by other data cited in the paper. “I am deeply skeptical about 1-hour glucose being more prognostic than all other glycemic tests. … Associations of glucose tests with incident diabetes are inherently a circular analysis since diabetes is defined by elevations in those same tests. It is helpful to look at progression of diabetes, but these analyses are not simple and doing them well, especially with head-to-head comparisons against different glycemic tests, is hard,” she said.
In her view, “lifestyle interventions and weight loss should be recommended in at-risk individuals, including those with overweight and obesity, hypertension, metabolic syndrome, etc., regardless of 1-hour glucose test results.”
She added, “If we want to prevent diabetes, we need to focus on preventing weight gain and obesity and managing cardiometabolic risk factors. Lifestyle interventions are not effective unless they are intensive and patients are highly adherent. We need to make lifestyle interventions available and affordable for patients. That is the big barrier. I don’t think more screening with 1-hour glucose is going to help.”
Dr. Bergman pointed out that the International Diabetes Federation represents more than 100 countries, including many that are middle- and low-income. “They need a simple, cost-effective tool for screening effectively. A1c is more expensive, and fasting glucose alone will underestimate disease prevalence. So, the IDF felt, after doing a fairly comprehensive due diligence, that the data warranted recommendation of the 1-hour glucose.”
He’s hoping other organizations like the ADA and the World Health Organization will sign on to bring this guidance into the primary care arena. This news organization reached out to ADA for comment, but their representative hadn’t responded by press time.
Dr. Bergman had no disclosures. Dr. Selvin was supported by the National Institutes of Health.
A version of this article appeared on Medscape.com.
A new position statement from the International Diabetes Federation advises using a 1-hour 75-g oral glucose tolerance test (OGTT) to improve identification of people at risk of developing type 2 diabetes.
“There are many, many people who may appear ‘normal’ if you use A1c or fasting glucose, but if you do a glucose tolerance test, they may have an abnormality after a glucose load. …The 1-hour plasma glucose has been found to be a more sensitive biomarker for the earlier identification of these high-risk individuals,” lead author Michael Bergman, MD, professor of medicine and population health at New York University Grossman School of Medicine, New York City, told this news organization in an interview.
Dr. Bergman presented the document, written by a 22-member international expert panel, on March 6, 2024, at the annual Advanced Technologies & Treatments for Diabetes meeting. It was simultaneously published in Diabetes Research and Clinical Practice.
This is the International Diabetes Federation’s (IDF’s) proposed screening algorithm for “intermediate hyperglycemia” and type 2 diabetes:
- At-risk high-risk individuals are first screened with a validated questionnaire such as the FINDRISK or the American Diabetes Association’s (ADA’s) risk screening tool.
- People identified as high risk should undergo laboratory screening with a 1-hour 75-g OGTT (although a 2-hour OGTT, fasting glucose, or A1c, as currently recommended by several organizations, is still considered acceptable).
- People with a 1-hour plasma glucose value at or above 155 mg (8.6 mmol/L) are considered to have intermediate hyperglycemia and should be prescribed lifestyle intervention and referred to a diabetes prevention program.
- Those with a 1-hour value greater than or equal to 209 mg/dL (11.6 mmol/L) are considered to have type 2 diabetes and should have a repeat test to confirm the diagnosis, with referral for further evaluation and treatment.
The new guidance is based on increasing evidence that the 1-hour test is a better predictor than other tests, including the 2-hour OGTT, of progression to type 2 diabetes and its associated complications, in a variety of populations. The document cites data showing that a plasma glucose of 155 mg/dl or greater on the 1-hour post-75-g test can identify people with undiagnosed type 2 diabetes or who are at increased risk but who have “normal” glucose tolerance as defined by an A1c < 5.7% (38.8 mmol/mol), a fasting plasma glucose < 100 mg/dL (5.6 mmol/L), or a 2-hour value below 140 mg/dL (7.8 mmol/L).
However, even though a 1-hour test may be more convenient than the traditional 2-hour test, incorporating OGTT into busy clinical practice may still pose logistical problems and may not improve ultimate outcomes, Elizabeth Selvin, PhD, MPH, of the Johns Hopkins University Bloomberg School of Public Health, Baltimore, Maryland, told this news organization. “I worry that emphasizing a burdensome test is not going to improve diabetes screening or diabetes prevention. Doing more 1-hour glucose screening is not going to get more people into diabetes prevention programs.”
When an audience member raised the logistics concern during the session Q&A, Dr. Bergman replied, “It’s no different than recommending colonoscopy or a mammogram. … I tell patients that we want to learn more about what is going on before we make a firm diagnosis. … I’ve done more than a hundred 1-hour glucose tolerance tests and have not seen one patient who refused because of inconvenience. Everything depends on the way we discuss things with patients.”
Recommendation Based on Emerging Evidence
“Intermediate hyperglycemia” is the World Health Organization’s term to refer to either impaired fasting glucose (IFG) or impaired glucose tolerance (IGT). These conditions are often called “prediabetes,” especially in the United States, but that term has been controversial because not everyone with IFG and IGT will go on to develop type 2 diabetes, said Dr. Bergman, who is also director of the NYU Langone Diabetes Prevention Program.
“With ‘prediabetes’ you’re labeling someone with a disease they may not develop. It’s not normal, but it’s not diabetes, so it’s an intermediate state,” he explained.
The statement provides a detailed summary of the data from 19 studies supporting use of a 1-hour plasma glucose of ≥ 155 mg/dL (8.6 mmol/L) to diagnose intermediate hyperglycemia and type 2 diabetes, including links between that level of glycemia and worsened metabolic and atherogenic profiles, risk for microvascular and macrovascular complications and mortality, and identification of risks for obstructive sleep apnea, cystic fibrosis-related diabetes mellitus, fatty liver disease, and premature mortality.
One major problem with current testing, Dr. Bergman said, is that “there is a huge disconnect between A1c and glucose values. … Only about 30% of individuals with an abnormal A1c will have an abnormal fasting glucose, and the inverse is also true. There’s a big mismatch between the two.”
Current guidelines suggest using both A1c and fasting glucose, but Dr. Bergman said that even then “you’re still missing about 20-30% who have IGT. … Part of the problem is that the criteria we use for defining abnormal fasting and 2-hour levels are too high. … 140 [mg/dL] for the 2-hour is too high and 100 [mg/dL] fasting is too high. … And that’s one of the reasons why many people progress to type 2 diabetes, because we’re using screening thresholds that are nonphysiologic.”
But Dr. Selvin disagrees, pointing to her own work showing that “using a combination of fasting glucose and A1c for screening does an excellent job at identifying high-risk individuals.”
She’s also unconvinced by other data cited in the paper. “I am deeply skeptical about 1-hour glucose being more prognostic than all other glycemic tests. … Associations of glucose tests with incident diabetes are inherently a circular analysis since diabetes is defined by elevations in those same tests. It is helpful to look at progression of diabetes, but these analyses are not simple and doing them well, especially with head-to-head comparisons against different glycemic tests, is hard,” she said.
In her view, “lifestyle interventions and weight loss should be recommended in at-risk individuals, including those with overweight and obesity, hypertension, metabolic syndrome, etc., regardless of 1-hour glucose test results.”
She added, “If we want to prevent diabetes, we need to focus on preventing weight gain and obesity and managing cardiometabolic risk factors. Lifestyle interventions are not effective unless they are intensive and patients are highly adherent. We need to make lifestyle interventions available and affordable for patients. That is the big barrier. I don’t think more screening with 1-hour glucose is going to help.”
Dr. Bergman pointed out that the International Diabetes Federation represents more than 100 countries, including many that are middle- and low-income. “They need a simple, cost-effective tool for screening effectively. A1c is more expensive, and fasting glucose alone will underestimate disease prevalence. So, the IDF felt, after doing a fairly comprehensive due diligence, that the data warranted recommendation of the 1-hour glucose.”
He’s hoping other organizations like the ADA and the World Health Organization will sign on to bring this guidance into the primary care arena. This news organization reached out to ADA for comment, but their representative hadn’t responded by press time.
Dr. Bergman had no disclosures. Dr. Selvin was supported by the National Institutes of Health.
A version of this article appeared on Medscape.com.
A new position statement from the International Diabetes Federation advises using a 1-hour 75-g oral glucose tolerance test (OGTT) to improve identification of people at risk of developing type 2 diabetes.
“There are many, many people who may appear ‘normal’ if you use A1c or fasting glucose, but if you do a glucose tolerance test, they may have an abnormality after a glucose load. …The 1-hour plasma glucose has been found to be a more sensitive biomarker for the earlier identification of these high-risk individuals,” lead author Michael Bergman, MD, professor of medicine and population health at New York University Grossman School of Medicine, New York City, told this news organization in an interview.
Dr. Bergman presented the document, written by a 22-member international expert panel, on March 6, 2024, at the annual Advanced Technologies & Treatments for Diabetes meeting. It was simultaneously published in Diabetes Research and Clinical Practice.
This is the International Diabetes Federation’s (IDF’s) proposed screening algorithm for “intermediate hyperglycemia” and type 2 diabetes:
- At-risk high-risk individuals are first screened with a validated questionnaire such as the FINDRISK or the American Diabetes Association’s (ADA’s) risk screening tool.
- People identified as high risk should undergo laboratory screening with a 1-hour 75-g OGTT (although a 2-hour OGTT, fasting glucose, or A1c, as currently recommended by several organizations, is still considered acceptable).
- People with a 1-hour plasma glucose value at or above 155 mg (8.6 mmol/L) are considered to have intermediate hyperglycemia and should be prescribed lifestyle intervention and referred to a diabetes prevention program.
- Those with a 1-hour value greater than or equal to 209 mg/dL (11.6 mmol/L) are considered to have type 2 diabetes and should have a repeat test to confirm the diagnosis, with referral for further evaluation and treatment.
The new guidance is based on increasing evidence that the 1-hour test is a better predictor than other tests, including the 2-hour OGTT, of progression to type 2 diabetes and its associated complications, in a variety of populations. The document cites data showing that a plasma glucose of 155 mg/dl or greater on the 1-hour post-75-g test can identify people with undiagnosed type 2 diabetes or who are at increased risk but who have “normal” glucose tolerance as defined by an A1c < 5.7% (38.8 mmol/mol), a fasting plasma glucose < 100 mg/dL (5.6 mmol/L), or a 2-hour value below 140 mg/dL (7.8 mmol/L).
However, even though a 1-hour test may be more convenient than the traditional 2-hour test, incorporating OGTT into busy clinical practice may still pose logistical problems and may not improve ultimate outcomes, Elizabeth Selvin, PhD, MPH, of the Johns Hopkins University Bloomberg School of Public Health, Baltimore, Maryland, told this news organization. “I worry that emphasizing a burdensome test is not going to improve diabetes screening or diabetes prevention. Doing more 1-hour glucose screening is not going to get more people into diabetes prevention programs.”
When an audience member raised the logistics concern during the session Q&A, Dr. Bergman replied, “It’s no different than recommending colonoscopy or a mammogram. … I tell patients that we want to learn more about what is going on before we make a firm diagnosis. … I’ve done more than a hundred 1-hour glucose tolerance tests and have not seen one patient who refused because of inconvenience. Everything depends on the way we discuss things with patients.”
Recommendation Based on Emerging Evidence
“Intermediate hyperglycemia” is the World Health Organization’s term to refer to either impaired fasting glucose (IFG) or impaired glucose tolerance (IGT). These conditions are often called “prediabetes,” especially in the United States, but that term has been controversial because not everyone with IFG and IGT will go on to develop type 2 diabetes, said Dr. Bergman, who is also director of the NYU Langone Diabetes Prevention Program.
“With ‘prediabetes’ you’re labeling someone with a disease they may not develop. It’s not normal, but it’s not diabetes, so it’s an intermediate state,” he explained.
The statement provides a detailed summary of the data from 19 studies supporting use of a 1-hour plasma glucose of ≥ 155 mg/dL (8.6 mmol/L) to diagnose intermediate hyperglycemia and type 2 diabetes, including links between that level of glycemia and worsened metabolic and atherogenic profiles, risk for microvascular and macrovascular complications and mortality, and identification of risks for obstructive sleep apnea, cystic fibrosis-related diabetes mellitus, fatty liver disease, and premature mortality.
One major problem with current testing, Dr. Bergman said, is that “there is a huge disconnect between A1c and glucose values. … Only about 30% of individuals with an abnormal A1c will have an abnormal fasting glucose, and the inverse is also true. There’s a big mismatch between the two.”
Current guidelines suggest using both A1c and fasting glucose, but Dr. Bergman said that even then “you’re still missing about 20-30% who have IGT. … Part of the problem is that the criteria we use for defining abnormal fasting and 2-hour levels are too high. … 140 [mg/dL] for the 2-hour is too high and 100 [mg/dL] fasting is too high. … And that’s one of the reasons why many people progress to type 2 diabetes, because we’re using screening thresholds that are nonphysiologic.”
But Dr. Selvin disagrees, pointing to her own work showing that “using a combination of fasting glucose and A1c for screening does an excellent job at identifying high-risk individuals.”
She’s also unconvinced by other data cited in the paper. “I am deeply skeptical about 1-hour glucose being more prognostic than all other glycemic tests. … Associations of glucose tests with incident diabetes are inherently a circular analysis since diabetes is defined by elevations in those same tests. It is helpful to look at progression of diabetes, but these analyses are not simple and doing them well, especially with head-to-head comparisons against different glycemic tests, is hard,” she said.
In her view, “lifestyle interventions and weight loss should be recommended in at-risk individuals, including those with overweight and obesity, hypertension, metabolic syndrome, etc., regardless of 1-hour glucose test results.”
She added, “If we want to prevent diabetes, we need to focus on preventing weight gain and obesity and managing cardiometabolic risk factors. Lifestyle interventions are not effective unless they are intensive and patients are highly adherent. We need to make lifestyle interventions available and affordable for patients. That is the big barrier. I don’t think more screening with 1-hour glucose is going to help.”
Dr. Bergman pointed out that the International Diabetes Federation represents more than 100 countries, including many that are middle- and low-income. “They need a simple, cost-effective tool for screening effectively. A1c is more expensive, and fasting glucose alone will underestimate disease prevalence. So, the IDF felt, after doing a fairly comprehensive due diligence, that the data warranted recommendation of the 1-hour glucose.”
He’s hoping other organizations like the ADA and the World Health Organization will sign on to bring this guidance into the primary care arena. This news organization reached out to ADA for comment, but their representative hadn’t responded by press time.
Dr. Bergman had no disclosures. Dr. Selvin was supported by the National Institutes of Health.
A version of this article appeared on Medscape.com.
Smoking Cessation Before Age 40 Years Brings Great Benefits
Chronic smoking remains a major cause of premature mortality on a global scale. Despite intensified efforts to combat this scourge, a quarter of deaths among middle-aged adults in Europe and North America are attributed to it. However, over the past decades, antismoking campaigns have borne fruit, and many smokers have quit before the age of 40 years, enabling some case-control studies.
Among those abstainers who made the right choice, the excess mortality attributable to smoking over a lifetime would be reduced by 90% compared with controls who continued smoking. The estimated benefit is clear, but the analysis lacks nuance. Is smoking cessation beneficial even at older ages? If so, is the effect measurable in terms of magnitude and speed of the effect? An article published online in The New England Journal of Medicine Evidence provided some answers to these questions.
Four-Cohort Meta-Analysis
The study was a meta-analysis of individual data collected within four national cohort studies that were linked to each country’s death registry. Two of these studies were nationally representative. The National Health Interview Survey involved a sample of US citizens living in the community, aged 20-79 years, who were included annually in the cohort between 1997 and 2018. The second, the Canadian Community Health Survey, included subjects in the same age group, with samples analyzed between 2000 and 2014.
In Norway, three cohort studies conducted between 1974 and 2003, in which participants aged 25-79 years were included, were combined to form the Norwegian Health Screening Survey. These were the Counties Study (1974-1988), the 40 Years Study (1985-1999), and the Cohort of Norway (1994-2003), respectively. The fourth cohort was established through recruitment via the UK Biobank, with adults aged 40-73 years invited to participate in the survey. The data analysis ultimately covered a relatively heterogeneous total population of 1.48 million adults, all from high-income countries and followed for 15 years. It relied on the Cox proportional hazards model applied to each study, considering smoker vs nonsmoker status, as well as the time elapsed since smoking cessation (less than 3 years, between 3 and 9 years, or at least 10 years). Statistical adjustments made in the context of multivariate Cox analysis considered age, education, alcohol consumption, and obesity.
Excess Mortality Confirmed
At the end of follow-up, 122,697 deaths were recorded. The comparison of smokers and nonsmokers confirmed smoking-related excess mortality, with adjusted hazard ratios (HRs) estimated at 2.80 for women and 2.70 for men. Smoking shortened life expectancy in the 40- to 79-year-age group by 12 years for women and 13 years for men, in terms of overall mortality. In terms of smoking-attributable specific mortality, the corresponding figures reached 24 and 26 years, respectively. Respiratory diseases ranked highest in both sexes (HR, 7.6 for women and 6.3 for men), followed by cardiovascular diseases (HR, 3.1 for women and 2.9 for men) and cancers (HR, 2.8 for women and 3.1 for men).
The Earlier, the Better
Smoking cessation halves overall excess mortality. Above all, quitting before age 40 years brings overall mortality back to the level of nonsmokers as early as the third year after quitting. The excess mortality decreases even more as the cessation period is prolonged, even after age 40 years. Thus, cessation ≥ 10 years in smokers aged 40-49 years almost cancels out overall excess mortality (-99% in women, -96% in men). The trend is almost as favorable in the older age group (50-59 years), with corresponding figures of -95% and -92%, respectively.
Long-term survival increases in the early years after cessation, especially if it occurs at a younger age, but the benefit remains tangible even in older smokers. Thus, cessation of less than 3 years, effective in patients aged 50-59 years, reduces overall excess mortality by 63% in women and 54% in men. In patients aged 60-79 years, the figures are -40% and -33%, respectively.
Naturally, the earlier the cessation, the greater the number of years gained. It is 12 years for cessation before age 40 years, reduced to 6 years for cessation between 40 and 49 years, and 2.5 years when it is even later (50-59 years). These quantitative results are approximate, given the methodology (a meta-analysis) and some heterogeneity in the studies, as well as the multitude of potential confounding factors that have not all been considered. Nevertheless, the results probably contain a kernel of truth, and their optimistic implications should be highlighted to encourage smokers to abstain, even older ones. Better late than never, even if the benefit of cessation is maximal when it occurs as early as possible, knowing that a minimum of 3 years of cessation would be sufficient to gain years of life.
This story was translated from JIM, which is part of the Medscape professional network, using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication. A version of this article appeared on Medscape.com.
Chronic smoking remains a major cause of premature mortality on a global scale. Despite intensified efforts to combat this scourge, a quarter of deaths among middle-aged adults in Europe and North America are attributed to it. However, over the past decades, antismoking campaigns have borne fruit, and many smokers have quit before the age of 40 years, enabling some case-control studies.
Among those abstainers who made the right choice, the excess mortality attributable to smoking over a lifetime would be reduced by 90% compared with controls who continued smoking. The estimated benefit is clear, but the analysis lacks nuance. Is smoking cessation beneficial even at older ages? If so, is the effect measurable in terms of magnitude and speed of the effect? An article published online in The New England Journal of Medicine Evidence provided some answers to these questions.
Four-Cohort Meta-Analysis
The study was a meta-analysis of individual data collected within four national cohort studies that were linked to each country’s death registry. Two of these studies were nationally representative. The National Health Interview Survey involved a sample of US citizens living in the community, aged 20-79 years, who were included annually in the cohort between 1997 and 2018. The second, the Canadian Community Health Survey, included subjects in the same age group, with samples analyzed between 2000 and 2014.
In Norway, three cohort studies conducted between 1974 and 2003, in which participants aged 25-79 years were included, were combined to form the Norwegian Health Screening Survey. These were the Counties Study (1974-1988), the 40 Years Study (1985-1999), and the Cohort of Norway (1994-2003), respectively. The fourth cohort was established through recruitment via the UK Biobank, with adults aged 40-73 years invited to participate in the survey. The data analysis ultimately covered a relatively heterogeneous total population of 1.48 million adults, all from high-income countries and followed for 15 years. It relied on the Cox proportional hazards model applied to each study, considering smoker vs nonsmoker status, as well as the time elapsed since smoking cessation (less than 3 years, between 3 and 9 years, or at least 10 years). Statistical adjustments made in the context of multivariate Cox analysis considered age, education, alcohol consumption, and obesity.
Excess Mortality Confirmed
At the end of follow-up, 122,697 deaths were recorded. The comparison of smokers and nonsmokers confirmed smoking-related excess mortality, with adjusted hazard ratios (HRs) estimated at 2.80 for women and 2.70 for men. Smoking shortened life expectancy in the 40- to 79-year-age group by 12 years for women and 13 years for men, in terms of overall mortality. In terms of smoking-attributable specific mortality, the corresponding figures reached 24 and 26 years, respectively. Respiratory diseases ranked highest in both sexes (HR, 7.6 for women and 6.3 for men), followed by cardiovascular diseases (HR, 3.1 for women and 2.9 for men) and cancers (HR, 2.8 for women and 3.1 for men).
The Earlier, the Better
Smoking cessation halves overall excess mortality. Above all, quitting before age 40 years brings overall mortality back to the level of nonsmokers as early as the third year after quitting. The excess mortality decreases even more as the cessation period is prolonged, even after age 40 years. Thus, cessation ≥ 10 years in smokers aged 40-49 years almost cancels out overall excess mortality (-99% in women, -96% in men). The trend is almost as favorable in the older age group (50-59 years), with corresponding figures of -95% and -92%, respectively.
Long-term survival increases in the early years after cessation, especially if it occurs at a younger age, but the benefit remains tangible even in older smokers. Thus, cessation of less than 3 years, effective in patients aged 50-59 years, reduces overall excess mortality by 63% in women and 54% in men. In patients aged 60-79 years, the figures are -40% and -33%, respectively.
Naturally, the earlier the cessation, the greater the number of years gained. It is 12 years for cessation before age 40 years, reduced to 6 years for cessation between 40 and 49 years, and 2.5 years when it is even later (50-59 years). These quantitative results are approximate, given the methodology (a meta-analysis) and some heterogeneity in the studies, as well as the multitude of potential confounding factors that have not all been considered. Nevertheless, the results probably contain a kernel of truth, and their optimistic implications should be highlighted to encourage smokers to abstain, even older ones. Better late than never, even if the benefit of cessation is maximal when it occurs as early as possible, knowing that a minimum of 3 years of cessation would be sufficient to gain years of life.
This story was translated from JIM, which is part of the Medscape professional network, using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication. A version of this article appeared on Medscape.com.
Chronic smoking remains a major cause of premature mortality on a global scale. Despite intensified efforts to combat this scourge, a quarter of deaths among middle-aged adults in Europe and North America are attributed to it. However, over the past decades, antismoking campaigns have borne fruit, and many smokers have quit before the age of 40 years, enabling some case-control studies.
Among those abstainers who made the right choice, the excess mortality attributable to smoking over a lifetime would be reduced by 90% compared with controls who continued smoking. The estimated benefit is clear, but the analysis lacks nuance. Is smoking cessation beneficial even at older ages? If so, is the effect measurable in terms of magnitude and speed of the effect? An article published online in The New England Journal of Medicine Evidence provided some answers to these questions.
Four-Cohort Meta-Analysis
The study was a meta-analysis of individual data collected within four national cohort studies that were linked to each country’s death registry. Two of these studies were nationally representative. The National Health Interview Survey involved a sample of US citizens living in the community, aged 20-79 years, who were included annually in the cohort between 1997 and 2018. The second, the Canadian Community Health Survey, included subjects in the same age group, with samples analyzed between 2000 and 2014.
In Norway, three cohort studies conducted between 1974 and 2003, in which participants aged 25-79 years were included, were combined to form the Norwegian Health Screening Survey. These were the Counties Study (1974-1988), the 40 Years Study (1985-1999), and the Cohort of Norway (1994-2003), respectively. The fourth cohort was established through recruitment via the UK Biobank, with adults aged 40-73 years invited to participate in the survey. The data analysis ultimately covered a relatively heterogeneous total population of 1.48 million adults, all from high-income countries and followed for 15 years. It relied on the Cox proportional hazards model applied to each study, considering smoker vs nonsmoker status, as well as the time elapsed since smoking cessation (less than 3 years, between 3 and 9 years, or at least 10 years). Statistical adjustments made in the context of multivariate Cox analysis considered age, education, alcohol consumption, and obesity.
Excess Mortality Confirmed
At the end of follow-up, 122,697 deaths were recorded. The comparison of smokers and nonsmokers confirmed smoking-related excess mortality, with adjusted hazard ratios (HRs) estimated at 2.80 for women and 2.70 for men. Smoking shortened life expectancy in the 40- to 79-year-age group by 12 years for women and 13 years for men, in terms of overall mortality. In terms of smoking-attributable specific mortality, the corresponding figures reached 24 and 26 years, respectively. Respiratory diseases ranked highest in both sexes (HR, 7.6 for women and 6.3 for men), followed by cardiovascular diseases (HR, 3.1 for women and 2.9 for men) and cancers (HR, 2.8 for women and 3.1 for men).
The Earlier, the Better
Smoking cessation halves overall excess mortality. Above all, quitting before age 40 years brings overall mortality back to the level of nonsmokers as early as the third year after quitting. The excess mortality decreases even more as the cessation period is prolonged, even after age 40 years. Thus, cessation ≥ 10 years in smokers aged 40-49 years almost cancels out overall excess mortality (-99% in women, -96% in men). The trend is almost as favorable in the older age group (50-59 years), with corresponding figures of -95% and -92%, respectively.
Long-term survival increases in the early years after cessation, especially if it occurs at a younger age, but the benefit remains tangible even in older smokers. Thus, cessation of less than 3 years, effective in patients aged 50-59 years, reduces overall excess mortality by 63% in women and 54% in men. In patients aged 60-79 years, the figures are -40% and -33%, respectively.
Naturally, the earlier the cessation, the greater the number of years gained. It is 12 years for cessation before age 40 years, reduced to 6 years for cessation between 40 and 49 years, and 2.5 years when it is even later (50-59 years). These quantitative results are approximate, given the methodology (a meta-analysis) and some heterogeneity in the studies, as well as the multitude of potential confounding factors that have not all been considered. Nevertheless, the results probably contain a kernel of truth, and their optimistic implications should be highlighted to encourage smokers to abstain, even older ones. Better late than never, even if the benefit of cessation is maximal when it occurs as early as possible, knowing that a minimum of 3 years of cessation would be sufficient to gain years of life.
This story was translated from JIM, which is part of the Medscape professional network, using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication. A version of this article appeared on Medscape.com.
How Does Snoring Affect Cardiovascular Health?
Snoring is a common disorder that affects 20%-40% of the general population. The mechanism of snoring is the vibration of anatomical structures in the pharyngeal airways. The flutter of the soft palate explains the harsh aspect of the snoring sound, which occurs during natural sleep or drug-induced sleep. The presentation of snoring may vary throughout the night or between nights, with a subjective, and therefore inconsistent, assessment of its loudness.
Objective evaluation of snoring is important for clinical decision-making and predicting the effect of therapeutic interventions. It also provides information regarding the site and degree of upper airway obstruction. Snoring is one of the main features of sleep-disordered breathing, including hypopnea events, which reflect partial upper airway obstruction.
Obstructive sleep apnea (OSA) is characterized by episodes of complete (apnea) or partial (hypopnea) collapse of the upper airways with associated oxygen desaturation or awakening from sleep. Most patients with OSA snore loudly almost every night. However, in the Sleep Heart Health Study, one-third of participants with OSA reported no snoring, while one-third of snoring participants did not meet the criteria for OSA. Therefore, subjective assessments of snoring (self-reported) may not be sufficiently reliable to assess its potential impact on cardiovascular (CV) health outcomes.
CV Effects
OSA has been hypothesized as a modifiable risk factor for CV diseases (CVD), including hypertension, coronary artery disease (CAD), atrial fibrillation, heart failure, and stroke, primarily because of the results of traditional observational studies. Snoring is reported as a symptom of the early stage of OSA and has also been associated with a higher risk for CVD. However, establishing causality based on observational studies is difficult because of residual confounding from unknown or unmeasured factors and reverse causality (i.e., the scenario in which CVD increases the risk for OSA or snoring). A Mendelian randomization study, using the natural random allocation of genetic variants as instruments capable of producing results analogous to those of randomized controlled trials, suggested that OSA and snoring increase the risk for hypertension and CAD, with associations partly driven by body mass index (BMI). Conversely, no evidence was found that CVD causally influenced OSA or snoring.
Snoring has been associated with multiple subclinical markers of CV pathology, including high blood pressure, and loud snoring can interfere with restorative sleep and contribute to the risk for hypertension and other adverse outcomes in snorers. However, evidence on the associations between snoring and CV health outcomes remains limited and is primarily based on subjective assessments of snoring or small clinical samples with objective assessments of snoring for only 1 night.
Snoring and Hypertension
A study of 12,287 middle-aged patients (age, 50 years) who were predominantly males (88%) and generally overweight (BMI, 28 kg/m2) determined the prevalence of snoring and its association with the prevalence of hypertension using objective evaluation of snoring over multiple nights and multiple daytime blood pressure measurements. The findings included the following observations:
An increase in snoring duration was associated with a 3-mmHg increase in systolic (SBP) and a 4 mmHg increase in diastolic blood pressure (DBP) in patients with frequent and regular snoring, compared with those with infrequent snoring, regardless of age, BMI, sex, and estimated apnea/hypopnea index.
The association between severe OSA alone and blood pressure had an effect size similar to that of the association between snoring alone and blood pressure. In a model where OSA severity was classified and snoring duration was stratified into quartiles, severe OSA without snoring was associated with 3.6 mmHg higher SBP and 3.5 mmHg higher DBP, compared with the absence of snoring or OSA. Participants without OSA but with intense snoring (4th quartile) had 3.8 mmHg higher SBP and 4.5 mmHg higher DBP compared with participants without nighttime apnea or snoring.
Snoring was significantly associated with uncontrolled hypertension. There was a 20% increase in the probability of uncontrolled hypertension in subjects aged > 50 years with obesity and a 98% increase in subjects aged ≤ 50 years with normal BMI.
Duration of snoring was associated with an 87% increase in the likelihood of uncontrolled hypertension.
Implications for Practice
This study indicates that 15% of a predominantly overweight male population snore for > 20% of the night and about 10% of these subjects without nighttime apnea snore for > 12% of the night.
Regular nighttime snoring is associated with elevated blood pressure and uncontrolled hypertension, regardless of the presence or severity of OSA.
Physicians must be aware of the potential consequences of snoring on the risk for hypertension, and these results highlight the need to consider snoring in clinical care and in the management of sleep problems, especially in the context of managing arterial hypertension.
This story was translated from Univadis Italy, which is part of the Medscape professional network, using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication. A version of this article appeared on Medscape.com.
Snoring is a common disorder that affects 20%-40% of the general population. The mechanism of snoring is the vibration of anatomical structures in the pharyngeal airways. The flutter of the soft palate explains the harsh aspect of the snoring sound, which occurs during natural sleep or drug-induced sleep. The presentation of snoring may vary throughout the night or between nights, with a subjective, and therefore inconsistent, assessment of its loudness.
Objective evaluation of snoring is important for clinical decision-making and predicting the effect of therapeutic interventions. It also provides information regarding the site and degree of upper airway obstruction. Snoring is one of the main features of sleep-disordered breathing, including hypopnea events, which reflect partial upper airway obstruction.
Obstructive sleep apnea (OSA) is characterized by episodes of complete (apnea) or partial (hypopnea) collapse of the upper airways with associated oxygen desaturation or awakening from sleep. Most patients with OSA snore loudly almost every night. However, in the Sleep Heart Health Study, one-third of participants with OSA reported no snoring, while one-third of snoring participants did not meet the criteria for OSA. Therefore, subjective assessments of snoring (self-reported) may not be sufficiently reliable to assess its potential impact on cardiovascular (CV) health outcomes.
CV Effects
OSA has been hypothesized as a modifiable risk factor for CV diseases (CVD), including hypertension, coronary artery disease (CAD), atrial fibrillation, heart failure, and stroke, primarily because of the results of traditional observational studies. Snoring is reported as a symptom of the early stage of OSA and has also been associated with a higher risk for CVD. However, establishing causality based on observational studies is difficult because of residual confounding from unknown or unmeasured factors and reverse causality (i.e., the scenario in which CVD increases the risk for OSA or snoring). A Mendelian randomization study, using the natural random allocation of genetic variants as instruments capable of producing results analogous to those of randomized controlled trials, suggested that OSA and snoring increase the risk for hypertension and CAD, with associations partly driven by body mass index (BMI). Conversely, no evidence was found that CVD causally influenced OSA or snoring.
Snoring has been associated with multiple subclinical markers of CV pathology, including high blood pressure, and loud snoring can interfere with restorative sleep and contribute to the risk for hypertension and other adverse outcomes in snorers. However, evidence on the associations between snoring and CV health outcomes remains limited and is primarily based on subjective assessments of snoring or small clinical samples with objective assessments of snoring for only 1 night.
Snoring and Hypertension
A study of 12,287 middle-aged patients (age, 50 years) who were predominantly males (88%) and generally overweight (BMI, 28 kg/m2) determined the prevalence of snoring and its association with the prevalence of hypertension using objective evaluation of snoring over multiple nights and multiple daytime blood pressure measurements. The findings included the following observations:
An increase in snoring duration was associated with a 3-mmHg increase in systolic (SBP) and a 4 mmHg increase in diastolic blood pressure (DBP) in patients with frequent and regular snoring, compared with those with infrequent snoring, regardless of age, BMI, sex, and estimated apnea/hypopnea index.
The association between severe OSA alone and blood pressure had an effect size similar to that of the association between snoring alone and blood pressure. In a model where OSA severity was classified and snoring duration was stratified into quartiles, severe OSA without snoring was associated with 3.6 mmHg higher SBP and 3.5 mmHg higher DBP, compared with the absence of snoring or OSA. Participants without OSA but with intense snoring (4th quartile) had 3.8 mmHg higher SBP and 4.5 mmHg higher DBP compared with participants without nighttime apnea or snoring.
Snoring was significantly associated with uncontrolled hypertension. There was a 20% increase in the probability of uncontrolled hypertension in subjects aged > 50 years with obesity and a 98% increase in subjects aged ≤ 50 years with normal BMI.
Duration of snoring was associated with an 87% increase in the likelihood of uncontrolled hypertension.
Implications for Practice
This study indicates that 15% of a predominantly overweight male population snore for > 20% of the night and about 10% of these subjects without nighttime apnea snore for > 12% of the night.
Regular nighttime snoring is associated with elevated blood pressure and uncontrolled hypertension, regardless of the presence or severity of OSA.
Physicians must be aware of the potential consequences of snoring on the risk for hypertension, and these results highlight the need to consider snoring in clinical care and in the management of sleep problems, especially in the context of managing arterial hypertension.
This story was translated from Univadis Italy, which is part of the Medscape professional network, using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication. A version of this article appeared on Medscape.com.
Snoring is a common disorder that affects 20%-40% of the general population. The mechanism of snoring is the vibration of anatomical structures in the pharyngeal airways. The flutter of the soft palate explains the harsh aspect of the snoring sound, which occurs during natural sleep or drug-induced sleep. The presentation of snoring may vary throughout the night or between nights, with a subjective, and therefore inconsistent, assessment of its loudness.
Objective evaluation of snoring is important for clinical decision-making and predicting the effect of therapeutic interventions. It also provides information regarding the site and degree of upper airway obstruction. Snoring is one of the main features of sleep-disordered breathing, including hypopnea events, which reflect partial upper airway obstruction.
Obstructive sleep apnea (OSA) is characterized by episodes of complete (apnea) or partial (hypopnea) collapse of the upper airways with associated oxygen desaturation or awakening from sleep. Most patients with OSA snore loudly almost every night. However, in the Sleep Heart Health Study, one-third of participants with OSA reported no snoring, while one-third of snoring participants did not meet the criteria for OSA. Therefore, subjective assessments of snoring (self-reported) may not be sufficiently reliable to assess its potential impact on cardiovascular (CV) health outcomes.
CV Effects
OSA has been hypothesized as a modifiable risk factor for CV diseases (CVD), including hypertension, coronary artery disease (CAD), atrial fibrillation, heart failure, and stroke, primarily because of the results of traditional observational studies. Snoring is reported as a symptom of the early stage of OSA and has also been associated with a higher risk for CVD. However, establishing causality based on observational studies is difficult because of residual confounding from unknown or unmeasured factors and reverse causality (i.e., the scenario in which CVD increases the risk for OSA or snoring). A Mendelian randomization study, using the natural random allocation of genetic variants as instruments capable of producing results analogous to those of randomized controlled trials, suggested that OSA and snoring increase the risk for hypertension and CAD, with associations partly driven by body mass index (BMI). Conversely, no evidence was found that CVD causally influenced OSA or snoring.
Snoring has been associated with multiple subclinical markers of CV pathology, including high blood pressure, and loud snoring can interfere with restorative sleep and contribute to the risk for hypertension and other adverse outcomes in snorers. However, evidence on the associations between snoring and CV health outcomes remains limited and is primarily based on subjective assessments of snoring or small clinical samples with objective assessments of snoring for only 1 night.
Snoring and Hypertension
A study of 12,287 middle-aged patients (age, 50 years) who were predominantly males (88%) and generally overweight (BMI, 28 kg/m2) determined the prevalence of snoring and its association with the prevalence of hypertension using objective evaluation of snoring over multiple nights and multiple daytime blood pressure measurements. The findings included the following observations:
An increase in snoring duration was associated with a 3-mmHg increase in systolic (SBP) and a 4 mmHg increase in diastolic blood pressure (DBP) in patients with frequent and regular snoring, compared with those with infrequent snoring, regardless of age, BMI, sex, and estimated apnea/hypopnea index.
The association between severe OSA alone and blood pressure had an effect size similar to that of the association between snoring alone and blood pressure. In a model where OSA severity was classified and snoring duration was stratified into quartiles, severe OSA without snoring was associated with 3.6 mmHg higher SBP and 3.5 mmHg higher DBP, compared with the absence of snoring or OSA. Participants without OSA but with intense snoring (4th quartile) had 3.8 mmHg higher SBP and 4.5 mmHg higher DBP compared with participants without nighttime apnea or snoring.
Snoring was significantly associated with uncontrolled hypertension. There was a 20% increase in the probability of uncontrolled hypertension in subjects aged > 50 years with obesity and a 98% increase in subjects aged ≤ 50 years with normal BMI.
Duration of snoring was associated with an 87% increase in the likelihood of uncontrolled hypertension.
Implications for Practice
This study indicates that 15% of a predominantly overweight male population snore for > 20% of the night and about 10% of these subjects without nighttime apnea snore for > 12% of the night.
Regular nighttime snoring is associated with elevated blood pressure and uncontrolled hypertension, regardless of the presence or severity of OSA.
Physicians must be aware of the potential consequences of snoring on the risk for hypertension, and these results highlight the need to consider snoring in clinical care and in the management of sleep problems, especially in the context of managing arterial hypertension.
This story was translated from Univadis Italy, which is part of the Medscape professional network, using several editorial tools, including AI, as part of the process. Human editors reviewed this content before publication. A version of this article appeared on Medscape.com.
Study Sounds Alert About GLP-1 RA Use and Aspiration Risk
TOPLINE:
Patients on weekly glucagon-like peptide-1 receptor agonists (GLP-1 RAs) have high residual gastric content, a major risk factor for aspiration under anesthesia, despite following fasting guidelines before undergoing elective procedures.
METHODOLOGY:
- The increasing use of GLP-1 RAs to manage weight and hyperglycemia has sparked safety concerns because of the drugs’ association with slow gastric emptying, a major risk factor for aspiration under anesthesia.
- This cross-sectional study used gastric ultrasonography to examine the link between GLP-1 RA use and the prevalence of increased residual gastric content.
- All 124 participants (median age, 56 years; 60% women) — half of whom received once-weekly GLP-1 RAs such as semaglutide, dulaglutide, or tirzepatide — adhered to the guideline-recommended fasting duration before undergoing elective procedures under anesthesia.
- The primary outcome focused on identifying increased residual gastric content, defined by the presence of solids, thick liquids, or > 1.5 mL/kg of clear liquids on ultrasound.
- An exploratory analysis examined the association between the duration of GLP-1 RA discontinuation and increased residual gastric content.
TAKEAWAY:
- The adjusted prevalence of increased residual gastric content was 30.5% (95% CI, 9.9%-51.2%) higher in participants who received GLP-1 RA than those who did not.
- Most patients took their last dose of GLP-1 RA within 5 days before their procedure, but elevated residual gastric content persisted even after 7 days of GLP-1 RA discontinuation.
- There was also no significant association between the type of GLP-1 RA used and the prevalence of increased residual gastric content.
IN PRACTICE:
“We expect healthcare professionals will encounter these classes of drugs with increasing frequency in the perioperative period. Perioperative physicians, including anesthesiologists, surgeons, and primary care physicians, should be well-informed about the safety implications of GLP-1 RA drugs,” the authors wrote.
SOURCE:
The study was led by Sudipta Sen, MD, from the Department of Anesthesiology, Critical Care and Pain Medicine, McGovern Medical School, University of Texas Health Center at Houston, Houston, Texas, and published online in JAMA Surgery.
LIMITATIONS:
Residual gastric content, the primary outcome, served as a proxy for aspiration risk and does not have an exact threshold of volume associated with increased risk. The study did not directly evaluate aspiration events. The authors also acknowledged potential bias from unmeasured confounders owing to the observational nature of this study. A small sample size limited the ability to detect a risk difference for each additional day of drug discontinuation before surgery.
DISCLOSURES:
One of the authors reported receiving a grant from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Patients on weekly glucagon-like peptide-1 receptor agonists (GLP-1 RAs) have high residual gastric content, a major risk factor for aspiration under anesthesia, despite following fasting guidelines before undergoing elective procedures.
METHODOLOGY:
- The increasing use of GLP-1 RAs to manage weight and hyperglycemia has sparked safety concerns because of the drugs’ association with slow gastric emptying, a major risk factor for aspiration under anesthesia.
- This cross-sectional study used gastric ultrasonography to examine the link between GLP-1 RA use and the prevalence of increased residual gastric content.
- All 124 participants (median age, 56 years; 60% women) — half of whom received once-weekly GLP-1 RAs such as semaglutide, dulaglutide, or tirzepatide — adhered to the guideline-recommended fasting duration before undergoing elective procedures under anesthesia.
- The primary outcome focused on identifying increased residual gastric content, defined by the presence of solids, thick liquids, or > 1.5 mL/kg of clear liquids on ultrasound.
- An exploratory analysis examined the association between the duration of GLP-1 RA discontinuation and increased residual gastric content.
TAKEAWAY:
- The adjusted prevalence of increased residual gastric content was 30.5% (95% CI, 9.9%-51.2%) higher in participants who received GLP-1 RA than those who did not.
- Most patients took their last dose of GLP-1 RA within 5 days before their procedure, but elevated residual gastric content persisted even after 7 days of GLP-1 RA discontinuation.
- There was also no significant association between the type of GLP-1 RA used and the prevalence of increased residual gastric content.
IN PRACTICE:
“We expect healthcare professionals will encounter these classes of drugs with increasing frequency in the perioperative period. Perioperative physicians, including anesthesiologists, surgeons, and primary care physicians, should be well-informed about the safety implications of GLP-1 RA drugs,” the authors wrote.
SOURCE:
The study was led by Sudipta Sen, MD, from the Department of Anesthesiology, Critical Care and Pain Medicine, McGovern Medical School, University of Texas Health Center at Houston, Houston, Texas, and published online in JAMA Surgery.
LIMITATIONS:
Residual gastric content, the primary outcome, served as a proxy for aspiration risk and does not have an exact threshold of volume associated with increased risk. The study did not directly evaluate aspiration events. The authors also acknowledged potential bias from unmeasured confounders owing to the observational nature of this study. A small sample size limited the ability to detect a risk difference for each additional day of drug discontinuation before surgery.
DISCLOSURES:
One of the authors reported receiving a grant from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
TOPLINE:
Patients on weekly glucagon-like peptide-1 receptor agonists (GLP-1 RAs) have high residual gastric content, a major risk factor for aspiration under anesthesia, despite following fasting guidelines before undergoing elective procedures.
METHODOLOGY:
- The increasing use of GLP-1 RAs to manage weight and hyperglycemia has sparked safety concerns because of the drugs’ association with slow gastric emptying, a major risk factor for aspiration under anesthesia.
- This cross-sectional study used gastric ultrasonography to examine the link between GLP-1 RA use and the prevalence of increased residual gastric content.
- All 124 participants (median age, 56 years; 60% women) — half of whom received once-weekly GLP-1 RAs such as semaglutide, dulaglutide, or tirzepatide — adhered to the guideline-recommended fasting duration before undergoing elective procedures under anesthesia.
- The primary outcome focused on identifying increased residual gastric content, defined by the presence of solids, thick liquids, or > 1.5 mL/kg of clear liquids on ultrasound.
- An exploratory analysis examined the association between the duration of GLP-1 RA discontinuation and increased residual gastric content.
TAKEAWAY:
- The adjusted prevalence of increased residual gastric content was 30.5% (95% CI, 9.9%-51.2%) higher in participants who received GLP-1 RA than those who did not.
- Most patients took their last dose of GLP-1 RA within 5 days before their procedure, but elevated residual gastric content persisted even after 7 days of GLP-1 RA discontinuation.
- There was also no significant association between the type of GLP-1 RA used and the prevalence of increased residual gastric content.
IN PRACTICE:
“We expect healthcare professionals will encounter these classes of drugs with increasing frequency in the perioperative period. Perioperative physicians, including anesthesiologists, surgeons, and primary care physicians, should be well-informed about the safety implications of GLP-1 RA drugs,” the authors wrote.
SOURCE:
The study was led by Sudipta Sen, MD, from the Department of Anesthesiology, Critical Care and Pain Medicine, McGovern Medical School, University of Texas Health Center at Houston, Houston, Texas, and published online in JAMA Surgery.
LIMITATIONS:
Residual gastric content, the primary outcome, served as a proxy for aspiration risk and does not have an exact threshold of volume associated with increased risk. The study did not directly evaluate aspiration events. The authors also acknowledged potential bias from unmeasured confounders owing to the observational nature of this study. A small sample size limited the ability to detect a risk difference for each additional day of drug discontinuation before surgery.
DISCLOSURES:
One of the authors reported receiving a grant from the National Institutes of Health. The authors declared no conflicts of interest.
A version of this article appeared on Medscape.com.
Does worsening metabolic syndrome increase the risk of developing cancer?
The conditions that comprise metabolic syndrome (high blood pressure, high blood sugar, increased abdominal adiposity, and high cholesterol and triglycerides) have been associated with an increased risk of diseases, including heart disease, stroke, and type 2 diabetes, wrote Li Deng, PhD, of Capital Medical University, Beijing, China, and colleagues.
A systematic review and meta-analysis published in Diabetes Care in 2012 showed an association between the presence of metabolic syndrome and an increased risk of various cancers including liver, bladder, pancreatic, breast, and colorectal.
More recently, a 2019 study published in Diabetes showed evidence of increased risk for certain cancers (pancreatic, kidney, uterine, cervical) but no increased risk for cancer overall.
However, the reasons for this association between metabolic syndrome and cancer remain unclear, and the effect of the fluctuating nature of metabolic syndrome over time on long-term cancer risk has not been explored, the researchers wrote.
What Does New Study Add to Other Research on Metabolic Syndrome and Cancer Risk?
In the new study, published in Cancer on March 11 (doi: 10.1002/cncr.35235), 44,115 adults in China were separated into four trajectories based on metabolic syndrome scores for the period from 2006 to 2010. The scores were based on clinical evidence of metabolic syndrome, defined using the International Diabetes Federation criteria of central obesity and the presence of at least two other factors including increased triglycerides, decreased HDL cholesterol, high blood pressure (or treatment for previously diagnosed hypertension), and increased fasting plasma glucose (or previous diagnosis of type 2 diabetes).
The average age of the participants was 49 years. The four trajectories of metabolic syndrome were low-stable (10.56% of participants), moderate-low (40.84%), moderate-high (41.46%), and elevated-increasing (7.14%), based on trends from the individuals’ initial physical exams on entering the study.
Over a median follow-up period of 9.4 years (from 2010 to 2021), 2,271 cancer diagnoses were reported in the study population. Those with an elevated-increasing metabolic syndrome trajectory had 1.3 times the risk of any cancer compared with those in the low-stable group. Risk for breast cancer, endometrial cancer, kidney cancer, colorectal cancer, and liver cancer in the highest trajectory group were 2.1, 3.3, 4.5, 2.5, and 1.6 times higher, respectively, compared to the lowest group. The increased risk in the elevated-trajectory group for all cancer types persisted when the low-stable, moderate-low, and moderate-high trajectory pattern groups were combined.
The researchers also examined the impact of chronic inflammation and found that individuals with persistently high metabolic syndrome scores and concurrent chronic inflammation had the highest risks of breast, endometrial, colon, and liver cancer. However, individuals with persistently high metabolic syndrome scores and no concurrent chronic inflammation had the highest risk of kidney cancer.
What Are the Limitations of This Research?
The researchers of the current study acknowledged the lack of information on other causes of cancer, including dietary habits, hepatitis C infection, and Helicobacter pylori infection. Other limitations include the focus only on individuals from a single community of mainly middle-aged men in China that may not generalize to other populations.
Also, the metabolic syndrome trajectories did not change much over time, which may be related to the short 4-year study period.
What Is the Takeaway Message for Clinical Practice?
The results suggest that monitoring and managing metabolic syndrome could help reduce cancer risk, the researchers concluded.
“This research suggests that proactive and continuous management of metabolic syndrome may serve as an essential strategy in preventing cancer,” senior author Han-Ping Shi, MD, PhD, of Capital Medical University in Beijing, said in a press release accompanying the study.
More research is needed to assess the impact of these interventions on cancer risk, he noted. However, the data from the current study can guide future research that may lead to more targeted treatments and more effective preventive strategies, he said in a statement.
The study was supported by the National Key Research and Development Program of China. The researchers had no financial conflicts to disclose.
The conditions that comprise metabolic syndrome (high blood pressure, high blood sugar, increased abdominal adiposity, and high cholesterol and triglycerides) have been associated with an increased risk of diseases, including heart disease, stroke, and type 2 diabetes, wrote Li Deng, PhD, of Capital Medical University, Beijing, China, and colleagues.
A systematic review and meta-analysis published in Diabetes Care in 2012 showed an association between the presence of metabolic syndrome and an increased risk of various cancers including liver, bladder, pancreatic, breast, and colorectal.
More recently, a 2019 study published in Diabetes showed evidence of increased risk for certain cancers (pancreatic, kidney, uterine, cervical) but no increased risk for cancer overall.
However, the reasons for this association between metabolic syndrome and cancer remain unclear, and the effect of the fluctuating nature of metabolic syndrome over time on long-term cancer risk has not been explored, the researchers wrote.
What Does New Study Add to Other Research on Metabolic Syndrome and Cancer Risk?
In the new study, published in Cancer on March 11 (doi: 10.1002/cncr.35235), 44,115 adults in China were separated into four trajectories based on metabolic syndrome scores for the period from 2006 to 2010. The scores were based on clinical evidence of metabolic syndrome, defined using the International Diabetes Federation criteria of central obesity and the presence of at least two other factors including increased triglycerides, decreased HDL cholesterol, high blood pressure (or treatment for previously diagnosed hypertension), and increased fasting plasma glucose (or previous diagnosis of type 2 diabetes).
The average age of the participants was 49 years. The four trajectories of metabolic syndrome were low-stable (10.56% of participants), moderate-low (40.84%), moderate-high (41.46%), and elevated-increasing (7.14%), based on trends from the individuals’ initial physical exams on entering the study.
Over a median follow-up period of 9.4 years (from 2010 to 2021), 2,271 cancer diagnoses were reported in the study population. Those with an elevated-increasing metabolic syndrome trajectory had 1.3 times the risk of any cancer compared with those in the low-stable group. Risk for breast cancer, endometrial cancer, kidney cancer, colorectal cancer, and liver cancer in the highest trajectory group were 2.1, 3.3, 4.5, 2.5, and 1.6 times higher, respectively, compared to the lowest group. The increased risk in the elevated-trajectory group for all cancer types persisted when the low-stable, moderate-low, and moderate-high trajectory pattern groups were combined.
The researchers also examined the impact of chronic inflammation and found that individuals with persistently high metabolic syndrome scores and concurrent chronic inflammation had the highest risks of breast, endometrial, colon, and liver cancer. However, individuals with persistently high metabolic syndrome scores and no concurrent chronic inflammation had the highest risk of kidney cancer.
What Are the Limitations of This Research?
The researchers of the current study acknowledged the lack of information on other causes of cancer, including dietary habits, hepatitis C infection, and Helicobacter pylori infection. Other limitations include the focus only on individuals from a single community of mainly middle-aged men in China that may not generalize to other populations.
Also, the metabolic syndrome trajectories did not change much over time, which may be related to the short 4-year study period.
What Is the Takeaway Message for Clinical Practice?
The results suggest that monitoring and managing metabolic syndrome could help reduce cancer risk, the researchers concluded.
“This research suggests that proactive and continuous management of metabolic syndrome may serve as an essential strategy in preventing cancer,” senior author Han-Ping Shi, MD, PhD, of Capital Medical University in Beijing, said in a press release accompanying the study.
More research is needed to assess the impact of these interventions on cancer risk, he noted. However, the data from the current study can guide future research that may lead to more targeted treatments and more effective preventive strategies, he said in a statement.
The study was supported by the National Key Research and Development Program of China. The researchers had no financial conflicts to disclose.
The conditions that comprise metabolic syndrome (high blood pressure, high blood sugar, increased abdominal adiposity, and high cholesterol and triglycerides) have been associated with an increased risk of diseases, including heart disease, stroke, and type 2 diabetes, wrote Li Deng, PhD, of Capital Medical University, Beijing, China, and colleagues.
A systematic review and meta-analysis published in Diabetes Care in 2012 showed an association between the presence of metabolic syndrome and an increased risk of various cancers including liver, bladder, pancreatic, breast, and colorectal.
More recently, a 2019 study published in Diabetes showed evidence of increased risk for certain cancers (pancreatic, kidney, uterine, cervical) but no increased risk for cancer overall.
However, the reasons for this association between metabolic syndrome and cancer remain unclear, and the effect of the fluctuating nature of metabolic syndrome over time on long-term cancer risk has not been explored, the researchers wrote.
What Does New Study Add to Other Research on Metabolic Syndrome and Cancer Risk?
In the new study, published in Cancer on March 11 (doi: 10.1002/cncr.35235), 44,115 adults in China were separated into four trajectories based on metabolic syndrome scores for the period from 2006 to 2010. The scores were based on clinical evidence of metabolic syndrome, defined using the International Diabetes Federation criteria of central obesity and the presence of at least two other factors including increased triglycerides, decreased HDL cholesterol, high blood pressure (or treatment for previously diagnosed hypertension), and increased fasting plasma glucose (or previous diagnosis of type 2 diabetes).
The average age of the participants was 49 years. The four trajectories of metabolic syndrome were low-stable (10.56% of participants), moderate-low (40.84%), moderate-high (41.46%), and elevated-increasing (7.14%), based on trends from the individuals’ initial physical exams on entering the study.
Over a median follow-up period of 9.4 years (from 2010 to 2021), 2,271 cancer diagnoses were reported in the study population. Those with an elevated-increasing metabolic syndrome trajectory had 1.3 times the risk of any cancer compared with those in the low-stable group. Risk for breast cancer, endometrial cancer, kidney cancer, colorectal cancer, and liver cancer in the highest trajectory group were 2.1, 3.3, 4.5, 2.5, and 1.6 times higher, respectively, compared to the lowest group. The increased risk in the elevated-trajectory group for all cancer types persisted when the low-stable, moderate-low, and moderate-high trajectory pattern groups were combined.
The researchers also examined the impact of chronic inflammation and found that individuals with persistently high metabolic syndrome scores and concurrent chronic inflammation had the highest risks of breast, endometrial, colon, and liver cancer. However, individuals with persistently high metabolic syndrome scores and no concurrent chronic inflammation had the highest risk of kidney cancer.
What Are the Limitations of This Research?
The researchers of the current study acknowledged the lack of information on other causes of cancer, including dietary habits, hepatitis C infection, and Helicobacter pylori infection. Other limitations include the focus only on individuals from a single community of mainly middle-aged men in China that may not generalize to other populations.
Also, the metabolic syndrome trajectories did not change much over time, which may be related to the short 4-year study period.
What Is the Takeaway Message for Clinical Practice?
The results suggest that monitoring and managing metabolic syndrome could help reduce cancer risk, the researchers concluded.
“This research suggests that proactive and continuous management of metabolic syndrome may serve as an essential strategy in preventing cancer,” senior author Han-Ping Shi, MD, PhD, of Capital Medical University in Beijing, said in a press release accompanying the study.
More research is needed to assess the impact of these interventions on cancer risk, he noted. However, the data from the current study can guide future research that may lead to more targeted treatments and more effective preventive strategies, he said in a statement.
The study was supported by the National Key Research and Development Program of China. The researchers had no financial conflicts to disclose.
FROM CANCER
COVID Virus Can Remain in the Body Over a Year
Scientists at the University of California, San Francisco, have discovered that remnants of the COVID-19 virus can linger in blood and tissue for more than a year after a person is first infected.
In their research on long COVID, the scientists found COVID antigens in the blood for up to 14 months after infection, and in tissue samples for more than 2 years after infection.
“These two studies provide some of the strongest evidence so far that COVID antigens can persist in some people, even though we think they have normal immune responses,” Michael Peluso, MD, an infectious disease researcher in the UCSF School of Medicine, who led both studies, said in a statement.
Scientists don’t know what causes long COVID, in which symptoms of the illness persist months or years after recovery. The most common symptoms are extreme fatigue, shortness of breath, loss of smell, and muscle aches.
The UCSF research team examined blood samples from 171 infected people and found the COVID “spike” protein was still present up to 14 months after infection in some people. The antigens were found more often in people who were hospitalized with COVID or who reported being very sick but were not hospitalized.
Researchers next looked at the UCSF Long COVID Tissue Bank, which contains samples donated by patients with and without long COVID.
They found portions of viral RNA in the tissue up to 2 years after people were infected, though there was no evidence of reinfection. Those viral fragments were found in connective tissue where immune cells are, suggesting that the fragments caused the immune system to attack, according to the researchers.
The UCSF team is running clinical trials to find out if monoclonal antibodies or antiviral drugs can remove the virus.
The findings were presented in Denver this week at the Conference on Retroviruses and Opportunistic Infections.
A version of this article appeared on WebMD.com.
Scientists at the University of California, San Francisco, have discovered that remnants of the COVID-19 virus can linger in blood and tissue for more than a year after a person is first infected.
In their research on long COVID, the scientists found COVID antigens in the blood for up to 14 months after infection, and in tissue samples for more than 2 years after infection.
“These two studies provide some of the strongest evidence so far that COVID antigens can persist in some people, even though we think they have normal immune responses,” Michael Peluso, MD, an infectious disease researcher in the UCSF School of Medicine, who led both studies, said in a statement.
Scientists don’t know what causes long COVID, in which symptoms of the illness persist months or years after recovery. The most common symptoms are extreme fatigue, shortness of breath, loss of smell, and muscle aches.
The UCSF research team examined blood samples from 171 infected people and found the COVID “spike” protein was still present up to 14 months after infection in some people. The antigens were found more often in people who were hospitalized with COVID or who reported being very sick but were not hospitalized.
Researchers next looked at the UCSF Long COVID Tissue Bank, which contains samples donated by patients with and without long COVID.
They found portions of viral RNA in the tissue up to 2 years after people were infected, though there was no evidence of reinfection. Those viral fragments were found in connective tissue where immune cells are, suggesting that the fragments caused the immune system to attack, according to the researchers.
The UCSF team is running clinical trials to find out if monoclonal antibodies or antiviral drugs can remove the virus.
The findings were presented in Denver this week at the Conference on Retroviruses and Opportunistic Infections.
A version of this article appeared on WebMD.com.
Scientists at the University of California, San Francisco, have discovered that remnants of the COVID-19 virus can linger in blood and tissue for more than a year after a person is first infected.
In their research on long COVID, the scientists found COVID antigens in the blood for up to 14 months after infection, and in tissue samples for more than 2 years after infection.
“These two studies provide some of the strongest evidence so far that COVID antigens can persist in some people, even though we think they have normal immune responses,” Michael Peluso, MD, an infectious disease researcher in the UCSF School of Medicine, who led both studies, said in a statement.
Scientists don’t know what causes long COVID, in which symptoms of the illness persist months or years after recovery. The most common symptoms are extreme fatigue, shortness of breath, loss of smell, and muscle aches.
The UCSF research team examined blood samples from 171 infected people and found the COVID “spike” protein was still present up to 14 months after infection in some people. The antigens were found more often in people who were hospitalized with COVID or who reported being very sick but were not hospitalized.
Researchers next looked at the UCSF Long COVID Tissue Bank, which contains samples donated by patients with and without long COVID.
They found portions of viral RNA in the tissue up to 2 years after people were infected, though there was no evidence of reinfection. Those viral fragments were found in connective tissue where immune cells are, suggesting that the fragments caused the immune system to attack, according to the researchers.
The UCSF team is running clinical trials to find out if monoclonal antibodies or antiviral drugs can remove the virus.
The findings were presented in Denver this week at the Conference on Retroviruses and Opportunistic Infections.
A version of this article appeared on WebMD.com.