User login
Long-term cannabis use linked to dementia risk factors
A large prospective, longitudinal study showed long-term cannabis users had an intelligence quotient (IQ) decline from age 18 to midlife (mean, 5.5 IQ points), poorer learning and processing speed, compared with childhood, and self-reported memory and attention problems. Long-term cannabis users also showed hippocampal atrophy at midlife (age 45), which combined with mild midlife cognitive deficits, all known risk factors for dementia.
“Long-term cannabis users – people who have used cannabis from 18 or 19 years old and continued using through midlife – showed cognitive deficits, compared with nonusers. They also showed more severe cognitive deficits, compared with long-term alcohol users and long-term tobacco users. But people who used infrequently or recreationally in midlife did not show as severe cognitive deficits. Cognitive deficits were confined to cannabis users,” lead investigator Madeline Meier, PhD, associate professor of psychology, Arizona State University, Tempe, said in an interview.
“Long-term cannabis users had smaller hippocampal volume, but we also found that smaller hippocampal volume did not explain the cognitive deficits among the long-term cannabis users,” she added.
The study was recently published online in the American Journal of Psychiatry.
Growing use in Boomers
Long-term cannabis use has been associated with memory problems. Studies examining the impact of cannabis use on the brain have shown conflicting results. Some suggest regular use in adolescence is associated with altered connectivity and reduced volume of brain regions involved in executive functions such as memory, learning, and impulse control compared with those who do not use cannabis.
Others found no significant structural differences between the brains of cannabis users and nonusers.
An earlier, large longitudinal study in New Zealand found that persistent cannabis use (with frequent use starting in adolescence) was associated with a loss of an average of six (or up to eight) IQ points measured in mid-adulthood.
Cannabis use is increasing among Baby Boomers – a group born between 1946 and 1964 – who used cannabis at historically high rates as young adults, and who now use it at historically high rates in midlife and as older adults.
To date, case-control studies, which are predominantly in adolescents and young adults, have found that cannabis users show subtle cognitive deficits and structural brain differences, but it is unclear whether these differences in young cannabis users might be larger in midlife and in older adults who have longer histories of use.
The study included a representative cohort of 1,037 individuals in Dunedin, New Zealand, born between April 1972 and March 1973, and followed from age 3 to 45.
Cannabis use and dependence were assessed at ages 18, 21, 26, 32, 38, and 45. IQ was assessed at ages 7, 9, 11, and 45. Specific neuropsychological functions and hippocampal volume were assessed at age 45.
“Most of the previous research has focused on adolescent and young-adult cannabis users. What we’re looking at here is long-term cannabis users in midlife, and we’re finding that long-term users show cognitive deficits. But we’re not just looking at a snapshot of people in midlife, we’re also doing a longitudinal comparison – comparing them to themselves in childhood. We saw that long-term cannabis users showed a decline in IQ from childhood to adulthood,” said Dr. Meier.
Participants in the study are members of the Dunedin Longitudinal Study, a representative birth cohort (n = 1,037; 91% of eligible births; 52% male) born between April 1972 and March 1973 in Dunedin, New Zealand, who participated in the first assessment at age 3.
This cohort matched socioeconomic status (SES), key health indicators, and demographics. Assessments were carried out at birth and ages 3, 5, 7, 9, 11, 13, 15, 18, 21, 26, 32, 38, and 45. IQ was assessed at ages 7, 9, 11, and 45. Specific neuropsychological functions and hippocampal volume were assessed at age 45.
Shrinking hippocampal volume
Cannabis use, cognitive function, and hippocampal volume were assessed comparing long-term cannabis users (n = 84) against five distinct groups:
- Lifelong cannabis nonusers (n = 196) – to replicate the control group most often reported in the case-control literature
- Midlife recreational cannabis users (n = 65) – to determine if cognitive deficits and structural brain differences are apparent in nonproblem users – the majority of cannabis users
- Long-term tobacco users (n = 75)
- Long-term alcohol users (n = 57) – benchmark comparisons for any cannabis findings and to disentangle potential cannabis effects from tobacco and alcohol effects
- Cannabis quitters (n = 58) – to determine whether differences are apparent after cessation
Tests were conducted on dose-response associations using continuously measured persistence of cannabis use, rigorously adjusting for numerous confounders derived from multiple longitudinal waves and data sources.
The investigators also tested whether associations between continuously measured persistence of cannabis use and cognitive deficits were mediated by hippocampal volume differences.
The hippocampus was the area of focus because it has a high density of cannabinoid receptors and is also instrumental for learning and memory, which is one of the most consistently impaired cognitive domains in cannabis users, and has been the brain region that most consistently emerges as smaller in cannabis users relative to controls. Structural MRI was done at age 45 for 875 participants (93% of age 45 participants).
Of 997 cohort members still alive at age 45, 938 (94.1%) were assessed at age 45. Age 45 participants did not differ significantly from other participants on childhood SES, childhood self-control, or childhood IQ. Cognitive functioning among midlife recreational cannabis users was similar to representative cohort norms, suggesting that infrequent recreational cannabis use in midlife is unlikely to compromise cognitive functioning.
However, long-term cannabis users did not perform significantly worse on any test than cannabis quitters. Cannabis quitters showed subtle cognitive deficits that may explain inconsistent findings on the benefits of cessation.
Smaller hippocampal volume is thought to be a possible mediator of cannabis-related cognitive deficits because the hippocampus is rich in CB1 receptors and is involved in learning and memory.
Long-term cannabis users had smaller bilateral volume in total hippocampus and 5 of 12 structurally and functionally distinct subregions (tail, hippocampal amygdala transition area, CA1, molecular layer, and dentate gyrus), and significantly smaller volumes than midlife recreational cannabis users in the left and right hippocampus, and 3 of 12 subfields (tail, CA1, and molecular layer), compared with non-users, consistent with case-control studies.
More potent
“If you’ve been using cannabis very long term and now are in midlife, you might want to consider quitting. Quitting is associated with slightly better cognitive performance in midlife. We also need to watch for risk of dementia. We know that people who show cognitive deficits at midlife are at elevated risk for later life dementia. And the deficits we saw among long-term cannabis users (although fairly mild), they were in the range in terms of effect size of what we see among people in other studies who have gone on to develop dementia in later life,” said Dr. Meier.
The study findings conflict with those of other studies, including one by the same research group, which compared the cognitive functioning of twins who were discordant for cannabis use and found little evidence of cannabis-related cognitive deficits. Because long-term cannabis users also use tobacco, alcohol, and other illicit drugs, disentangling cannabis effects from other substances is challenging.
“Long-term cannabis users tend to be long-term polysubstance users, so it’s hard to isolate,” said Dr. Meier.
Additionally, some group sizes were small, raising concerns about low statistical power.
“Group sizes were small but we didn’t rely only on those group comparisons; however, we did find statistical differences. We also tested highly statistically powered dose-response associations between persistence of cannabis use over ages 18-45 and each of our outcomes (IQ, learning, and processing speed in midlife) while adjusting possible alternate explanations such as low childhood IQ, other substance use, [and] socioeconomic backgrounds.
“These dose-response associations used large sample sizes, were highly powered, and took into account a number of alternative explanations. These two different approaches showed very similar findings and one bolstered the other,” said Dr. Meier.
The study’s results were based on individuals who began using cannabis in the 1980s or ‘90s, but the concentration of tetrahydrocannabinol (THC) has risen in recent years.
“When the study began, THC concentration was approximately 4%. Over the last decade we have seen it go up to 12% or even higher. A recent study surveying U.S. dispensaries found 20% THC. If THC accounts for impairment, then the effects can be larger [with higher concentrations]. One of the challenges in the U.S. is that there are laws prohibiting researchers from testing cannabis, so we have to rely on product labels, which we know are unreliable,” said Dr. Meier.
A separate report is forthcoming with results of exploratory analyses of associations between long-term cannabis use and comprehensive MRI measures of global and regional gray and white matter.
The data will also be used to answer a number of different questions about cognitive deficits, brain structure, aging preparedness, social preparedness (strength of social networks), financial and health preparedness, and biological aging (the pace of aging relative to chronological age) in long-term cannabis users, Dr. Meier noted.
‘Fantastic’ research
Commenting on the research for this news organization , Andrew J. Saxon, MD, professor, department of psychiatry & behavioral sciences at University of Washington, Seattle, and a member of the American Psychiatric Association’s Council on Addiction Psychiatry, said the study “provides more evidence that heavy and regular cannabis use is not benign behavior.”
“It’s a fantastic piece of research in which they enrolled participants at birth and have followed them up to age 45. In most of the other research that has been done, we have no idea what their baseline was. What’s so remarkable here is that they can clearly demonstrate the loss of IQ points from childhood to age 45,” said Dr. Saxon.
“It is clear that, in people using cannabis long term, cognition is impaired. It would be good to have a better handle on how much cognitive function can be regained if you quit, because that could be a motivator for quitting in people where cannabis is having an adverse effect on their lives,” he added.
On the issue of THC potency, Dr. Saxon said that, while it’s true the potency of cannabis is increasing in terms of THC concentrations, the question is: “Do people who use cannabis use a set amount or do they imbibe until they achieve the state of altered consciousness that they’re seeking? Although there has been some research in the area of self-regulation and cannabis potency, we do not yet have the answers to determine if there is any causation,” said Dr. Saxon.
Dr. Meier and Dr. Saxon reported no relevant financial conflicts of interest.
A version of this article first appeared on Medscape.com.
A large prospective, longitudinal study showed long-term cannabis users had an intelligence quotient (IQ) decline from age 18 to midlife (mean, 5.5 IQ points), poorer learning and processing speed, compared with childhood, and self-reported memory and attention problems. Long-term cannabis users also showed hippocampal atrophy at midlife (age 45), which combined with mild midlife cognitive deficits, all known risk factors for dementia.
“Long-term cannabis users – people who have used cannabis from 18 or 19 years old and continued using through midlife – showed cognitive deficits, compared with nonusers. They also showed more severe cognitive deficits, compared with long-term alcohol users and long-term tobacco users. But people who used infrequently or recreationally in midlife did not show as severe cognitive deficits. Cognitive deficits were confined to cannabis users,” lead investigator Madeline Meier, PhD, associate professor of psychology, Arizona State University, Tempe, said in an interview.
“Long-term cannabis users had smaller hippocampal volume, but we also found that smaller hippocampal volume did not explain the cognitive deficits among the long-term cannabis users,” she added.
The study was recently published online in the American Journal of Psychiatry.
Growing use in Boomers
Long-term cannabis use has been associated with memory problems. Studies examining the impact of cannabis use on the brain have shown conflicting results. Some suggest regular use in adolescence is associated with altered connectivity and reduced volume of brain regions involved in executive functions such as memory, learning, and impulse control compared with those who do not use cannabis.
Others found no significant structural differences between the brains of cannabis users and nonusers.
An earlier, large longitudinal study in New Zealand found that persistent cannabis use (with frequent use starting in adolescence) was associated with a loss of an average of six (or up to eight) IQ points measured in mid-adulthood.
Cannabis use is increasing among Baby Boomers – a group born between 1946 and 1964 – who used cannabis at historically high rates as young adults, and who now use it at historically high rates in midlife and as older adults.
To date, case-control studies, which are predominantly in adolescents and young adults, have found that cannabis users show subtle cognitive deficits and structural brain differences, but it is unclear whether these differences in young cannabis users might be larger in midlife and in older adults who have longer histories of use.
The study included a representative cohort of 1,037 individuals in Dunedin, New Zealand, born between April 1972 and March 1973, and followed from age 3 to 45.
Cannabis use and dependence were assessed at ages 18, 21, 26, 32, 38, and 45. IQ was assessed at ages 7, 9, 11, and 45. Specific neuropsychological functions and hippocampal volume were assessed at age 45.
“Most of the previous research has focused on adolescent and young-adult cannabis users. What we’re looking at here is long-term cannabis users in midlife, and we’re finding that long-term users show cognitive deficits. But we’re not just looking at a snapshot of people in midlife, we’re also doing a longitudinal comparison – comparing them to themselves in childhood. We saw that long-term cannabis users showed a decline in IQ from childhood to adulthood,” said Dr. Meier.
Participants in the study are members of the Dunedin Longitudinal Study, a representative birth cohort (n = 1,037; 91% of eligible births; 52% male) born between April 1972 and March 1973 in Dunedin, New Zealand, who participated in the first assessment at age 3.
This cohort matched socioeconomic status (SES), key health indicators, and demographics. Assessments were carried out at birth and ages 3, 5, 7, 9, 11, 13, 15, 18, 21, 26, 32, 38, and 45. IQ was assessed at ages 7, 9, 11, and 45. Specific neuropsychological functions and hippocampal volume were assessed at age 45.
Shrinking hippocampal volume
Cannabis use, cognitive function, and hippocampal volume were assessed comparing long-term cannabis users (n = 84) against five distinct groups:
- Lifelong cannabis nonusers (n = 196) – to replicate the control group most often reported in the case-control literature
- Midlife recreational cannabis users (n = 65) – to determine if cognitive deficits and structural brain differences are apparent in nonproblem users – the majority of cannabis users
- Long-term tobacco users (n = 75)
- Long-term alcohol users (n = 57) – benchmark comparisons for any cannabis findings and to disentangle potential cannabis effects from tobacco and alcohol effects
- Cannabis quitters (n = 58) – to determine whether differences are apparent after cessation
Tests were conducted on dose-response associations using continuously measured persistence of cannabis use, rigorously adjusting for numerous confounders derived from multiple longitudinal waves and data sources.
The investigators also tested whether associations between continuously measured persistence of cannabis use and cognitive deficits were mediated by hippocampal volume differences.
The hippocampus was the area of focus because it has a high density of cannabinoid receptors and is also instrumental for learning and memory, which is one of the most consistently impaired cognitive domains in cannabis users, and has been the brain region that most consistently emerges as smaller in cannabis users relative to controls. Structural MRI was done at age 45 for 875 participants (93% of age 45 participants).
Of 997 cohort members still alive at age 45, 938 (94.1%) were assessed at age 45. Age 45 participants did not differ significantly from other participants on childhood SES, childhood self-control, or childhood IQ. Cognitive functioning among midlife recreational cannabis users was similar to representative cohort norms, suggesting that infrequent recreational cannabis use in midlife is unlikely to compromise cognitive functioning.
However, long-term cannabis users did not perform significantly worse on any test than cannabis quitters. Cannabis quitters showed subtle cognitive deficits that may explain inconsistent findings on the benefits of cessation.
Smaller hippocampal volume is thought to be a possible mediator of cannabis-related cognitive deficits because the hippocampus is rich in CB1 receptors and is involved in learning and memory.
Long-term cannabis users had smaller bilateral volume in total hippocampus and 5 of 12 structurally and functionally distinct subregions (tail, hippocampal amygdala transition area, CA1, molecular layer, and dentate gyrus), and significantly smaller volumes than midlife recreational cannabis users in the left and right hippocampus, and 3 of 12 subfields (tail, CA1, and molecular layer), compared with non-users, consistent with case-control studies.
More potent
“If you’ve been using cannabis very long term and now are in midlife, you might want to consider quitting. Quitting is associated with slightly better cognitive performance in midlife. We also need to watch for risk of dementia. We know that people who show cognitive deficits at midlife are at elevated risk for later life dementia. And the deficits we saw among long-term cannabis users (although fairly mild), they were in the range in terms of effect size of what we see among people in other studies who have gone on to develop dementia in later life,” said Dr. Meier.
The study findings conflict with those of other studies, including one by the same research group, which compared the cognitive functioning of twins who were discordant for cannabis use and found little evidence of cannabis-related cognitive deficits. Because long-term cannabis users also use tobacco, alcohol, and other illicit drugs, disentangling cannabis effects from other substances is challenging.
“Long-term cannabis users tend to be long-term polysubstance users, so it’s hard to isolate,” said Dr. Meier.
Additionally, some group sizes were small, raising concerns about low statistical power.
“Group sizes were small but we didn’t rely only on those group comparisons; however, we did find statistical differences. We also tested highly statistically powered dose-response associations between persistence of cannabis use over ages 18-45 and each of our outcomes (IQ, learning, and processing speed in midlife) while adjusting possible alternate explanations such as low childhood IQ, other substance use, [and] socioeconomic backgrounds.
“These dose-response associations used large sample sizes, were highly powered, and took into account a number of alternative explanations. These two different approaches showed very similar findings and one bolstered the other,” said Dr. Meier.
The study’s results were based on individuals who began using cannabis in the 1980s or ‘90s, but the concentration of tetrahydrocannabinol (THC) has risen in recent years.
“When the study began, THC concentration was approximately 4%. Over the last decade we have seen it go up to 12% or even higher. A recent study surveying U.S. dispensaries found 20% THC. If THC accounts for impairment, then the effects can be larger [with higher concentrations]. One of the challenges in the U.S. is that there are laws prohibiting researchers from testing cannabis, so we have to rely on product labels, which we know are unreliable,” said Dr. Meier.
A separate report is forthcoming with results of exploratory analyses of associations between long-term cannabis use and comprehensive MRI measures of global and regional gray and white matter.
The data will also be used to answer a number of different questions about cognitive deficits, brain structure, aging preparedness, social preparedness (strength of social networks), financial and health preparedness, and biological aging (the pace of aging relative to chronological age) in long-term cannabis users, Dr. Meier noted.
‘Fantastic’ research
Commenting on the research for this news organization , Andrew J. Saxon, MD, professor, department of psychiatry & behavioral sciences at University of Washington, Seattle, and a member of the American Psychiatric Association’s Council on Addiction Psychiatry, said the study “provides more evidence that heavy and regular cannabis use is not benign behavior.”
“It’s a fantastic piece of research in which they enrolled participants at birth and have followed them up to age 45. In most of the other research that has been done, we have no idea what their baseline was. What’s so remarkable here is that they can clearly demonstrate the loss of IQ points from childhood to age 45,” said Dr. Saxon.
“It is clear that, in people using cannabis long term, cognition is impaired. It would be good to have a better handle on how much cognitive function can be regained if you quit, because that could be a motivator for quitting in people where cannabis is having an adverse effect on their lives,” he added.
On the issue of THC potency, Dr. Saxon said that, while it’s true the potency of cannabis is increasing in terms of THC concentrations, the question is: “Do people who use cannabis use a set amount or do they imbibe until they achieve the state of altered consciousness that they’re seeking? Although there has been some research in the area of self-regulation and cannabis potency, we do not yet have the answers to determine if there is any causation,” said Dr. Saxon.
Dr. Meier and Dr. Saxon reported no relevant financial conflicts of interest.
A version of this article first appeared on Medscape.com.
A large prospective, longitudinal study showed long-term cannabis users had an intelligence quotient (IQ) decline from age 18 to midlife (mean, 5.5 IQ points), poorer learning and processing speed, compared with childhood, and self-reported memory and attention problems. Long-term cannabis users also showed hippocampal atrophy at midlife (age 45), which combined with mild midlife cognitive deficits, all known risk factors for dementia.
“Long-term cannabis users – people who have used cannabis from 18 or 19 years old and continued using through midlife – showed cognitive deficits, compared with nonusers. They also showed more severe cognitive deficits, compared with long-term alcohol users and long-term tobacco users. But people who used infrequently or recreationally in midlife did not show as severe cognitive deficits. Cognitive deficits were confined to cannabis users,” lead investigator Madeline Meier, PhD, associate professor of psychology, Arizona State University, Tempe, said in an interview.
“Long-term cannabis users had smaller hippocampal volume, but we also found that smaller hippocampal volume did not explain the cognitive deficits among the long-term cannabis users,” she added.
The study was recently published online in the American Journal of Psychiatry.
Growing use in Boomers
Long-term cannabis use has been associated with memory problems. Studies examining the impact of cannabis use on the brain have shown conflicting results. Some suggest regular use in adolescence is associated with altered connectivity and reduced volume of brain regions involved in executive functions such as memory, learning, and impulse control compared with those who do not use cannabis.
Others found no significant structural differences between the brains of cannabis users and nonusers.
An earlier, large longitudinal study in New Zealand found that persistent cannabis use (with frequent use starting in adolescence) was associated with a loss of an average of six (or up to eight) IQ points measured in mid-adulthood.
Cannabis use is increasing among Baby Boomers – a group born between 1946 and 1964 – who used cannabis at historically high rates as young adults, and who now use it at historically high rates in midlife and as older adults.
To date, case-control studies, which are predominantly in adolescents and young adults, have found that cannabis users show subtle cognitive deficits and structural brain differences, but it is unclear whether these differences in young cannabis users might be larger in midlife and in older adults who have longer histories of use.
The study included a representative cohort of 1,037 individuals in Dunedin, New Zealand, born between April 1972 and March 1973, and followed from age 3 to 45.
Cannabis use and dependence were assessed at ages 18, 21, 26, 32, 38, and 45. IQ was assessed at ages 7, 9, 11, and 45. Specific neuropsychological functions and hippocampal volume were assessed at age 45.
“Most of the previous research has focused on adolescent and young-adult cannabis users. What we’re looking at here is long-term cannabis users in midlife, and we’re finding that long-term users show cognitive deficits. But we’re not just looking at a snapshot of people in midlife, we’re also doing a longitudinal comparison – comparing them to themselves in childhood. We saw that long-term cannabis users showed a decline in IQ from childhood to adulthood,” said Dr. Meier.
Participants in the study are members of the Dunedin Longitudinal Study, a representative birth cohort (n = 1,037; 91% of eligible births; 52% male) born between April 1972 and March 1973 in Dunedin, New Zealand, who participated in the first assessment at age 3.
This cohort matched socioeconomic status (SES), key health indicators, and demographics. Assessments were carried out at birth and ages 3, 5, 7, 9, 11, 13, 15, 18, 21, 26, 32, 38, and 45. IQ was assessed at ages 7, 9, 11, and 45. Specific neuropsychological functions and hippocampal volume were assessed at age 45.
Shrinking hippocampal volume
Cannabis use, cognitive function, and hippocampal volume were assessed comparing long-term cannabis users (n = 84) against five distinct groups:
- Lifelong cannabis nonusers (n = 196) – to replicate the control group most often reported in the case-control literature
- Midlife recreational cannabis users (n = 65) – to determine if cognitive deficits and structural brain differences are apparent in nonproblem users – the majority of cannabis users
- Long-term tobacco users (n = 75)
- Long-term alcohol users (n = 57) – benchmark comparisons for any cannabis findings and to disentangle potential cannabis effects from tobacco and alcohol effects
- Cannabis quitters (n = 58) – to determine whether differences are apparent after cessation
Tests were conducted on dose-response associations using continuously measured persistence of cannabis use, rigorously adjusting for numerous confounders derived from multiple longitudinal waves and data sources.
The investigators also tested whether associations between continuously measured persistence of cannabis use and cognitive deficits were mediated by hippocampal volume differences.
The hippocampus was the area of focus because it has a high density of cannabinoid receptors and is also instrumental for learning and memory, which is one of the most consistently impaired cognitive domains in cannabis users, and has been the brain region that most consistently emerges as smaller in cannabis users relative to controls. Structural MRI was done at age 45 for 875 participants (93% of age 45 participants).
Of 997 cohort members still alive at age 45, 938 (94.1%) were assessed at age 45. Age 45 participants did not differ significantly from other participants on childhood SES, childhood self-control, or childhood IQ. Cognitive functioning among midlife recreational cannabis users was similar to representative cohort norms, suggesting that infrequent recreational cannabis use in midlife is unlikely to compromise cognitive functioning.
However, long-term cannabis users did not perform significantly worse on any test than cannabis quitters. Cannabis quitters showed subtle cognitive deficits that may explain inconsistent findings on the benefits of cessation.
Smaller hippocampal volume is thought to be a possible mediator of cannabis-related cognitive deficits because the hippocampus is rich in CB1 receptors and is involved in learning and memory.
Long-term cannabis users had smaller bilateral volume in total hippocampus and 5 of 12 structurally and functionally distinct subregions (tail, hippocampal amygdala transition area, CA1, molecular layer, and dentate gyrus), and significantly smaller volumes than midlife recreational cannabis users in the left and right hippocampus, and 3 of 12 subfields (tail, CA1, and molecular layer), compared with non-users, consistent with case-control studies.
More potent
“If you’ve been using cannabis very long term and now are in midlife, you might want to consider quitting. Quitting is associated with slightly better cognitive performance in midlife. We also need to watch for risk of dementia. We know that people who show cognitive deficits at midlife are at elevated risk for later life dementia. And the deficits we saw among long-term cannabis users (although fairly mild), they were in the range in terms of effect size of what we see among people in other studies who have gone on to develop dementia in later life,” said Dr. Meier.
The study findings conflict with those of other studies, including one by the same research group, which compared the cognitive functioning of twins who were discordant for cannabis use and found little evidence of cannabis-related cognitive deficits. Because long-term cannabis users also use tobacco, alcohol, and other illicit drugs, disentangling cannabis effects from other substances is challenging.
“Long-term cannabis users tend to be long-term polysubstance users, so it’s hard to isolate,” said Dr. Meier.
Additionally, some group sizes were small, raising concerns about low statistical power.
“Group sizes were small but we didn’t rely only on those group comparisons; however, we did find statistical differences. We also tested highly statistically powered dose-response associations between persistence of cannabis use over ages 18-45 and each of our outcomes (IQ, learning, and processing speed in midlife) while adjusting possible alternate explanations such as low childhood IQ, other substance use, [and] socioeconomic backgrounds.
“These dose-response associations used large sample sizes, were highly powered, and took into account a number of alternative explanations. These two different approaches showed very similar findings and one bolstered the other,” said Dr. Meier.
The study’s results were based on individuals who began using cannabis in the 1980s or ‘90s, but the concentration of tetrahydrocannabinol (THC) has risen in recent years.
“When the study began, THC concentration was approximately 4%. Over the last decade we have seen it go up to 12% or even higher. A recent study surveying U.S. dispensaries found 20% THC. If THC accounts for impairment, then the effects can be larger [with higher concentrations]. One of the challenges in the U.S. is that there are laws prohibiting researchers from testing cannabis, so we have to rely on product labels, which we know are unreliable,” said Dr. Meier.
A separate report is forthcoming with results of exploratory analyses of associations between long-term cannabis use and comprehensive MRI measures of global and regional gray and white matter.
The data will also be used to answer a number of different questions about cognitive deficits, brain structure, aging preparedness, social preparedness (strength of social networks), financial and health preparedness, and biological aging (the pace of aging relative to chronological age) in long-term cannabis users, Dr. Meier noted.
‘Fantastic’ research
Commenting on the research for this news organization , Andrew J. Saxon, MD, professor, department of psychiatry & behavioral sciences at University of Washington, Seattle, and a member of the American Psychiatric Association’s Council on Addiction Psychiatry, said the study “provides more evidence that heavy and regular cannabis use is not benign behavior.”
“It’s a fantastic piece of research in which they enrolled participants at birth and have followed them up to age 45. In most of the other research that has been done, we have no idea what their baseline was. What’s so remarkable here is that they can clearly demonstrate the loss of IQ points from childhood to age 45,” said Dr. Saxon.
“It is clear that, in people using cannabis long term, cognition is impaired. It would be good to have a better handle on how much cognitive function can be regained if you quit, because that could be a motivator for quitting in people where cannabis is having an adverse effect on their lives,” he added.
On the issue of THC potency, Dr. Saxon said that, while it’s true the potency of cannabis is increasing in terms of THC concentrations, the question is: “Do people who use cannabis use a set amount or do they imbibe until they achieve the state of altered consciousness that they’re seeking? Although there has been some research in the area of self-regulation and cannabis potency, we do not yet have the answers to determine if there is any causation,” said Dr. Saxon.
Dr. Meier and Dr. Saxon reported no relevant financial conflicts of interest.
A version of this article first appeared on Medscape.com.
FROM THE AMERICAN JOURNAL OF PSYCHIATRY
New injectable gel can deliver immune cells directly to cancer tumors
A simple, two-ingredient gel may boost the fighting power of a groundbreaking cancer treatment, say Stanford University engineers.
The gel – made from water and a plant-based polymer – delivers targeted T cells adjacent to a cancer growth, taking aim at solid tumors.
It’s the latest development in CAR T-cell therapy, a type of immunotherapy that involves collecting the patient’s T cells, reengineering them to be stronger, and returning them to the patient’s body.
Results have been promising in blood cancers, such as leukemia and lymphoma, but less so in solid tumors, such as brain, breast, or kidney cancer, according to the National Cancer Institute.
The gel “is a really exciting step forward,” says Abigail Grosskopf, a PhD candidate at Stanford (Calif.) University, who is the lead study author, “because it can change the delivery of these cells and expand this kind of treatment to other cancers.”
CAR T-cell therapy: Limits in solid tumors
Currently available CAR T-cell therapies are administered by intravenous infusion. But that doesn’t do much against tumors in specific locations because the cells enter the bloodstream and flow throughout the body. The cancer-fighting effort exhausts the T cells, weakening their ability to infiltrate dense tumors.
CAR T cells need cytokines to tell them when to attack, Ms. Grosskopf explains. If delivered through an IV drip, the number of cytokines required to destroy a solid tumor would be toxic to other, healthy parts of the body.
So
In their study, which was published in Science Advances, the injections wiped out mouse tumors in 12 days. The gel degraded harmlessly a few weeks later.
A “leaky pen” that fights cancer
The reason a gel works better than a liquid is because of its staying power, says Ms. Grosskopf, who compares the method to a leaky pen.
The gel acts as the “pen,” releasing activated CAR T cells at regular intervals to attack the cancerous growth. Whereas liquid dissipates quickly, the gel’s structure is strong enough to stay in place for weeks, Ms. Grosskopf says. Plus, it’s biocompatible and harmless within the body, she adds.
More preclinical studies are needed before human clinical trials can occur, Ms. Grosskopf says.
“Not only could this be a way to deliver T cells and cytokines,” Ms. Grosskopf says, “but it may be used for other targeted therapy cancer drugs that are in development. So we see this as running parallel to those efforts.”
Taking an even broader view, the gel could have applications across medical specialties, such as slow-release delivery of vaccines.
A version of this article first appeared on Medscape.com.
A simple, two-ingredient gel may boost the fighting power of a groundbreaking cancer treatment, say Stanford University engineers.
The gel – made from water and a plant-based polymer – delivers targeted T cells adjacent to a cancer growth, taking aim at solid tumors.
It’s the latest development in CAR T-cell therapy, a type of immunotherapy that involves collecting the patient’s T cells, reengineering them to be stronger, and returning them to the patient’s body.
Results have been promising in blood cancers, such as leukemia and lymphoma, but less so in solid tumors, such as brain, breast, or kidney cancer, according to the National Cancer Institute.
The gel “is a really exciting step forward,” says Abigail Grosskopf, a PhD candidate at Stanford (Calif.) University, who is the lead study author, “because it can change the delivery of these cells and expand this kind of treatment to other cancers.”
CAR T-cell therapy: Limits in solid tumors
Currently available CAR T-cell therapies are administered by intravenous infusion. But that doesn’t do much against tumors in specific locations because the cells enter the bloodstream and flow throughout the body. The cancer-fighting effort exhausts the T cells, weakening their ability to infiltrate dense tumors.
CAR T cells need cytokines to tell them when to attack, Ms. Grosskopf explains. If delivered through an IV drip, the number of cytokines required to destroy a solid tumor would be toxic to other, healthy parts of the body.
So
In their study, which was published in Science Advances, the injections wiped out mouse tumors in 12 days. The gel degraded harmlessly a few weeks later.
A “leaky pen” that fights cancer
The reason a gel works better than a liquid is because of its staying power, says Ms. Grosskopf, who compares the method to a leaky pen.
The gel acts as the “pen,” releasing activated CAR T cells at regular intervals to attack the cancerous growth. Whereas liquid dissipates quickly, the gel’s structure is strong enough to stay in place for weeks, Ms. Grosskopf says. Plus, it’s biocompatible and harmless within the body, she adds.
More preclinical studies are needed before human clinical trials can occur, Ms. Grosskopf says.
“Not only could this be a way to deliver T cells and cytokines,” Ms. Grosskopf says, “but it may be used for other targeted therapy cancer drugs that are in development. So we see this as running parallel to those efforts.”
Taking an even broader view, the gel could have applications across medical specialties, such as slow-release delivery of vaccines.
A version of this article first appeared on Medscape.com.
A simple, two-ingredient gel may boost the fighting power of a groundbreaking cancer treatment, say Stanford University engineers.
The gel – made from water and a plant-based polymer – delivers targeted T cells adjacent to a cancer growth, taking aim at solid tumors.
It’s the latest development in CAR T-cell therapy, a type of immunotherapy that involves collecting the patient’s T cells, reengineering them to be stronger, and returning them to the patient’s body.
Results have been promising in blood cancers, such as leukemia and lymphoma, but less so in solid tumors, such as brain, breast, or kidney cancer, according to the National Cancer Institute.
The gel “is a really exciting step forward,” says Abigail Grosskopf, a PhD candidate at Stanford (Calif.) University, who is the lead study author, “because it can change the delivery of these cells and expand this kind of treatment to other cancers.”
CAR T-cell therapy: Limits in solid tumors
Currently available CAR T-cell therapies are administered by intravenous infusion. But that doesn’t do much against tumors in specific locations because the cells enter the bloodstream and flow throughout the body. The cancer-fighting effort exhausts the T cells, weakening their ability to infiltrate dense tumors.
CAR T cells need cytokines to tell them when to attack, Ms. Grosskopf explains. If delivered through an IV drip, the number of cytokines required to destroy a solid tumor would be toxic to other, healthy parts of the body.
So
In their study, which was published in Science Advances, the injections wiped out mouse tumors in 12 days. The gel degraded harmlessly a few weeks later.
A “leaky pen” that fights cancer
The reason a gel works better than a liquid is because of its staying power, says Ms. Grosskopf, who compares the method to a leaky pen.
The gel acts as the “pen,” releasing activated CAR T cells at regular intervals to attack the cancerous growth. Whereas liquid dissipates quickly, the gel’s structure is strong enough to stay in place for weeks, Ms. Grosskopf says. Plus, it’s biocompatible and harmless within the body, she adds.
More preclinical studies are needed before human clinical trials can occur, Ms. Grosskopf says.
“Not only could this be a way to deliver T cells and cytokines,” Ms. Grosskopf says, “but it may be used for other targeted therapy cancer drugs that are in development. So we see this as running parallel to those efforts.”
Taking an even broader view, the gel could have applications across medical specialties, such as slow-release delivery of vaccines.
A version of this article first appeared on Medscape.com.
FROM SCIENCE ADVANCES
U.S. life expectancy dropped by 2 years in 2020: Study
according to a new study.
The study, published in medRxiv, said U.S. life expectancy went from 78.86 years in 2019 to 76.99 years in 2020, during the thick of the global COVID-19 pandemic. Though vaccines were widely available in 2021, the U.S. life expectancy was expected to keep going down, to 76.60 years.
In “peer countries” – Austria, Belgium, Denmark, England and Wales, Finland, France, Germany, Israel, Italy, the Netherlands, New Zealand, Northern Ireland, Norway, Portugal, Scotland, South Korea, Spain, Sweden, and Switzerland – life expectancy went down only 0.57 years from 2019 to 2020 and increased by 0.28 years in 2021, the study said. The peer countries now have a life expectancy that’s 5 years longer than in the United States.
“The fact the U.S. lost so many more lives than other high-income countries speaks not only to how we managed the pandemic, but also to more deeply rooted problems that predated the pandemic,” said Steven H. Woolf, MD, one of the study authors and a professor of family medicine and population health at Virginia Commonwealth University, Richmond, according to Reuters.
“U.S. life expectancy has been falling behind other countries since the 1980s, and the gap has widened over time, especially in the last decade.”
Lack of universal health care, income and educational inequality, and less-healthy physical and social environments helped lead to the decline in American life expectancy, according to Dr. Woolf.
The life expectancy drop from 2019 to 2020 hit Black and Hispanic people hardest, according to the study. But the drop from 2020 to 2021 affected White people the most, with average life expectancy among them going down about a third of a year.
Researchers looked at death data from the National Center for Health Statistics, the Human Mortality Database, and overseas statistical agencies. Life expectancy for 2021 was estimated “using a previously validated modeling method,” the study said.
A version of this article first appeared on WebMD.com.
according to a new study.
The study, published in medRxiv, said U.S. life expectancy went from 78.86 years in 2019 to 76.99 years in 2020, during the thick of the global COVID-19 pandemic. Though vaccines were widely available in 2021, the U.S. life expectancy was expected to keep going down, to 76.60 years.
In “peer countries” – Austria, Belgium, Denmark, England and Wales, Finland, France, Germany, Israel, Italy, the Netherlands, New Zealand, Northern Ireland, Norway, Portugal, Scotland, South Korea, Spain, Sweden, and Switzerland – life expectancy went down only 0.57 years from 2019 to 2020 and increased by 0.28 years in 2021, the study said. The peer countries now have a life expectancy that’s 5 years longer than in the United States.
“The fact the U.S. lost so many more lives than other high-income countries speaks not only to how we managed the pandemic, but also to more deeply rooted problems that predated the pandemic,” said Steven H. Woolf, MD, one of the study authors and a professor of family medicine and population health at Virginia Commonwealth University, Richmond, according to Reuters.
“U.S. life expectancy has been falling behind other countries since the 1980s, and the gap has widened over time, especially in the last decade.”
Lack of universal health care, income and educational inequality, and less-healthy physical and social environments helped lead to the decline in American life expectancy, according to Dr. Woolf.
The life expectancy drop from 2019 to 2020 hit Black and Hispanic people hardest, according to the study. But the drop from 2020 to 2021 affected White people the most, with average life expectancy among them going down about a third of a year.
Researchers looked at death data from the National Center for Health Statistics, the Human Mortality Database, and overseas statistical agencies. Life expectancy for 2021 was estimated “using a previously validated modeling method,” the study said.
A version of this article first appeared on WebMD.com.
according to a new study.
The study, published in medRxiv, said U.S. life expectancy went from 78.86 years in 2019 to 76.99 years in 2020, during the thick of the global COVID-19 pandemic. Though vaccines were widely available in 2021, the U.S. life expectancy was expected to keep going down, to 76.60 years.
In “peer countries” – Austria, Belgium, Denmark, England and Wales, Finland, France, Germany, Israel, Italy, the Netherlands, New Zealand, Northern Ireland, Norway, Portugal, Scotland, South Korea, Spain, Sweden, and Switzerland – life expectancy went down only 0.57 years from 2019 to 2020 and increased by 0.28 years in 2021, the study said. The peer countries now have a life expectancy that’s 5 years longer than in the United States.
“The fact the U.S. lost so many more lives than other high-income countries speaks not only to how we managed the pandemic, but also to more deeply rooted problems that predated the pandemic,” said Steven H. Woolf, MD, one of the study authors and a professor of family medicine and population health at Virginia Commonwealth University, Richmond, according to Reuters.
“U.S. life expectancy has been falling behind other countries since the 1980s, and the gap has widened over time, especially in the last decade.”
Lack of universal health care, income and educational inequality, and less-healthy physical and social environments helped lead to the decline in American life expectancy, according to Dr. Woolf.
The life expectancy drop from 2019 to 2020 hit Black and Hispanic people hardest, according to the study. But the drop from 2020 to 2021 affected White people the most, with average life expectancy among them going down about a third of a year.
Researchers looked at death data from the National Center for Health Statistics, the Human Mortality Database, and overseas statistical agencies. Life expectancy for 2021 was estimated “using a previously validated modeling method,” the study said.
A version of this article first appeared on WebMD.com.
FROM MEDRXIV
Infectious disease pop quiz: Clinical challenge #23 for the ObGyn
What are the most common organisms that cause chorioamnionitis and puerperal endometritis?
Continue to the answer...
Chorioamnionitis and puerperal endometritis are polymicrobial, mixed aerobic-anaerobic infections. The dominant organisms are anaerobic gram-negative bacilli (Bacteroides and Prevotella species); anaerobic gram-positive cocci (Peptococcus species and Peptostreptococcus species); aerobic gram-negative bacilli (principally, Escherichia coli, Klebsiella pneumoniae, and Proteus species); and aerobic gram-positive cocci (enterococci, staphylococci, and group B streptococci).
- Duff P. Maternal and perinatal infections: bacterial. In: Landon MB, Galan HL, Jauniaux ERM, et al. Gabbe’s Obstetrics: Normal and Problem Pregnancies. 8th ed. Elsevier; 2021:1124-1146.
- Duff P. Maternal and fetal infections. In: Resnik R, Lockwood CJ, Moore TJ, et al. Creasy & Resnik’s Maternal-Fetal Medicine: Principles and Practice. 8th ed. Elsevier; 2019:862-919.
What are the most common organisms that cause chorioamnionitis and puerperal endometritis?
Continue to the answer...
Chorioamnionitis and puerperal endometritis are polymicrobial, mixed aerobic-anaerobic infections. The dominant organisms are anaerobic gram-negative bacilli (Bacteroides and Prevotella species); anaerobic gram-positive cocci (Peptococcus species and Peptostreptococcus species); aerobic gram-negative bacilli (principally, Escherichia coli, Klebsiella pneumoniae, and Proteus species); and aerobic gram-positive cocci (enterococci, staphylococci, and group B streptococci).
What are the most common organisms that cause chorioamnionitis and puerperal endometritis?
Continue to the answer...
Chorioamnionitis and puerperal endometritis are polymicrobial, mixed aerobic-anaerobic infections. The dominant organisms are anaerobic gram-negative bacilli (Bacteroides and Prevotella species); anaerobic gram-positive cocci (Peptococcus species and Peptostreptococcus species); aerobic gram-negative bacilli (principally, Escherichia coli, Klebsiella pneumoniae, and Proteus species); and aerobic gram-positive cocci (enterococci, staphylococci, and group B streptococci).
- Duff P. Maternal and perinatal infections: bacterial. In: Landon MB, Galan HL, Jauniaux ERM, et al. Gabbe’s Obstetrics: Normal and Problem Pregnancies. 8th ed. Elsevier; 2021:1124-1146.
- Duff P. Maternal and fetal infections. In: Resnik R, Lockwood CJ, Moore TJ, et al. Creasy & Resnik’s Maternal-Fetal Medicine: Principles and Practice. 8th ed. Elsevier; 2019:862-919.
- Duff P. Maternal and perinatal infections: bacterial. In: Landon MB, Galan HL, Jauniaux ERM, et al. Gabbe’s Obstetrics: Normal and Problem Pregnancies. 8th ed. Elsevier; 2021:1124-1146.
- Duff P. Maternal and fetal infections. In: Resnik R, Lockwood CJ, Moore TJ, et al. Creasy & Resnik’s Maternal-Fetal Medicine: Principles and Practice. 8th ed. Elsevier; 2019:862-919.
Weight gain may exacerbate structural damage in knee OA
researchers reported at the OARSI 2022 World Congress.
Using data from the Osteoarthritis Initiative (OAI), researchers from the University of California found that a greater than 5% increase in body weight over 4 years was associated with a 29% increased risk for medial joint space narrowing (JSN), compared with controls (P = .038). There was also a 34% increased risk for developing frequent knee pain (P = .009)
Conversely, weight loss appeared to offer some protection from structural damage in knee OA, Gabby B. Joseph, PhD, a specialist in radiology and biomedical imaging, said at the congress, sponsored by the Osteoarthritis Research Society International.
Indeed, individuals who had achieved a weight loss of more than 5% at 4-year follow up were less likely to have a worsened Kellgren and Lawrence (KL) grade than those whose body weight remained the same (odds ratio, 0.69, P = .009).
Weight loss was also associated with a higher change of experiencing resolution in knee pain over 12 months, with an OR of 1.40 (P = .019).
Importance of weight change in OA
“We know that weight loss has beneficial effects on knee OA symptoms, such as pain relief and improvement in physical function,” commented Xingzhong Jin, PhD, an NHMRC Early Career Fellow at the Centre for Big Data Research in Health at the University of New South Wales, Sydney.
“But what is unclear is whether weight loss could slow down structural degradation in the joint in the long run,” he said in an interview. “These findings mean that weight control is clearly very important for knee OA, in terms of improving symptoms as well as preventing structural progression.”
He added: “The evidence on hip OA is less clear. As most of the knowledge in this space was generated from people with knee OA, this work is an important contribution to knowledge around the care of people with hip OA.”
Why look at weight change effects in OA?
“Obesity is a modifiable risk factor for osteoarthritis,” Dr. Joseph said at the start of her virtual presentation. Indeed, patients with obesity are more than twice as likely to develop knee OA than their normal weight counterparts.
Although there have been various studies looking at weight loss and weight gain in OA, most have focused on weight loss rather than gain, and OA in the knee rather than the hip, she explained.
The aim of the present study, therefore, was to take a closer look at the possible effect of both weight gain and weight loss in people with hip or knee OA in terms of radiographic outcomes (KL grade change, medial JSN), symptomatic outcomes (knee pain and resolution at 12 months), and the need for joint replacement.
“The clinical implications are to develop targeted long-term strategies for site-specific informed recommendations to prevent joint degeneration,” Dr. Joseph said.
Using data on nearly 3,000 individuals from the OAI, Dr Joseph and collaborators classified people with OA into one of three groups: those with at least a 5% gain in weight, (n = 714), those with no (–3% to 3%) change in weight (n = 1,553), and those with at least a 5% loss in weight over a 4-year period.
The results, which were published in Arthritis Care & Research, also revealed no differences in the rate of total hip or knee arthroplasties between the groups, and no differences between the weight gain and weight loss groups and controls in term of hip radiographic or symptomatic changes.
“Why are there differing effects of weight change in the knee versus the hip? This could be multifactorial, but there could be a few things going on,” said Dr. Joseph. “First, the joint structure is clearly different between the knee and the hip. The knee is a hinge joint. The hip is a ball and socket joint malalignment could affect these in different ways.”
Additionally, “the knee also has thicker cartilage, the hip has thinner cartilage again, and the loading patterns may be different in these joints.”
There were also differences in the rate of progression between the knee and the hip, “this was especially noticeable for the radiographic progression,” Dr. Joseph said, with rates being higher in the knee.
Noting that the study is limited by its retrospective design, Dr. Joseph concluded: “We don’t know why these people lost or gained weight. So, this would be something that would be more apparent in a prospective study.
“Also, there were no MRI outcomes, as MRI imaging was not available in the hip in the OAI, but clearly morphology T1 and T2 would be useful to assess as outcomes here as well.”
The OAI is a public-private partnership funded by the National Institutes of Health and initial support from Merck, Novartis, GlaxoSmithKline and Pfizer. Dr. Joseph and Dr. Jin reported having no conflicts of interest to disclose.
researchers reported at the OARSI 2022 World Congress.
Using data from the Osteoarthritis Initiative (OAI), researchers from the University of California found that a greater than 5% increase in body weight over 4 years was associated with a 29% increased risk for medial joint space narrowing (JSN), compared with controls (P = .038). There was also a 34% increased risk for developing frequent knee pain (P = .009)
Conversely, weight loss appeared to offer some protection from structural damage in knee OA, Gabby B. Joseph, PhD, a specialist in radiology and biomedical imaging, said at the congress, sponsored by the Osteoarthritis Research Society International.
Indeed, individuals who had achieved a weight loss of more than 5% at 4-year follow up were less likely to have a worsened Kellgren and Lawrence (KL) grade than those whose body weight remained the same (odds ratio, 0.69, P = .009).
Weight loss was also associated with a higher change of experiencing resolution in knee pain over 12 months, with an OR of 1.40 (P = .019).
Importance of weight change in OA
“We know that weight loss has beneficial effects on knee OA symptoms, such as pain relief and improvement in physical function,” commented Xingzhong Jin, PhD, an NHMRC Early Career Fellow at the Centre for Big Data Research in Health at the University of New South Wales, Sydney.
“But what is unclear is whether weight loss could slow down structural degradation in the joint in the long run,” he said in an interview. “These findings mean that weight control is clearly very important for knee OA, in terms of improving symptoms as well as preventing structural progression.”
He added: “The evidence on hip OA is less clear. As most of the knowledge in this space was generated from people with knee OA, this work is an important contribution to knowledge around the care of people with hip OA.”
Why look at weight change effects in OA?
“Obesity is a modifiable risk factor for osteoarthritis,” Dr. Joseph said at the start of her virtual presentation. Indeed, patients with obesity are more than twice as likely to develop knee OA than their normal weight counterparts.
Although there have been various studies looking at weight loss and weight gain in OA, most have focused on weight loss rather than gain, and OA in the knee rather than the hip, she explained.
The aim of the present study, therefore, was to take a closer look at the possible effect of both weight gain and weight loss in people with hip or knee OA in terms of radiographic outcomes (KL grade change, medial JSN), symptomatic outcomes (knee pain and resolution at 12 months), and the need for joint replacement.
“The clinical implications are to develop targeted long-term strategies for site-specific informed recommendations to prevent joint degeneration,” Dr. Joseph said.
Using data on nearly 3,000 individuals from the OAI, Dr Joseph and collaborators classified people with OA into one of three groups: those with at least a 5% gain in weight, (n = 714), those with no (–3% to 3%) change in weight (n = 1,553), and those with at least a 5% loss in weight over a 4-year period.
The results, which were published in Arthritis Care & Research, also revealed no differences in the rate of total hip or knee arthroplasties between the groups, and no differences between the weight gain and weight loss groups and controls in term of hip radiographic or symptomatic changes.
“Why are there differing effects of weight change in the knee versus the hip? This could be multifactorial, but there could be a few things going on,” said Dr. Joseph. “First, the joint structure is clearly different between the knee and the hip. The knee is a hinge joint. The hip is a ball and socket joint malalignment could affect these in different ways.”
Additionally, “the knee also has thicker cartilage, the hip has thinner cartilage again, and the loading patterns may be different in these joints.”
There were also differences in the rate of progression between the knee and the hip, “this was especially noticeable for the radiographic progression,” Dr. Joseph said, with rates being higher in the knee.
Noting that the study is limited by its retrospective design, Dr. Joseph concluded: “We don’t know why these people lost or gained weight. So, this would be something that would be more apparent in a prospective study.
“Also, there were no MRI outcomes, as MRI imaging was not available in the hip in the OAI, but clearly morphology T1 and T2 would be useful to assess as outcomes here as well.”
The OAI is a public-private partnership funded by the National Institutes of Health and initial support from Merck, Novartis, GlaxoSmithKline and Pfizer. Dr. Joseph and Dr. Jin reported having no conflicts of interest to disclose.
researchers reported at the OARSI 2022 World Congress.
Using data from the Osteoarthritis Initiative (OAI), researchers from the University of California found that a greater than 5% increase in body weight over 4 years was associated with a 29% increased risk for medial joint space narrowing (JSN), compared with controls (P = .038). There was also a 34% increased risk for developing frequent knee pain (P = .009)
Conversely, weight loss appeared to offer some protection from structural damage in knee OA, Gabby B. Joseph, PhD, a specialist in radiology and biomedical imaging, said at the congress, sponsored by the Osteoarthritis Research Society International.
Indeed, individuals who had achieved a weight loss of more than 5% at 4-year follow up were less likely to have a worsened Kellgren and Lawrence (KL) grade than those whose body weight remained the same (odds ratio, 0.69, P = .009).
Weight loss was also associated with a higher change of experiencing resolution in knee pain over 12 months, with an OR of 1.40 (P = .019).
Importance of weight change in OA
“We know that weight loss has beneficial effects on knee OA symptoms, such as pain relief and improvement in physical function,” commented Xingzhong Jin, PhD, an NHMRC Early Career Fellow at the Centre for Big Data Research in Health at the University of New South Wales, Sydney.
“But what is unclear is whether weight loss could slow down structural degradation in the joint in the long run,” he said in an interview. “These findings mean that weight control is clearly very important for knee OA, in terms of improving symptoms as well as preventing structural progression.”
He added: “The evidence on hip OA is less clear. As most of the knowledge in this space was generated from people with knee OA, this work is an important contribution to knowledge around the care of people with hip OA.”
Why look at weight change effects in OA?
“Obesity is a modifiable risk factor for osteoarthritis,” Dr. Joseph said at the start of her virtual presentation. Indeed, patients with obesity are more than twice as likely to develop knee OA than their normal weight counterparts.
Although there have been various studies looking at weight loss and weight gain in OA, most have focused on weight loss rather than gain, and OA in the knee rather than the hip, she explained.
The aim of the present study, therefore, was to take a closer look at the possible effect of both weight gain and weight loss in people with hip or knee OA in terms of radiographic outcomes (KL grade change, medial JSN), symptomatic outcomes (knee pain and resolution at 12 months), and the need for joint replacement.
“The clinical implications are to develop targeted long-term strategies for site-specific informed recommendations to prevent joint degeneration,” Dr. Joseph said.
Using data on nearly 3,000 individuals from the OAI, Dr Joseph and collaborators classified people with OA into one of three groups: those with at least a 5% gain in weight, (n = 714), those with no (–3% to 3%) change in weight (n = 1,553), and those with at least a 5% loss in weight over a 4-year period.
The results, which were published in Arthritis Care & Research, also revealed no differences in the rate of total hip or knee arthroplasties between the groups, and no differences between the weight gain and weight loss groups and controls in term of hip radiographic or symptomatic changes.
“Why are there differing effects of weight change in the knee versus the hip? This could be multifactorial, but there could be a few things going on,” said Dr. Joseph. “First, the joint structure is clearly different between the knee and the hip. The knee is a hinge joint. The hip is a ball and socket joint malalignment could affect these in different ways.”
Additionally, “the knee also has thicker cartilage, the hip has thinner cartilage again, and the loading patterns may be different in these joints.”
There were also differences in the rate of progression between the knee and the hip, “this was especially noticeable for the radiographic progression,” Dr. Joseph said, with rates being higher in the knee.
Noting that the study is limited by its retrospective design, Dr. Joseph concluded: “We don’t know why these people lost or gained weight. So, this would be something that would be more apparent in a prospective study.
“Also, there were no MRI outcomes, as MRI imaging was not available in the hip in the OAI, but clearly morphology T1 and T2 would be useful to assess as outcomes here as well.”
The OAI is a public-private partnership funded by the National Institutes of Health and initial support from Merck, Novartis, GlaxoSmithKline and Pfizer. Dr. Joseph and Dr. Jin reported having no conflicts of interest to disclose.
FROM OARSI 2022
FDA recommends switch to partially or fully disposable duodenoscope models
Health care facilities and providers should complete the transition to fully disposable duodenoscopes and those with disposable components, the U.S. Food and Drug Administration announced this week after an analysis of postmarket surveillance studies was completed.
The FDA’s directive updates its April 2020 recommendations on the subject. It cites concerns about cleaning fixed endcap duodenoscopes and the increasing availability of models that eliminate the need for reprocessing.
The announcement highlighted the potential for a dramatic difference in between-patient contamination risk, reducing it “by half or more as compared to reusable, or fixed endcaps.”
“Interim results from one duodenoscope model with a removable component show a contamination rate of just 0.5%, as compared to older duodenoscope models which had contamination rates as high as 6%,” the FDA writes.
Duodenoscopes are used in more than 500,000 procedures each year in the United States and are key in assessing and treating diseases and conditions of the pancreas and bile ducts.
Upgrade to new models to decrease infections
Manufacturers no longer market fixed endcap models in the United States, but some health care facilities continue to use them. The FDA recommends that all fixed endcap models be replaced.
The FDA says some manufacturers are offering replacement programs to upgrade to a model with a disposable component at no cost.
Two fully disposable models and five with disposable components have been cleared by the FDA. (One model is no longer marketed and thus not listed here.)
Fully Disposable:
Ambu Innovation GmbH, Duodenoscope model aScope Duodeno
Boston Scientific Corporation, EXALT Model D Single-Use Duodenoscope
Disposable Components:
Fujifilm Corporation, Duodenoscope model ED-580XT
Olympus Medical Systems, Evis Exera III Duodenovideoscope Olympus TJF-Q190V
Pentax Medical, Duodenoscope model ED34-i10T2
Pentax Medical, Duodenoscope model ED32-i10
Additionally, the failure to correctly reprocess a duodenoscope could result in tissue or fluid from one patient transferring to a subsequent patient.
“In rare cases, this can lead to patient-to-patient disease transmission,” the FDA says.
Postmarket surveillance studies
In 2015, the FDA ordered three manufacturers of reusable devices (Fujifilm, Olympus, and Pentax) to conduct postmarket surveillance studies to determine contamination rates after reprocessing.
In 2019, the FDA also ordered postmarket surveillance studies to the makers of duodenoscopes with disposable endcaps to verify that the new designs reduce the contamination rate.
The final results of the fixed endcap design indicate that contamination rates were as high as 6.6% with high-concern organisms after contamination. High-concern organisms are those more often associated with disease, such as E coli and Pseudomonas contamination.
“As a result, Pentax and Olympus are withdrawing their fixed endcap duodenoscopes from the market, and Fujifilm has completed withdrawal of its fixed endcap duodenoscope,” the FDA writes.
Studies are not yet complete for duodenoscopes with removable components. As of August 12, 2021, the Fujifilm ED-580XT duodenoscope with a removable cap had 57% of the samples required. Interim results indicate that no samples tested positive for enough low-concern organisms to indicate a reprocessing failure, and only 0.5% tested positive for high-concern organisms.
In addition to the contamination risk sampling, each manufacturer was ordered to do postmarket surveillance studies to evaluate whether staff could understand and follow the manufacturer’s reprocessing instructions in real-world health care settings.
According to the FDA, the results showed that users frequently had difficulty understanding and following the manufacturers’ instructions and were not able to successfully complete reprocessing with the older models.
However, the newer models had high user success rates for understanding instructions and correctly performing reprocessing tasks, the FDA says.
A version of this article first appeared on Medscape.com.
AGA supports FDA’s continued efforts to reduce the risk of disease transmission by duodenoscopes. Through the AGA Center for GI Innovation and Technology, AGA continues to support innovation in medical technology. To get up to date on past challenges with scope infections and future directions, check out AGA’s Innovation in Duodenoscope Design program, consisting of articles, webinars, and podcasts with leading experts in this space.
Health care facilities and providers should complete the transition to fully disposable duodenoscopes and those with disposable components, the U.S. Food and Drug Administration announced this week after an analysis of postmarket surveillance studies was completed.
The FDA’s directive updates its April 2020 recommendations on the subject. It cites concerns about cleaning fixed endcap duodenoscopes and the increasing availability of models that eliminate the need for reprocessing.
The announcement highlighted the potential for a dramatic difference in between-patient contamination risk, reducing it “by half or more as compared to reusable, or fixed endcaps.”
“Interim results from one duodenoscope model with a removable component show a contamination rate of just 0.5%, as compared to older duodenoscope models which had contamination rates as high as 6%,” the FDA writes.
Duodenoscopes are used in more than 500,000 procedures each year in the United States and are key in assessing and treating diseases and conditions of the pancreas and bile ducts.
Upgrade to new models to decrease infections
Manufacturers no longer market fixed endcap models in the United States, but some health care facilities continue to use them. The FDA recommends that all fixed endcap models be replaced.
The FDA says some manufacturers are offering replacement programs to upgrade to a model with a disposable component at no cost.
Two fully disposable models and five with disposable components have been cleared by the FDA. (One model is no longer marketed and thus not listed here.)
Fully Disposable:
Ambu Innovation GmbH, Duodenoscope model aScope Duodeno
Boston Scientific Corporation, EXALT Model D Single-Use Duodenoscope
Disposable Components:
Fujifilm Corporation, Duodenoscope model ED-580XT
Olympus Medical Systems, Evis Exera III Duodenovideoscope Olympus TJF-Q190V
Pentax Medical, Duodenoscope model ED34-i10T2
Pentax Medical, Duodenoscope model ED32-i10
Additionally, the failure to correctly reprocess a duodenoscope could result in tissue or fluid from one patient transferring to a subsequent patient.
“In rare cases, this can lead to patient-to-patient disease transmission,” the FDA says.
Postmarket surveillance studies
In 2015, the FDA ordered three manufacturers of reusable devices (Fujifilm, Olympus, and Pentax) to conduct postmarket surveillance studies to determine contamination rates after reprocessing.
In 2019, the FDA also ordered postmarket surveillance studies to the makers of duodenoscopes with disposable endcaps to verify that the new designs reduce the contamination rate.
The final results of the fixed endcap design indicate that contamination rates were as high as 6.6% with high-concern organisms after contamination. High-concern organisms are those more often associated with disease, such as E coli and Pseudomonas contamination.
“As a result, Pentax and Olympus are withdrawing their fixed endcap duodenoscopes from the market, and Fujifilm has completed withdrawal of its fixed endcap duodenoscope,” the FDA writes.
Studies are not yet complete for duodenoscopes with removable components. As of August 12, 2021, the Fujifilm ED-580XT duodenoscope with a removable cap had 57% of the samples required. Interim results indicate that no samples tested positive for enough low-concern organisms to indicate a reprocessing failure, and only 0.5% tested positive for high-concern organisms.
In addition to the contamination risk sampling, each manufacturer was ordered to do postmarket surveillance studies to evaluate whether staff could understand and follow the manufacturer’s reprocessing instructions in real-world health care settings.
According to the FDA, the results showed that users frequently had difficulty understanding and following the manufacturers’ instructions and were not able to successfully complete reprocessing with the older models.
However, the newer models had high user success rates for understanding instructions and correctly performing reprocessing tasks, the FDA says.
A version of this article first appeared on Medscape.com.
AGA supports FDA’s continued efforts to reduce the risk of disease transmission by duodenoscopes. Through the AGA Center for GI Innovation and Technology, AGA continues to support innovation in medical technology. To get up to date on past challenges with scope infections and future directions, check out AGA’s Innovation in Duodenoscope Design program, consisting of articles, webinars, and podcasts with leading experts in this space.
Health care facilities and providers should complete the transition to fully disposable duodenoscopes and those with disposable components, the U.S. Food and Drug Administration announced this week after an analysis of postmarket surveillance studies was completed.
The FDA’s directive updates its April 2020 recommendations on the subject. It cites concerns about cleaning fixed endcap duodenoscopes and the increasing availability of models that eliminate the need for reprocessing.
The announcement highlighted the potential for a dramatic difference in between-patient contamination risk, reducing it “by half or more as compared to reusable, or fixed endcaps.”
“Interim results from one duodenoscope model with a removable component show a contamination rate of just 0.5%, as compared to older duodenoscope models which had contamination rates as high as 6%,” the FDA writes.
Duodenoscopes are used in more than 500,000 procedures each year in the United States and are key in assessing and treating diseases and conditions of the pancreas and bile ducts.
Upgrade to new models to decrease infections
Manufacturers no longer market fixed endcap models in the United States, but some health care facilities continue to use them. The FDA recommends that all fixed endcap models be replaced.
The FDA says some manufacturers are offering replacement programs to upgrade to a model with a disposable component at no cost.
Two fully disposable models and five with disposable components have been cleared by the FDA. (One model is no longer marketed and thus not listed here.)
Fully Disposable:
Ambu Innovation GmbH, Duodenoscope model aScope Duodeno
Boston Scientific Corporation, EXALT Model D Single-Use Duodenoscope
Disposable Components:
Fujifilm Corporation, Duodenoscope model ED-580XT
Olympus Medical Systems, Evis Exera III Duodenovideoscope Olympus TJF-Q190V
Pentax Medical, Duodenoscope model ED34-i10T2
Pentax Medical, Duodenoscope model ED32-i10
Additionally, the failure to correctly reprocess a duodenoscope could result in tissue or fluid from one patient transferring to a subsequent patient.
“In rare cases, this can lead to patient-to-patient disease transmission,” the FDA says.
Postmarket surveillance studies
In 2015, the FDA ordered three manufacturers of reusable devices (Fujifilm, Olympus, and Pentax) to conduct postmarket surveillance studies to determine contamination rates after reprocessing.
In 2019, the FDA also ordered postmarket surveillance studies to the makers of duodenoscopes with disposable endcaps to verify that the new designs reduce the contamination rate.
The final results of the fixed endcap design indicate that contamination rates were as high as 6.6% with high-concern organisms after contamination. High-concern organisms are those more often associated with disease, such as E coli and Pseudomonas contamination.
“As a result, Pentax and Olympus are withdrawing their fixed endcap duodenoscopes from the market, and Fujifilm has completed withdrawal of its fixed endcap duodenoscope,” the FDA writes.
Studies are not yet complete for duodenoscopes with removable components. As of August 12, 2021, the Fujifilm ED-580XT duodenoscope with a removable cap had 57% of the samples required. Interim results indicate that no samples tested positive for enough low-concern organisms to indicate a reprocessing failure, and only 0.5% tested positive for high-concern organisms.
In addition to the contamination risk sampling, each manufacturer was ordered to do postmarket surveillance studies to evaluate whether staff could understand and follow the manufacturer’s reprocessing instructions in real-world health care settings.
According to the FDA, the results showed that users frequently had difficulty understanding and following the manufacturers’ instructions and were not able to successfully complete reprocessing with the older models.
However, the newer models had high user success rates for understanding instructions and correctly performing reprocessing tasks, the FDA says.
A version of this article first appeared on Medscape.com.
AGA supports FDA’s continued efforts to reduce the risk of disease transmission by duodenoscopes. Through the AGA Center for GI Innovation and Technology, AGA continues to support innovation in medical technology. To get up to date on past challenges with scope infections and future directions, check out AGA’s Innovation in Duodenoscope Design program, consisting of articles, webinars, and podcasts with leading experts in this space.
Unraveling primary ovarian insufficiency
In the presentation of secondary amenorrhea, pregnancy is the No. 1 differential diagnosis. Once this has been excluded, an algorithm is initiated to determine the etiology, including an assessment of the hypothalamic-pituitary-ovarian axis. While the early onset of ovarian failure can be physically and psychologically disrupting, the effect on fertility is an especially devastating event. Previously identified by terms including premature ovarian failure and premature menopause, “primary ovarian insufficiency” (POI) is now the preferred designation. This month’s article will address the diagnosis, evaluation, and management of POI.
The definition of POI is the development of primary hypogonadism before the age of 40 years. Spontaneous POI occurs in approximately 1 in 250 women by age 35 years and 1 in 100 by age 40 years. After excluding pregnancy, the clinician should determine signs and symptoms that can lead to expedited and cost-efficient testing.
Consequences
POI is an important risk factor for bone loss and osteoporosis, especially in young women who develop ovarian dysfunction before they achieve peak adult bone mass. At the time of diagnosis of POI, a bone density test (dual-energy x-ray absorptiometry) should be obtained. Women with POI may also develop depression and anxiety as well as experience an increased risk for cardiovascular morbidity and mortality, possibly related to endothelial dysfunction.
Young women with spontaneous POI are at increased risk of developing autoimmune adrenal insufficiency (AAI), a potentially fatal disorder. Consequently, to diagnose AAI, serum adrenal cortical and 21-hydroxylase antibodies should be measured in all women who have a karyotype of 46,XX and experience spontaneous POI. Women with AAI have a 50% risk of developing adrenal insufficiency. Despite initial normal adrenal function, women with positive adrenal cortical antibodies should be followed annually.
Causes (see table for a more complete list)
Iatrogenic
Known causes of POI include chemotherapy/radiation often in the setting of cancer treatment. The three most commonly used drugs, cyclophosphamide, cisplatin, and doxorubicin, cause POI by inducing death and/or accelerated activation of primordial follicles and increased atresia of growing follicles. The most damaging agents are alkylating drugs. A cyclophosphamide equivalent dose calculator has been established for ovarian failure risk stratification from chemotherapy based on the cumulative dose of alkylating agents received.
One study estimated the radiosensitivity of the oocyte to be less than 2 Gy. Based upon this estimate, the authors calculated the dose of radiotherapy that would result in immediate and permanent ovarian failure in 97.5% of patients as follows:
- 20.3 Gy at birth
- 18.4 Gy at age 10 years
- 16.5 Gy at age 20 years
- 14.3 Gy at age 30 years
Genetic
Approximately 10% of cases are familial. A family history of POI raises concern for a fragile X premutation. Fragile X syndrome is an X-linked form of intellectual disability that is one of the most common causes of mental retardation worldwide. There is a strong relationship between age at menopause, including POI, and premutations for fragile X syndrome. The American College of Obstetricians and Gynecologists recommends that women with POI or an elevated follicle-stimulating hormone (FSH) level before age 40 years without known cause be screened for FMR1 premutations. Approximately 6% of cases of POI are associated with premutations in the FMR1 gene.
Turner syndrome is one of the most common causes of POI and results from the lack of a second X chromosome. The most common chromosomal defect in humans, TS occurs in up to 1.5% of conceptions, 10% of spontaneous abortions, and 1 of 2,500 live births.
Serum antiadrenal and/or anti–21-hydroxylase antibodies and antithyroid antiperoxidase antibodies, can aid in the diagnosis of adrenal gland, ovary, and thyroid autoimmune causes, which is found in 4% of women with spontaneous POI. Testing for the presence of 21-hydroxylase autoantibodies or adrenal autoantibodies is sufficient to make the diagnosis of autoimmune oophoritis in women with proven spontaneous POI.
The etiology of POI remains unknown in approximately 75%-90% of cases. However, studies using whole exome or whole genome sequencing have identified genetic variants in approximately 30%-35% of these patients.
Risk factors
Factors that are thought to play a role in determining the age of menopause, include genetics (e.g., FMR1 premutation and mosaic Turner syndrome), ethnicity (earlier among Hispanic women and later in Japanese American women when compared with White women), and smoking (reduced by approximately 2 years ).
Regarding ovarian aging, the holy grail of the reproductive life span is to predict menopause. While the definitive age eludes us, anti-Müllerian hormone levels appear to show promise. An ultrasensitive anti-Müllerian hormone assay (< 0.01 ng/mL) predicted a 79% probability of menopause within 12 months for women aged 51 and above; the probability was 51% for women below age 48.
Diagnosis
The three P’s of secondary amenorrhea are physiological, pharmacological, or pathological and can guide the clinician to a targeted evaluation. Physiological causes are pregnancy, the first 6 months of continuous breastfeeding (from elevated prolactin), and natural menopause. Pharmacological etiologies, excluding hormonal treatment that suppresses ovulation (combined oral contraceptives, gonadotropin-releasing hormone agonist/antagonist, or danazol), include agents that inhibit dopamine thereby increasing serum prolactin, such as metoclopramide; phenothiazine antipsychotics, such as haloperidol; and tardive dystonia dopamine-depleting medications, such as reserpine. Pathological causes include pituitary adenomas, thyroid disease, functional hypothalamic amenorrhea from changes in weight, exercise regimen, and stress.
Management
About 50%-75% of women with 46,XX spontaneous POI experience intermittent ovarian function and 5%-10% of women remain able to conceive. Anecdotally, a 32-year-old woman presented to me with primary infertility, secondary amenorrhea, and suspected POI based on vasomotor symptoms and elevated FSH levels. Pelvic ultrasound showed a hemorrhagic cyst, suspicious for a corpus luteum. Two weeks thereafter she reported a positive home urine human chorionic gonadotropin test and ultimately delivered twins. Her diagnosis of POI with amenorrhea remained postpartum.
Unless there is an absolute contraindication, estrogen therapy should be prescribed to women with POI to reduce the risk of osteoporosis, cardiovascular disease, and urogenital atrophy as well as to maintain sexual health and quality of life. For those with an intact uterus, women should receive progesterone because of the risk of endometrial hyperplasia from unopposed estrogen. Rather than oral estrogen, the use of transdermal or vaginal delivery of estrogen is a more physiological approach and provides lower risks of venous thromboembolism and gallbladder disease. Of note, standard postmenopausal hormone therapy, which has a much lower dose of estrogen than combined estrogen-progestin contraceptives, does not provide effective contraception. Per ACOG, systemic hormone treatment should be prescribed until age 50-51 years to all women with POI.
For fertility, women with spontaneous POI can be offered oocyte or embryo donation. The uterus does not age reproductively, unlike oocytes, therefore women can achieve reasonable pregnancy success rates through egg donation despite experiencing menopause.
Future potential options
Female germline stem cells have been isolated from neonatal mice and transplanted into sterile adult mice, who then were able to produce offspring. In a second study, oogonial stem cells were isolated from neonatal and adult mouse ovaries; pups were subsequently born from the oocytes. Further experiments are needed before the implications for humans can be determined.
Emotionally traumatic for most women, POI disrupts life plans, hopes, and dreams of raising a family. The approach to the patient with POI involves the above evidence-based testing along with empathy from the health care provider.
Dr. Trolice is director of The IVF Center in Winter Park, Fla., and professor of obstetrics and gynecology at the University of Central Florida, Orlando.
In the presentation of secondary amenorrhea, pregnancy is the No. 1 differential diagnosis. Once this has been excluded, an algorithm is initiated to determine the etiology, including an assessment of the hypothalamic-pituitary-ovarian axis. While the early onset of ovarian failure can be physically and psychologically disrupting, the effect on fertility is an especially devastating event. Previously identified by terms including premature ovarian failure and premature menopause, “primary ovarian insufficiency” (POI) is now the preferred designation. This month’s article will address the diagnosis, evaluation, and management of POI.
The definition of POI is the development of primary hypogonadism before the age of 40 years. Spontaneous POI occurs in approximately 1 in 250 women by age 35 years and 1 in 100 by age 40 years. After excluding pregnancy, the clinician should determine signs and symptoms that can lead to expedited and cost-efficient testing.
Consequences
POI is an important risk factor for bone loss and osteoporosis, especially in young women who develop ovarian dysfunction before they achieve peak adult bone mass. At the time of diagnosis of POI, a bone density test (dual-energy x-ray absorptiometry) should be obtained. Women with POI may also develop depression and anxiety as well as experience an increased risk for cardiovascular morbidity and mortality, possibly related to endothelial dysfunction.
Young women with spontaneous POI are at increased risk of developing autoimmune adrenal insufficiency (AAI), a potentially fatal disorder. Consequently, to diagnose AAI, serum adrenal cortical and 21-hydroxylase antibodies should be measured in all women who have a karyotype of 46,XX and experience spontaneous POI. Women with AAI have a 50% risk of developing adrenal insufficiency. Despite initial normal adrenal function, women with positive adrenal cortical antibodies should be followed annually.
Causes (see table for a more complete list)
Iatrogenic
Known causes of POI include chemotherapy/radiation often in the setting of cancer treatment. The three most commonly used drugs, cyclophosphamide, cisplatin, and doxorubicin, cause POI by inducing death and/or accelerated activation of primordial follicles and increased atresia of growing follicles. The most damaging agents are alkylating drugs. A cyclophosphamide equivalent dose calculator has been established for ovarian failure risk stratification from chemotherapy based on the cumulative dose of alkylating agents received.
One study estimated the radiosensitivity of the oocyte to be less than 2 Gy. Based upon this estimate, the authors calculated the dose of radiotherapy that would result in immediate and permanent ovarian failure in 97.5% of patients as follows:
- 20.3 Gy at birth
- 18.4 Gy at age 10 years
- 16.5 Gy at age 20 years
- 14.3 Gy at age 30 years
Genetic
Approximately 10% of cases are familial. A family history of POI raises concern for a fragile X premutation. Fragile X syndrome is an X-linked form of intellectual disability that is one of the most common causes of mental retardation worldwide. There is a strong relationship between age at menopause, including POI, and premutations for fragile X syndrome. The American College of Obstetricians and Gynecologists recommends that women with POI or an elevated follicle-stimulating hormone (FSH) level before age 40 years without known cause be screened for FMR1 premutations. Approximately 6% of cases of POI are associated with premutations in the FMR1 gene.
Turner syndrome is one of the most common causes of POI and results from the lack of a second X chromosome. The most common chromosomal defect in humans, TS occurs in up to 1.5% of conceptions, 10% of spontaneous abortions, and 1 of 2,500 live births.
Serum antiadrenal and/or anti–21-hydroxylase antibodies and antithyroid antiperoxidase antibodies, can aid in the diagnosis of adrenal gland, ovary, and thyroid autoimmune causes, which is found in 4% of women with spontaneous POI. Testing for the presence of 21-hydroxylase autoantibodies or adrenal autoantibodies is sufficient to make the diagnosis of autoimmune oophoritis in women with proven spontaneous POI.
The etiology of POI remains unknown in approximately 75%-90% of cases. However, studies using whole exome or whole genome sequencing have identified genetic variants in approximately 30%-35% of these patients.
Risk factors
Factors that are thought to play a role in determining the age of menopause, include genetics (e.g., FMR1 premutation and mosaic Turner syndrome), ethnicity (earlier among Hispanic women and later in Japanese American women when compared with White women), and smoking (reduced by approximately 2 years ).
Regarding ovarian aging, the holy grail of the reproductive life span is to predict menopause. While the definitive age eludes us, anti-Müllerian hormone levels appear to show promise. An ultrasensitive anti-Müllerian hormone assay (< 0.01 ng/mL) predicted a 79% probability of menopause within 12 months for women aged 51 and above; the probability was 51% for women below age 48.
Diagnosis
The three P’s of secondary amenorrhea are physiological, pharmacological, or pathological and can guide the clinician to a targeted evaluation. Physiological causes are pregnancy, the first 6 months of continuous breastfeeding (from elevated prolactin), and natural menopause. Pharmacological etiologies, excluding hormonal treatment that suppresses ovulation (combined oral contraceptives, gonadotropin-releasing hormone agonist/antagonist, or danazol), include agents that inhibit dopamine thereby increasing serum prolactin, such as metoclopramide; phenothiazine antipsychotics, such as haloperidol; and tardive dystonia dopamine-depleting medications, such as reserpine. Pathological causes include pituitary adenomas, thyroid disease, functional hypothalamic amenorrhea from changes in weight, exercise regimen, and stress.
Management
About 50%-75% of women with 46,XX spontaneous POI experience intermittent ovarian function and 5%-10% of women remain able to conceive. Anecdotally, a 32-year-old woman presented to me with primary infertility, secondary amenorrhea, and suspected POI based on vasomotor symptoms and elevated FSH levels. Pelvic ultrasound showed a hemorrhagic cyst, suspicious for a corpus luteum. Two weeks thereafter she reported a positive home urine human chorionic gonadotropin test and ultimately delivered twins. Her diagnosis of POI with amenorrhea remained postpartum.
Unless there is an absolute contraindication, estrogen therapy should be prescribed to women with POI to reduce the risk of osteoporosis, cardiovascular disease, and urogenital atrophy as well as to maintain sexual health and quality of life. For those with an intact uterus, women should receive progesterone because of the risk of endometrial hyperplasia from unopposed estrogen. Rather than oral estrogen, the use of transdermal or vaginal delivery of estrogen is a more physiological approach and provides lower risks of venous thromboembolism and gallbladder disease. Of note, standard postmenopausal hormone therapy, which has a much lower dose of estrogen than combined estrogen-progestin contraceptives, does not provide effective contraception. Per ACOG, systemic hormone treatment should be prescribed until age 50-51 years to all women with POI.
For fertility, women with spontaneous POI can be offered oocyte or embryo donation. The uterus does not age reproductively, unlike oocytes, therefore women can achieve reasonable pregnancy success rates through egg donation despite experiencing menopause.
Future potential options
Female germline stem cells have been isolated from neonatal mice and transplanted into sterile adult mice, who then were able to produce offspring. In a second study, oogonial stem cells were isolated from neonatal and adult mouse ovaries; pups were subsequently born from the oocytes. Further experiments are needed before the implications for humans can be determined.
Emotionally traumatic for most women, POI disrupts life plans, hopes, and dreams of raising a family. The approach to the patient with POI involves the above evidence-based testing along with empathy from the health care provider.
Dr. Trolice is director of The IVF Center in Winter Park, Fla., and professor of obstetrics and gynecology at the University of Central Florida, Orlando.
In the presentation of secondary amenorrhea, pregnancy is the No. 1 differential diagnosis. Once this has been excluded, an algorithm is initiated to determine the etiology, including an assessment of the hypothalamic-pituitary-ovarian axis. While the early onset of ovarian failure can be physically and psychologically disrupting, the effect on fertility is an especially devastating event. Previously identified by terms including premature ovarian failure and premature menopause, “primary ovarian insufficiency” (POI) is now the preferred designation. This month’s article will address the diagnosis, evaluation, and management of POI.
The definition of POI is the development of primary hypogonadism before the age of 40 years. Spontaneous POI occurs in approximately 1 in 250 women by age 35 years and 1 in 100 by age 40 years. After excluding pregnancy, the clinician should determine signs and symptoms that can lead to expedited and cost-efficient testing.
Consequences
POI is an important risk factor for bone loss and osteoporosis, especially in young women who develop ovarian dysfunction before they achieve peak adult bone mass. At the time of diagnosis of POI, a bone density test (dual-energy x-ray absorptiometry) should be obtained. Women with POI may also develop depression and anxiety as well as experience an increased risk for cardiovascular morbidity and mortality, possibly related to endothelial dysfunction.
Young women with spontaneous POI are at increased risk of developing autoimmune adrenal insufficiency (AAI), a potentially fatal disorder. Consequently, to diagnose AAI, serum adrenal cortical and 21-hydroxylase antibodies should be measured in all women who have a karyotype of 46,XX and experience spontaneous POI. Women with AAI have a 50% risk of developing adrenal insufficiency. Despite initial normal adrenal function, women with positive adrenal cortical antibodies should be followed annually.
Causes (see table for a more complete list)
Iatrogenic
Known causes of POI include chemotherapy/radiation often in the setting of cancer treatment. The three most commonly used drugs, cyclophosphamide, cisplatin, and doxorubicin, cause POI by inducing death and/or accelerated activation of primordial follicles and increased atresia of growing follicles. The most damaging agents are alkylating drugs. A cyclophosphamide equivalent dose calculator has been established for ovarian failure risk stratification from chemotherapy based on the cumulative dose of alkylating agents received.
One study estimated the radiosensitivity of the oocyte to be less than 2 Gy. Based upon this estimate, the authors calculated the dose of radiotherapy that would result in immediate and permanent ovarian failure in 97.5% of patients as follows:
- 20.3 Gy at birth
- 18.4 Gy at age 10 years
- 16.5 Gy at age 20 years
- 14.3 Gy at age 30 years
Genetic
Approximately 10% of cases are familial. A family history of POI raises concern for a fragile X premutation. Fragile X syndrome is an X-linked form of intellectual disability that is one of the most common causes of mental retardation worldwide. There is a strong relationship between age at menopause, including POI, and premutations for fragile X syndrome. The American College of Obstetricians and Gynecologists recommends that women with POI or an elevated follicle-stimulating hormone (FSH) level before age 40 years without known cause be screened for FMR1 premutations. Approximately 6% of cases of POI are associated with premutations in the FMR1 gene.
Turner syndrome is one of the most common causes of POI and results from the lack of a second X chromosome. The most common chromosomal defect in humans, TS occurs in up to 1.5% of conceptions, 10% of spontaneous abortions, and 1 of 2,500 live births.
Serum antiadrenal and/or anti–21-hydroxylase antibodies and antithyroid antiperoxidase antibodies, can aid in the diagnosis of adrenal gland, ovary, and thyroid autoimmune causes, which is found in 4% of women with spontaneous POI. Testing for the presence of 21-hydroxylase autoantibodies or adrenal autoantibodies is sufficient to make the diagnosis of autoimmune oophoritis in women with proven spontaneous POI.
The etiology of POI remains unknown in approximately 75%-90% of cases. However, studies using whole exome or whole genome sequencing have identified genetic variants in approximately 30%-35% of these patients.
Risk factors
Factors that are thought to play a role in determining the age of menopause, include genetics (e.g., FMR1 premutation and mosaic Turner syndrome), ethnicity (earlier among Hispanic women and later in Japanese American women when compared with White women), and smoking (reduced by approximately 2 years ).
Regarding ovarian aging, the holy grail of the reproductive life span is to predict menopause. While the definitive age eludes us, anti-Müllerian hormone levels appear to show promise. An ultrasensitive anti-Müllerian hormone assay (< 0.01 ng/mL) predicted a 79% probability of menopause within 12 months for women aged 51 and above; the probability was 51% for women below age 48.
Diagnosis
The three P’s of secondary amenorrhea are physiological, pharmacological, or pathological and can guide the clinician to a targeted evaluation. Physiological causes are pregnancy, the first 6 months of continuous breastfeeding (from elevated prolactin), and natural menopause. Pharmacological etiologies, excluding hormonal treatment that suppresses ovulation (combined oral contraceptives, gonadotropin-releasing hormone agonist/antagonist, or danazol), include agents that inhibit dopamine thereby increasing serum prolactin, such as metoclopramide; phenothiazine antipsychotics, such as haloperidol; and tardive dystonia dopamine-depleting medications, such as reserpine. Pathological causes include pituitary adenomas, thyroid disease, functional hypothalamic amenorrhea from changes in weight, exercise regimen, and stress.
Management
About 50%-75% of women with 46,XX spontaneous POI experience intermittent ovarian function and 5%-10% of women remain able to conceive. Anecdotally, a 32-year-old woman presented to me with primary infertility, secondary amenorrhea, and suspected POI based on vasomotor symptoms and elevated FSH levels. Pelvic ultrasound showed a hemorrhagic cyst, suspicious for a corpus luteum. Two weeks thereafter she reported a positive home urine human chorionic gonadotropin test and ultimately delivered twins. Her diagnosis of POI with amenorrhea remained postpartum.
Unless there is an absolute contraindication, estrogen therapy should be prescribed to women with POI to reduce the risk of osteoporosis, cardiovascular disease, and urogenital atrophy as well as to maintain sexual health and quality of life. For those with an intact uterus, women should receive progesterone because of the risk of endometrial hyperplasia from unopposed estrogen. Rather than oral estrogen, the use of transdermal or vaginal delivery of estrogen is a more physiological approach and provides lower risks of venous thromboembolism and gallbladder disease. Of note, standard postmenopausal hormone therapy, which has a much lower dose of estrogen than combined estrogen-progestin contraceptives, does not provide effective contraception. Per ACOG, systemic hormone treatment should be prescribed until age 50-51 years to all women with POI.
For fertility, women with spontaneous POI can be offered oocyte or embryo donation. The uterus does not age reproductively, unlike oocytes, therefore women can achieve reasonable pregnancy success rates through egg donation despite experiencing menopause.
Future potential options
Female germline stem cells have been isolated from neonatal mice and transplanted into sterile adult mice, who then were able to produce offspring. In a second study, oogonial stem cells were isolated from neonatal and adult mouse ovaries; pups were subsequently born from the oocytes. Further experiments are needed before the implications for humans can be determined.
Emotionally traumatic for most women, POI disrupts life plans, hopes, and dreams of raising a family. The approach to the patient with POI involves the above evidence-based testing along with empathy from the health care provider.
Dr. Trolice is director of The IVF Center in Winter Park, Fla., and professor of obstetrics and gynecology at the University of Central Florida, Orlando.
Assay-guided chemo in recurrent glioma linked to longer survival
New research suggests that chemotherapy treatments for recurrent high-grade gliomas indicated by an assay-guided tool called ChemoID can boost median survival, compared with physician choice.
The randomized, phase 3 trial results were presented at the annual meeting of the American Association for Cancer Research.
Over a median follow-up of 9 months, median overall survival in the ChemoID group was 12.5 months (95% confidence interval, 10.2-14.7), compared with 9 months (95% CI, 4.2-13.8) in the group whose treatments were chosen by physicians (P = .010).
“While the prognosis is very dismal, we’re still providing a 3.5-month benefit in the guided arm versus physician choice,” said study coauthor Jagan Valluri, PhD, professor of cellular biology and integrative medicine at Marshall University, Huntington, W. Va.
As Dr. Valluri noted, patients with recurrent high-grade gliomas typically have failed radiation and are left with poor prognoses. Fewer than one in four patients respond to chemotherapy at this point, he said, and the response is inconsistent from patient to patient.
“We developed ChemoID since cancer is very unique,” he said, “and any kind of chemotherapy should be tailored to each individual patient on a case-by-case basis.”
The ChemoID tool, a proprietary assay, tests the response of patient cells to various chemotherapy treatments. A test costs $3,500, and some insurers cover it, Dr. Valluri said.
For the new study, researchers randomly assigned 50 patients with grade III/IV recurrent glioma to be treated with chemotherapy chosen by physicians or chemotherapy recommended by the ChemoID tool.
Risk of death in the ChemoID group was lower than in the physician-guided group (hazard ratio, 0.44; 95% CI, 0.24-0.81; P = .008), and median progression-free survival was higher in the ChemoID group (10.1 months vs. 3.5 months; 95% CI, 4.8-15.4 vs. 1.9-5.1; HR, 0.25; 95% CI, 0.14-0.44; P < .001).
“We want the treating physician to have actionable tools in front of them before they treat the patient,” Dr. Valluri said. “We want this assay to become mainstream and part of the standard care workup.”
The study is funded by Cordgenics, where Dr. Valluri serves as chief operating officer.
New research suggests that chemotherapy treatments for recurrent high-grade gliomas indicated by an assay-guided tool called ChemoID can boost median survival, compared with physician choice.
The randomized, phase 3 trial results were presented at the annual meeting of the American Association for Cancer Research.
Over a median follow-up of 9 months, median overall survival in the ChemoID group was 12.5 months (95% confidence interval, 10.2-14.7), compared with 9 months (95% CI, 4.2-13.8) in the group whose treatments were chosen by physicians (P = .010).
“While the prognosis is very dismal, we’re still providing a 3.5-month benefit in the guided arm versus physician choice,” said study coauthor Jagan Valluri, PhD, professor of cellular biology and integrative medicine at Marshall University, Huntington, W. Va.
As Dr. Valluri noted, patients with recurrent high-grade gliomas typically have failed radiation and are left with poor prognoses. Fewer than one in four patients respond to chemotherapy at this point, he said, and the response is inconsistent from patient to patient.
“We developed ChemoID since cancer is very unique,” he said, “and any kind of chemotherapy should be tailored to each individual patient on a case-by-case basis.”
The ChemoID tool, a proprietary assay, tests the response of patient cells to various chemotherapy treatments. A test costs $3,500, and some insurers cover it, Dr. Valluri said.
For the new study, researchers randomly assigned 50 patients with grade III/IV recurrent glioma to be treated with chemotherapy chosen by physicians or chemotherapy recommended by the ChemoID tool.
Risk of death in the ChemoID group was lower than in the physician-guided group (hazard ratio, 0.44; 95% CI, 0.24-0.81; P = .008), and median progression-free survival was higher in the ChemoID group (10.1 months vs. 3.5 months; 95% CI, 4.8-15.4 vs. 1.9-5.1; HR, 0.25; 95% CI, 0.14-0.44; P < .001).
“We want the treating physician to have actionable tools in front of them before they treat the patient,” Dr. Valluri said. “We want this assay to become mainstream and part of the standard care workup.”
The study is funded by Cordgenics, where Dr. Valluri serves as chief operating officer.
New research suggests that chemotherapy treatments for recurrent high-grade gliomas indicated by an assay-guided tool called ChemoID can boost median survival, compared with physician choice.
The randomized, phase 3 trial results were presented at the annual meeting of the American Association for Cancer Research.
Over a median follow-up of 9 months, median overall survival in the ChemoID group was 12.5 months (95% confidence interval, 10.2-14.7), compared with 9 months (95% CI, 4.2-13.8) in the group whose treatments were chosen by physicians (P = .010).
“While the prognosis is very dismal, we’re still providing a 3.5-month benefit in the guided arm versus physician choice,” said study coauthor Jagan Valluri, PhD, professor of cellular biology and integrative medicine at Marshall University, Huntington, W. Va.
As Dr. Valluri noted, patients with recurrent high-grade gliomas typically have failed radiation and are left with poor prognoses. Fewer than one in four patients respond to chemotherapy at this point, he said, and the response is inconsistent from patient to patient.
“We developed ChemoID since cancer is very unique,” he said, “and any kind of chemotherapy should be tailored to each individual patient on a case-by-case basis.”
The ChemoID tool, a proprietary assay, tests the response of patient cells to various chemotherapy treatments. A test costs $3,500, and some insurers cover it, Dr. Valluri said.
For the new study, researchers randomly assigned 50 patients with grade III/IV recurrent glioma to be treated with chemotherapy chosen by physicians or chemotherapy recommended by the ChemoID tool.
Risk of death in the ChemoID group was lower than in the physician-guided group (hazard ratio, 0.44; 95% CI, 0.24-0.81; P = .008), and median progression-free survival was higher in the ChemoID group (10.1 months vs. 3.5 months; 95% CI, 4.8-15.4 vs. 1.9-5.1; HR, 0.25; 95% CI, 0.14-0.44; P < .001).
“We want the treating physician to have actionable tools in front of them before they treat the patient,” Dr. Valluri said. “We want this assay to become mainstream and part of the standard care workup.”
The study is funded by Cordgenics, where Dr. Valluri serves as chief operating officer.
FROM AACR 2022
Persistent problem: High C-section rates plague the South
All along, Julia Maeda knew she wanted to have her baby naturally. For her, that meant in a hospital, vaginally, without an epidural for pain relief.
This was her first pregnancy. And although she is a nurse, she was working with cancer patients at the time, not with laboring mothers or babies. “I really didn’t know what I was getting into,” said Ms. Maeda, now 32. “I didn’t do much preparation.”
Her home state of Mississippi has the highest cesarean section rate in the United States – nearly 4 in 10 women who give birth there deliver their babies via C-section. Almost 2 weeks past her due date in 2019, Ms. Maeda became one of them after her doctor came to her bedside while she was in labor.
“‘You’re not in distress, and your baby is not in distress – but we don’t want you to get that way, so we need to think about a C-section,’” she recalled her doctor saying. “I was totally defeated. I just gave in.”
C-sections are sometimes necessary and even lifesaving, but public health experts have long contended that too many performed in the U.S. aren’t. They argue it is major surgery accompanied by significant risk and a high price tag.
Overall, 31.8% of all births in the U.S. were C-sections in 2020, just a slight tick up from 31.7% the year before, according to the latest data from the Centers for Disease Control and Prevention. But that’s close to the peak in 2009, when it was 32.9%. And the rates are far higher in many states, especially across the South.
These high C-section rates have persisted – and in some states, such as Alabama and Kentucky, even grown slightly – despite continual calls to reduce them. And although the pandemic presented new challenges for pregnant women, research suggests that the U.S. C-section rate was unaffected by COVID. Instead, obstetricians and other health experts say the high rate is an intractable problem.
Some states, such as California and New Jersey, have reduced their rates through a variety of strategies, including sharing C-section data with doctors and hospitals. But change has proved difficult elsewhere, especially in the South and in Texas, where women are generally less healthy heading into their pregnancies and maternal and infant health problems are among the highest in the United States.
“We have to restructure how we think about C-sections,” said Veronica Gillispie-Bell, MD, an ob.gyn. who is medical director of the Louisiana Perinatal Quality Collaborative, Kenner, La., a group of 43 birthing hospitals focused on lowering Louisiana’s C-section rate. “It’s a lifesaving technique, but it’s also not without risks.”
She said C-sections, like any operation, create scar tissue, including in the uterus, which may complicate future pregnancies or abdominal surgeries. C-sections also typically lead to an extended hospital stay and recovery period and increase the chance of infection. Babies face risks, too. In rare cases, they can be nicked or cut during an incision.
Although C-sections are sometimes necessary, public health leaders say these surgeries have been overused in many places. Black women, particularly, are more likely to give birth by C-section than any other racial group in the country. Often, hospitals and even regions have wide, unexplained variations in rates.
“If you were delivering in Miami-Dade County, you had a 75% greater chance of having a cesarean than in northern Florida,” said William Sappenfield, MD, an ob.gyn. and epidemiologist at the University of South Florida, Tampa, who has studied the state’s high C-section rate.
Some physicians say their rates are driven by mothers who request the procedure, not by doctors. But Rebekah Gee, MD, an ob.gyn. at Louisiana State University Healthcare Network, New Orleans, and former secretary of the Louisiana Department of Health, said she saw C-section rates go dramatically up at 4 and 5 p.m. – around the time when doctors tend to want to go home.
She led several initiatives to improve birth outcomes in Louisiana, including leveling Medicaid payment rates to hospitals for vaginal deliveries and C-sections. In most places, C-sections are significantly more expensive than vaginal deliveries, making high C-section rates not only a concern for expectant mothers but also for taxpayers.
Medicaid pays for 60% of all births in Louisiana, according to KFF, and about half of all births in most Southern states, compared with 42% nationally. That’s one reason some states – including Louisiana, Tennessee, and Minnesota – have tried to tackle high C-section rates by changing how much Medicaid pays for them. But payment reform alone isn’t enough, Dr. Gee said.
“There was a guy in central Louisiana who was doing more C-sections and early elective deliveries than anyone in the U.S.,” she said. “When you have a culture like that, it’s hard to shift from it.”
Linda Schwimmer, president and CEO of the New Jersey Health Care Quality Institute, said many hospitals and doctors don’t even know their C-section rates. Sharing this data with doctors and hospitals – and making it public – made some providers uncomfortable, she said, but it ultimately worked. New Jersey’s C-section rate among first-time, low-risk mothers dropped from 33.1% in 2013 to 26.7% 6 years later once the state began sharing these data, among other initiatives.
The New Jersey Health Care Quality Institute and other groups like it around the country focus on reducing a subset of C-sections called “nulliparous, term, singleton, vertex” C-sections, or surgeries on first-time, full-term moms giving birth to a single infant who is positioned head-down in the uterus.
NTSV C-sections are important to track because women who have a C-section during their first pregnancy face a 90% chance of having another in subsequent pregnancies. Across the U.S., the rate for these C-sections was 25.9% in 2020 and 25.6% in 2019.
Elliott Main, MD, a maternal-fetal specialist at Stanford (Calif.) University and the medical director of the California Maternal Quality Care Collaborative, coauthored a paper, published in JAMA last year, that outlines interventions the collaborative took that lowered California’s NTSV C-Section rate from 26.0% in 2014 to 22.8% in 2019. Nationally, the rate was unchanged during that period.
Allowing women to labor for longer stretches of time before resorting to surgery is important, he said.
The cervix must be 10 cm dilated before a woman gives birth. The threshold for “active labor” used to be when the cervix was dilated at least 4 cm. In more recent years, though, the onset of active labor has been changed to 5-6 cm.
“People show up at the hospital too early,” said Toni Hill, president of the Mississippi Midwives Alliance. “If you show up to the hospital at 2-3 centimeters, you can be at 2-3 centimeters for weeks. I don’t even consider that labor.”
Too often, she said, women at an early stage of labor end up being induced and deliver via C-section.
“It’s almost like, at this point, C-sections are being handed out like lollipops,” said LA’Patricia Washington, a doula based in Jackson, Miss. Doulas are trained, nonmedical workers who help parents before, during, and after delivery.
Ms. Washington works with a nonprofit group, the Jackson Safer Childbirth Experience, that pays for doulas to help expectant mothers in the region. Some state Medicaid programs, such as New Jersey’s, reimburse for services by doulas because research shows they can reduce C-section rates. California has been trying to roll out the same benefit for its Medicaid members.
In 2020, when Julia Maeda became pregnant again, she paid out-of-pocket for a doula to attend the birth. The experience of having her son via C-section the previous year had been “emotionally and psychologically traumatic,” Ms. Maeda said.
She told her ob.gyn. that she wanted a VBAC, short for “vaginal birth after cesarean.” But, she said, “he just shook his head and said, ‘That’s not a good idea.’”
She had VBAC anyway. Ms. Maeda credits her doula with making it happen.
“Maybe just her presence relayed to the nursing staff that this was something I was serious about,” Ms. Maeda said. “They want you to have your baby during business hours. And biology doesn’t work that way.”
KHN (Kaiser Health News) is a national newsroom that produces in-depth journalism about health issues. Together with Policy Analysis and Polling, KHN is one of the three major operating programs at KFF (Kaiser Family Foundation). KFF is an endowed nonprofit organization providing information on health issues to the nation.
All along, Julia Maeda knew she wanted to have her baby naturally. For her, that meant in a hospital, vaginally, without an epidural for pain relief.
This was her first pregnancy. And although she is a nurse, she was working with cancer patients at the time, not with laboring mothers or babies. “I really didn’t know what I was getting into,” said Ms. Maeda, now 32. “I didn’t do much preparation.”
Her home state of Mississippi has the highest cesarean section rate in the United States – nearly 4 in 10 women who give birth there deliver their babies via C-section. Almost 2 weeks past her due date in 2019, Ms. Maeda became one of them after her doctor came to her bedside while she was in labor.
“‘You’re not in distress, and your baby is not in distress – but we don’t want you to get that way, so we need to think about a C-section,’” she recalled her doctor saying. “I was totally defeated. I just gave in.”
C-sections are sometimes necessary and even lifesaving, but public health experts have long contended that too many performed in the U.S. aren’t. They argue it is major surgery accompanied by significant risk and a high price tag.
Overall, 31.8% of all births in the U.S. were C-sections in 2020, just a slight tick up from 31.7% the year before, according to the latest data from the Centers for Disease Control and Prevention. But that’s close to the peak in 2009, when it was 32.9%. And the rates are far higher in many states, especially across the South.
These high C-section rates have persisted – and in some states, such as Alabama and Kentucky, even grown slightly – despite continual calls to reduce them. And although the pandemic presented new challenges for pregnant women, research suggests that the U.S. C-section rate was unaffected by COVID. Instead, obstetricians and other health experts say the high rate is an intractable problem.
Some states, such as California and New Jersey, have reduced their rates through a variety of strategies, including sharing C-section data with doctors and hospitals. But change has proved difficult elsewhere, especially in the South and in Texas, where women are generally less healthy heading into their pregnancies and maternal and infant health problems are among the highest in the United States.
“We have to restructure how we think about C-sections,” said Veronica Gillispie-Bell, MD, an ob.gyn. who is medical director of the Louisiana Perinatal Quality Collaborative, Kenner, La., a group of 43 birthing hospitals focused on lowering Louisiana’s C-section rate. “It’s a lifesaving technique, but it’s also not without risks.”
She said C-sections, like any operation, create scar tissue, including in the uterus, which may complicate future pregnancies or abdominal surgeries. C-sections also typically lead to an extended hospital stay and recovery period and increase the chance of infection. Babies face risks, too. In rare cases, they can be nicked or cut during an incision.
Although C-sections are sometimes necessary, public health leaders say these surgeries have been overused in many places. Black women, particularly, are more likely to give birth by C-section than any other racial group in the country. Often, hospitals and even regions have wide, unexplained variations in rates.
“If you were delivering in Miami-Dade County, you had a 75% greater chance of having a cesarean than in northern Florida,” said William Sappenfield, MD, an ob.gyn. and epidemiologist at the University of South Florida, Tampa, who has studied the state’s high C-section rate.
Some physicians say their rates are driven by mothers who request the procedure, not by doctors. But Rebekah Gee, MD, an ob.gyn. at Louisiana State University Healthcare Network, New Orleans, and former secretary of the Louisiana Department of Health, said she saw C-section rates go dramatically up at 4 and 5 p.m. – around the time when doctors tend to want to go home.
She led several initiatives to improve birth outcomes in Louisiana, including leveling Medicaid payment rates to hospitals for vaginal deliveries and C-sections. In most places, C-sections are significantly more expensive than vaginal deliveries, making high C-section rates not only a concern for expectant mothers but also for taxpayers.
Medicaid pays for 60% of all births in Louisiana, according to KFF, and about half of all births in most Southern states, compared with 42% nationally. That’s one reason some states – including Louisiana, Tennessee, and Minnesota – have tried to tackle high C-section rates by changing how much Medicaid pays for them. But payment reform alone isn’t enough, Dr. Gee said.
“There was a guy in central Louisiana who was doing more C-sections and early elective deliveries than anyone in the U.S.,” she said. “When you have a culture like that, it’s hard to shift from it.”
Linda Schwimmer, president and CEO of the New Jersey Health Care Quality Institute, said many hospitals and doctors don’t even know their C-section rates. Sharing this data with doctors and hospitals – and making it public – made some providers uncomfortable, she said, but it ultimately worked. New Jersey’s C-section rate among first-time, low-risk mothers dropped from 33.1% in 2013 to 26.7% 6 years later once the state began sharing these data, among other initiatives.
The New Jersey Health Care Quality Institute and other groups like it around the country focus on reducing a subset of C-sections called “nulliparous, term, singleton, vertex” C-sections, or surgeries on first-time, full-term moms giving birth to a single infant who is positioned head-down in the uterus.
NTSV C-sections are important to track because women who have a C-section during their first pregnancy face a 90% chance of having another in subsequent pregnancies. Across the U.S., the rate for these C-sections was 25.9% in 2020 and 25.6% in 2019.
Elliott Main, MD, a maternal-fetal specialist at Stanford (Calif.) University and the medical director of the California Maternal Quality Care Collaborative, coauthored a paper, published in JAMA last year, that outlines interventions the collaborative took that lowered California’s NTSV C-Section rate from 26.0% in 2014 to 22.8% in 2019. Nationally, the rate was unchanged during that period.
Allowing women to labor for longer stretches of time before resorting to surgery is important, he said.
The cervix must be 10 cm dilated before a woman gives birth. The threshold for “active labor” used to be when the cervix was dilated at least 4 cm. In more recent years, though, the onset of active labor has been changed to 5-6 cm.
“People show up at the hospital too early,” said Toni Hill, president of the Mississippi Midwives Alliance. “If you show up to the hospital at 2-3 centimeters, you can be at 2-3 centimeters for weeks. I don’t even consider that labor.”
Too often, she said, women at an early stage of labor end up being induced and deliver via C-section.
“It’s almost like, at this point, C-sections are being handed out like lollipops,” said LA’Patricia Washington, a doula based in Jackson, Miss. Doulas are trained, nonmedical workers who help parents before, during, and after delivery.
Ms. Washington works with a nonprofit group, the Jackson Safer Childbirth Experience, that pays for doulas to help expectant mothers in the region. Some state Medicaid programs, such as New Jersey’s, reimburse for services by doulas because research shows they can reduce C-section rates. California has been trying to roll out the same benefit for its Medicaid members.
In 2020, when Julia Maeda became pregnant again, she paid out-of-pocket for a doula to attend the birth. The experience of having her son via C-section the previous year had been “emotionally and psychologically traumatic,” Ms. Maeda said.
She told her ob.gyn. that she wanted a VBAC, short for “vaginal birth after cesarean.” But, she said, “he just shook his head and said, ‘That’s not a good idea.’”
She had VBAC anyway. Ms. Maeda credits her doula with making it happen.
“Maybe just her presence relayed to the nursing staff that this was something I was serious about,” Ms. Maeda said. “They want you to have your baby during business hours. And biology doesn’t work that way.”
KHN (Kaiser Health News) is a national newsroom that produces in-depth journalism about health issues. Together with Policy Analysis and Polling, KHN is one of the three major operating programs at KFF (Kaiser Family Foundation). KFF is an endowed nonprofit organization providing information on health issues to the nation.
All along, Julia Maeda knew she wanted to have her baby naturally. For her, that meant in a hospital, vaginally, without an epidural for pain relief.
This was her first pregnancy. And although she is a nurse, she was working with cancer patients at the time, not with laboring mothers or babies. “I really didn’t know what I was getting into,” said Ms. Maeda, now 32. “I didn’t do much preparation.”
Her home state of Mississippi has the highest cesarean section rate in the United States – nearly 4 in 10 women who give birth there deliver their babies via C-section. Almost 2 weeks past her due date in 2019, Ms. Maeda became one of them after her doctor came to her bedside while she was in labor.
“‘You’re not in distress, and your baby is not in distress – but we don’t want you to get that way, so we need to think about a C-section,’” she recalled her doctor saying. “I was totally defeated. I just gave in.”
C-sections are sometimes necessary and even lifesaving, but public health experts have long contended that too many performed in the U.S. aren’t. They argue it is major surgery accompanied by significant risk and a high price tag.
Overall, 31.8% of all births in the U.S. were C-sections in 2020, just a slight tick up from 31.7% the year before, according to the latest data from the Centers for Disease Control and Prevention. But that’s close to the peak in 2009, when it was 32.9%. And the rates are far higher in many states, especially across the South.
These high C-section rates have persisted – and in some states, such as Alabama and Kentucky, even grown slightly – despite continual calls to reduce them. And although the pandemic presented new challenges for pregnant women, research suggests that the U.S. C-section rate was unaffected by COVID. Instead, obstetricians and other health experts say the high rate is an intractable problem.
Some states, such as California and New Jersey, have reduced their rates through a variety of strategies, including sharing C-section data with doctors and hospitals. But change has proved difficult elsewhere, especially in the South and in Texas, where women are generally less healthy heading into their pregnancies and maternal and infant health problems are among the highest in the United States.
“We have to restructure how we think about C-sections,” said Veronica Gillispie-Bell, MD, an ob.gyn. who is medical director of the Louisiana Perinatal Quality Collaborative, Kenner, La., a group of 43 birthing hospitals focused on lowering Louisiana’s C-section rate. “It’s a lifesaving technique, but it’s also not without risks.”
She said C-sections, like any operation, create scar tissue, including in the uterus, which may complicate future pregnancies or abdominal surgeries. C-sections also typically lead to an extended hospital stay and recovery period and increase the chance of infection. Babies face risks, too. In rare cases, they can be nicked or cut during an incision.
Although C-sections are sometimes necessary, public health leaders say these surgeries have been overused in many places. Black women, particularly, are more likely to give birth by C-section than any other racial group in the country. Often, hospitals and even regions have wide, unexplained variations in rates.
“If you were delivering in Miami-Dade County, you had a 75% greater chance of having a cesarean than in northern Florida,” said William Sappenfield, MD, an ob.gyn. and epidemiologist at the University of South Florida, Tampa, who has studied the state’s high C-section rate.
Some physicians say their rates are driven by mothers who request the procedure, not by doctors. But Rebekah Gee, MD, an ob.gyn. at Louisiana State University Healthcare Network, New Orleans, and former secretary of the Louisiana Department of Health, said she saw C-section rates go dramatically up at 4 and 5 p.m. – around the time when doctors tend to want to go home.
She led several initiatives to improve birth outcomes in Louisiana, including leveling Medicaid payment rates to hospitals for vaginal deliveries and C-sections. In most places, C-sections are significantly more expensive than vaginal deliveries, making high C-section rates not only a concern for expectant mothers but also for taxpayers.
Medicaid pays for 60% of all births in Louisiana, according to KFF, and about half of all births in most Southern states, compared with 42% nationally. That’s one reason some states – including Louisiana, Tennessee, and Minnesota – have tried to tackle high C-section rates by changing how much Medicaid pays for them. But payment reform alone isn’t enough, Dr. Gee said.
“There was a guy in central Louisiana who was doing more C-sections and early elective deliveries than anyone in the U.S.,” she said. “When you have a culture like that, it’s hard to shift from it.”
Linda Schwimmer, president and CEO of the New Jersey Health Care Quality Institute, said many hospitals and doctors don’t even know their C-section rates. Sharing this data with doctors and hospitals – and making it public – made some providers uncomfortable, she said, but it ultimately worked. New Jersey’s C-section rate among first-time, low-risk mothers dropped from 33.1% in 2013 to 26.7% 6 years later once the state began sharing these data, among other initiatives.
The New Jersey Health Care Quality Institute and other groups like it around the country focus on reducing a subset of C-sections called “nulliparous, term, singleton, vertex” C-sections, or surgeries on first-time, full-term moms giving birth to a single infant who is positioned head-down in the uterus.
NTSV C-sections are important to track because women who have a C-section during their first pregnancy face a 90% chance of having another in subsequent pregnancies. Across the U.S., the rate for these C-sections was 25.9% in 2020 and 25.6% in 2019.
Elliott Main, MD, a maternal-fetal specialist at Stanford (Calif.) University and the medical director of the California Maternal Quality Care Collaborative, coauthored a paper, published in JAMA last year, that outlines interventions the collaborative took that lowered California’s NTSV C-Section rate from 26.0% in 2014 to 22.8% in 2019. Nationally, the rate was unchanged during that period.
Allowing women to labor for longer stretches of time before resorting to surgery is important, he said.
The cervix must be 10 cm dilated before a woman gives birth. The threshold for “active labor” used to be when the cervix was dilated at least 4 cm. In more recent years, though, the onset of active labor has been changed to 5-6 cm.
“People show up at the hospital too early,” said Toni Hill, president of the Mississippi Midwives Alliance. “If you show up to the hospital at 2-3 centimeters, you can be at 2-3 centimeters for weeks. I don’t even consider that labor.”
Too often, she said, women at an early stage of labor end up being induced and deliver via C-section.
“It’s almost like, at this point, C-sections are being handed out like lollipops,” said LA’Patricia Washington, a doula based in Jackson, Miss. Doulas are trained, nonmedical workers who help parents before, during, and after delivery.
Ms. Washington works with a nonprofit group, the Jackson Safer Childbirth Experience, that pays for doulas to help expectant mothers in the region. Some state Medicaid programs, such as New Jersey’s, reimburse for services by doulas because research shows they can reduce C-section rates. California has been trying to roll out the same benefit for its Medicaid members.
In 2020, when Julia Maeda became pregnant again, she paid out-of-pocket for a doula to attend the birth. The experience of having her son via C-section the previous year had been “emotionally and psychologically traumatic,” Ms. Maeda said.
She told her ob.gyn. that she wanted a VBAC, short for “vaginal birth after cesarean.” But, she said, “he just shook his head and said, ‘That’s not a good idea.’”
She had VBAC anyway. Ms. Maeda credits her doula with making it happen.
“Maybe just her presence relayed to the nursing staff that this was something I was serious about,” Ms. Maeda said. “They want you to have your baby during business hours. And biology doesn’t work that way.”
KHN (Kaiser Health News) is a national newsroom that produces in-depth journalism about health issues. Together with Policy Analysis and Polling, KHN is one of the three major operating programs at KFF (Kaiser Family Foundation). KFF is an endowed nonprofit organization providing information on health issues to the nation.
Overuse of surveillance in bladder cancer, despite guidelines
(NMIBC), a new study concludes.
These cancers are associated with low rates of recurrence, progression, and bladder cancer–specific death, so current clinical practice guidelines recommend against frequent monitoring and testing.
However, the study authors found that patients with a low grade Ta NMIBC diagnosis underwent a median of three cystoscopies per year, and many also received a median of two imagine scans (CT or MRI) as well as 2-3 urine-based tests.
“These data suggest a need for ongoing efforts to limit overuse of treatment and surveillance, which may in turn mitigate associated increases in the costs of care,” write the authors, led by Kelly K. Bree, MD, from the department of urology, University of Texas MD Anderson Cancer Center, Houston. Bladder cancer has the highest lifetime treatment cost of all malignancies, they point out.
The study was published online in JAMA Network Open.
Higher value and more evidence-based
The impact of increased surveillance of this patient cohort has broad implications for patients and the health care system in general, say experts writing in an accompanying editorial.
“It has been well established that workup for NMIBC can have negative consequences for the physical and psychological health of patients,” note Grayden S. Cook, BS, and Jeffrey M. Howard, MD, PhD, both from University of Texas Southwestern Medical Center, Dallas.
“Many of these patients undergo frequent CT imaging of the urinary tract, which carries a high dose of radiation as well as the potential for financial toxic effects (that is, detrimental consequences to the patient because of health care costs),” they write.
Additionally, patient distress is a factor, as they may experience preprocedural anxiety, physical discomfort during procedures, and worry about disease progression, they point out.
“The impact of these patterns is substantial and may have negative consequences for both patients and the health care system,” they conclude. “Thus, it is imperative to move forward with initiatives that provide higher value and are more evidence-based and patient-centered.”
Study finds frequent surveillance
The American Urological Association (AUA)/Society of Urologic Oncologists (SUO), the European Association of Urology, and the International Bladder Cancer Group have made an effort to de-escalate surveillance and treatment for patients with low-grade Ta disease, while at the same time maintaining appropriate surveillance for high-grade aggressive disease.
However, the new study found that in practice, such patients undergo frequent testing.
The study involved 13,054 patients with low-grade Ta NMIBC. Most of the participants were male (73.5%), with a median age of 76 years, and had no or few comorbidities (71.2%).
Most patients had undergone cystoscopy, and rates increased over time: from 79.3% of patients in 2004 to 81.5% of patients in 2013 (P = .007). Patients underwent a median of 3.0 cystoscopies per year following their diagnosis, and upper-tract imaging was performed in most patients.
The use of kidney ultrasonography also rose from 19% of patients in 2004 to 23.2% in 2013, as did retrograde pyelography (20.9% in 2004 vs. 24.2% in 2013). Conversely, the use of intravenous pyelography declined (from 14.5% in 2004 to 1.7% in 2012), but there was an increase in CT and MRI in all years except 2010 (from 30.4% of patients in 2004 to 47% of patients in 2013; P < .001). The rate of urine-based testing also significantly increased during the study period (from 44.8% in 2004 to 54.9% in 2013; P < .001), with patients undergoing between two to three tests per year.
Adherence to current guidelines remained similar during the study time frame. For example, 55.2% of patients received two cystoscopies per year in 2004-2008, compared with 53.8% in 2009-2013 (P = .11), suggesting that there was an overuse of all surveillance testing modalities.
As for treatment, 17.2% received intravesical immunotherapy with bacillus Calmette-Guérin, 6.1% were treated with intravesical chemotherapy (excluding receipt of a single perioperative dose). Disease recurrence within this cohort was 1.7%, and only 0.4% experienced disease progression.
When looking at the cost, the total median expenditures at 1 year after diagnosis increased by 60% during the study period, from $34,792 in 2004 to $53,986 in 2013. Higher costs were seen among patients who experienced a recurrence versus no recurrence ($76,669 vs. $53,909).
The study was supported by a grant from the U.S. Department of Defense Peer Reviewed Cancer Research Program. Several of the authors have disclosed relationships with industry, as noted in the original article. Editorialists Mr. Cook and Dr. Howard have disclosed no relevant financial relationships.
A version of this article first appeared on Medscape.com.
(NMIBC), a new study concludes.
These cancers are associated with low rates of recurrence, progression, and bladder cancer–specific death, so current clinical practice guidelines recommend against frequent monitoring and testing.
However, the study authors found that patients with a low grade Ta NMIBC diagnosis underwent a median of three cystoscopies per year, and many also received a median of two imagine scans (CT or MRI) as well as 2-3 urine-based tests.
“These data suggest a need for ongoing efforts to limit overuse of treatment and surveillance, which may in turn mitigate associated increases in the costs of care,” write the authors, led by Kelly K. Bree, MD, from the department of urology, University of Texas MD Anderson Cancer Center, Houston. Bladder cancer has the highest lifetime treatment cost of all malignancies, they point out.
The study was published online in JAMA Network Open.
Higher value and more evidence-based
The impact of increased surveillance of this patient cohort has broad implications for patients and the health care system in general, say experts writing in an accompanying editorial.
“It has been well established that workup for NMIBC can have negative consequences for the physical and psychological health of patients,” note Grayden S. Cook, BS, and Jeffrey M. Howard, MD, PhD, both from University of Texas Southwestern Medical Center, Dallas.
“Many of these patients undergo frequent CT imaging of the urinary tract, which carries a high dose of radiation as well as the potential for financial toxic effects (that is, detrimental consequences to the patient because of health care costs),” they write.
Additionally, patient distress is a factor, as they may experience preprocedural anxiety, physical discomfort during procedures, and worry about disease progression, they point out.
“The impact of these patterns is substantial and may have negative consequences for both patients and the health care system,” they conclude. “Thus, it is imperative to move forward with initiatives that provide higher value and are more evidence-based and patient-centered.”
Study finds frequent surveillance
The American Urological Association (AUA)/Society of Urologic Oncologists (SUO), the European Association of Urology, and the International Bladder Cancer Group have made an effort to de-escalate surveillance and treatment for patients with low-grade Ta disease, while at the same time maintaining appropriate surveillance for high-grade aggressive disease.
However, the new study found that in practice, such patients undergo frequent testing.
The study involved 13,054 patients with low-grade Ta NMIBC. Most of the participants were male (73.5%), with a median age of 76 years, and had no or few comorbidities (71.2%).
Most patients had undergone cystoscopy, and rates increased over time: from 79.3% of patients in 2004 to 81.5% of patients in 2013 (P = .007). Patients underwent a median of 3.0 cystoscopies per year following their diagnosis, and upper-tract imaging was performed in most patients.
The use of kidney ultrasonography also rose from 19% of patients in 2004 to 23.2% in 2013, as did retrograde pyelography (20.9% in 2004 vs. 24.2% in 2013). Conversely, the use of intravenous pyelography declined (from 14.5% in 2004 to 1.7% in 2012), but there was an increase in CT and MRI in all years except 2010 (from 30.4% of patients in 2004 to 47% of patients in 2013; P < .001). The rate of urine-based testing also significantly increased during the study period (from 44.8% in 2004 to 54.9% in 2013; P < .001), with patients undergoing between two to three tests per year.
Adherence to current guidelines remained similar during the study time frame. For example, 55.2% of patients received two cystoscopies per year in 2004-2008, compared with 53.8% in 2009-2013 (P = .11), suggesting that there was an overuse of all surveillance testing modalities.
As for treatment, 17.2% received intravesical immunotherapy with bacillus Calmette-Guérin, 6.1% were treated with intravesical chemotherapy (excluding receipt of a single perioperative dose). Disease recurrence within this cohort was 1.7%, and only 0.4% experienced disease progression.
When looking at the cost, the total median expenditures at 1 year after diagnosis increased by 60% during the study period, from $34,792 in 2004 to $53,986 in 2013. Higher costs were seen among patients who experienced a recurrence versus no recurrence ($76,669 vs. $53,909).
The study was supported by a grant from the U.S. Department of Defense Peer Reviewed Cancer Research Program. Several of the authors have disclosed relationships with industry, as noted in the original article. Editorialists Mr. Cook and Dr. Howard have disclosed no relevant financial relationships.
A version of this article first appeared on Medscape.com.
(NMIBC), a new study concludes.
These cancers are associated with low rates of recurrence, progression, and bladder cancer–specific death, so current clinical practice guidelines recommend against frequent monitoring and testing.
However, the study authors found that patients with a low grade Ta NMIBC diagnosis underwent a median of three cystoscopies per year, and many also received a median of two imagine scans (CT or MRI) as well as 2-3 urine-based tests.
“These data suggest a need for ongoing efforts to limit overuse of treatment and surveillance, which may in turn mitigate associated increases in the costs of care,” write the authors, led by Kelly K. Bree, MD, from the department of urology, University of Texas MD Anderson Cancer Center, Houston. Bladder cancer has the highest lifetime treatment cost of all malignancies, they point out.
The study was published online in JAMA Network Open.
Higher value and more evidence-based
The impact of increased surveillance of this patient cohort has broad implications for patients and the health care system in general, say experts writing in an accompanying editorial.
“It has been well established that workup for NMIBC can have negative consequences for the physical and psychological health of patients,” note Grayden S. Cook, BS, and Jeffrey M. Howard, MD, PhD, both from University of Texas Southwestern Medical Center, Dallas.
“Many of these patients undergo frequent CT imaging of the urinary tract, which carries a high dose of radiation as well as the potential for financial toxic effects (that is, detrimental consequences to the patient because of health care costs),” they write.
Additionally, patient distress is a factor, as they may experience preprocedural anxiety, physical discomfort during procedures, and worry about disease progression, they point out.
“The impact of these patterns is substantial and may have negative consequences for both patients and the health care system,” they conclude. “Thus, it is imperative to move forward with initiatives that provide higher value and are more evidence-based and patient-centered.”
Study finds frequent surveillance
The American Urological Association (AUA)/Society of Urologic Oncologists (SUO), the European Association of Urology, and the International Bladder Cancer Group have made an effort to de-escalate surveillance and treatment for patients with low-grade Ta disease, while at the same time maintaining appropriate surveillance for high-grade aggressive disease.
However, the new study found that in practice, such patients undergo frequent testing.
The study involved 13,054 patients with low-grade Ta NMIBC. Most of the participants were male (73.5%), with a median age of 76 years, and had no or few comorbidities (71.2%).
Most patients had undergone cystoscopy, and rates increased over time: from 79.3% of patients in 2004 to 81.5% of patients in 2013 (P = .007). Patients underwent a median of 3.0 cystoscopies per year following their diagnosis, and upper-tract imaging was performed in most patients.
The use of kidney ultrasonography also rose from 19% of patients in 2004 to 23.2% in 2013, as did retrograde pyelography (20.9% in 2004 vs. 24.2% in 2013). Conversely, the use of intravenous pyelography declined (from 14.5% in 2004 to 1.7% in 2012), but there was an increase in CT and MRI in all years except 2010 (from 30.4% of patients in 2004 to 47% of patients in 2013; P < .001). The rate of urine-based testing also significantly increased during the study period (from 44.8% in 2004 to 54.9% in 2013; P < .001), with patients undergoing between two to three tests per year.
Adherence to current guidelines remained similar during the study time frame. For example, 55.2% of patients received two cystoscopies per year in 2004-2008, compared with 53.8% in 2009-2013 (P = .11), suggesting that there was an overuse of all surveillance testing modalities.
As for treatment, 17.2% received intravesical immunotherapy with bacillus Calmette-Guérin, 6.1% were treated with intravesical chemotherapy (excluding receipt of a single perioperative dose). Disease recurrence within this cohort was 1.7%, and only 0.4% experienced disease progression.
When looking at the cost, the total median expenditures at 1 year after diagnosis increased by 60% during the study period, from $34,792 in 2004 to $53,986 in 2013. Higher costs were seen among patients who experienced a recurrence versus no recurrence ($76,669 vs. $53,909).
The study was supported by a grant from the U.S. Department of Defense Peer Reviewed Cancer Research Program. Several of the authors have disclosed relationships with industry, as noted in the original article. Editorialists Mr. Cook and Dr. Howard have disclosed no relevant financial relationships.
A version of this article first appeared on Medscape.com.
FROM JAMA NETWORK OPEN