No Harmful Effects of Donor Iron Deficiency on Blood Quality or Donor Wellbeing
Findings of randomized study are good news for national blood supply, validate current donation practices
(Washington, Sept. 7) – Most of the U.S. blood supply comes from about 5 million altruistic volunteers who donate blood regularly. Frequent blood donation is known to result in iron deficiency in donors, which has led to concerns that the practice may not only diminish the quality of donated blood but also harm blood donors’ health and quality of life. Reassuringly, a new study published in Blood found that donor iron deficiency produced no harmful effects on either the quality of donated blood or the wellbeing of frequent blood donors.
Red blood cells require iron, an essential dietary element, to transport oxygen throughout the body. Iron deficiency may lead to anemia (a shortage of red blood cells) and associated symptoms, including fatigue, depression, and brain fog. Although a 2012 study found that roughly 35% of regular blood donors become iron deficient after repeated blood donations, few studies have evaluated the effects on donor health or on the quality of donated blood.
“The results of our study offer robust evidence that the existing criteria for blood donation both preserve the quality of the blood supply and protect the wellbeing of adults who are frequent blood donors,” said Eldad A. Hod, MD, an associate professor of pathology and cell biology at Columbia University’s Irving Medical Center in New York, and the study’s first author.
Dr. Hod added that he hoped the study’s findings encourage more people to become blood donors. According to a 2019 study, less than 6% of people over 18 years old in the United States had given blood within the previous year. An increase in the number of donors would improve the blood supply and reduce reliance on those who donate frequently, he said. The COVID-19 pandemic has exacerbated an existing shortage of donated blood in the United States, leading to delays in planned blood transfusions and cancellations of surgery due to a lack of blood products.
The Donor Iron Deficiency Study screened 983 frequent blood donors ages 18 to 75. Of the 110 iron deficient, but non-anemic, donors, 79 were ultimately randomized. Donors who were too anemic to donate blood, pregnant, or had conditions that put them at risk for anaphylactic shock with iron treatment (e.g., severe allergies) were excluded.
Following screening to confirm their eligibility, all participants donated one standard unit of blood. These donors also completed questionnaires about their physical and mental health and quality of life, and completed tests of their cognitive ability, including memory, attentiveness, and ability to process information.
Participants returned about six weeks later to complete quality of life and cognitive testing again, together with a test to determine whether the blood they had donated on their previous visit met the U.S. Food and Drug Administration’s (FDA’s) quality standard for donated blood. After completing these tests, they were randomly assigned to receive an intravenous infusion of either iron or a placebo substance (saline solution). The study was double blinded, meaning that neither the participants nor members of the study team knew who had received the iron infusion and who had received the placebo.
Four to six months later, participants returned to make a second blood donation and complete another round of quality of life and cognitive testing. Six weeks after this second donation, they undertook a second test to determine whether the quality of this donated blood met the FDA standard and then completed a final assessment of quality of life and cognition. The primary endpoint of the study was the change in participants’ results on the first blood-quality test when all donors had low iron levels, and the second test, performed after the donors had been randomly assigned to receive treatment with intravenous iron or a placebo.
While Dr. Hod and colleagues predicted that donated blood would fail to meet the FDA’s quality standard and that donors’ wellbeing would improve with additional iron, the study’s results defied expectations. Among those who were randomly assigned to receive intravenous iron, the treatment corrected their iron deficiency, while participants who received the placebo remained iron deficient, with many developing anemia. Nevertheless, the donated blood from the iron-deficient donors who received placebo still met the FDA’s quality standards, and their scores on both cognitive function tests and quality-of-life measures showed no significant differences at any time point from those of the donors whose iron deficiency had been corrected with intravenous iron.
“This finding is good news because it not only shows that blood donated by frequent donors remains of high quality, but also that those who donate blood regularly are not being significantly harmed by doing so,” explained Dr. Hod.
One limitation of the study’s findings is that people under age 18 were excluded from participation. About 10% of all blood donors are in this age group, according to Dr. Hod. Adolescents need iron for brain development, a process that isn’t complete until they reach their mid-20s, he said. “We don’t know to what extent our results can be generalized to the under-18 population,” he said. “We hope to look at this age group in our next study.”
The study was funded by the National Heart, Lung, and Blood Institute.
Blood (www.bloodjournal.org), the most cited peer-reviewed publication in the field of hematology, is available weekly in print and online. Blood is a journal of the American Society of Hematology (ASH) (www.hematology.org).
Kira Sampson, American Society of Hematology
[email protected]; 202-499-1796