Radiation hormesis is the hypothesis that low doses of ionizing radiation (within the region of and just above natural background levels) are beneficial, stimulating the activation of repair mechanisms that protect against disease, that are not activated in absence of ionizing radiation. The reserve repair mechanisms are hypothesized to be sufficiently effective when stimulated as to not only cancel the detrimental effects of ionizing radiation but also inhibit disease not related to radiation exposure (see hormesis).[1] [2] [3] [4] It has been a mainstream concept since at least 2009.[5]
While the effects of high and acute doses of ionising radiation are easily observed and understood in humans (e.g. Japanese atomic bomb survivors), the effects of low-level radiation are very difficult to observe and highly controversial. This is because the baseline cancer rate is already very high and the risk of developing cancer fluctuates 40% because of individual life style and environmental effects,[6] [7] obscuring the subtle effects of low-level radiation. An acute effective dose of 100 millisieverts may increase cancer risk by ~0.8%. However, children are particularly sensitive to radioactivity, with childhood leukemias and other cancers increasing even within natural and man-made background radiation levels (under 4 mSv cumulative with 1 mSv being an average annual dose from terrestrial and cosmic radiation, excluding radon which primarily doses the lung).[8] [9] There is limited evidence that exposures around this dose level will cause negative subclinical health impacts to neural development.[10] Students born in regions of higher Chernobyl fallout performed worse in secondary school, particularly in mathematics. "Damage is accentuated within families (i.e., siblings comparison) and among children born to parents with low education..." who often don't have the resources to overcome this additional health challenge.[11]
Hormesis remains largely unknown to the public. Government and regulatory bodies disagree on the existence of radiation hormesis and research points to the "severe problems and limitations" with the use of hormesis in general as the "principal dose-response default assumption in a risk assessment process charged withensuring public health protection."[12]
Quoting results from a literature database research, the Académie des Sciences – Académie nationale de Médecine (French Academy of Sciences – National Academy of Medicine) stated in their 2005 report concerning the effects of low-level radiation that many laboratory studies have observed radiation hormesis.[13] [14] However, they cautioned that it is not yet known if radiation hormesis occurs outside the laboratory, or in humans.[15]
Reports by the United States National Research Council and the National Council on Radiation Protection and Measurements and the United Nations Scientific Committee on the Effects of Atomic Radiation (UNSCEAR) argue[16] that there is no evidence for hormesis in humans and in the case of the National Research Council hormesis is outright rejected as a possibility. Therefore, estimating linear no-threshold model (LNT) continues to be the model generally used by regulatory agencies for human radiation exposure.
Radiation hormesis proposes that radiation exposure comparable to and just above the natural background level of radiation is not harmful but beneficial, while accepting that much higher levels of radiation are hazardous. Proponents of radiation hormesis typically claim that radio-protective responses in cells and the immune system not only counter the harmful effects of radiation but additionally act to inhibit spontaneous cancer not related to radiation exposure. Radiation hormesis stands in stark contrast to the more generally accepted linear no-threshold model (LNT), which states that the radiation dose-risk relationship is linear across all doses, so that small doses are still damaging, albeit less so than higher ones. Opinion pieces on chemical and radiobiological hormesis appeared in the journals Nature[1] and Science[3] in 2003.
Assessing the risk of radiation at low doses (<100 mSv) and low dose rates (<0.1 mSv.min−1) is highly problematic and controversial.[17] [18] While epidemiological studies on populations of people exposed to an acute dose of high level radiation such as Japanese atomic bomb survivors have robustly upheld the LNT (mean dose ~210 mSv),[19] studies involving low doses and low dose rates have failed to detect any increased cancer rate.[18] This is because the baseline cancer rate is already very high (~42 of 100 people will be diagnosed in their lifetime) and it fluctuates ~40% because of lifestyle and environmental effects,[7] [20] obscuring the subtle effects of low level radiation. Epidemiological studies may be capable of detecting elevated cancer rates as low as 1.2 to 1.3 i.e. 20% to 30% increase. But for low doses (1–100 mSv) the predicted elevated risks are only 1.001 to 1.04 and excess cancer cases, if present, cannot be detected due to confounding factors, errors and biases.[20] [21] [22]
In particular, variations in smoking prevalence or even accuracy in reporting smoking cause wide variation in excess cancer and measurement error bias. Thus, even a large study of many thousands of subjects with imperfect smoking prevalence information will fail to detect the effects of low level radiation than a smaller study that properly compensates for smoking prevalence.[23] Given the absence of direct epidemiological evidence, there is considerable debate as to whether the dose-response relationship <100 mSv is supralinear, linear (LNT), has a threshold, is sub-linear, or whether the coefficient is negative with a sign change, i.e. a hormetic response.
The radiation adaptive response seems to be a main origin of the potential hormetic effect. The theoretical studies indicate that the adaptive response is responsible for the shape of dose-response curve and can transform the linear relationship (LNT) into the hormetic one.[24] [25]
While most major consensus reports and government bodies currently adhere to LNT,[26] the 2005 French Academy of Sciences-National Academy of Medicine's report concerning the effects of low-level radiation rejected LNT as a scientific model of carcinogenic risk at low doses.[15]
Using LNT to estimate the carcinogenic effect at doses of less than 20 mSv is not justified in the light of current radiobiologic knowledge.
They consider there to be several dose-effect relationships rather than only one, and that these relationships have many variables such as target tissue, radiation dose, dose rate and individual sensitivity factors. They request that further study is required on low doses (less than 100 mSv) and very low doses (less than 10 mSv) as well as the impact of tissue type and age. The Academy considers the LNT model is only useful for regulatory purposes as it simplifies the administrative task. Quoting results from literature research,[13] [14] they furthermore claim that approximately 40% of laboratory studies on cell cultures and animals indicate some degree of chemical or radiobiological hormesis, and state:
...its existence in the laboratory is beyond question and its mechanism of action appears well understood.
They go on to outline a growing body of research that illustrates that the human body is not a passive accumulator of radiation damage but it actively repairs the damage caused via a number of different processes, including:[15] [18]
Furthermore, increased sensitivity to radiation induced cancer in the inherited condition Ataxia-telangiectasia like disorder, illustrates the damaging effects of loss of the repair gene Mre11h resulting in the inability to fix DNA double-strand breaks.[27]
The BEIR-VII report argued that, "the presence of a true dose threshold demands totally error-free DNA damage response and repair." The specific damage they worry about is double strand breaks (DSBs) and they continue, "error-prone nonhomologous end joining (NHEJ) repair in postirradiation cellular response, argues strongly against a DNA repair-mediated low-dose threshold for cancer initiation". Recent research observed that DSBs caused by CAT scans are repaired within 24 hours and DSBs may be more efficiently repaired at low doses, suggesting that the risk of ionizing radiation at low doses may not be directly proportional to the dose.[28] [29] However, it is not known if low-dose ionizing radiation stimulates the repair of DSBs not caused by ionizing radiation i.e. a hormetic response. Radon gas in homes is the largest source of radiation dose for most individuals and it is generally advised that the concentration be kept below 150 Bq/m³ (4 pCi/L).[30] A recent retrospective case-control study of lung cancer risk showed substantial cancer rate reduction between 50 and 123 Bq per cubic meter relative to a group at zero to 25 Bq per cubic meter.[31] This study is cited as evidence for hormesis, but a single study all by itself cannot be regarded as definitive. Other studies into the effects of domestic radon exposure have not reported a hormetic effect; including for example the respected "Iowa Radon Lung Cancer Study" of Field et al. (2000), which also used sophisticated radon exposure dosimetry.[32] In addition, Darby et al. (2005) argue that radon exposure is negatively correlated with the tendency to smoke and environmental studies need to accurately control for this; people living in urban areas where smoking rates are higher usually have lower levels of radon exposure due to the increased prevalence of multi-story dwellings.[33] When doing so, they found a significant increase in lung cancer amongst smokers exposed to radon at doses as low as 100 to 199 Bq m−3 and warned that smoking greatly increases the risk posed by radon exposure i.e. reducing the prevalence of smoking would decrease deaths caused by radon.[33] [34] However, the discussion about the opposite experimental results is still going on,[35] especially the popular US and German studies have found some hormetic effects.[36] [37]
Furthermore, particle microbeam studies show that passage of even a single alpha particle (e.g. from radon and its progeny) through cell nuclei is highly mutagenic,[38] and that alpha radiation may have a higher mutagenic effect at low doses (even if a small fraction of cells are hit by alpha particles) than predicted by linear no-threshold model, a phenomenon attributed to bystander effect.[39] However, there is currently insufficient evidence at hand to suggest that the bystander effect promotes carcinogenesis in humans at low doses.[40]
Radiation hormesis has not been accepted by either the United States National Research Council, or the National Council on Radiation Protection and Measurements (NCRP).[41] In May 2018, the NCRP published the report of an interdisciplinary group of radiation experts who critically reviewed 29 high-quality epidemiologic studies of populations exposed to radiation in the low dose and low dose-rate range, mostly published within the last 10 years. The group of experts concluded:
The recent epidemiologic studies support the continued use of the LNT model for radiation protection. This is in accord with judgments by other national and international scientific committees, based on somewhat older data, that no alternative dose-response relationship appears more pragmatic or prudent for radiation protection purposes than the LNT model.
In addition, the United Nations Scientific Committee on the Effects of Atomic Radiation (UNSCEAR) wrote in its 2000 report:[42]
Until the [...] uncertainties on low-dose response are resolved, the Committee believes that an increase in the risk of tumour induction proportionate to the radiation dose is consistent with developing knowledge and that it remains, accordingly, the most scientifically defensible approximation of low-dose response. However, a strictly linear dose response should not be expected in all circumstances.
This is a reference to the fact that very low doses of radiation have only marginal impacts on individual health outcomes. It is therefore difficult to detect the 'signal' of decreased or increased morbidity and mortality due to low-level radiation exposure in the 'noise' of other effects. The notion of radiation hormesis has been rejected by the National Research Council's (part of the National Academy of Sciences) 16-year-long study on the Biological Effects of Ionizing Radiation. "The scientific research base shows that there is no threshold of exposure below which low levels of ionizing radiation can be demonstrated to be harmless or beneficial. The health risks – particularly the development of solid cancers in organs – rise proportionally with exposure" says Richard R. Monson, associate dean for professional education and professor of epidemiology, Harvard School of Public Health, Boston.[43]
Kerala's monazite sand (containing a third of the world's economically recoverable reserves of radioactive thorium) emits about 8 microsieverts per hour of gamma radiation, 80 times the dose rate equivalent in London, but a decade-long study of 69,985 residents published in Health Physics in 2009 "showed no excess cancer risk from exposure to terrestrial gamma radiation. The excess relative risk of cancer excluding leukemia was estimated to be −0.13 per Gy (95% CI: −0.58, 0.46)", indicating no statistically significant positive or negative relationship between background radiation levels and cancer risk in this sample.[44]
Studies in cell cultures can be useful for finding mechanisms for biological processes, but they also can be criticized for not effectively capturing the whole of the living organism.
A study by E. I. Azzam suggested that pre-exposure to radiation causes cells to turn on protection mechanisms.[45] A different study by de Toledo and collaborators has shown that irradiation with gamma rays increases the concentration of glutathione, an antioxidant found in cells.[46]
In 2011, an in vitro study led by S. V. Costes showed in time-lapse images a strongly non-linear response of certain cellular repair mechanisms called radiation-induced foci (RIF). The study found that low doses of radiation prompted higher rates of RIF formation than high doses, and that after low-dose exposure RIF continued to form after the radiation had ended. Measured rates of RIF formation were 15 RIF/Gy at 2 Gy, and 64 RIF/Gy at 0.1 Gy. These results suggest that low dose levels of ionizing radiation may not increase cancer risk directly proportional to dose and thus contradict the linear-no-threshold standard model.[47] Mina Bissell, a world-renowned breast-cancer researcher and collaborator in this study stated: "Our data show that at lower doses of ionizing radiation, DNA repair mechanisms work much better than at higher doses. This non-linear DNA damage response casts doubt on the general assumption that any amount of ionizing radiation is harmful and additive."
An early study on mice exposed to low dose of radiation daily (0.11 R per day) suggest that they may outlive control animals.[48] A study by Otsuka and collaborators found hormesis in animals.[49] Miyachi conducted a study on mice and found that a 200 mGy X-ray dose protects mice against both further X-ray exposure and ozone gas.[50] In another rodent study, Sakai and collaborators found that (1 mGy/h) gamma irradiation prevents the development of cancer (induced by chemical means, injection of methylcholanthrene).[51]
In a 2006 paper,[52] a dose of 1 Gy was delivered to the cells (at constant rate from a radioactive source) over a series of lengths of time. These were between 8.77 and 87.7 hours, the abstract states for a dose delivered over 35 hours or more (low dose rate) no transformation of the cells occurred. Also for the 1 Gy dose delivered over 8.77 to 18.3 hours that the biological effect (neoplastic transformation) was about "1.5 times less than that measured at high dose rate in previous studies with a similar quality of [X-ray] radiation". Likewise it has been reported that fractionation of gamma irradiation reduces the likelihood of a neoplastic transformation.[53] Pre-exposure to fast neutrons and gamma rays from Cs-137 is reported to increase the ability of a second dose to induce a neoplastic transformation.[54]
Caution must be used in interpreting these results, as it noted in the BEIR VII report, these pre-doses can also increase cancer risk:
However, 75 mGy/d cannot be accurately described as a low dose rate – it is equivalent to over 27 sieverts per year. The same study on dogs showed no increase in cancer nor reduction in life expectancy for dogs irradiated at 3 mGy/d.[55]
In long-term study of Chernobyl disaster liquidators[56] was found that: "During current research paradoxically longer telomeres were found among persons, who have received heavier long-term irradiation." and "Mortality due to oncologic diseases was lower than in general population in all age groups that may reflect efficient health care of this group." Though in conclusion interim results were ignored and conclusion followed LNT hypothesis: "The signs of premature aging were found in Chernobyl disaster clean-up workers; moreover, aging process developed in heavier form and at younger age in humans, who underwent greater exposure to ionizing radiation."
In an Australian study which analyzed the association between solar UV exposure and DNA damage, the results indicated that although the frequency of cells with chromosome breakage increased with increasing sun exposure, the misrepair of DNA strand breaks decreased as sun exposure was heightened.[57]
The health of the inhabitants of radioactive apartment buildings in Taiwan has received prominent attention. In 1982, more than 20,000 tons of steel was accidentally contaminated with cobalt-60, and much of this radioactive steel was used to build apartments and exposed thousands of Taiwanese to gamma radiation levels of up to >1000 times background (average 47.7 mSv, maximum 2360 mSv excess cumulative dose). The radioactive contamination was discovered in 1992.
A seriously flawed 2004 study compared the building's younger residents with the much older general population of Taiwan and determined that the younger residents were less likely to have been diagnosed with cancer than older people; this was touted as evidence of a radiation hormesis effect.[58] [59] (Older people have much higher cancer rates even in the absence of excess radiation exposure.)
In the years shortly after exposure, the total number cancer cases have been reported to be either lower than the society-wide average or slightly elevated.[60] [61] Leukaemia and thyroid cancer were substantially elevated. When a lower rate of "all cancers" was found, it was thought to be due to the exposed residents having a higher socioeconomic status, and thus overall healthier lifestyle. Additionally, Hwang, et al. cautioned in 2006 that leukaemia was the first cancer type found to be elevated amongst the survivors of the Hiroshima and Nagasaki bombings, so it could be decades before any increase in more common cancer types is seen.
Besides the excess risks of leukaemia and thyroid cancer, a later publication notes various DNA anomalies and other health effects among the exposed population:[62]
There have been several reports concerning the radiation effects on the exposed population, including cytogenetic analysis that showed increased micronucleus frequencies in peripheral lymphocytes in the exposed population, increases in acentromeric and single or multiple centromeric cytogenetic damages, and higher frequencies of chromosomal translocations, rings and dicentrics. Other analyses have shown persistent depression of peripheral leucocytes and neutrophils, increased eosinophils, altered distributions of lymphocyte subpopulations, increased frequencies of lens opacities, delays in physical development among exposed children, increased risk of thyroid abnormalities, and late consequences in hematopoietic adaptation in children.People living in these buildings also experienced infertility.[63]
Intentional exposure to water and air containing increased amounts of radon is perceived as therapeutic, and "radon spas" can be found in United States, Czechia, Poland, Germany, Austria and other countries.
Given the uncertain effects of low-level and very-low-level radiation, there is a pressing need for quality research in this area. An expert panel convened at the 2006 Ultra-Low-Level Radiation Effects Summit at Carlsbad, New Mexico, proposed the construction of an Ultra-Low-Level Radiation laboratory.[64] The laboratory, if built, will investigate the effects of almost no radiation on laboratory animals and cell cultures, and it will compare these groups to control groups exposed to natural radiation levels. Precautions would be made, for example, to remove potassium-40 from the food of laboratory animals. The expert panel believes that the Ultra-Low-Level Radiation laboratory is the only experiment that can explore with authority and confidence the effects of low-level radiation; that it can confirm or discard the various radiobiological effects proposed at low radiation levels e.g. LNT, threshold and radiation hormesis.[65]
The first preliminary results of the effects of almost no-radiation on cell cultures was reported by two research groups in 2011 and 2012; researchers in the US studied cell cultures protected from radiation in a steel chamber 650 meters underground at the Waste Isolation Pilot Plant in Carlsbad, New Mexico[66] and researchers in Europe proposed an experiment design to study the effects of almost no-radiation on mouse cells (pKZ1 transgenic chromosomal inversion assay), but did not carry out the experiment.[67]