A vast number of communication networks interconnect societies worldwide, and cellular wireless technology networks make up an increasing fraction of this number. The presence of radiofrequency (RF) waves from wireless technologies has become ubiquitous. Mobile telephony (construction and operation of telephones or telephonic systems) is relied on by > 1.4 billion people, or around 20% of the world's population. Given that the public is frequently reminded that we are all surrounded by ever-present electromagnetic fields (EMFs), which some call "electro-smog," it is not surprising that some individuals and groups express concern about possible health effects from low-level, chronic exposure to a variety of RF sources. To help address this concern, the World Health Organization (WHO) convened a "Workshop on Base Stations and Wireless Networks" as part of its "International EMF Project" to discuss the state of the science in RF health effects. In this article we provide a summary of several key points addressed at the workshop.
RF waves have long been used for different types of wireless broadcast, such as for wireless Morse code, radio, television, and so on. The radio-wave spectrum spans the frequency range from about 0.5 MHz in the AM radio band up to about 30,000 MHz in the radar band. RF-emitting devices have become commonplace in homes, offices, and schools. Table 1 lists examples of RF sources contributing to the radio-wave background in almost every modern-day location. The actual RF level from each source depends on the details of the exposure location (i.e., distance from the antenna), but the last column lists whether the source is more ubiquitous and universal (U) or more limited and local (L); a "+" indicates that this source would typically contribute a significant fraction of ambient RF background levels. Even for individuals in the vicinity of transmitting antennas, surveys of RF levels report results that are far below the applicable exposure guidelines both in the United States (Burch et al. 2006; Tell and Mantiply 1980) and in Europe (Foster K, in press).
In addition to the increasing prevalence of cellular telephones, there has been continuing expansion of wireless Internet access, such as WiFi, into homes, schools, workplaces, and public areas. "WiFi" is an abbreviation for "Wireless Fidelity," and is used generically when referring to any type of wireless technology that supports local, over-the-air computer communication via a wireless local-area network (WLAN). Typically, the transmission frequency is approximately 2.4 GHz, and WiFi provides data-transmission rates in the range of 1–50 Mbps (megabytes per second). "WiMAX" is a long-range version of WiFi. Cellular wireless technology is now capable of delivering voice, text, images, music, and other data to consumers everywhere, and it relies on an extensive network of fixed antennas, or base stations, for relaying information using RF signals. The number of base stations required increases with greater mobile phone use (requiring extensive micro-cell or pico-cell distributed antenna systems in urban areas), with market competition (enabling more operators to provide services), and with new technological capabilities (e.g., 3G). 3G (or 3-G) is short for "third-generation" mobile telephone technology. The services associated with 3G provide the ability to transfer both voice data (a telephone call) and nonvoice data (e.g., downloading information, exchanging email, and instant messaging).
Table 1.
|
Table 2.
|
Table 3.
|
Table 4.
|
Table 5.
|
Table 6.
|
Figure 1. The causal chain leading from an exposure to disease has multiple steps, each of which may or may not trigger the next step. For RF interactions with molecules, cell structures, or tissues, the transduction mechanism is a crucial first link in the causal chain. By definition, the electric and magnetic fields in RF waves can exert force on electrically charged particles.
|
The public, regulators, and scientists have questioned whether there are possible health consequences of this mushrooming mobile phone technology, particularly because the handset operates in close proximity to the human body and because large numbers of base station antennas are required. Although the RF levels produced by base stations at consumer locations are much lower than those from use of the phone handset, the more continuous exposure from base stations has produced a greater public concern.
The total electromagnetic energy available, in terms of effective radiated power from an RF source (or antenna), varies widely according to source type, as shown in Table 2. The visible-light example (light bulb) is provided for comparison, but its energy output is primarily in the infrared and visible portion of the electromagnetic spectrum. Among RF sources, cellular telephone base stations are at the low end when considering the strength of the source of RF power.
Radiofrequency exposure is typically quantified as RF energy flux per unit area, for example, watts of RF energy crossing a square meter of area (W/m2). Alternatively, the intensity of radiowaves can be given in terms of electric field intensity, where the units are volts per meter (V/m). These two metrics are mathematically related to each other when considering locations many wavelengths distant from the antenna (or the RF source). That is, the energy flux per unit area (S) is proportional to the square of the electric field intensity (E):
S (W/m2) = [E(V/m)]2/[377(V2/W)]. [1]
For example, RF energy of 1 W/m2 is equal to 19.4 V/m, and 10 W/m2 is equal to 61.4 V/m (because of the squared dependence between S and E).
The relevant RF energy flux (in terms of potential health impacts) is at exposure points where people may intercept the RF energy, and is measured in power per square meter of surface area. A comparison of energy fluxes in this regard is given in Table 3, which compares both RF and non-RF sources. It can be seen that more energetic electromagnetic waves (visible light, infrared waves) are normally present at energy flux levels more intense than the maximum allowable RF intensities in the cell telephone band. In fact, our body surfaces radiate sufficient infrared energy that they are easily seen by "night vision" cameras. Because of their warm temperature, our bodies also emit RF energy in the microwave band (~ 30–300 GHz) at about 0.003 W/m2.
Table 3 also illustrates that the amount of electromagnetic energy that is present due to cellular telephones and cellular base stations is quite small in comparison to both electromagnetic energy sources generally and RF sources in particular.
Within the home and office environment, a variety of other sources of RF energy are used. Table 4 lists frequency ranges and maximum output powers from typical device classes in home and office environments (Kühn et al. 2006). The peak output power represents the maximum peak output of the investigated device classes. The International Commission on Non-Ionizing Radiation Protection (ICNIRP 1998) identified allowable public exposure levels for electric field (E-field) over these frequencies ranging from approximately 30 V/m to 60 V/m.
Spot measurement data often show that, where a particular RF source is the focus of concern, other, less visually obvious sources may give greater contributions to exposure. The data also show that exposures vary greatly, even at similar distances from base stations, illustrating the dramatic effect of the local environment on RF signals through physical processes such as reflection, diffraction, and mutual interference of signal elements traveling through multiple, different paths (Ardoino et al. 2004; Mann et al. 2006).
For assessing occupation RF exposure in the context of base station antennas, a "compliance boundary" can be used, defined so that for personnel outside the boundary, RF levels are low enough to be in compliance with relevant safety standards. The size and shape of a given compliance boundary varies with frequency, with type of antenna, and with antenna power output. For a typical base-station antenna running at 25 W, the compliance boundary has the shape of a cylinder with a diameter of 3 m, and a height corresponding to the antenna height plus about 0.5 m. The height is centered on the antenna, and the cylinder wall begins about 0.1 m behind the radiative front of the antenna and extends to about 2.9 m in front of the antenna. This virtual space encloses the volume where the RF signal may be in excess of ICNIRP occupational standards, but for all distances outside this cylinder, RF levels are low enough to be considered safe (Mild et al. 2006). By comparison, in occupations such as "plastic sealers," RF levels can be considerably higher than in the close vicinity of base antennas (Wilen et al. 2004).
Both measurement surveys and theoretical predictions show that RF levels from base stations and wireless technologies generally decrease with distance from the device (with focused antenna arrays, maximum ground level RF is 50–300 m from the antenna base). That is, the greater the distance from the antenna, the lower the RF. Under conditions typical for public exposure to base stations and for wireless consumer devices, the RF energy fluxes are > 100-fold below international RF guidelines for public locations. However, in very close proximity to base-station antenna elements under occupational conditions (i.e., when performing maintenance on an operating antenna), or immediately adjacent to wireless local area network (LAN) and Bluetooth transmitters, there is the possibility that RF absorption limits for the general public may be exceeded (Kühn et al. 2006). Thus, there is a need to ensure that, under normal operating conditions, these devices comply with the international limits. In the case of nonoccupational exposure to RF from base stations, the most common circumstance is that the contribution of base stations to a person's total RF exposure is minimal. ("Bluetooth" is a term generally designating digital wireless communication among personal-computer-associated devices—i.e., "digital enhanced cordless telecommunication" between laptops, personal computers, personal digital assistants, cell phones, printers, digital cameras, etc. The name "Bluetooth" refers to the 10th-century king of Denmark, King Harold Bluetooth, whose diplomacy led warring parties to negotiate with each other. The inventors of the Bluetooth technology thought this a fitting name for a technology that allowed different devices to talk to each other.)
Cellular telephone radio waves transmit information that is encoded into electromagnetic waves by means of "modulation," which refers to the patterns of change in the frequency and/or amplitude of the RF carrier wave. As cellular telephone technology has advanced, the modulation patterns have become increasingly complex, and the question arises as to whether a high-frequency modulated RF wave might have greater potential for health effects than a pure sinusoidal RF wave. The applicability of fundamental physics to all systems, and particularly to biology, permits one to conclude that modulation is unlikely to lead to unexpected RF interactions with ions, molecules, cells, and organisms—i.e., interactions substantially different from unmodulated RF (Valberg 2006).
Modulation introduces a spread of frequencies into the RF signal, but the frequency bandwidth of the net RF signal generally remains a small fraction of the central, carrier frequency. This means that the most representative frequency range for modulated electromagnetic waves is that of the (high-frequency) RF carrier, not the (low-frequency) modulation pattern. Even though the power of the RF signal may vary in step with the modulation frequency, the transmitted RF spectrum contains no electromagnetic waves at the modulation frequency. Characteristics of the bandwidth, carrier wave, and modulation depth for some typical RF sources are summarized in Table 5.
As Table 5 illustrates, parameters of potential biological significance include the frequency content of the signal (ratio of modulation frequency to carrier wave frequency), the ratio of peak-to-average RF wave amplitude, the central frequency of the RF (carrier wave), and the modulation frequency (typically ~ 0–10 kHz).
Tumorigenicity studies in laboratory animals provide some insight as to the biological effects of RF modulation and of RF exposure generally. Elder recently reviewed 36 publications that reported tumorigenicity assays in rodent species, after RF exposure in the frequency range applicable to mobile telephony {Elder JA, personal communication; [these studies are in the WHO EMF Database (WHO 2006a)]}. Table 6 summarizes the results of Elder's compilation of the animal tumorigenicity literature, grouped by type of RF modulation and with each result at a different frequency, different modulation, or different power level counted as a separate "test," resulting in 68 separate tests in the 36 publications. The species were primarily mouse and rat, and the RF frequencies ranged from 435 MHz to 9,400 MHz. Table 6 also lists the modulation types tested by the investigators.
Table 6 reveals a preponderance of null results, and Elder observed that the more recent, better-designed studies were overall negative. For the seven positive outcomes, the authors of the studies were not able to conclude that any given positive result fully met criteria of validity such as dose response, consistency, and reproducibility. It should be noted that, when testing at a p < 0.05 significance level, for 68 tests, about three to four positive results would be expected by chance alone. These data do not tend to support the idea that modulated RF is more potent than non-modulated RF, because three results reaching p < 0.05 appeared in 15 tests of nonmodulated RF (CW only), i.e., 20%, whereas the remaining 4 results reaching p < 0.05 appeared in 63 tests of modulated RF, i.e., 6%. Overall, the weight of evidence in animals exposed for extended periods, up to lifetime exposures of 2 years, at a variety of frequencies and modulations, suggests that exposure to modulated RF does not increase risk of tumor development (Elder JA, personal communication; Dasenbrock 2005).
Mechanistic considerations are central to examination of the role of modulation in biological effects of RF because living organisms rely on the same physical laws that govern all systems (Durney and Christensen 2000). Physics forms the basis of chemistry, which forms the basis of biology, which forms the basis of medicine. Hence, even though moving up this progression is marked by an increase in complexity, each successive layer must obey the fundamental laws found to be valid for the layer below. The most fundamental level rests on the laws of physics, which have been exhaustively validated by experimentation and through internal consistency. The principles behind radiofrequency waves—namely, Maxwell's laws of electromagnetism—are accepted to be invariant in time and space, and their accuracy in describing the interactions between electromagnetic fields and matter underlies the functioning of virtually all technology. No exceptions have been found despite constant challenges. Likewise, physics has been found to be valid in complex systems, encompassing chemistry, biology, technology, and medicine (Polk and Postow 1996). Simple conservation laws (e.g., energy, motion, charge, momentum) are universally applicable, and biology is no exception.
Any biological interaction mechanisms capable of detecting the difference between a modulated RF signal and an nonmodulated RF signal must be either be a) fast enough to respond to and detect changes in the central RF frequency, or b) sensitive to the RF power changes occurring at the modulation frequency. For a), scientists have not been able to identify biological structures capable of the necessary high-frequency RF tuning or bandwidth discrimination. For b), being sensitive to the power changes in the signal would require a biological structure that is nonlinear at low power levels (i.e., that can "rectify" the RF), which has not as yet been identified by biologists or anatomists. If "rectification" by biological structures were to occur, it remains difficult to envision how a tiny amount of modulated RF power, occurring at "nonthermal" levels below existing exposure RF standards, would lead to deleterious effects on biological systems. Living systems have considerable thermal output, overall thermal inertia, and efficient thermal regulation. Those nonthermal mechanisms that rely on nonlinear responses (e.g., breakdown of the cell membrane) have high RF–electric-field thresholds. RF levels capable of electroporation would in themselves produce hazardous tissue heating (Adair et al. 1997).
For RF energy to change physiological function, initiate dysfunction, or cause the onset of disease in humans or animals there must exist a mechanism by which the physical forces exerted by electric and magnetic fields on charged particles alter molecules, chemical reactions, cell membranes, or biological structures (Parkinson 1985). RF is a physical not a chemical agent, and the biological plausibility of initiating a process that leads to adverse health effects must be assessed with this in mind. The initial physical step is illustrated in the following causal chain by which RF interaction effects could occur:
RF ⇒ Matter (physics) ⇒
Molecules (chemistry) ⇒
Organisms (biology) ⇒ Disease
This process is illustrated in more detail in Figure 1. Biological processes in living organisms include many interactions among electric charges (on ions, molecules, proteins, and membranes). Hence, it is clearly possible that exposure to RF, the electromagnetic fields of which can exert forces on fixed and moving charges, might have the potential to modulate biological function. For RF to cause or exacerbate disease in humans, the RF electric fields would have to trigger an initial transduction step, and then also begin a cascade of sequential steps that leads to a disease outcome. As Figure 1 illustrates, the causal chain would begin with human exposure to RF. To complete the first step, RF would interact with biological molecules (or structures) in such a way as to alter their size, shape, charge, chemical state, or energy. In this initial "transduction" step, some absorption of RF energy must occur or there can be no effect.
Figure 1 shows that, for observable biological (and possibly health adverse) effects to follow transduction, sequential events at the molecular, cellular, and tissue level are required. Each step can lead to a variety of possible end points, and only certain outcomes lead farther down the causal chain. The outcome "Progress toward disease" in the upper right corner (Figure 1) is only one of many possible outcomes, and it requires specific triggering of intermediate steps. There are multiple points in the causal chain where the signal produced by a weak preceding step might be within normal variations and, therefore, may well have no further functional consequences beyond this point in the causal chain.
For low levels of RF (nonthermal levels) the first, mechanistic or transduction step in the multistep pathway has eluded intensive scientific search (Veyret 2006). Even the well-known "hearing effect" for pulsed high-level microwaves (Aran et al. 2004; Elder and Chou 2003) is based on small thermal fluctuations, even though the time-averaged thermal input is nil. Because the health and viability of the human body depends in a fundamental way on the normal structure and function of large molecules (e.g., proteins, nucleic acids, carbohydrates, and lipids), any proposed RF mechanism must predict how RF could interfere with or modify the normal synthesis, function, or degradation of these molecules. The RF interaction mechanism(s) would then predict thresholds of exposure effectiveness in terms amplitudes, frequencies, time of onset, intermittency of exposure, homogeneity/heterogeneity of amplitudes and frequencies, exposure duration, transients, polarization, and so on. Moreover, mechanistic understanding would address the possibility that children may react to RF more effectively than adults. However, our current understanding does not bear out this hypothesis (Christ and Kuster 2005).
The magnitudes of endogenous forces that are known to act at the cellular level to modify protein structures have been measured and can serve as a basis for comparison to forces produced by RF fields (Titushkin and Cho 2006; Valberg 2006). Thus, the mechanistic constraints on having observable, force-generated effects on biological systems can be appreciated by ranking electric field forces from RF signals relative to forces normally generated (or sensed) by molecules in living systems. Using the electric field strength in tissue corresponding to a specific absorption rate (SAR) of 2 W/kg, which is the maximum level allowed for cell telephones (averaged over 10 g of tissue), we can estimate the associated internal electric field to be ~ 45 V/m at 1 GHz. In evaluating possible bioeffects of this electric field, a useful unit of force is the piconewton, which is one-trillionth (10–12) of a newton (N) of force. The weight of 1 cm3 of water (mass = 1 g) is ~ 0.01 N. The mass of a human cell, approximately 10 µm in size would be 10–9 g, and would weigh (in air) approximately 10–11 N, or 10 pN. Considering a protein molecule that has 100 unbalanced electric charges (+ or –), we can calculate the maximum force of 45 V/m on this molecule to be ~ 0.0007 piconewtons (pN). This is > 10,000-fold smaller than the smallest forces known to modify protein molecule function. For example, the electric force on a 100-fold-charged protein molecule embedded in the cell membrane would be about 160 pN, because of the electric field from the normal resting cell-membrane potential (~ 70 mV over 7 nm). That is, typical cell membrane voltages result in robust electric-field forces on cell-membrane proteins, which are known to modify function, such as open and close ion channels. Forces 10,000 times smaller can be expected to be without effect.
Examination of the magnitude of the possible biophysical interactions (thermal, photon, force) of electromagnetic fields with living matter shows that, under modulated RF exposure conditions allowed by the current safety limits, there does not appear to be an overlooked hazard specific to RF modulation, with the possible exception of RF in the form of very short, high-intensity pulses, which are far more energetic than any pulses encountered in cellular telephone technology (Adair 2003; Challis 2005; Foster and Repacholi 2004; National Council for Radiation Protection 2003). For RF levels below the established standards (modulated or not), scientific research has not identified reproducible and plausible mechanisms by which biological effects can be caused in living systems. Because of the safety factors built into guideline levels, even RF somewhat exceeding permissible levels would yield amounts of thermal energy absorbed that are within the adaptive capacity of the body, and would not be likely to lead to disease (Ebert et al. 2005).
Consideration of the possible mechanisms by which mobile-telephony RF can interact with the body does not reveal a means by which modulated RF, specifically, could lead to adverse effects.
Cancer. Most cellular studies indicate an absence of effects on DNA, and generally biological data, including results from animal studies, do not demonstrate an increased risk of cancer from exposure to RF fields (Vijayalaxmi and Obe 2004). Reassuringly, with respect to base station RF fields per se, the levels of exposure are very low compared with those at which biological effects of any kind have been observed. However, because some biological experiments continue to suggest new possibilities for RF interactions, more scientific research is called for. It has been pointed out that, if one were proposing new chemicals for commerce or new pharmaceuticals, the quality of most available RF studies would not be acceptable for registration with the responsible authorities (Dasenbrock 2005).
The ongoing INTERPHONE collaboration is a multicenter, comprehensive study on mobile phones and cancer. It is coordinated by the International Agency on Research on Cancer (IARC), a specialized cancer agency of the WHO, and researchers in 13 countries are taking part using a common protocol. The INTERPHONE protocol is a population-based, case–control study correlating head and neck tumors with mobile-phone use by persons 30–59 years of age who reside in the study regions. Exposure assessment is reliable because it is based on individual records of cell phone use. Because of pooling of data from all participating centers, the study is statistically powerful.
For example, the risk of acoustic neuroma in relation to mobile phone use has been assessed via six population-based, shared-protocol, case–control studies in four Nordic countries and the United Kingdom. The authors concluded that there was no association of risk with duration of use, lifetime cumulative hours of use, or number of calls, for phone use overall or for analogue or digital phones separately (Schoemaker et al. 2005). Recent results from INTERPHONE have reported lack of brain tumor or acoustic neuroma risk in Japan (Takebayashi et al. 2006), Germany (Berg et al. 2006; Schuz et al. 2006), and in a meta-analysis of five Northern European countries (Lahkola et al. 2007).
There have also been other studies of mobile telephone users, particularly on brain tumors (and less often on other cancers and on symptoms). Results of these studies to date give no consistent or convincing evidence of a causal relation between RF exposure and any adverse health effect (Ahlbom 2006; Ahlbom et al. 2004; Lonn et al. 2005). A 4-year British survey released in 2006 showed no link between regular, long-term use of cell phones and the most common type of brain tumor, glioma (Hepworth et al. 2006). A German study did find an elevated risk of glioma in long-term users, but the increase was not statistically significant. The authors concluded that no overall increased risk of glioma or meningioma was observed among these cellular phone users (Schuz et al. 2006).
Another approach that has been used is to compare temporal trends in disease rates with temporal trends in prevalence of cell phone use. For example, trends in acoustic neuroma incidence in England and Wales were found not to lag behind trends in cell phone use in a correlated fashion (Nelson et al. 2006). Because increases in acoustic neuroma incidence predate or parallel rates of cell phone use, they likely reflect changes in reporting and diagnosis, and the temporal trends go counter to what would be expected if RF exposure played a role.
Several epidemiologic studies of potential cancer risk have used proximity to commercial broadcast transmission towers as the measure of RF exposure. However, individual RF exposures are not necessarily related to distance. None of these epidemiologic studies has provided sound evidence that RF exposure from the transmitters increased the risk of cancer or any other health effect [see summary by Jauchem (2003)]. The reporting of cancer "clusters" around RF broadcast transmitters and mobile phone base stations has heightened concern among the general public, but given the random nature of the distribution of cancers in the population it is not surprising statistically that such clusters should appear. Also, given the ubiquity of base stations, one would expect that a base station being near existing cancer clusters is a likely occurrence. Hence, reliable scientific evidence on how the distribution of cancer and other diseases in the population might be related to environmental factors (e.g., cellular telephone RF exposures) can best be obtained through carefully planned and executed epidemiologic studies such as INTERPHONE.
Noncancer health effects. Some changes have been reported in relation to cardiovascular function, but these findings have been in operators of broadcasting stations (Vangelova et al. 2006). From the weight of evidence and the very low exposure levels associated with base stations, there is no clear evidence of any adverse effect from such exposures (Feychting 2005; Feychting et al. 2005).
Potential neurologic and behavioral effects. Laboratory studies with volunteers have investigated whether low-level exposure to RF fields associated with mobile phones can affect brain function and behavior. Reported reactions to assumed RF exposure include a wide variety of nonspecific symptoms. Most commonly reported symptoms are sleeplessness, fatigue, dizziness, digestive disturbances, and concentration difficulties. By and large, well-controlled and -conducted double-blind studies have shown that symptoms are not correlated with RF exposure. There are also some indications that these symptoms may be caused by preexisting conditions such as stress reactions resulting from worrying about perceived RF health effects rather than the RF exposure per se. To date, only subtle and transient effects have been reported, and any implications for health remain unclear and unlikely (Cosquer et al. 2005). Exposures used in these studies are similar to those to the head from mobile phone use, rather than to the much lower RF levels associated with general public exposure from base stations.
Reviews of the evidence on electromagnetic hypersensitivity have been conducted (Fox 2006). An extensive systematic search identified relevant blind or double-blind provocation studies of individuals potentially hypersensitive to the presence of EMF. A meta-analysis found no evidence of an improved ability to detect EMF in "hypersensitive" participants. That is, it was concluded that weak electromagnetic fields are not likely to be causative factors for neurological symptoms (Rubin et al. 2005, 2006a, 2006b). An investigation into possible differences in blood cells between patients reporting EMF hypersensitivity and normal patients did not find any differences in lymphocyte response to RF from GSM mobile telephones (Markova et al. 2005). Other investigators have likewise concluded that "based on the limited studies available, there is no valid evidence for an association between impaired well-being and exposure to mobile phone radiation" (Seitz et al. 2005).
However, it is important to recognize the plight of people suffering from "hypersensitivity reactions." The WHO recently issued a fact sheet about people reporting nonspecific symptoms that they relate to RF fields from base stations and other EMF devices. Details can be found at WHO (2005). Moreover, E. Fox in the United Kingdom is continuing to analyze possible electromagnetic hypersensitivity reactions, and a report on the findings has been published (Eltiti et al. 2007).
Summary on RF health effects. The accumulated evidence does not establish the existence of adverse short- or long-term health effects from the signals produced by base station and local wireless networks. In fact, for similar RF exposure intensities (watts per square meter), the body absorbs about 5 times more of the RF energy from FM radio and television frequencies (around 100 MHz) than from base station frequencies (around 1–2 GHz). It is reassuring to note that radio and TV broadcast stations have been in operation for > 50 years, and health statistics have not demonstrated adverse health consequences.
Development of national and international RF guidelines. The heath-effect guidelines of ICNIRP in the mobile telephony frequency spectrum range from about 40–60 V/m (4.3–10 W/m2) (ICNIRP 1998). The ICNIRP guidelines have been widely accepted (> 30 countries worldwide) and, for example, are consistent with Health Canada (1999), U.S. [American National Standards Institute/Institute of Electrical and Electronics Engineers (2006), Federal Communications Commission (2006)], UK [National Radiation Protection Board/Health Protection Agency (NRPB/HPA 2004a, 2004b)], and Australian [Australian Radiation Protection and Nuclear Safety Agency (ARPANSA 2002)] standards. However, some countries and regions have adopted more stringent guidelines without specifically justifying them on the basis of available scientific evidence. In contrast to the ICNIRP levels, the following are some examples of these more restrictive guidelines, in the mobile telephone frequency range (Baumann 2006; Vecchia 2006):
- ICNIRP Guidelines: 40–60 V/m or 4.3–10 W/m2
- "Italian Exposure Limit": 6 V/m
- "Paris Charter": 2 V/m, 24-hr average, indoors
- "Salzburg Protection Value": 1 W/m2
- "Swiss Regulation": 4–5 V/m at full power.
The issues that most often drive more localized RF guidelines are not established health risk per se, but rather risk perceptions (Siegrist et al. 2005). In this regard, the "Precautionary Principle" is often invoked—"better safe than sorry"—part of which involves taking "protective measures without having to wait until the reality or seriousness of those risks becomes apparent." One expression of how "protective measures" might be applied to RF levels is:
The proposed [RF] standard also recommends that it is generally sensible to minimize exposure which is unnecessary or incidental to achievement of service objectives or process requirements, provided this does not introduce other risks and it can be readily achieved at modest expense. (ARPANSA 2001)
The term "modest expense" implies some type of cost–benefit analysis. Appropriate application of the Precautionary Principle requires that the policies be tailored such that the time, effort, expense, and risk of any "protective measures" be commensurate with what the society expends on other public risks of similar magnitude. However, if scientific research is not able to establish "apparent risks" in a quantitative way, making such a calculation is problematic.
If a basis for precautionary limits cannot be provided, then the danger behind promoting arbitrary limits and superfluous safety factors is that reliance on logical, science-based policy will be undermined in favor of unreasoned, fear-based, poorly thought out actions. Such actions, rather than providing reassurance, will likely trigger concerns, amplify unwarranted anxieties, and likely divert scarce resources into areas yielding little or no public health benefit (Barnett et al. 2006; Wiedemann and Schutz 2005). Despite unavoidable uncertainty and other limitations of the scientific method, science remains our best source of knowledge about how the world works and how we can rely on natural laws to understand interactions between the environment and living things.
Despite reassuring scientific evidence, some people perceive risks from mobile telephony RF exposure as likely and possibly severe. Several reasons for public fear have been proposed, including media announcements of new and unconfirmed scientific studies, leading to a feeling of uncertainty and a perception that there may be unknown or undiscovered hazards. Risk perception cannot be understood as a monolithic concept; rather, communicating "risk" is a unique challenge because it focuses on three major elements that are difficult to convey to a lay public: complexity, uncertainty, and ambiguity (Renn 2004, 2006).
The first barrier to easy risk communication is complexity. In everyday life, people perceive causal connections to be simple. In most scientific assessments of risk, causal connections are highly complex. Causes and effects are not obvious, and there are many intervening variables that obscure possible relationships. Often facility with quantitative methods is required, and appreciation of the fact that "the devil is in the details." Even though most scientists agree that health impacts of mobile telephony RF are unlikely (but not impossible), a lay audience will often assume that complexity is used to "pull the wool over their eyes."
The second barrier to explaining risk is uncertainty. Most risks have a hypothetical component, and we cannot say with certainty that x exposure will cause (has caused) adverse outcome y. For example, science can rarely make definitive statements about what has "caused" a particular disease like cancer. The best we can do is to calculate probabilities linked to different causal pathways. In terms of risk communication, conveying the different types of uncertainty to a lay audience poses major challenges. People misunderstand uncertainties as being an indicator for bad science or sloppy risk assessment. In the face of uncertainty, a lay audience will typically resort to the simplistic strategy: "Better safe than sorry."
The third component of risk that is difficult to communicate is ambiguity. Ambiguity refers to the existence of divergent or contested interpretations and perspectives on the severity associated with a potential health threat. Not only can scientific interpretations have an ambiguous component, but ambiguity also emerges with regard to selection of appropriate values, priorities, assumptions, ethics, distributions of risk, and quality of life parameters to be applied. In the face of "duelling experts," a lay audience will often assume that the correct interpretation is midway between the two perspectives offered, because a lay audience has limited ability to recognize fringe positions.
To deal with these problems in risk communication, the public needs to be provided access to accurate information and education on scientific consensus positions. First, the solution to complexity is not a simple prohibition of a given technology. Rather, it must be recognized that scientific inquiry can test for harmful effects, but it can never prove that something is safe. This asymmetrical relationship is difficult to communicate, but one approach is to encourage people to realize that absolute safety is not a requirement that they can or do impose on any societal activity (e.g., public transportation, food supply, medical procedures, prescription drugs). Second, science deals with uncertainty by incorporating safety factors. That is, safety guidelines are not bright boundary lines between "good" and "bad," but rather they incorporate an adequate margin of safety so that even exceeding the exposure guideline would not lead one to anticipate ill effects. Third, in dealing with ambiguity, lay audiences need to be educated that the scientific consensus does not necessarily lie midway between the opinions they are offered. Hypothetical risks need to be recognized at such, and in a scientific, secular world, explanation patterns that are counterfactual, paranormal, and metaphysical are not valid. Although the natural human ability to sympathize with the victims of disease is to be admired, the associated desire to pin the blame for inexplicable diseases on a nearby environmental factor is misguided. That is, when individuals contract diseases, and when we have little idea as to what specifically caused the disease in question, that ignorance somehow sounds like evidence in favor of focusing on "unknown" risks such as RF or EMF. But this lack of knowledge is not evidence.
Perhaps the most important element of risk communication is to assure audiences that RF standards have been and continue to be under ongoing scrutiny. Large numbers of scientists, medical doctors, and public health professionals of disparate orientations and areas of expertise sift existing data and contribute new data in an ongoing risk assessment effort. The vast majority of human cancers are likely caused by unavoidable environmental exposures (e.g., viruses, diet, lifestyle, sunlight, background ionizing radiation) or to processes inherent to life itself (e.g., genetic instability, copying errors in DNA, endogenous hormones, creation of mutagens and free radical molecules by metabolism of food, production of reactive chemicals for microbicidal defense) (Gotay 2005; Henderson et al. 1991; McKean-Cowdin et al. 2000; Wogan et al. 2004). In scientific risk assessment, one compares the ability of the exposure of interest to increase risk above these baseline, natural processes.
As outlined above, our best scientific understanding indicates that there are no health consequences of base-station RF exposure, and no adverse effects are foreseen at the RF levels typical of cellular telephone technology. This viewpoint is not only consistent with the conclusions of the WHO workshop on "Base Stations and Wireless Networks," but it is also consistent with numerous other public health reviews on the safety of wireless technologies, although most public health agencies continue to favor "additional research." Some of these blue-ribbon, consensus-group conclusions are mentioned below.
ICNIRP (1998) has developed guidelines to protect human health from exposure to EMF across the RF spectrum. These ICNIRP guidelines have been adopted by > 30 countries. Certain countries have instituted standards limiting emissions from cellular telephone base stations that are significantly below recommended ICNIRP limits. Such additional restrictions are not based on any known health effects, but rather tend to be either a precautionary measure or an "as low as reasonably achievable" (ALARA) measure that requires base station transmissions to be no more than required for providing a good service.
Several groups in Great Britain have evaluated potential health effects of RF. The Advisory Group on Non-Ionizing Radiation (2003) updated the year 2000 report of the Independent Expert Group on Mobile Phones (2000) and concluded that "exposures due to living near to base stations are extremely low, and the overall evidence indicates that they are unlikely to pose a risk to health." (Advisory Group on Non-Ionizing Radiation 2003).
The UK Health Protection Agency (formerly the National Radiation Protection Board) (NRPB/HPA 2004a) also has concluded that RF energy can potentially cause health effects only if people are exposed to RF levels significantly exceeding international limits. That is, they recommended that exposure to EMFs (0–300 GHz) in the UK be based on the 1998 guidelines issued by ICNIRP (NRPB/HPA 2004a).
In a specific review of cellular telephone technology (NRPB/HPA 2004b), the agency proposed that even though "there is a lack of hard information showing that the mobile phone systems in use are damaging to health," they continued to endorse a "precautionary approach" to the use of mobile phone technologies.
The Health Council of the Netherlands (Health Council 2002) has prepared a report on the potential risks of EM fields from mobile telephones. The report concluded that "the EM field of a mobile telephone does not constitute a health hazard, according to the present state of scientific knowledge." Moreover, the review committee concluded that "the scientific information concerning non-thermal effects discussed in this report provides no reason to apply the precautionary principle and lower the SAR limits for partial body exposure" (Health Council 2002). A 2005 Health Council update concluded that "the Committee therefore disagrees … that a connection has been found between living in the proximity of a base station and the occurrence of cancer" (Health Council 2005).
ARPANSA prepared a fact sheet titled "What about base stations and telecommunication towers–are there any health effects?" ARPANSA concluded that "the weight of national and international scientific opinion is that there is no substantiated evidence that RF emissions associated with living near a mobile phone base station or telecommunications tower poses a health risk" (ARPANSA 2003a). ARPANSA also evaluated the potential for risk to children and concluded that "the balance of evidence does not indicate a risk to the health of people, including children, living in the vicinity of base stations where the exposure levels are only small fractions of the ARPANSA Standard" (ARPANSA 2003b).
The Royal Society of Canada has an "Expert Panel on Potential Health Risks of Radiofrequency Fields from Wireless Telecommunication Devices," and their most recent update (2004) notes that "all of the authoritative reviews completed within the last two years have concluded that there is no clear evidence of adverse health effects associated with RF fields" (Royal Society of Canada 2004).
The advice of the U.S. Health Physics Society (a professional society of specialists in radiation safety) is that there is no reason to believe that cellular base station towers could constitute a potential health hazard to nearby residents or students (Health Physics Society 2006).
At present, the only established effects that can result from excessive exposure to RF energy are related to tissue heating. Although RF energy can be absorbed by living organisms to some degree at any frequency, available data do not demonstrate adverse health consequences at exposure levels below internationally accepted limits, which do not allow significant heating. In summary, none of the recent research or reviews of research have concluded that permissible RF exposure levels from mobile phones and their base stations lead to adverse health consequences.
Although scientists generally assign low priority to conducting research on base stations or other wireless technologies having such weak RF signals, some gaps in knowledge still exist (Repacholi 1998). Research recommended to fill these gaps can be found in the WHO RF research agenda (WHO 2006b).