Introduction
Interstellar travel became technologically plausible in the 1950s, when the energy release of thermonuclear fusion was observed in the first hydrogen bombs. First studies were based on the idea of a pulse drive, directly propelled by the explosions of atomic bombs behind the craft (Dyson, Reference Dyson1965, Reference Dyson1968), evolving into a direct fusion rocket (Bond & Martin, Reference Bond and Martin1978). These designs were manned interstellar arks with masses of the order of 10 million tonnes and speeds of 10 % the speed of light.
Classical rockets, both chemical and nuclear, suffer from the limitations imposed by Tsiolkovsky's rocket equation (Plastino & Muzzio, Reference Plastino and Muzzio1992): if a rocket carries its own fuel, the ratio of total rocket mass versus final velocity is an exponential function, making high speeds extremely expensive. A different method, which does not require the fuel to be accelerated with the ship, has been proposed by Johannes Kepler (Reference Kepler1604). After observing a comet, he suggested that the cometary tail points away from the sun due to a ‘breeze’, and proposed to ‘provide ships or sails adapted to the heavenly breezes, and there will be some who will brave even that void.’. James Clerk Maxwell predicted that radiation carries momentum and exerts pressure: ‘Hence in a Medium in which waves are propagated there is a pressure in the direction normal to the wave, and numerically equal to the energy in unit of volume’ (Maxwell, Reference Maxwell1873; Maxwell & Harman, Reference Maxwell and Harman1990).
Redding (Reference Redding1967) noted that there was no obvious way to decelerate the spacecraft at the target star system. Only recently, Heller & Hippke (Reference Heller and Hippke2017) and Heller et al. (Reference Heller, Hippke and Kervella2017) suggested to decelerate using the stellar radiation and gravitation in a maneuver they referred to as photogravitational assist. A project by the ‘Breakthrough Initiatives’Footnote 1 provides monetary support (of order 100 m USD) for research on gram-scale robotic spacecrafts, using a light sail for propulsion (Lubin, Reference Lubin2016; Popkin, Reference Popkin2017).
Between ‘Project Orion’, and the ‘nanocraft concept’, there is a factor of 1013 in weight. The smaller weight results in lower build- and launch costs, a benefit that could make such a mission affordable within the current century. When we compare the early studies with the most recent concept, we have to distinguish that the main purpose of interstellar travel shifted from colonization of exoplanets with human (biological) settlers to unmanned research probes, taking spectroscopic and photographic measurements of the putative biological environment on potentially habitable exoplanets. Software and hardware engineering has made sufficient progress since the 1960s that such probes can be highly autonomous. Consequently, the required mass for probes can be reduced.
Our benefit from autonomous interstellar probes is purely in the information they send back to us. Thus we shall seek to maximize the amount of information we can obtain from them. A major issue is that these probes are designed to be very light-weight, and thus limited in terms of power. While traditional, fusion-based concepts proposed the use of high (MW, Milne et al., Reference Milne, Lamontagne and Freeland2016) power at GHz frequencies for data transfer, small sailing probes cannot have a fusion reactor on board and will have to rely on photovoltaic energy, which delivers of order kW km−2 surface area. In the current era of high-resolution video, a high data rate to transfer spectacular observations of an alien world could be important for the public reception of such a mission, and thus its financial funding. It is therefore crucial to optimize interstellar communication, precisely the data rate, to maximize the data volume of scientific and public data.
In this work, paper I of the series, our contributions are: (1) to introduce the variables in the framework of data transfer between telescopes; (2) to assess limiting factors such as extinction, noise and technological constraints; and (3) to calculate optimal frequencies and achievable data rates for exemplary cases.
Method to calculate data rates
The free-space photon flux F received from a telescope at distance d can be calculated as (Kaushal et al., Reference Kaushal, Jain and Kar2017):
where P t is the transmitted power, f the photon frequency, and h the Planck's constant ( ≈ 6.626 × 10−34 J s). The (half) opening angle of the diverging light beam is θ d = Q Rλ/D t (in radians) with Q R ≈ 1.22 for a diffraction limited circular transmitting telescope of diameter D t (Rayleigh, Reference Rayleigh1879), and λ = c/f with c as the speed of light in vacuum (299 792 458 m s−1). In a receiving telescope with aperture A R = π D 2r/4 we obtain the flux
This assumes a uniform plane-wave illumination. A telescope with central obscuration and plane-wave Gaussian-beam illumination has been calculated by Klein & Degnan (Reference Klein and Degnan1974), and the flux loss from pointing errors by Marshall (Reference Marshall1987); but these secondary effects will be neglected here. For a laserbeam, the narrower ‘waist’ leads to an intensity pattern with a characteristic angular beam size given by (Duarte, Reference Duarte2015; Tellis & Marcy, Reference Tellis and Marcy2015) θ L = Q L(2/π)λ/D, or θ L/θ d ≈ 0.5, which leads to a tightening of the beam. Note that a laserbeam shape is not maintained in systems where laser light is broadened with a beam expander and then focused with a telescope, and so we neglect this possibility here.
The widely used approximationFootnote 2 of the diffraction-limited aperture, θ ≈ λ/D, leads to an overestimate of the received photon flux on the receiver side by ≈ 49%. This can be verified by setting Q R = 1 versus Q R = 1.22 numerically. The considerable difference comes from the fact that θ enters the equation through the inverse square law. The precise value, θ = 1.2196..λ/D comes from the Fraunhofer diffraction where this number is the first zero of the order-one Bessel function of the first kind, J 1(x)/π.
Several factors will constrain the achievable data rates. Regarding the loss of photons, the most important are interstellar extinction (section ‘Loss of photons from extinction’), of which we denote the surviving fraction as 0 < S E < 1. For ground-based telescopes, atmospheric transmission allows for the reception of another fraction of photons, 0 < S A < 1 (section ‘Loss of photons from atmospheric transmission’). The receiver efficiency is denoted as 0 < η < 1. Technological constraints on the telescopes will be denoted as Q (section ‘Technological limits of telescopes’). Other small factors, such as scintillation and scattering (section ‘Scintillation and scattering of photons’), might play a role and can be included in calculations in a similar manner, but we neglect them here for brevity. The major noise sources are atmospheric sky background (section ‘Atmospheric sky background’), zodiacal light (section ‘Background light from zodiacal light’), and others (sections ‘Background light from the target star and celestial bodies’ and ‘Instrumental noise’).
Channel capacity for a coherent wave
We now define the theoretical maximum data rates based on frequency, signal and noise. For completeness, we will first discuss the optimum case where the number of photons received is sufficiently large to form a coherent wave. While this might not be realistic for most schemes of interstellar communication (section ‘Results’), it is useful to define the classical upper bound. The maximum rate at which information can be transmitted over a communications channel is (Shannon, Reference Shannon1949):
where C is the channel capacity (in bits per second), B is the bandwidth of the channel (in Hertz), S is the average signal power (in Watt) and N is the average Gaussian noise (in Watt). The bandwidth is the difference between the highest (f H) and lowest (f L) frequency in a continuous set of frequencies.
To compare data rates for different frequencies, we can approximate bandwidth with frequency by taking a constant fractional bandwidth, b. With f C as the centre frequency, we can define b = (f H − f L)/f C. With a value of, e.g. b = 0.1, we can approximate B ≈ c/λ (in Hz).
Channel capacity is proportional to frequency and to the logarithm of S/N. These relations suggest that the frequency should be increased to the practical maximum, and that the signal power should merely be increased to overcome noise, with little benefit beyond.
If the received number of photons (after extinction and other losses) is sufficiently large to form a coherent wave, we can plug equation (2) (as the signal S in photons per second) into (3) and define the noise equally in photons per second:
where N γ is the number of photons (γ) from noise per second (physical and instrumental). Then, the data signaling rate for the case of a continuous wave, DSRc, can be conveniently calculated in units of bits/s if P is in Watt.
Intuitively, one would assume that at least one photon is required to transfer one bit of information, but this is incorrect: More than one bit can be encoded per photon. This is done with a modulation scheme to define an alphabet, often using a combination of polarization, phase, frequency and amplitude modulation (e.g., Jones, Reference Jones and Shostak1995). Each symbol of such an alphabet can encode several bits, scaling with the logarithm to base 2 of the number of members. This is called spectral efficiency and is measured in (bits/s)/Hz. Modulation rate, spectral efficiency and data rate can be increased for a constant bandwidth at the cost of an exponential rise in signal-to-noise ratio (SNR), or, for a constant noise level, in an exponential increase in P.
For the extreme case of communication with negligible losses (e.g., d → 0), equation (3) suggests the use of infinitely high bandwidth. However, infinite frequencies (and infinite capacity) are unphysical. In the classical sense, the limit comes from the fact that an increase in bandwidth also increases noise power (Shannon's power efficiency limit). A noiseless channel has infinite capacity: with equation (3) we have C = Blog2(1 + ∞) = ∞. However, in reality noise is never zero because photons are quantized (section ‘Instrumental noise’). Then, the capacity of Shannon's limit becomes (Chitode, Reference Chitode2009, pp. 5–117):
In the framework of quantum state propagation, any transmission system can exchange only a limited (quantized) amount of information in a given time frame (Yuen & Shapiro, Reference Yuen and Shapiro1978), and is thus limited by physical resources (Bekenstein, Reference Bekenstein1981). Therefore, increasing frequency to infinity does not increase data rate to infinity (Giovannetti et al., Reference Giovannetti, Guha, Lloyd, Maccone, Shapiro and Yuen2004b).
The photon-limited case
The limit for equation (3) only applies if the number of photons is sufficiently large to form a coherent wave. In many schemes for interstellar communication (section ‘Results’), the data rate is photon-limited. Then, Holevo's bound (Holevo, Reference Holevo1973) establishes the upper limit to the amount of information which can be transmitted with a quantum state. It applies independently from the frequency of the wave, and assumes that a number (quantity) of modes can be used per photon, which originate from the photons' dimensions, namely polarization, frequency and time of arrival. The inverse of this quantity, M, is the number of photons per mode. Then, as shown by Giovannetti et al. (Reference Giovannetti, Guha, Lloyd, Maccone, Shapiro and Yuen2004b), the ultimate quantum limit of bits per photon can be expressed as:
where η is the receiver efficiency and g(x) = (1 + x)log2(1 + x) − xlog2x so that g(x) is a functionFootnote 3 of η × M. In the presence of thermal noise, it was conjectured (Giovannetti et al., Reference Giovannetti, Guha, Lloyd, Maccone and Shapiro2004a) and recently proven (Giovannetti et al., Reference Giovannetti, García-Patrón, Cerf and Holevo2014) that the capacity is:
where N M is the average number of noise photons per mode. It is an open question if the maximum can fully, or only approximately be achieved in practice (Guha & Wilde, Reference Guha and Wilde2012; Wilde et al., Reference Wilde, Guha, Tan and Lloyd2012). The achievable capacity is shown for a wide range of modes in Fig. 1. It is clear that even large numbers of modes and small fractional noise increase the number of bits per photon only within a factor of a few.
We can multiply equations (2) and (7) to calculate the data rate for the photon limited case of two communicating telescopes:
where DSRγ is in units of bits per second when P is in Watt. It assumes that the free path loss caused by η, d, S E, S A is known and accounted for in the encoding scheme. Variations and uncertainties in the number of received photons can be treated as an additional noise source, but optimal encoding schemes will be neglected in this paper. In the following sections, we will discuss the values in these equations.
Signal losses
Loss of photons from extinction
From the infrared (IR) to the ultraviolet (UV), extinction is caused by the scattering of radiation by dust, while at wavelengths shorter than the Lyman limit (91.2 nm), extinction is dominated by photo-ionization of atoms (Ryter, Reference Ryter1996). For short interstellar distances, extinction in the optical is small, ≈ 0.1 mag within 100 pc, 0.05 − 0.15 mag out to 200 pc (Vergely et al., Reference Vergely, Ferrero, Egret and Koeppen1998). It is much larger towards the galactic centre, E(B − V) ≈ 3 at A(V) > 44 mag at 550 nm (Porquet et al., Reference Porquet, Grosso, Predehl, Hasinger, Yusef-Zadeh, Aschenbach, Trap, Melia, Warwick, Goldwurm, Bélanger, Tanaka, Genzel, Dodds-Eden, Sakano and Ferrando2008; Fritz et al., Reference Fritz, Gillessen, Dodds-Eden, Lutz, Genzel, Raab, Ott, Pfuhl, Eisenhauer and Yusef-Zadeh2011), an attenuation by a factor of 10−18. Another prominent feature in measured extinction curves is a ‘bump’ in the UV at 217.5 nm (Stecher, Reference Stecher1965, Reference Stecher1969), where extinction is about an order of magnitude higher. It is attributed to organic carbon and amorphous silicates present in the grains (Bradley et al., Reference Bradley, Dai, Erni, Browning, Graham, Weber, Smith, Hutcheon, Ishii, Bajt, Floss, Stadermann and Sandford2005). Other features are the water ice absorption at 3.1 μm and the 10 and 18 μm silicate absorption.
While higher frequencies have higher channel capacities for coherent waves, and allow for tighter beams (at a given telescope size), they also generally suffer from higher extinction between UV and IR. To analyse this trade-off (section ‘Interstellar Extinction’), we use the synthetic extinction curve presented in Draine (Reference Draine2003a,Reference Draineb,Reference Drainec) which covers wavelengths from 1 cm (30 GHz) to 1 Å (12.4 keV). We scale this curve for different distances using A(V) = 1.8 mag per kpc in the galactic plane (Whittet, Reference Whittet2003), equivalent to E(B − V ) = 0.28 mag per kpc (Dias et al., Reference Dias, Alessi, Moitinho and Lépine2002). For the highest extinction values towards the galactic centre, where E(B − V) ≈ 3, we use measurements for the optical and infrared (IR) (Fritz et al., Reference Fritz, Gillessen, Dodds-Eden, Lutz, Genzel, Raab, Ott, Pfuhl, Eisenhauer and Yusef-Zadeh2011) and the UV (Valencic et al., Reference Valencic, Clayton and Gordon2004; McJunkin et al., Reference McJunkin, France, Schindhelm, Herczeg, Schneider and Brown2016) and interpolate in between individual data points with a spline. This extinction curves covers the wavelength range 0.1 − 27 μm. While extinction is typically given in astronomical magnitudes, we convert these to the fraction of photons received over distance (S E), and show examples in Fig. 2.
Loss of photons from atmospheric transmission
The earth is surrounded by an atmosphere (Forbes, Reference Forbes1842), which is essential for almost all life on this planet (Canfield et al., Reference Canfield, Poulton and Narbonne2007), but of the greatest annoyance for almost all astronomers (Kuiper, Reference Kuiper1950). For a space telescope there is no loss of photons from a surrounding cloud of gas, dust and water, so that the surviving fractions of photons is S A = 1. On earth, atmospheric transmission depends on the wavelength and varying characteristics, such as the content of water vapour in the air. As an example, we use a transmission curve S A(λ) for Mauna Kea with a water vapour column of 1 mm, which represents excellent observing conditions, and occurs in the 20% of the best nights of an average year (Lord, Reference Lord1992; Guharay et al., Reference Guharay, Nath, Pant, Pande, Russell and Pandey2009). This curve covers the wavelength range of 200 nm–10 cm (3 GHz). Figure 3 shows the part up to 10 mm (30 GHz), after which transmission reaches near unity.
Transmission is zero for all practical purposes for wavelengths below 291 nm, above 20 m, and between 30 and 200 μm. In the optical and IR, transmission is highly variable due to numerous absorption lines from water, carbon dioxide, ozone and other gases. When communicating with photons in a narrow (nm) bandwidth, as is common with lasers, the exact wavelength must be chosen carefully, because transmission fluctuates rapidly. For example, S A = 0.98 at λ = 934.36 nm, but S A = 0.22 at λ = 934.52 nm, a spectral distance of only 0.16 nm. Under good atmospheric conditions, transmission can be close to unity for many wavelengths in the optical and near- to mid-IR.
For brevity, we neglect other atmospheric effects such as scattering and turbulence (‘seeing’, Coulman et al., Reference Coulman, Vernin and Fuchs1995) which is a variation of the optical refractive index and enlarges the point spread function of the telescope, if not corrected for with adaptive optics.
Technological limits of telescopes
The angular beam size is limited to Q R ≥ 1.22 (Rayleigh limit), or Q L ≥ 1 for a laserbeam. Technology may place a stricter limit. We have examined the angular resolution of current (earth 2017) space telescopes for different wavelengths. As can be seen in Fig. 4, Q real/Q R is an exponential function for wavelengths λ < 300 nm, indicating the technological difficulty to focus wavelengths in UV and shorter.
For diffraction-limited telescope mirrors, the polished surfaces need to have surface smoothness <λ/4 (Danjon & Couder, Reference Danjon and Couder1935), which makes the production of telescopes for UV, X-ray and γ-ray increasingly difficult. Additionally, the refractive index of all known materials is close to 1 at high (keV) energies, making it difficult to focus photons efficiently and avoid absorption (Aristov & Shabel'nikov, Reference Aristov and Shabel'nikov2008).
With today's technology, resolution in the milli-arcsec regime is possible at optical wavelengths, but X-rays are limited to angular resolutions of 20 arcsec (Salmaso et al., Reference Salmaso, Basso, Brizzolari, Civitani, Ghigo, Pareschi, Spiga, Tagliaferri and Vecchi2014), a difference of four orders of magnitude. For example, the Swift X-Ray satellite has an angular resolution of 18 arcsec at λ = 1 nm (1.5 keV) from a 30 cm aperture (Burrows et al., Reference Burrows, Hill, Nousek, Kennea, Wells, Osborne, Abbey, Beardmore, Mukerjee, Short, Chincarini, Campana, Citterio, Moretti, Pagani, Tagliaferri, Giommi, Capalbi, Tamburelli, Angelini, Cusumano, Bräuninger, Burkert and Hartner2005), while the diffraction limit would be 1.22λ/D = 8 × 10−4 arcsec, so that Q real/Q R = 4 × 10−5. Technology is believed to eventually achieve sub-arcsec resolution at X-rays, but at the expense of large designs, with focal lengths of 105 km (Gorenstein, Reference Gorenstein, Citterio and O’Dell2004).
Technological limits of the receiver
Photon energy depends on wavelength, E = hc/λ, which should make it easier to detect higher energy photons in theory. In practice, single photon detection with high quantum efficiency is possible throughout a wide range of wavelengths, from X-Rays (Tanguay et al., Reference Tanguay, Yun, Kim and Cunningham2013, Reference Tanguay, Yun, Kim and Cunningham2015) to microwaves (Poudel et al., Reference Poudel, McDermott and Vavilov2012; Wong & Vavilov, Reference Wong and Vavilov2015). Interestingly, even the human eye can detect single photons in the visible light (Tinsley et al., Reference Tinsley, Molodtsov, Prevedel, Wartmann, Espigulé-Pons, Lauwers and Vaziri2016).
We will neglect a possible wavelength-dependence in quantum efficiency of photon detectors in this paper. This is supported by the much stronger influence from technological limits in focusing beams (section ‘Technological limits of telescopes’), and the influence of interstellar extinction on photon throughput (section ‘Loss of photons from extinction’), so that detector differences (of a few per cent) will be negligible for most practical cases.
Noise
Noise sources can be astrophysical (scattering of the signal, background light) or instrumental (shot noise and read noise). For ground-based telescopes, the total noise has been measured (section ‘Atmospheric sky background’), for space-based telescopes, it will be discussed in sections ‘Background light from zodiacal light', ‘Background light from galactic and extragalactic sources', ‘Scintillation and scattering of photons', ‘Background light from the target star and celestial bodies', and ‘Instrumental noise'.
Atmospheric sky background
For a telescope located on earth, the total sky background which enters as noise into the receiver can be measured by observing a (maximally) empty sky area. Naturally, it includes all sources: terrestrial, solar system and interstellar.
Precise raw sky emission data are available for many observatory sites, and as in section ‘Loss of photons from atmospheric transmission’ we use Mauna Kea as an example. The measurements are for the sky background only and do not include the emission from a telescope or sensor (which has been subtracted out). The data were manufactured with a synthetic sky transmission (Lord, Reference Lord1992) subtracted from unity. This gives an emissivity which is then multiplied by a blackbody function with a temperature of 273 K (Guharay et al., Reference Guharay, Nath, Pant, Pande, Russell and Pandey2009). The authors added emission spectra based on observations from Mauna Kea, and the dark sky continuum mostly from zodiacal light. Finally, the curve has been scaled to produce 18.2 mag arcsec−2 in the H band, as observed on Mauna Kea by Maihara et al. (Reference Maihara, Iwamuro, Yamashita, Hall, Cowie, Tokunaga and Pickles1993). The resolution of the final data product is 0.1 nmFootnote 4.
These values are in agreement with measurements from the darkest observatory sites on earth, which have an optical sky background minimum of 22 mag arcsec−2 (Smith et al., Reference Smith, Warner, Orellana, Munizaga, Sanhueza, Bogglio, Cartier, Gibbs, Barnes, Manning and Partridge2008), corresponding to an optical flux of a few γ arcsec−2 s−1 from unresolved sky sources, air glow and zodiacal light.
The sky background at Mauna Kea is shown in Fig. 5 and covers the band from 300 nm to 30 μm. Similarly to the transmission (section ‘Loss of photons from atmospheric transmission’), background levels vary by up to three orders of magnitude over few nm. Generally, the flux is ≈ γ nm−1 arcsec−2 m−2 in the optical and near infrared (NIR), with a steep increases for λ > 2.5 μm and reaches 107 γ nm−1 arcsec−2 m−2 at 10 μm.
This indicates that earth-based interstellar communication is favourable for λ < 2.5 μm. For telescopes on other planets, we would need to know precisely the exoplanet atmospheres, exozodiacal dust, etc. which may result in a different noise structure; a detailed discussion is beyond the scope of this paper.
Background light from zodiacal light
Space telescopes are not affected by the strong atmospheric light. However, they still collect undesired photons. The strongest source is sunlight which is scattered off of dust grains in the solar system, an effect called zodiacal light. In the ecliptic plane, it can be as bright as 1.5 × 10−6 ergs s−1cm−2Å−1. It is faintest at heliocentric longitude 130° − 170° away from the sun because of larger scattering angles, and at low ecliptic latitudes <30° because of the minimum in the interplanetary dust column density at levels <10−7 ergs s−1cm−2Å−1 (Bernstein et al., Reference Bernstein, Freedman and Madore2002). The scattering strength only weakly depends on wavelength and closely resembles the solar spectrum between 150 nm and 10 μm (Leinert et al., Reference Leinert, Richter, Pitz and Planck1981; Matsuura et al., Reference Matsuura, Matsumoto, Matsuhara and Noda1995).
These levels contribute a flux of order 3 γ nm−1 arcsec−2 m−2 at 1 μm in the ecliptic, and decrease to 0.1 (0.03) photons at latitude 45° (90°). We show an all-sky map in Fig. 6 which makes it clear that the source's location on the sky is important, in addition to the wavelength.
Background light from galactic and extragalactic sources
The Galactic light comes from stars, starlight scattered from interstellar dust and line emission from the warm interstellar medium. Its levels are of order 10−9 ergs s−1cm−2Å−1 between 200 nm and 1 μm.
The mean flux of the optical extragalactic background light has been measured as 4.0 ± 2.5, 2.7 ± 1.4 and 2.2 ± 1.0 × 10−9 ergs s−1cm−2Å−1 at wavelengths of 300 nm, 550 nm and 800 nm (Bernstein et al., Reference Bernstein, Freedman and Madore2002).
Compared with the zodiacal light, these sources are weaker by two orders of magnitude and are only relevant if the source is near the ecliptic poles, where zodi is smallest; and for wavelengths outside the zodi-band of ≈ 0.3 − 300 μm (see Fig. 7).
Scintillation and scattering of photons
Extinction causes not only a loss of photons from absorption, but also scattering. The latter reduces the ‘prompt’ pulse height and produces an exponential tail (Howard et al., Reference Howard, Horowitz, Coldwell, Klein, Sung, Wolff, Caruso, Latham, Papaliolios, Stefanik, Zajac, Lemarchand and Meech2000).
Scatter broadening is well known from pulsars and magnetars. As an extreme example, magnetars close (0.1 pc) to the galactic centre with dispersion measures DM = 1778 pc cm−3 have their pulses broadened to 1.423 ± 0.32 s at 1.2 GHz, and 0.2 ± 0.07 ms at 18.95 GHz, following a power law with index −2.8 (Spitler et al., Reference Spitler, Lee, Eatough, Kramer, Karuppusamy, Bassa, Cognard, Desvignes, Lyne, Stappers, Bower, Cordes, Champion and Falcke2014). A single pulse which is broadened to a width of one second results in a very low data rate of the order of bit per second. Extrapolating with the power law indicates that nanosecond pulse widths (10−9 s) can be expected for frequencies >500 GHz (λ < 0.6 mm), and the broadening should become shorter than the wavelength at λ ≈ μm.
For these higher frequencies, the amplitude level of the scattering tail, and its length, is unknown in practice. Limits from the Crab pulsar show no detectable scattering tail at UV and optical wavelengths for an optical millisecond pulse width and E(B − V) = 0.52 (Sollerman et al., Reference Sollerman, Lundqvist, Lindler, Chevalier, Fransson, Gull, Pun and Sonneborn2000), consistent with the power-law scaling from radio observations. These results indicate that the impact of extinction is mainly on the absorption for frequencies >500 GHz (<0.6 mm), and not on pulse broadening. Therefore, we neglect this effect in our calculations, but suggest further research in this area.
Background light from the target star and celestial bodies
On the direct path, even modest-sized telescopes receive a relevant number of photons from nearby stars. For example, 5 × 1010 γ s−1 m−2 from α Cen A (distance 1.3 pc, Kervella et al., Reference Kervella, Mignard, Mérand and Thévenin2016, L = 1.522L ⊙). From Proxima Centauri (L = 1.38 × 10−4L ⊙), it is 4.25 × 106 γ sec−1 m−2, or 3.5 × 109 (3.5 × 105) γ sec−1 m−2 from a sun-like star in a distance of 10 light year (LY) (1000 LY). A coronograph or occulter could be used to block a significant part of this flux (10−9, Guyon et al., Reference Guyon, Pluzhnik, Kuchner, Collins and Ridgway2006; Liu et al., Reference Liu, Ren, Dou, Zhu, Zhang, Zhao, Wu and Chen2015). Additionally, a filter with a small band-pass, e.g. 1 nm, would reduce the flux further by >103. A good angular resolution of the receiving telescope would be helpful to separate the transmitter from the nearby target star. For comparison, a probe at a distance of 1 au from the star α Cen A would appear at an angular separation of 1.42 arcsec as seen from earth, resolvable even with small telescopes, assuming sufficient contrast.
The flux levels from reflected exoplanet light and exozodiacal dust is many orders of magnitude fainter than from the flux in the home solar system, and can thus be neglected.
Instrumental noise
Apart from a loss of photons from imperfect reflection or transmission in the receiver, the conversion from photons to electrons (e.g. with CCDs or photomultiplier tubes, which are analogue devices) causes a small, but non-zero amount of noise.
Even a perfect instrument will produce some noise. Fundamentally, this originates from the fact that photons and electrons are quantized (Einstein, Reference Einstein1905), so that only a finite number can be counted in a given time. This phenomenon is the shot noise (Schottky, Reference Schottky1918), and is correlated with the brightness of the target.
Results
A Starshot-like probe at α Centauri
We will now calculate exemplary quantitative data rates. Our default example is to maximize data rates with a probe at α Cen (d = 1.3 pc) and examine the influence of the variables presented in the previous section. Our standard example probe uses a telescope with a circular aperture D t = 1 m, through which it transmits with a power of P = 1000 W. The telescope quality Q R ≈ 1.22 for λ > 300 nm is of current (earth 2017) technology, and positioned in space. The hypothetical receiver has D r = 39 m, comparable with the upcoming generation of ‘extremely large telescopes’ (E-ELTs). It must be located in the southern hemisphere, e.g. in Chile, because α Cen is not observable from Mauna Kea's northern latitude which served as an example in previous sections. The total receiver efficiency is η = 0.5. It uses N = 105 modes, which could be done with a R = 100 000 spectrograph, 105 time slots, or a combination of both. The atmospheric sky background represents very good (20-percentile) conditions as described in section ‘Atmospheric sky background’.
From the transmitted P = 1000 W (2.2 × 1021 γ s−1 at λ = 429.6 nm), the theoretical flux near earth after free-space loss is 1860 γ s−1 in the receiver aperture. Interstellar extinction for this wavelength and distance is ≈ 0.3%, causing a loss of 6 photons, or a reduction to 1854 γ s−1. Sky transparency is 0.74, so that 1369 γ s−1 survive. This is the signal before receiver efficiency.
Regarding the total sky background, we assume that the filter width at the receiver has a bandpass of 1 nm, and the on-sky resolution is 1 arcsec. We neglect the photon flux from α Cen as it can be effectively suppressed (section ‘Background light from the target star and celestial bodies’), and is then negligible compared with the atmospheric background of 0.6 γ nm−1 s arcsec−2 m−2 (section ‘Loss of photons from atmospheric transmission’), resulting in 702 noise photons per second in the telescope. We will discuss the case of blended sources (probe and star) in the section ‘Blend of probe and star’. We also neglect the noise flux from the receiver itself. Following equation (7), the Holevo bound with our noise is then 1.81 bits per photon. This includes the receiver efficiency of η = 0.5.
We can now multiply the received photons with the encoding limit and estimate 1,369 γ s−1 at 1.81 bits γ−1 = 2480 bits s−1. This is also the peak value at λ = 429.6 nm in Fig. 8 (left), indicating that any other wavelength decreases the effective data rate. In practice, this is an upper bound; realistic data rates including sensor noise, margin for error, etc. will yield smaller data rates by a factor of a few.
The cut-off for λ < 290 nm comes from the atmospheric intransparency (Fig. 3). The decline in data rate towards longer wavelengths comes from two effects: the decrease in telescope focusing (section ‘Technological limits of telescopes’), and increasing atmospheric noise (Fig. 5). Individual atmospheric absorption lines can be clearly seen which should be avoided for communication (Figs. 3 and 5).
Space-based receiver
For the space-based analysis, we restrict the receiver size to D r = 10 m to make it more realistic for the current technological level. The optimal wavelength is now λ ≈ 300 nm which is limited by the telescope quality (Fig. 4). Noise levels are dominated by zodiacal light; α Cen is 42° from the ecliptic, resulting in noise levels of ≈ 0.1 γ nm−1 s arcsec−2 m−2 and a higher capacity of 2.83 bits photon−1. The signal decreases to 174 γ −1 for a maximum data rate of 494 bits s−1.
Power
At first approximation, data rate is a linear function of power, DSRγ ∝ P. This holds for constant capacity C th which however depends on the ratio of signal to noise, and thus decreases for decreasing signal. The effect is small for S ≫ N but becomes very considerable for N > S. As shown in Fig. 9, a capacity C th = 1 bits per photon is possible for N M ≤ 0.13 (noise photons per mode) in our standard example using M = 10−5 modes and receiver efficiency η = 0.5. Capacity is a logarithmic function of SNR, and the sweet spot appears between 0.1–5 bits photon−1, which is achievable for 10−6 < N M < 10 assuming 105 modes and η = 0.5.
Transmitter size
The transmitter size for a circular aperture scales as DSRγ ∝ D 2t assuming no technological limitations, which we identify as possible for current (earth 2017) technology at λ > 300 nm. Increasing the dish size to focus optical lasers is thus very beneficial for the data rate, and it is recommended to make the aperture as large and high-quality as possible.
Receiver size
The receiver size for a circular aperture scales as DSRγ ∝ D 2r, and we here relax the technological limitations: imperfect focusing will still collect all photons (signal), but collect more noise due to the larger beam width; the total effect is however much smaller. For a real application, this additional noise factor can be modelled.
Interstellar extinction
Extinction is largely irrelevant for the shortest interstellar distances, <1% in the optical to α Cen. Outside of the Lyman continuum (≈50 < λ < 91.2 nm), any frequency is equally suitable. The situation changes significantly for distances >200 pc, where optical extinction is >0.5 (compare Fig. 2). To examine the optimal choice of wavelength versus distance due to extinction, we have plotted the normalized photon rate in Fig. 10, and subtracted out the free-space loss. The optimal wavelength for space-based communication is limited by technology at 300 nm out to 200 pc, and increases to 3 μm for the longest paths in the galaxy. For an earth-based receiver, the lower limit is 420 nm due to limited atmospheric transmission and special care must be taken not to select a narrow absorption line.
In this calculation, we assumed uniform extinction of A(V) = 1.8 mag kpc−1 in the galactic plane (Whittet, Reference Whittet2003). In reality, however, the situation is much more complex. Extinction in the galactic plane can vary on small scales (because of individual molecular clouds), and on large (degree) scales (Schlafly et al., Reference Schlafly, Meisner, Stutz, Kainulainen, Peek, Tchernyshyov, Rix, Finkbeiner, Covey, Green, Bell, Burgett, Chambers, Draper, Flewelling, Hodapp, Kaiser, Magnier, Martin, Metcalfe, Wainscoat and Waters2016). Galactic communication with maximized data rates will require a precise measurement along each line of sight (communication path) to choose the best wavelength. If a civilization, or a club of civilizations, prefers to choose a single frequency for all distances, it will be at ≈3 μm. Then, long distance communication is near optimal (it would be prohibitive at shorter wavelengths), while data rates for short-distance communication are smaller by a factor of a few compared with individual optima.
Discussion
Assessment of achievable data rates
Achievable data rates are of order kbits s−1 KW−1 for a metre-sized probe at α Cen. For comparison, the NASA probe ‘New Horizons’ achieved a data rate of 1 kbits s−1 at P = 13 W from Pluto, and transmitted a total of 50 Gbits (5 × 1010 bits, buffered) over the course of 15 months. The transfer of an image as shown in Fig. 11 with a compressed volume of ≈400 kbits takes 7 min to transfer at 1 kbits s−1 for a P = 1 kW at α Cen, or days (to weeks with problematic SNR) at P = 1 W, which might be regarded as acceptable.
Photovoltaic energy is available at a level of kW m−2 at au distance from the star, so that a probe in orbit (perhaps decelerated by stellar photons, Heller & Hippke (Reference Heller and Hippke2017); Heller et al. (Reference Heller, Hippke and Kervella2017)) has no power issues for transmissions. A fly-by probe at 20% c, however, transverses the au distance in 17 min, translating into a data volume of order Mbits m−2 if the whole time were used for transmission (which is unrealistic, given that the target exoplanet is to be observed). Available photovoltaic energy decreases with the inverse square to the distance from the star, and by integrating over an exemplary trajectory with a closest encounter of 1 au to the star we can estimate the total collected photovoltaic energy, during the fly-by, of order kWh m−2. With this energy, perhaps stored on-board and used for later transmission, the probe can send a few Mbits m−2, i.e. a few high-quality images (Fig. 11). Alternative options would require an onboard energy source.
Onboard storage requirements
The data volume during the fly-by governs the size of the transmission buffer. ‘New Horizons’ carried a total of 16 GBs, which contained all the data it recorded during the fly-by, and which was transfered afterwards. The same scheme could be used for a fly-by at α Cen. If the probe starts to transmit after the observations, and transmits 1 bit s−1 (1 kbit s−1) for a total of 10 years remaining lifetime, it can transfer (and needs to store) a total of 31 Mbits (31 Gbits). Both are low numbers and can be stored with current (Earth 2017) technology at millimetre sizes and milligram masses.
Earth's rotation
A space-based receiver, for example at a Lagrangian point, can be used near-continuously. Earth-based telescopes, however, suffer from Earth's rotation (daylight) and weather. When ‘New Horizons’ encountered Pluto, the entire NASA Deep Space Network was online to ensure there were no breaks in reception. If the communication scheme with α Cen is the same, a large number of telescopes will be required. We can, however, replace (expensive) 39 m E-ELTs with a number of smaller telescopes. To replace one E-ELT in terms of aperture, ≈1500 telescopes with d = 1 m are required, or 24 telescopes with d = 8 m.
Laser line width, orbital motion, beam sizes
Transmitter and receiver are in relative motion, which results in a change of path length, as already noted by Messerschmitt (Reference Messerschmitt2013, Reference Messerschmitt2015). If the sender (receiver) is located on a planet which orbits a star, the Doppler shift will cause a shift in the sender (receiver) frequency. For example, earth's equatorial speed is 465.1 m s−1, or a frequency shift of 1.55 × 10−6. This is an order of magnitude smaller than current spectrographs (R = 100 000), but larger than typical high-power laser line-widths (350 MHz, or 6 × 10−6, Duarte (Reference Duarte1999)) by a factor of a few. Laser line width in the mHz range, although at low (10−12 W) power, have been demonstrated (Meiser et al., Reference Meiser, Ye, Carlson and Holland2009; Kessler et al., Reference Kessler, Hagemann, Grebing, Legero, Sterr, Riehle, Martin, Chen and Ye2012). For such small line widths, the shift would need to be modelled and compensated. Regarding noise per mode (atmospheric, zodiacal, etc.), very narrow line widths are preferred.
Narrow line widths may give rise to additional noise sources, namely instrumental frequency shifts in the sender and/or receiver, or a change in the interstellar scattering geometry, which may also result in non-Gaussian noise per sub-channel.
For the closest stars within a few pc, large optical telescopes (10–100 m) have diffraction-limited (adaptive optics) beam sizes smaller than typical orbits (au) of exoplanets. When using such tight beams in the transmitter, the position of the receiving telescope (e.g. on a planet in motion) needs to be known with high accuracy at the time of arrival of the photons (Sherwood et al., Reference Sherwood, Mumma, Donaldson, Mendell, Alred, Bell, Cintala, Crabb, Durrett, Finney, Franklin, French and Greenberg1992; Mankevich & Orlov, Reference Mankevich and Orlov2016).
Blend of probe and star
In the previous sections, we have neglected the noise flux from the target star. This is justified for sky-projected separations which allow for the use of coronographs, and suppress 10−9 (Guyon et al., Reference Guyon, Pluzhnik, Kuchner, Collins and Ridgway2006; Liu et al., Reference Liu, Ren, Dou, Zhu, Zhang, Zhao, Wu and Chen2015) of the starlight (4.25 × 106 γ s−1 m−2) at a separation of 1 au at Proxima Centauri. During most of the flight, the problem is much less severe because of the large proper motion of α Cen (3.7 arcsec yr−1, Kervella et al., Reference Kervella, Mignard, Mérand and Thévenin2016).
We can estimate data rates for this increased noise level within the Holevo bound for this situation, and get a capacity of the order of 10−5 bits s−1 W−1. Such a low data rate is insufficient for the transfer of images or other observational data, but may be sufficient for simple telemetry and onboard health status.
Current technological level and photon dimensions
The Holevo bound assumes the use of a number of modes to encode information into photons. The available modes in photons are their time of arrival (sometimes called the phase modulation), their frequency (or colour) and their polarization. Realizing 10−5 photons per mode will require many (>105) modes to encode the information. This can be done with a combination of colour, timing and polarization. Commonly used are time–frequency modulations. The usage of polarized light is less common, but might be beneficial for our case. Starlight is polarized only by a few percent (Fosalba et al., Reference Fosalba, Lazarian, Prunet, Tauber, Cecchini, Cortiglioni, Sault and Sbarra2002), so that the use of polarization, which is possible for lasers, can reduce noise levels by a factor of two.
We now examine currently available technology. For the sender, the shortest possible laser pulse length has decreased by 11 orders of magnitude during the last 50 years, from 100 μs in the free-running laser of Maiman (Reference Maiman1960) to 67 attoseconds (10−18 s, Zhao et al., Reference Zhao, Zhang, Chini, Wu, Wang and Chang2012). For a detailed history of the exponential improvements, see Agostini & Di Mauro (Reference Agostini and Di Mauro2004). While the pulse length is very short, the repetition rate is slower by many orders of magnitude.
The highest data rates are currently found in fibre-optic communication by sending pulses of light through an optical fibre, with a current record of the order of Tbits s−1 (1012 bits s−1) on one glass fibre (Maher et al., Reference Maher, Alvarado, Lavery and Bayvel2016). Commercial products are available with data rates 1–3 orders of magnitude below this value. The industry standard employs 100 channels with a channel spacing of 100 GHz (0.8 nm) between 1530 and 1612 nm with a typical bandwidth (frequency range) of 186–196 THz (International Telecommunication Union, 2012). Limiting factors are small bandwidth (82 nm, or b = 5%), the wavelength stability of lasers with thermal changes, signal degradation from nonlinear effects in optical fibres, inter-channel crosstalk and (clock) timing jitter.
On the receiver side, current photon-counting detectors can be relatively fast (timings below 10−10 s) and efficient (>90%) with a low dark count rate (<1 c.p.s.), but suffer from longer (10−7 s) reset times (Marsili et al., Reference Marsili, Verma, Stern, Harrington, Lita, Gerrits, Vayshenker, Baek, Shaw, Mirin and Nam2013). Classical photomultiplier tubes offer timings (and reset times) of 10−9 s (Dolgoshein et al., Reference Dolgoshein, Balagura, Buzhan, Danilov, Filatov, Garutti, Groll, Ilyin, Kantserov, Kaplin, Karakash, Kayumov, Klemin, Korbel, Meyer, Mizuk, Morgunov, Novikov, Pakhlov, Popova, Rusinov, Sefkow, Tarkovsky and Tikhomirov2006). Current photon detectors are fast enough to sample 10 GHz frequencies at the Nyquist limit (B < f/2, Nyquist, Reference Nyquist1928). These limits, however, are technological, and further improvements can be expected. The ideal instrument for high-mode communication would be a high-throughput, high-resolution spectrograph with low-noise, high-speed photon counters on each subchannel.
Bidirectional communication
The focus in this paper was the communication from a distant, small, power-limited probe towards home-base. The opposite way, perhaps to send new instructions, is comparably easier: Home-base has less stringent limits on aperture size and power. Telescope diameters might be larger by 1–2 orders of magnitude and power by several orders of magnitude. A major issue might be that the probe needs to ‘listen’ at the moment the photons arrive, and not spend the time sending, making observations, or in hibernation. A simple solution would be pre-arranged timeslots.
Comparison with the literature
In his ‘Roadmap to Interstellar Flight’, Lubin (Reference Lubin2016) recently approximated the communication flux as (his section 5.6, our notation) F = D 2P/(4d 2λ 2) which yields an overestimate by ≈11.7% compared with our equation (2).
In their ‘Search for nanosecond optical pulses’, Howard et al. (Reference Howard, Horowitz, Coldwell, Klein, Sung, Wolff, Caruso, Latham, Papaliolios, Stefanik, Zajac, Lemarchand and Meech2000) and Howard et al. (Reference Howard, Horowitz, Wilkinson, Coldwell, Groth, Jarosik, Latham, Stefanik, Willman, Wolff and Zajac2004) describe the received photon flux as (their equation (2), neglecting extinction; they set D = D r = D t). Numerically, this produces a received photon flux which is too high by ≈3.67×.
In their ‘Search for Optical Laser Emission’, Tellis & Marcy (Reference Tellis and Marcy2015) define the received photon flux in the same form as in our equation (2) (their equation (5)), but with an incorrect divisor of 2, resulting in 4× too many photons received.
The work by Horwath (Reference Horwath, Kingsley and Lemarchand1996) discusses beam widths and frequencies of interstellar laser communications, but neglects extinction and consequently proposes laser communication in the Lyman Hα line at 126 nm over distances of 3000 LY, which is impossible because of very high UV extinction (Fig. 2).
A more traditional interstellar radio communication design from α Cen has recently been published by Milne et al. (Reference Milne, Lamontagne and Freeland2016). It presents scenarios for antennas with sizes of 1–15 km on both sides, transmitting MW power at 32 GHz, achieving a data rate of Gbits s−1 (109 bits s−1). The antenna weight is mentioned as 40 000 kg, and the total space-ship weight is 107 kg. Clearly, if such masses and power can be sent to other stars, the question of communication will be trivial in comparison.
PyCom software package
We provide the Python-based software package PyCom as open source under the free MIT licenseFootnote 5. The repository provides function calls for the equations in this paper, a tutorial and scripts to reproduce all figures.
Conclusion
In this work (paper I of the series), we have set the framework of data transfer between telescopes, using the example of a light-weight, power-limited probe at α Cen. We have explored limiting factors such as extinction, noise and technological constraints. We have calculated optimal frequencies and achievable data rates.
The Holevo bound gives an upper limit of a few bits per photon for realistic signal and noise levels from a communication between a metre-sized probe at α Cen and a large (39 m) telescope on the Earth. The achievable data rate is of the order of bits per second per Watt. For a probe with a size of a few metres, and photovoltaic energy of KW m−2, power levels might be KW, resulting in data rates of the order of kbits s−1. The optimal wavelength for a communication with α Cen, at current technological levels, is 300 nm (space-based receiver) to 400 nm (earth-based) and increases with distance, due to extinction, to a maximum of ≈3 μm to the centre of the Galaxy at 8 kpc.
A critical requirement in this scheme is the coronagraphic suppression of the stellar background at the level of 10−9 within a few tenths of an arcsecond of the bright star, which has not been demonstrated yet. Further research on this topic is encouraged.
In paper II, the use of a stellar gravitational lens will be discussed. In paper III, we will relax technological constraints to explore the ultimate, most efficient interstellar communication scheme which yields insight into communication of more advanced life in the Universe, if it exists.
Acknowledgments
The author is grateful to René Heller, Duncan Forgan, John Learned and Tony Zee for helpful discussions, and to the Breakthrough Initiatives for an invitation to the Breakthrough Discuss 2017 conference at Stanford University.