Status, Challenges and Directions in Indirect Dark Matter Searches

Indirect searches for dark matter are based on detecting an anomalous flux of photons, neutrinos or cosmic-rays produced in annihilations or decays of dark matter candidates gravitationally accumulated in heavy cosmological objects, like galaxies, the Sun or the Earth. Additionally, evidence for dark matter that can also be understood as indirect can be obtained from early universe probes, like fluctuations of the Cosmic Microwave Background temperature, the primordial abundance of light elements or the Hydrogen 21-cm line. The techniques needed to detect these different signatures require very different types of detectors: air shower arrays, Cherenkov telescopes, neutrino telescopes, radio telescopes or particle detectors in balloons or satellites. Although many of these detectors were not originally intended to search for dark matter, they have proven to be unique complementary tools to the direct search efforts. In this review we summarize the current status of indirect searches for dark matter, mentioning also the challenges and limitations that these techniques encounter.


Introduction
The term "dark matter" was originally coined because of the need to explain the observed rotation curves of stars in galaxies and the peculiar velocities of galaxies within clusters. The velocities of these large gravitational systems were measured to be incompatible with the expectations based on Newtonian mechanics and the visible amount of matter in the system [1][2][3]. These observations seemed to provide indirect evidence for the existence of either a non-luminous matter component in the universe, or a deviation from the standard Newtonian mechanics as we use it to try to make sense of the universe at larges scales [4]. Additional matter is also needed to understand how the first galaxies could have formed from the small density perturbations imprinted in the cosmic microwave background [5] and to describe the large-scale structure of the universe as seen today [6,7]. Observations of gravitational lensing added to the list of evidence for the need of dark matter [8]. Although proposed modifications of gravity and other recent approaches [9][10][11] do not need dark matter at all, in this review we summarize the current experimental efforts aimed at probing indirectly the most popular "solution" to the dark matter problem: the particle solution.
The particle solution to the dark matter problem is based on the assumption that stable (or enough long-lived) relic particles exist, whose present density is determined by the thermal history of the early universe [13]. Incidentally, models needed to explain certain shortcomings of the Standard Model of particle physics do provide viable candidates for dark matter in the form of new weakly-interacting massive particles (WIMPs) in the GeV-O(100 TeV) mass range. If they are thermally produced in the early universe, the upper mass limit for WIMPs arises from theoretical arguments in order to preserve unitarity [14]. There is practically no lower limit on the mass of the dark matter candidates as long as they can perform their role of being abundant enough to solve the problem and they do not over close the Universe [15]. Higher masses than O(100 TeV) and/or stronger interactions can be accommodated in models where the dark matter candidates are not produced thermally [16]. So the dark matter zoo encompasses a wide class of particles with a wide range of interactions as illustrated in Figure 1. From a theoretical point of view, the connection between a cosmological problem (the need for additional matter at cosmological scales) and the possibility of explaining it with particle physics ideas is extremely compelling.
The term "indirect" appertaining to dark matter searches refers to those methods based on detecting the products of the annihilation or decays of dark matter candidates gravitationally accumulated in cosmological objects, in contrast with "direct" detection techniques where the interactions of dark matter particles from the local halo with a suitable target are searched for [17]. Still, these two techniques rely on detecting the dark matter that surrounds us. Usually not included in "indirect" searches is the use of early universe probes (Cosmic Microwave Background power spectrum, 21-cm astronomy and nucleosynthesis) to set constraints to the dark matter content and characteristics. There is no more indirect way of searching for something that using signatures that might have occurred billions of years ago, so we have included a short description of dark matter searches using those probes at the end of this review. Additionally, accelerator searches try to produce new particles in the controlled conditions of collider or fixed target interactions, that could be identified also as candidates for dark matter [18][19][20]. In any case, the problem remains to identify and characterize this new form of matter, if this is indeed the way Nature chose to function.
Any review about dark matter, be it about direct or indirect searches, faces at least two, I think insurmountable, problems: to be sufficiently up to date in a rapidly evolving field, and to give due credit to previous work by the innumerable scientists that have contributed to the field. I confess already here that I have given up in trying to solve these two problems. Instead, I have tried to provide a personal choice of topics and mention some general problems faced in the field of indirect dark matter searches that I think are relevant and that sometimes can have a bearing on the interpretation of results (the choice of halo model or the consideration of particle physics or astrophysical uncertainties for example). This means that I have not followed a historical approach (there are excellent reviews from that perspective [21,22]), neither I have tried to be complete in describing the theoretical models being usually, or less usually, considered (there are also excellent reviews on that [23][24][25]), or tried to be comprehensive in mentioning the results of every single experiment that exists in the field. I have also favoured citing published results than preprints or conference proceedings when possible, sometimes at the expense of not pointing the reader to the very latest results. With these caveats in mind, I hope this review is of use to some readers.

Dark matter in galactic halos
A viable particle candidate for dark matter must be stable in timescales comparable with the age of the Universe and have interactions with ordinary matter such it does not affect the evolution of the Universe in a way that leads to a large-scale structure that is incompatible with observations today. It must also be neutral, otherwise star formation and evolution would have been disturbed in ways incompatible with observations [26], and the formation of "dark atoms" in the interstellar medium should have already been revealed in spectroscopy studies of heavenly objects [27]. However, micro-charges, about a million times smaller than the charge on the electron, can still be accommodated if only a few percent of the dark matter would have such charge [28]. Aside these generic qualities, practically anything goes. Even though WIMPs are popular candidates because straightforward extensions of the Standard Model, like many flavours of super-symmetry, predict them [29], extremely light dark matter in the form of axion-like particles (ALPs) [30] or superheavy dark matter [31,32] are also well justified. We will therefore not use the term WIMP in this review since many of the experimental results summarized later can be reinterpreted in terms of The red, pink and blue colors represent hot dark matter (HDM), warm dark matter (WDM) and cold dark matter (CDM) candidates, respectively. Reprinted from [16] with permission from Elsevier. more general dark matter candidates. See [16,[33][34][35][36] and references therein for recent reviews of candidates and production scenarios beyond the vanilla WIMP miracle.
Accumulations of dark matter are expected in halos around galaxies, including the Milky Way, and in galaxy clusters. In order to detect signatures from these regions with high density of dark matter, the dark matter particles need to be able to annihilate between themselves, or decay. It is the stable standard model particles that result from these processes that constitute the potential signature for indirect searches. These can be neutrinos, photons, electrons, protons or even light nuclei like He and D (and their antiparticles).
The distribution of dark matter in the halos of galaxies is an important ingredient both in calculating expected observables and eventually in interpreting a signal. We know from N-body simulations of galaxy formation that galaxies are embedded in clumpy, rather complex dark matter halos that extend well beyond the visible galaxy [38]. There have been several proposals to describe the dark matter density distribution in galactic halos as a function of the distance from the centre of the galaxy [39][40][41][42][43][44]. The common feature of these profiles is a denser, typically spherically symmetric, region of dark matter in the centre of the galaxy, with decreasing density as the radial distance to the centre increases. Figure 2 shows a few examples of these profiles. Note that the slope of the density distribution varies with the distance, deviating from initial calculations in [39] that predicted a single power law distribution across the whole distance range. Where the different models diverge is in the predicted shape of the central region. Profiles obtained from N-body simulations of galaxy formation and evolution tend to predict a steep power-law type behaviour of the dark matter component in the central region [40,43], while profiles based on observational data (stellar velocity fields) tend to favour a more constant dark matter density near the core [41,42,44]. This discrepancy has been dubbed the core-cusp problem [45], and it is an unresolved issue (although see [46] for a recent claim that the cusp obtained in N-body simulations represents a numerical issue of the simulations themselves that do not reflect reality). For example, a recent fit to the motion of stars in the disk of the Milky Way [47] gives inconclusive evidence for a cored or cusped profile.  Following [49], a general parametrization of the dark halo in galaxies can written as 1 where r is the distance from the Galactic Centre, the parameters α, β and γ determine the shape of the halo and ρ 0 is a normalization constant depending on the model, i.e., depending on the choice of α, β and γ. Different combinations of these parameters can recover the halo profiles mentioned above. These profiles describe the smooth distribution of dark matter around galaxies. Any clumpy component must be added on top in order to describe the outcome of N-body simulations. Figure 2 shows that the different halo models proposed show agreement on the dark mater distribution at the location of the Solar System (about 8 kpc from the Galactic Centre) and beyond, but the predicted dark matter density distribution diverges considerably at distances near the centre of the galaxy [48]. This is of course also true when considering other galaxies and assuming the same type of profiles apply universally at any redshift. However universality might not be completely justified, and the shape of a dark halo can depend on local characteristics of any given galaxy. There are indications that halo profiles might depend on the size and surroundings of a chosen galaxy, since the original mass function of nearby satellites will affect the accretion of additional dark matter to the halo, causing the inner slope of the dark matter profile to deviate form a universal function like the ones in Equation 1 [50]. Indeed, recent results from the Gaia mission point to a more complex dark matter halo in the Milky Way than just a spherical distribution. Gaia will map the position and radial velocity of about one billion stars, providing and unprecedented 3-dimensional picture of our galaxy and invaluable data to precisely map the dark matter halo for the first time. Gaia's data release 2, along with data from the Sloan Digital Sky Survey(SDSS), have already provided indications that our dark matter halo has a non-trivial structure, with localized substructures due to the recent history of our Galaxy that might not yet be in equilibrium. Actually, the Gaia results seem to be in tension with the characteristics of the dark halo derived from previous stellar data [51][52][53][54][55]. The first implication of these results is that the Milky Way halo is more complex than initially thought, with local substructure that now can be mapped. This can have implications for direct dark matter searches whose expected signal depends on the assumed dark matter density in the solar neighborhood. But in the bigger picture, the Gaia results point to a dark matter halo which is strongly dependent on the specific history of the galaxy, determined by past accretion events from nearby satellite galaxies. Additionally, the integrated rate of supernova explosions through the evolutionary history of a galaxy can distort the distribution of dark matter in the halo due to the effect of the ejected expanding gas on the gravitational potential of the galaxy [56,57]. If dark matter halos strongly depend on the individual history of each galaxy, the assumption of universality of the dark matter profiles breaks down, with consequences for the interpretation of indirect dark matter searches as well.
Barring the caveats just mentioned, given a specific dark matter halo, the flux of cosmic-rays, gamma-rays or neutrinos arising from dark matter annihilations can be expressed as where i stands for the type of particle, σ A v , is the thermally averaged product of the dark matter self-annihilation cross-section times the dark matter velocity, dN i /dE i is the spectrum of species i produced by the annihilations, ρ DM is the dark matter density of the source, given by Equation 1 and m DM is the dark matter mass. The integral of the squared density along the line of sight (l.o.s.) is the so-called J-factor. It is source-specific, and encompasses our knowledge of the structure of the dark halo of the source in consideration. The J-factor depends on the halo profile chosen as well as on the distance to the source, and results from indirect dark matter searches must be given under the assumption of a specific halo model. Clearly, distances need to be estimated independently, which for far away objects can add to the final uncertainty of the result. J-factors are measured in GeV 2 cm −5 sr. As Figure 2 illustrates, using one parametrization of ρ DM or another can give results that differ by as much as orders of magnitude. Assuming a particle physics model which gives the expected particle spectrum, dN i /dE i , and a halo model, then an experimental measurement of dΦ i /dE i can be used to probe the two independent terms in Equation 2: σ A v versus the dark matter mass, m DM . Experimental efforts are usually kept as model-independent as possible, looking for signatures from generic candidates over a few benchmark masses that cover a wide mass range and, since the exact branching ratios of dark matter annihilations into different channels are unknown, analyses are typically optimized assuming 100% annihilation to a few characteristic channels. This brackets the expected spectrum from any realistic model which would produce an spectrum that is a weighted mixture of all the possible annihilation channels. If signals from dark matter decay instead of annihilation are to be probed, then Equation 2 needs to be slightly modified: the l.o.s. integral is over ρ and not ρ 2 (only one particle participates in a decay, while two are needed in an annihilation), only one m DM factor is needed in the denominator (for the same reason), and what is probed is 1/τ instead of σ A v , where τ is the candidate lifetime. In this case the units of the J-factor are GeVcm −2 sr.

Dark matter signatures from the cosmos
Since the expected signal from dark matter annihilation(decay) is proportional to ρ 2 DM (ρ DM ), natural sources to consider in indirect searches are regions in the Universe with an increased density of dark matter. Typically, our own Milky Way, close-by dwarf spheroid galaxies and galaxy clusters are considered in indirect searches. Dwarf spheroid galaxies are good candidates because they present high mass-to-light ratios, with little interstellar gas, and they seem therefore to be dominated by dark-matter. Additionally, there is no evidence of non-thermal emission from any of the 20+ known dwarf spheroids that could produce a background of high energy neutrinos, γ-rays or cosmic-rays so they are a clean target from the observational point of view. Even if they are small systems, with J-factors much smaller than our Galactic Centre, observations of several dwarves can be stacked in order to increase the sensitivity to a potential dark matter signal.
Each messenger has its quirks and features, though, both from the point of view of propagation and/or detection. In order to increase the sensitivity to specific sources that can be observed with different instruments, several collaborations have joined forces and obtained limits from their combined observations. Combining results from different experiments requires a consensus on the use of common J-factors for each target, statistical techniques and the treatment of the systematic uncertainties of the different instruments. The result of these efforts are usually stronger limits than those obtained by the individual collaborations. They also help unify analysis methods and the use of common external inputs (astrophysical quantities, cross sections, dark matter models...) across the community, which in the long run helps obtaining a coherent view of the field and makes it easier to compare, and even to combine, results from different messengers. One can really talk of a multi-messenger era in dark matter searches.

Gamma-rays
Gamma-rays propagate without deflection, so they point to their source, and are relatively easy to detect, the technology being available since long. Gamma-ray telescopes and gamma-ray detectors on space are able to point with great accuracy (about 0.1 degree for single photons above 10 GeV in Fermi-LAT) and the same for Cherenkov telescopes like MAGIC or H.E.S.S. for example (although at higher energies, above about 100 GeV) and they cover a wide range of photon energies, from 20 GeV in Fermi-LAT to few hundred TeV in Cherenkov telescopes. But absorption and backgrounds are always a concern when searching for new physics with photons from space. The Milky Way centre is a quite crowded region with γ-emitting objects and a diffuse component from unresolved sources and star formation regions. In this respect, the clean structure of dwarf spheroid galaxies is expected to provide a favourable environment for dark matter searches with gammas, and indeed strong limits on the dark matter annihilation cross section have been obtained from these objects. As an illustration of the results that can be obtained with these searches we show in Figure 3 the limits on the dark matter annihilation cross section as a function of dark matter mass. The left plot shows the results from the Fermi-LAT telescope [72], while the right plot illustrates the improvement that can be achieved when combining results from different experiments. In particular, the plot shows the combination of results from Fermi-LAT, HAWC, H.E.S.S., MAGIC and VERITAS [93]. Note the different energy reach of both plots, which reflects the different energy reach of the different detectors, Cherenkov telescopes contributing mainly to the high energy region.
The uncontroversial negative results from dwarf galaxies can be confronted with the status of dark matter searches with gammas from the Galactic Centre. An intriguing excess of gamma-rays in the few GeV region over the background expected from conventional sources has been claimed by several authors, both from EGRET data [94] and more recently from Fermi-LAT data [95]. But an excess is of course defined with respect to a background, and here is where things become tricky. The region around the Galactic Centre is one of the more complex in the sky, not least in gamma-rays. It is a star-forming region, surrounded with interstellar gas and with plenty of individual sources, like short-period pulsars and supernova remnants, as well as allegedly unresolved sources. Gamma-rays can be produced in point sources but also as a diffuse glow by electron bremsstrahlung, by inverse Compton scattering of electrons and positrons on the intense ambient radiation field in the region, or as products of the interactions of cosmic-rays with the interstellar gas. The calculation of the expected gamma-ray flux from all these processes, along with the galactic model used to track the diffusion of cosmic-rays in their way through the galaxy, is plagued with uncertainties which reflect in the predicted gamma ray emission from the Galactic Centre (which is the background when searching for dark matter) [96][97][98][99]. Indeed the EGRET excess was promptly explained by detailed calculations of the expected cosmic-ray flux without the need of new physics [100]. An additional important uncertainty in modeling the inner region of the Galaxy arises from the "Fermi bubbles" [101]. Their contribution to the gamma ray glow is difficult to estimate since information is lacking about their structure and emitted spectrum close to the Galactic Centre. Masking known point sources and subtracting the diffuse emission from the Galactic Centre in order to characterize the excess is therefore a really complicated task, plagued with unknowns.
But the location of the excess, within 1.5 • of the Galactic Centre, is indeed a place where an increased density of dark matter is expected. Given the unknowns in any dark matter model, the excess can of course be fitted as originating from dark matter annihilations, as indeed it has (see, e.g. [102][103][104][105] although the literature is vast on this topic). The main "result" of these claims is that a dark matter particle in the mass range of a few tens of GeV to a few hundred GeV, annihilating to soft (typically bb or τ + τ − ) or hard (typically W + W − or tt), channels can fit the data. The wide range of masses and annihilation channels proposed simply reflects the underlying uncertainties in estimating the background, as well as the rather unconstrained options that the dark matter solution provides. The allowed parameter space in dark matter models is ample enough to be able to fit the Fermi-LAT gamma ray excess without problem, even with the oversimplifying assumption of a single one-particle species annihilating through just one channel and taking into account limits from new particle searches in accelerators.
The authors in [97] have performed an extensive study of the Galactic Centre and used several models of the galaxy to estimate the expected gamma ray glow, finding that an excess over the expected gamma ray flux from the Galactic Centre can be obtained in all the models considered. But they also point out that the excess is still compatible, given the huge uncertainties, with the emission from standard astrophysical sources when considering contributions from cosmic-ray interactions in the interstellar medium or unresolved populations of pulsars. It is, of course, also compatible with ad-hoc dark matter models but, again, Occam's razor renders at this point the dark matter interpretation of the gamma ray excess from the Galactic Centre as, at most, a mere viable solution, but not the only one. A summary of this study, in comparison with the original claims of the excess, is shown in Figure 4. The figure shows the expected emission from the Galactic Centre obtained by the Fermi-LAT collaboration with a detailed modeling of the Galaxy (points tagged as "Sample") compared with the excess claimed by other authors, using different assumptions on the morphology of the Galaxy and background templates (so, in a strict sense, the curves are not really directly comparable). The shaded area represents the systematic uncertainty from the Fermi-LAT collaboration analysis, and it is obtained as the envelope of the predictions of the gamma-ray flux under different assumptions on diffuse gamma-ray emission, galaxy morphology, gas distributions in the galaxy, cosmic-ray distribution and propagation, etc but, in any case, without any new physics.

Neutrinos
Neutrinos also propagate undeflected and therefore point to their source, but they present additional challenges compared with photons. First, their extremely low cross section with matter means that large detectors and exposure times are needed to collect just an event from a given source, not to mention a handful of events that would be statistically significant by themselves. In this sense, the best bet of neutrino telescopes to function as dark matter detectors is to correlate their findings with other messenger or, ideally, more than one. This would, in turn, help interpreting any signal from gamma-rays or cosmic-rays, since an excess of events from a single messenger might be subject to interpretation as background from other processes rather than from dark matter, or it would need quite ad-hoc cooked theoretical models to explain why the dark sector annihilates or decays to some standard model particles but not others. Besides, neutrino detectors have a limited angular resolution, of the order of 1 o at O(100 GeV) neutrino energy, only reaching below one degree at TeV energies, and therefore pointing becomes challenging in searches for relatively low-mass dark matter candidates. Even in a perfect detector the angular resolution would be ultimately limited by the physics of the neutrino-nucleon cross section. Limited pointing resolution impacts the degree of background rejection, mainly atmospheric neutrinos, when pointing towards a source. On the other hand, neutrinos have the advantage of not loosing energy or being absorbed through their propagation over cosmic . The heart symbols represent analyses performed by the authors of [107] with public data, and not by the collaborations. Figure from [107] distances. Neither there is a significant background or foreground from astrophysical objects, so they remain an attractive signature. They are indeed the only possible messengers to use when looking for dark matter accumulated in dense objects like the Sun or the Earth. More on this below. The total number of signal events, µ s , expected at a neutrino telescope is given by, where T live is the exposure time, A eff ν α (Ω, E ν ) the detector effective area for neutrino flavour α, that depends on the detector response with respect to the observation angle and neutrino energy and Φ ν α (E ν ) is the neutrino flux for flavour α arising the annihilation processes. Note that the effective area of neutrino telescopes is not a fixed quantity, but it depends not only on neutrino energy and arrival direction, but also on the specific event selection performed. In the absence of a signal, the 90% confidence level limit on the number of signal events, µ 90 s , can be directly translated into a limit on the neutrino flux and, in turn, into a limit on either the velocity-averaged annihilation cross section σ A v or the dark matter lifetime as a function of dark matter mass, through Equation 2. As in the case with gamma-rays, sources with similar characteristics, where the signal can be expected to be also similar, can be stacked to increase the signal-to-noise ratio. This is typically done when considering nearby dwarf galaxies.
The current status of searches for dark matter with neutrinos from galaxies span a large chunk of dark matter masses as illustrated in Figure 5. The figure shows the current landscape of limits on the dark matter self-annihilation cross section as a function of dark matter mass obtained by the main players in the neutrino dark matter search industry, including many lowenergy experiments which are not the focus of this review [107]. Next-generation experiments are shown as dashed-lines, while the shaded areas show current limits, some of them obtained independently from the collaborations by the authors (marked with a heart). Results cover an impressive eleven orders of magnitude in dark matter mass, extending beyond the theoretical limit for a thermal relic, where results should be interpreted under different class of models than the vanilla WIMP scenario (see e.g. [33] and references therein).
In addition to our Galaxy and other objects at cosmological distances, the Sun and the centre of the Earth are also two possible sources of dark matter annihilations that could be detectable due to their proximity. The effect of the Sun gravitational attraction over the lifetime of the Solar System can result in the capture dark matter from the local halo through repeated scatterings of dark matter particles in orbits crossing the star [108][109][110][111][112]. The same applies for the Earth [113,114]. The only caveat is that only neutrinos from dark matter annihilations in the centre of these objects can escape their dense interior. So it is only neutrino telescopes that can pursue dark matter searches from the centre of the Sun and Earth (with some exceptions mentioned below).
The expected neutrino flux from dark matter annihilations inside the Sun depends, among other factors, on the capture rate of dark matter (which is proportional to the dark matter-nucleon cross section), and the annihilation rate (which is proportional to the velocity-averaged dark matter annihilation cross section). The evolution of the number density of dark matter particles, N DM , accumulated in the Sun interior follows the equation where Γ C is the capture rate per unit volume. The numerical factors take into account that annihilations consume two particles per interaction but there are only 1/2 of the particles available to form pairs. We have neglected a possible "evaporation" term, proportional to N DM , which is only relevant for very small dark matter masses, 4 GeV [111,115,116], below the range currently probed by neutrino telescopes due to their relatively high energy threshold (precisely a few GeV), but that can be an issue to take into account in future low-energy extensions of current projects. Equation 4 has an equilibrium solution given by which represents the steady amount of dark matter accumulated in the Sun if capture and annihilation have reached equilibrium. Since the Sun is around 4.6 Gyr old, it is usually assumed that equilibrium has indeed been achieved. In this case, the neutrino emission only depends on the capture rate, which is directly related to the dark matter-nucleon scattering cross section, σ DM−N , and this allows to set limits on σ DM−N from the measured (or rather, lack of) anomalous high-energy neutrino flux from the Sun. In models where equilibrium might not have been achieved, then an assumption on the dark matter self annihilation cross section is necessary to derive predictions on the neutrino signal from dark matter annihilations. The dark matter-nucleon cross section can be expressed in terms of a spin-dependent, σ SD , and spin-independent, σ SI , contributions [117] and, since the Sun is mainly composed of Hydrogen (75% of H and 24% of He in mass) [118], the capture of dark matter from the halo by the solar gravitational well occurs mainly through the spin-dependent scattering. Heavier elements constitute less than 2% of the mass of the Sun, but can still play a role when considering dark matter capture, since the spin-independent cross section is proportional to the square of the atomic mass number. These heavy elements can also take part in the spin-dependent capture process if dark matter presents momentum-dependent interactions with normal matter, giving rise to an increased capture rate in comparison to previous calculations, that can have a bearing in the interpretation of experimental results [119]. Even if searches for dark matter with neutrinos from the Sun are performed in the most model-independent way, there is of course the possibility to probe specific models, and both the main collaborations or external authors using public data have done so. These include limits on Kaluza-Klein dark matter arising in models of universal extra dimensions [120], inelastic dark matter [121], strongly interacting dark matter [122], or specific extensions of the MSSM [123][124][125], among others.
There are some theoretical scenarios that predict a flux of gamma-rays or cosmic-rays due to dark matter annihilations in the Sun. If the dark matter "world" is coupled to our usual world through long lived mediators that can escape the Sun interior and decay into Standard Model particles in-flight to the Earth, then other techniques can also search for dark matter accumulated in the Sun. In such models of secluded dark matter [126], an excess of gamma-rays or cosmic-rays from the direction of the Sun could also be interpreted as a signature of dark matter annihilations inside it [127][128][129][130]. Secluded dark matter models also present advantages for neutrino searches. Neutrinos loose energy on their way out of the dense solar interior, so the flux that would reach the Earth from annihilations in the solar core would be mainly low-energy (up to a few 100 GeV) independent of the mass of the dark matter annihilating inside the Sun. The correlation between dark matter mass and highest possible neutrino energy is therefore lost. But since the decay of the mediator in secluded models can occur outside the Sun, the neutrinos will reach the Earth without loosing energy. This is advantageous for neutrino telescopes, which usually have better sensitivities and pointing for higher neutrino energies, O(100 GeV) and above. These models add, of course, additional free parameters: the mediator mass, lifetime and decay branching ratios to different final states. Experimental results are therefore usually provided under different assumptions for the mediator mass and lifetime, for a given annihilation channel [131,132]. Figure 6 shows an example of the results that can be attained within secluded dark matter models. The figure shows limits on the mediator decay length as a function of the the annihilation rate in the Sun (left plot) and limits on the spin-independent dark matter-nucleon cross section as a function of dark matter (right plot) obtained with public data of IceCube and ANTARES [131]. Note that neutrino telescopes, IceCube in particular, provide the best limits on the spin-independent dark matter-nucleon cross section for dark matter masses above a few 100 GeV for the model described in the reference, improving even over direct search experiments like LUX, shown in the figure as the black dot-dashed line.
The Earth presents somewhat a different case, since the processes of capture of dark mater and annihilation in its core can not be assumed to have reached equilibrium. Additionally, the most abundant isotopes of the main components of the Earth inner core, mantle and crust, 56 Fe, 28 Si and 16 O [133], are spin-0 nuclei, so capture in the Earth is a good candidate to probe the σ SI component of the dark matter-nucleon cross section. From the experimental side, the mass of these isotopes produces mass resonances in the capture rate, so neutrino searches for dark matter annihilations from the centre of the Earth are going to be more sensitive to dark matter masses similar to the masses of those isotopes than to other masses. This is reflected in the shape of the limits obtained by IceCube, ANTARES and Super-Kamiokande, shown in Figure 7. The figure shows current limits on the spin-independent dark matter-nucleon cross section as a function of dark matter mass [134].
But perhaps of special interest is the recent discovery of a high-energy cosmic neutrino flux by IceCube [135][136][137], since it opens new possibilities for dark matter scenarios, specifically for annihilations or decays of very heavy dark matter, O(PeV), that is impossible to probe in accelerators. These are called top-down scenarios, where high energy neutrinos are produced from the decay of heavy relic particles, as opposed to bottom-up scenarios where cosmic neutrinos are expected to be produced from the decay of accelerated protons in a source or its surroundings. Additionally, extremely heavy (and extremely light) dark matter scenarios are becoming of interest since classic WIMP models have been disfavoured by accelerator results and the persistent negative results of direct searches. Indeed even the first two PeV cosmic neutrinos detected by IceCube were promptly interpreted within the framework of PeV mass decaying dark matter candidates (R-parity violating gravitinos, extra dimension fermions or hidden sector gauge bosons) [138]. Decay of such heavy particles into neutrinos would produce a monochromatic line, compatible with the IceCube events at the time. With currently an excess of neutrinos well beyond 7σ over the atmospheric neutrino background above 60 TeV, and no sources identified except for the blazar TXS 0506+056 [139,140], the interpretation of the IceCube excess as a signature of dark matter has spurred a lot of activity (e.g [141][142][143][144][145][146][147][148][149][150][151][152] among others). The most general approach is to assume a two-component flux where a fraction of the IceCube neutrino flux is assumed to originate from unidentified astrophysical sources, and a fraction of the flux comming from heavy dark matter decays. Note that dark matter decays would produce also a flux of high energy photons that can be recycled to lower energies through pair production and inverse Compton scattering on ambient radiation. So any model that explains the IceCube cosmic flux from heavy dark matter decay must be consistent with measurements of the gamma-ray sky. The low statistics of the IceCube data above 1 PeV do not allow for the moment to establish if the flux presents a cut-off or not. A sharp cut-off in the neutrino energy spectrum would favour the dark matter interpretation. Experimental results on the lifetime of heavy dark matter from IceCube [76], compared with limits from HAWC and Fermi-LAT, are shown in Figure 8. The IceCube result assumes a two-component fit, with the astrophysical component following a single power law. Lifetimes below 10 −27 − 10 −29 s are disfavoured at 90% CL, depending on the candidate mass.
The existence of a high-energy cosmic neutrino flux can be used to probe conventional (i.e. not extremely heavy, and even very low mass) dark matter in another, neat, way. Astrophysical neutrinos reach the Earth from far away, traversing dark halos of galaxies in their way, including the Milky Way. Interactions of these neutrinos with the dark matter could reduce the expected neutrino flux from specific directions, providing a unique way of probing the neutrino coupling  to the dark sector [144,145]. Although strongly model-dependent, these approaches can shed light on specific aspects of neutrino-dark matter interactions. An example of the type of results that can be obtained with these analyses is illustrated in Figure 9. The plot shows the result of a search for a depletion of cosmic neutrinos from the direction of the Galactic Centre due to neutrino-dark matter interactions. The results are cast in terms of the mass of a vector-like mediator of the interaction and the dark matter mass, as well as the logarithm of the coupling strength of the interaction (color code). Note that this method is sensitive to very low mass dark matter candidates, on the MeV-GeV range, and therefore complementary to other analyses based on the cosmic neutrino flux. The plot also shows that neutrino telescopes have access to an area of the parameter space, typically high mediator masses and high coupling constants, where limits from cosmological arguments based on analyses of the cosmic microwave background do not reach.

Cosmic-rays
Cosmic-rays (protons, electrons or light nuclei and their antiparticles) can also be produced in the hadronization of qq or W + W − pairs from dark matter annihilations or decays. But using cosmic-rays as probes of dark matter poses challenges as well. In order to have access to particle identification, charge separation and energy measurement one must deploy real particle physics detectors in space, ideally with calorimetry and a magnetic field for charged-particle tracking capabilities, a technical challenge by itself. But AMS [153], DAMPE [154] or CALET [155] have demonstrated the feasibility of this type of detectors and produced impressive physics results. The limitations of using cosmic-rays as dark matter probes arise, though, from fundamental physics. Cosmic-rays are charged particles (neutrons being too short lived to be of relevance) and propagation through the Galaxy can randomize their arrival directions and prevent accurate pointing to a given object. Additionally, cosmic-rays can also lose energy through inverse Compton scattering and synchrotron radiation during propagation, so their energy spectrum at Earth might differ from that where dark matter annihilations take place. But the detection of anti-matter (anti-protons, positrons or light anti-nuclei) still provides a good handle in dark matter searches since anti-matter from astrophysical sources is relatively rare at the Earth. Due to its high probability of annihilation with normal matter, any anti-matter source needs to be relatively close to the Earth. But this covers the Galactic Centre and Halo, so there are plenty of predictions of what a detection of an excess of anti-protons or positrons would mean in terms of dark matter. But even with anti-matter searches one has to deal with background from close-by sources, like pulsars, which can produce an (anti)cosmic-ray flux that can be indiscernible from a dark matter signal. Indeed the positron flux (compared to the electron flux) has been the source of controversy since results from the PAMELA satellite [156][157][158] and the ATIC [159] balloon experiment showed an excess of positrons at energies above ∼10 GeV over the expectation from standard galactic primary cosmic-ray propagation and interactions, including the production of secondary electrons and positrons [160,161]. AMS02 [162,163], CALET [164] and DAMPE [165] confirmed the excess and showed that it extends to the TeV region, but with a cutoff at around 1 TeV. The shape and normalization of the positron excess have given rise to a wealth of models explaining the excess from dark matter annihilations. Some of these solutions are quite contrived since, for example, the anti-proton flux does not show such a pronounced deviation from the expected background (there remains to be seen if there is an antiproton excess at all, see below) and therefore an explanation of the positron excess based on dark matter annihilations should prevent the same dark matter to produce a significant excess of anti-protons. These have been called leptophilic models, e.g. [166][167][168][169] and they are an example of the challenges posed by trying to explain spectral features by invoking multi-component contributions to the spectrum. A fit will always get better the more free parameters are added to the function to fit, without this reflecting on the physical relevance of the parameters. For anyone willing to wield Occam's razor, there is a simpler, standard astrophysical explanation to a positron excess in the cosmic-ray flux: it can indeed be produced by localized pulsar wind nebulae close enough to the Earth [170][171][172][173][174][175]. In a pulsar wind nebula, positrons are created by electromagnetic cascading from seed electrons that have been ejected from the rapidly spinning neutron star. This electron and positron "wind" eventually creates a shock when interacting with the older ejecta of the original supernova remnant. A fraction of the positrons might escape this complex system into the interstellar medium and become the galactic positron flux. Pulsars are, thus, antimatter sources in the Galaxy. There are known, close enough pulsars to the Earth that can provide the needed positron flux to explain the measured excess over the vanilla, point source free, background. Additionally, a pure dark matter explanation of the positron excess is also strongly disfavoured by the fact that it should have produced a signal in gamma-rays. The lack of a signal in Fermi-LAT searches for dark matter in dwarf spheroid galaxies have set a limit on the dark matter annihilation cross section into channels that would produce an e + e − flux which is incompatible with the measured excess [176,177]. Again, very few and very contrived models can escape the strong Fermi-LAT constrains and still be compatible with the measurements of PAMELA, ATIC, AMS, CALET and DAMPE.
We mentioned above antiprotons, and the story with them as a dark matter signal is not so simple as it might have been suggested. Antiprotons are produced in interactions of protons and light nuclei with the interstellar medium. So, given the cosmic-ray flux, the distribution of matter in the Galaxy, the anti-proton production cross section and a model for propagation, the standard background of anti-protons can be, in principle, calculated. Antiprotons from dark matter annihilations would be produced as final state hadrons in annihilations to qq or W + W − pairs, and they would appear as an excess over the standard background [178][179][180]. Figure 10 shows the result of looking for dark matter annihilations into antiprotons with AMS-02. Limits on the velocity-weighted annihilation cross section as a function of dark matter mass for annihilation into bb (left) and W + W − (right) are shown, along with the 1σ and 2σ sensitivity bands.
The thing is that the uncertainties in the ingredients just mentioned above that enter into the calculation of the standard background are larger than current measurements of the antiproton flux by AMS02. The strongest culprit is the uncertainty on the anti-proton production cross section in p-p and p-He collisions, which is just known to about 20%, slightly depending on energy, from recent measurements of NA61 and LHCb [181]. Additionally, uncertainties in the propagation in the Galaxy also play a determinant role. Anti-protons, as charged particles, diffuse in the galactic magnetic field and a complex transport equation, which must include diffusion, convection and reacceleration, with uncertainties in each term, needs to be solved numerically to estimate the flux at the Earth. Although there are recurrent claims of a dark matter signal in cosmic-rays from the global fit industry [182,183], the fact remains that, within current uncertainties on the cosmic-ray propagation and on the elementary production cross sections, the data remain compatible with the expected background from standard processes. The simple fact that different authors performing what in principle is a very similar analysis [183,185] reach contradicting conclusions on the significance of the anti-proton excess (4.7σ versus 2.2σ) points to the effect of the uncertainties involved in these type of calculations, mainly due to the cosmic-ray propagation in the galaxy.
A way out of this impasse might be the precise measurement of the anti-deuteron and antihelium fluxes, which should be correlated with the anti-proton flux. A comprehensive analysis of the different anti-matter cosmic-ray nuclei could help disentangle the standard or exotic origin of these fluxes [184,185].

Constrains from cosmology: CMB, BBN and the 21-cm Hydrogen line
There are additional, independent, ways to infer the presence and characteristics of dark matter based on its imprint on the early evolution of the Universe, be it on the Cosmic Microwave Background (CMB), on the production of primordial light elements through Big-Bang Nucleosynthesis (BBN) or on the 21-cm power spectrum of Hydrogen (see [186] for a review and references therein). The annihilation of dark matter present in the primordial soup injects additional energy in the form of photons, leptons or hadrons into the expanding universe. These will heat, ionize and excite the medium leaving imprints in the expected angular power spectrum that, assuming a given dark matter candidate, can be calculated and compared with observations. Turning the argument around: the consistency of the ΛCDM model with observations provides a strong constrain on the characteristics of any new dark matter particle species. These methods have the advantage that they are free from astrophysical uncertainties like halo models of galaxies or dark matter distribution in clusters. They also probe a quite ample range of the history of the Universe, from the recombination era (z=1000) to the end of reionization (z∼6).
Going from higher to lower redshifts, the first effect of dark matter annihilations to consider is on the temperature and polarization power spectra of the CMB. The CMB conveys detailed information about the last scattering surface and, if dark matter decayed or annihilated in the early universe, the extra energy deposited in the plasma could delay recombination or smooth out the primordial density fluctuations. Such an effect would be visible in the CMB data today [187][188][189]. However, not all the energy released in the annihilation of two dark matter particles is absorbed into the surrounding medium and it is usual to express results from CMB analyses in terms of an efficiency factor, f (z), which represents the fraction of the annihilation energy that goes into heating and ionizing the surrounding plasma. This factor depends both on redshift and the specific dark matter model through the branching ratios to particles that couple to the plasma, typically photons and electrons and positrons, but it lies in the range 0.001-1 [190].
The redshift dependence of f (z) is less acute than its dependence on the dark matter model, since the universe is evolving fast, and the effects of injecting energy into it are relevant only during a limited z range, typically between 1000 and 600, so one can approximate f (z) with an effective constant value f eff which practically only depends on the properties of the dark matter model assumed. Following [190], the energy released per unit volume from dark matter annihilations as the universe expands can be written as where ρ crit the critical density of the Universe today, m DM is the mass of the dark matter particle, and σ A v is the thermally-averaged annihilation cross-section times the dark matter velocity. Although results from CMB observations are sometimes given as a function of f eff σ A v , detailed calculations of the effect of annihilations assuming a simple dark matter model can be performed in order to break the degeneracy and express the results as limits to the σ A v as a function of the dark matter mass. This has the advantage of presenting results which are directly comparable with other indirect search techniques mentioned in this review, while not being more model dependent. On the contrary, rather less model dependent since, as mentioned above, source modeling does not enter here aside from the usual assumptions of ΛCDM cosmology. There is one caveat, though, when comparing limits on σ A v from indirect experiments, probing dark matter accumulated in galactic halos, with limits on σ A v from CMB. Note that the direct comparison works under the assumption of s-wave annihilations. For higher order contributions or more complex models, with velocity-dependent annihilation cross section for example, the comparison of limits should be taken with care since they probe very different dark matter velocity regimes. A thermalized 100 GeV particle in a 3000 K plasma (the temperature of the Universe at recombination) has a typical velocity of the order of 100 m/s, while dark matter particles in galactic halos have typical velocities of 200 km/s. The left plot of Figure 11 shows limits on σ A v as a function of dark matter mass, where a detailed calculation of the effects in the CMB anisotropy induced by the injection of photons, electrons, muons and W's from dark matter annihilations in the expanding plasma has been used to calculate f eff [191][192][193].  [191]. (Right) 95% confidence level limits on the velocity-weighted dark matter annihilation cross section as a function of dark matter mass for different annihilation channels, obtained from an analysis on the primordial abundances of light elements from big-bang nucleosynthesis. Figure from [195].
Photons, electrons and positrons couple readily to the plasma, but quarks or gluons will hadronize into strongly interacting particles, typically mesons and baryons (and their antiparticles). Short-lived particles (with lifetimes much shorter than the strong interaction rate at the temperature of the surroundings) will decay, reinforcing the energy dumping into the plasma and inducing photo-dissociation of already formed light elements. But stable particles, like protons (or neutrons for all purposes) will become part of the expanding plasma, ready to participate in nucleosynthesis. Since light-element synthesis depends on the baryon-to-photon ratio, additional stable baryons in the plasma due to dark matter annihilations, added to the photo-dissociation mentioned above and pion-exchange interactions that will drive proton-neutron conversion, can change the rate of certain reactions and produce primordial element abundances incompatible with observations. Again, turning the argument around, the observed abundance of primordial elements can serve to set limits on the amount of dark matter annihilations during the nucleosynthesis era [194][195][196][197]. The main uncertainties of this procedure are the uncertainties in measuring the primordial element abundances, a difficult measurement due to the intrinsic contamination of non-primordial elements in any source considered. Measurements of some elements have reached a precision of less than 1% ( 4 He or D), but others like 7 Li remain at about 20% [198]. Even with these caveats, BBN arguments allow to set quite stringent limits on σ A v as a function of dark matter mass, as it is shown in the right plot of Figure 11.
After recombination, the Universe enters a phase where primordial atoms have been formed and the ambient radiation does not have enough energy to ionize them. This is called the "Dark Ages", and it lasts until about z∼200, when the first large structures start to form by gravitational growth from the primordial fluctuations. As matter collapses into structures it heats up, emitting ultraviolet radiation that can interact with the surrounding gas, specifically inducing a strong absorption at the resonant frequency of the hyper-fine transition between the triplet and the singlet levels of the hydrogen ground state. These two levels are separated by 5.9 × 10 −6 eV, which corresponds to a wavelength of 21.1 cm and a frequency of 1420 MHz. Today, the redshifted absorption signal appears typically in the 30-200 MHz range and opens the possibility to use radio telescopes as instruments to study the early Universe. The 21-cm line has been used to measure rotation curves of galaxies in a complementary way to optical measurements. But the primordial 21-cm sky follows the matter distribution at the reionization era and provides additional information to CMB cosmology [199]. At about z∼6 reionization ends and the whole universe remains ionized until the present epoch. But up to redshifts about z∼50, the 21-cm signal is expected to be negligible. The presence of dark matter annihilations or decays can change this timeline, providing a continuous source of energetic electrons, positrons and gamma-rays that could start the reionization at an earlier time and leave a measurable effect in the 21-cm map [200][201][202][203][204]. Ideally, one could even distinguish the time evolution of dark matter annihilations, which is different from decays. Since the annihilation rate is proportional to N 2 DM , annihilations can dominate at early times, when the cosmic expansion has not diluted the number density, but they could reignite at a later stage, when dark matter has collapsed into primordial halos, providing a local enhancement of the dark matter density. Since decays are just proportional to N DM , the energy injection into the surrounding medium is uniform with time [205,206]. Given the wide energy spectrum of the annihilation/decay products of dark matter (up to the dark matter mass) and the resonance character of ionization, only a small fraction of the injected energy from dark matter can be used for ionization. This limits the dark matter models that can be probed by 21-cm line cosmology to low-mass, typically O(MeV), candidates. Only a tiny fraction of the spectrum of annihilations of heavier candidates is effective for ionization, so the strength of the method decreases rapidly with dark matter mass [207][208][209][210][211].
As we showed for the case of the CMB, the energy injected into the intergalactic medium due to dark matter annihilations can be related to the annihilation cross section and candidate mass. There are several ways to express this, as fraction of ionized Hydrogen as a function of redshift, as the temperature of the Hydrogen gas as function of redshift or, what I think gives a more visual connection to the dark matter characteristics, as the rate of energy deposition from annihilations as a function of redshift [212], where N DM,0 is the current number density of dark matter particles, N b,0 is the current baryon number density, f abs is the fraction of the energy released by the dark matter annihilations that goes into ionization and, as in Equation 6, m DM is the mass of the dark matter particle and σ A v is the thermally-averaged annihilation cross-section times the dark matter velocity. The calculation of the left hand side of Equation 7 is not easy since the injection of energy occurs in a complex and expanding medium, but once a model for the evolution of the intergalactic medium is in place, observations of the 21-cm primordial intensity in the sky (proportional to the injected energy from dark matter annihilations) can be used to constrain σ A v as a function of m DM . The main foreground comes from the contribution from later processes, like radio emission from synchrotron processes in our galaxy or from other galaxies and clusters of galaxies, which has to be carefully dealt with.
There are several radio observatories in operation that have the sensitivity to set competitive limits on dark matter annihilations and, indeed, observations from the EDGES telescope [213] have been used to set limits on dark matter characteristics. The EDGES telescope measured the brightness temperature of the 21-cm line relative to the CMB, finding that the absorption profile is consistent with expectations. But it also found that the relative amplitude of the absorption is more than a factor of two larger than the predictions from standard cosmology at a 3.8σ confidence level. Since we are talking of a relative measurement of T 21cm with respect to T CMB , the discrepancy can be interpreted either as the primordial gas at z∼20 being colder than expected, or the background radiation being hotter than expected (T 21cm and T CMB denote, respectively, the temperature of the ambient gas and of the CMB at the end of the Dark Ages). Standard explanations assuming increased radiation from the early stars and/or a slightly different decoupling redshift between matter and radiation can (more or less) accommodate the EDGES measurement, since there are still large uncertainties on the early star formation history of the Universe. But the result from EDGES can also be interpreted in terms of dark matter. Indeed, interactions of dark matter with surrounding baryons or the injection of energy due to dark matter decays or annihilations can change the thermal history of the hydrogen gas and accommodate the observation by EDGES [202,[214][215][216]. The contribution of dark matter annihilation products to heating the background radiation can lead to a suppression of the observed absorption (even erasing it if the energy injection is too large), so the observation of an anomalous absorption level in the 21-cm line sky can be translated into bounds on dark matter annihilations or decays. On the other hand, interactions between dark matter and surrounding baryons can lower the ambient gas temperature, and can explain the observed feature if the dark matter species is light (below a few GeV) and its interaction cross section lies above a few 10 −21 cm 2 [216]. Figure 12 shows an example of the constraints obtained from the EDGES observation for two annihilation channels, and assuming that the energy deposited from the annihilations is absorbed by the surrounding plasma with some delay [202]. This needs to be modeled and adds a source of uncertainty in the calculations but it is justified for some annihilation products that can penetrate de gas before fully interacting with it. In other cases, like short-lived particles, a "prompt" energy deposition approximation is more accurate. See [202] for additional limits obtained under the prompt energy deposition approximation. Note that the limits reach lower masses than other indirect searches, but the caveat mentioned above about the different average velocities of dark matter today (as probed by annihilations in the halo of our Galaxy or nearby galaxies) and at higher redshifts (as probed by the 21-cm line) remains when trying to compare results.
Constrains on the dark matter annihilation cross section can also be obtained by using radio telescopes to target specific sources, in a similar way that it is done with gamma-ray and neutrino telescopes. The technique is based on looking for synchrotron radio emission from the charged products (mainly e − and e + ) of dark matter annihilations in the halo of a suitable galaxy. The author in [217] used Draco as early as 2002 to set limits on the dark matter velocity averaged self annihilation cross section from VLA observations. More recently, the LOFAR collaboration has carried out a competitive search for dark matter in the dwarf galaxy Canis Venatici [218]. The results of this technique depend on the diffusion of the charged particles once emitted and on the profile of the local magnetic field, two quantities that are not well known, so the results are ineluctably model-dependent. But with reasonable assumptions on the structure and magnetic field strength from a given source, this method yields competitive results to gamma-ray, cosmic-ray or neutrino dark matter searches from dwarf galaxies as illustrated in Figure 13.

Discussion
Even that severe constraints from accelerator experiments have closed a big chunk of the supersymmetric parameter space, which was originally a favourite to provide a dark matter candidate Fig. 13 Limits on the velocity-weighted dark matter annihilation cross section as a function of dark matter mass from LOFAR, assuming annihilation into e + e − . Results are dependent on the galactic diffusion model used and assumed magnetic field strengths, and results are shown for three different models (blue dotted, black full and red dashed lines). Figure from [218]. See reference for details in the diffusion models used. with the right mass and interaction strength, the possibilities for particle candidates remain quite wide: models with very low or very high masses and/or very weak or strong interactions are still viable and escape current accelerator constraints. The aim of this review is to give a glimpse on the signatures looked for in indirect searches for dark matter, accompanying them with examples of recent results and hopefully conveying to the reader that indirect searches for dark matter is a complex and multidisciplinary field, currently driven by observations. Different messengers suffer from different backgrounds and systematics, and any potential future signal will have to be evaluated by correlating the results of all experiments.
The sensitivity of neutrino telescopes, gamma-ray telescopes and cosmic-ray detectors has improved enormously, sometimes by orders of magnitude, in the last decade, but the fact remains that no experiment has reported any positive signal so far. Searches for the products of dark matter annihilations in our galaxy or nearby galaxies with gamma-ray telescopes provide stringent limits to the dark matter annihilation cross section, mainly at relatively low dark matter masses, in the region of a few GeV to ∼100 GeV, but they suffer from complex backgrounds. Dwarf spheroid galaxies are a relatively background free sources that are commonly used in these kind of searches. Neutrino telescopes provide a complementary search approach, currently more sensitive to high dark matter masses, O(100 GeV) and above, while searches with charged cosmic-rays suffer from uncertainties from propagation and relevant cross sections, like antiproton production cross section, needed to understand the production mechanisms at the source and as a background. But they also provide competitive results, comparable to the two other techniques just mentioned. Large radio arrays have recently joined the effort of searching for annihilations of dark matter in dwarf galaxies by looking for anomalous radio emission from synchrotron radiation of e + and e − . There are new astrophysical uncertainties involved in this kind of technique due to the needed knowledge of the magnetic field of the source and the propagation of the electrons. But results look promising, and they are competitive with the results from gamma-rays, neutrinos and cosmic-rays under certain assumptions of the structure of the source. So these four, let us call them source-based, searches provide really complementary results and, taken together, will really help characterizing any signal.
On top of the source-based searches, early universe cosmology presents an independent way to search for dark matter, based on the effects of dark matter annihilations (or decay) in the evolution of the early Universe. Additional energy deposition on the primordial plasma, from the CMB era to the first formation of structures, must not cause a distortion of the large-scale structure of the Universe as we see it today. Quite stringent limits on the annihilation or decay characteristics of dark matter can be extracted from such requirement.
In the big picture of things the question remains: is the dark sector really so simple as one stable particle species, while the visible sector comprises several fundamental particles and families? Or, even more radically: is really the solution to the dark matter problem a particlebased one? From a pure experimental point of view, there is currently no evidence for or against either of these options. The particle solution was historically adopted since extensions of the Standard Model of particle physics predict the existence of new particles, and some of them have the characteristics needed of a dark matter candidate (the "WIMP miracle"), and in order to reduce the parameter space, experimental searches are simply interpreted in terms of a single particle solution. The constraining power of the results presented in this review would be lost if several candidates would have to be taken into account. Not least, the hierarchy of the masses, strength of interactions and annihilation branching ratios of the different candidates would be unknown. This poses a limitation from the experimental point of view, but it should not stop theorists from exploring such scenarios [219][220][221][222][223][224][225].
As for the second question above, there are proposed ways out from the dark matter conundrum without the need of dark matter. Modified Newtonian Dynamics (MOND) [9], and extensions thereof, is the more completely developed approach so far, but it requires a modification of Newtonian gravity (or General Relativity for that matter), theories that are understood to be universally valid and, therefore, any proposed modification is, at a minimum, philosophically difficult to swallow. MOND really solves the problem it set out to solve: explaining the flat rotation curves of galaxies taking into account only the visible baryonic matter. It tends to fail though when trying to describe complex halos with velocity streams perpendicular to the galactic plane [226], and has serious difficulties with early universe cosmology, the formation of large scale structure and, specifically, the speed of gravitational waves [227,228]. Not that ΛCDM cosmology is free from problems (a singularity at t = 0 and the need for an ad-hoc cosmological constant for example. Not to mention the need for a new type of dark matter itself). It is not the aim to review here the advantages of ΛCDM cosmology over alternative theories of gravity or vice-versa, but I would just like to finish reminding the reader that the lack of any experimental signal so far from dark matter searches really forces us to keep all our options open.