In the series of recent posts I’ve made about the Milky Way, I missed an important reply made in the comments by Francois Hammer, one of the eminent scientists doing the work. I was on to writing the next post when he wrote it, and simply didn’t see it until yesterday. Dr. Hammer has some important things to say that are both illustrative of the specific topic and also of how science should work. I wanted to highlight his concerns with their own post, so, with his permission, I cut & paste his comments below, making this, in effect, a guest post by Francois Hammer.
There are two aspects weโd like to mention, as they may help to clarify part of the debate: 1- When saying โGaia is great, but has its limits. It is really optimized for nearby stars (within a few kpc). Outside of that, the statisticsโฆ leave something to be desired. Is it safe to push out beyond 20 kpc?โ, one may wonder whether the significance of Gaia data has been really understood. In the Eilers et al. 2019 DR2 rotation curve, you may see points with small error bar up to 21-22 kpc. Gaia DR3 provides proper motion (systematics) uncertainties that are 2 times smaller than from Gaia DR2, so it can easily goes to 25 kpc or more. The gain in quality for parallaxes is indeed smaller (30% gain). However, our results cannot be affected by distance estimates, since the large number of stars with parallax estimates in Wang et al. (2023) is giving the same rotation curve than that from (a lower number of) RGB stars with spectrophotometric distances (Ou et al. 2023), i.e., following Eilers et al. 2019. And both show a Keplerian decline, which was already noticeable with DR2 results from Eilers et al 2019. The latter authors said in their conclusions: โWe do see a mild but significant deviation from the straightly declining circular velocity curve at Rโ19โ21 kpc of ฮvโ15 km sโ1.โ Our work using Gaia DR3 is nothing else than having a factor 2 better in accounting for systematics, and then being able to resolve what looks like a Keplerian decrease of the rotation curve. We may also mention here that one of us participated to an unprecedented study of the kinematics LMC (Gaia Collaboration 2021, Luriโs paper), which is at 50 kpc. Unless one proves everything that people has done about the LMC and MW is wrong, and that the data are too uncertain to conclude anything about what happens at R=17-25 kpc, the above clarifications about Gaia accuracy are truly necessary for people reading your blog. 2- The argument that the result โviolates a gazillion well-established constraints.โ has to be taken with some caution, since otherwise, no one can do any progress in the field. In fact, the problem with many probes (so-called โsatellitesโ) in the MW halo, is the fact that one cannot guarantee whether or not their orbits are at equilibrium with the MW potential. This is the reverse for the MW disk, for which stars are rotating in the disk, and, e.g., at 25 kpc, they have likely experience 7-8 orbits since the last merger (Gaia-Sausage-Enceladus), about 9 billion years go. In other words, the mass provided by a system mostly at equilibrium, likely supersedes masses provided by systems that equilibrium conditions are not secured. An interesting example of this is given by globular clusters (GCs). If taken as an ensemble of 156 GCs (from Baumgardt catalog), just by removing Pyxis and Terzan 8, the MW mass inside 50 kpc passes from 5.5 to 2.1 10^11 Msun. This is likely because these two GCs may have come quite recently, meaning that their initial kinetic energy is still contributing to their total energy. A similar mass overestimate could happen if one accounts the LMC or Leo I as MW satellites at equilibrium with the MW potential.โจSo we agree that near 25 kpc the disk of the MW may show signs of less-equilibrium, or sign of slightly less circular orbits due to different phenomenas discussed in the blog. However, why taking into account objects for which there is no proof they are at equilibrium as being the true measurements? In our work, we have considerably focused in understanding and expanding the whole contribution of systematics, which may comes from Gaia data, but also from assumptions about stellar profile (i.e., deviations from exponential profiles), from the Sun distance and proper motion and so on. You may find a description in Ou et al.โs Figure 5 and Jiao et al.โs Figure 4, both showing that systematics cannot gives much more than 10% error on circular velocity estimates. This is an area where we are considered by the Local Group community as being quite conservative, and following Gaia specialists with who we have worked to deliver the EDR3 catalog of dwarf galaxy motions (Li, Hammer, Babusiaux et al 2021) up to about 150 kpc. Jiao et al. paper main contribution is the fair accounting of systematics, which analysis shows error bars that are much larger than those from other sources of errors especially in MW outskirts (see Fig. 2).
Francois Hammer, 24 September 2023
The image at top is Fig. 2 from Jiao et al. illustrating their assessment of the rotation curve and its systematic uncertainties.
I am primarily an extragalactic astronomer – someone who studies galaxies outside our own. Our home Galaxy is a subject in its own right. Naturally, I became curious how the Milky Way appeared in the light of the systematic behaviors we have learned from external galaxies. I first wrote a paper about it in 2008; in the process I realized that I could use the RAR to infer the distribution of stellar mass from the terminal velocities observed in interstellar gas. That’s not necessary in external galaxies, where we can measure the light distribution, but we don’t get a view of the whole Galaxy from our location within it. Still, it wasn’t my field, so it wasn’t until 2015/16 that I did the exercise in detail. Shortly after that, the folks who study the supermassive black hole at the center of the Galaxy provided a very precise constraint on the distance there. That was the one big systematic uncertainty in my own work up to that point, but I had guessed well enough, so it didn’t make a big change. Still, I updated the model to the new distance in 2018, and provided its details on my model page so anyone could use it. Then Gaia data started to pour in, which was overwhelming, but I found I really didn’t need to do any updating: the second data release indicated a declining rotation curve at exactly the rate the model predicted: -1.7 km/s/kpc. So far so good.
I call it the RAR model because it only involves the radial force. All I did was assume that the Milky Way was a typical spiral galaxy that followed the RAR, and ask what the mass distribution of the stars needed to be to match the observed terminal velocities. This is a purely empirical exercise that should work regardless of the underlying cause of the RAR, be it MOND or something else. Of course, MOND is the only theory that explicitly predicted the RAR ahead of time, but we’ve gone to great lengths to establish that the RAR is present empirically whether we know about MOND or not. If we accept that the cause of the RAR is MOND, which is the natural interpretation, then MOND over-predicts the vertical motions by a bit. That may be an important clue, either into how MOND works (it doesn’t necessarily follow the most naive assumption) or how something else might cause the observed MONDian phenomenology, or it could just be another systematic uncertainty of the sort that always plagues astronomy. Here I will focus on the RAR model, highlighting specific radial ranges where the details of the RAR model provide insight that can’t be obtained in other ways.
The RAR Milky Way model was fit to the terminal velocity data (in grey) over the radial range 3 < R < 8 kpc. Everything outside of that range is a prediction. It is not a prediction limited to that skinny blue line, as I have to extrapolate the mass distribution of the Milky Way to arbitrarily large radii. If there is a gradient in the mass-to-light ratio, or even if I guess a little wrong in the extrapolation, it’ll go off at some point. It shouldn’t be far off, as V(R) is mostly fixed by the enclosed mass. Mostly. If there is something else out there, it’ll be higher (like the cyan line including an estimate of the coronal gas in the plot that goes out to 130 kpc). If there is a bit less than the extrapolation, it’ll be lower.
The RAR model Milky Way (blue line) together with the terminalvelocities to which it was fit (light grey points), VVV data in the inner 2.2 kpc (dark grey squares), and the Zhou et al. (2023) realization of the Gaia DR3 data. Also shown are the number of stars per bin from Gaia (right axis).
From 8 to 19 kpc, the Gaia data as realized by Zhao et al. fall bang on the model. They evince exactly the slowly declining rotation curve that was predicted. That’s pretty good for an extrapolation from R < 8 kpc. I’m not aware of any other model that did this well in advance of the observation. Indeed, I can’t think of a way to even make a prediction with a dark matter model. I’ve tried this – a lot – and it is as easy to come up with a model whose rotation curve is rising as one that is falling. There’s nothing in the dark matter paradigm that is predictive at this level of detail.
Beyond R > 19 kpc, the match of the model and Zhou et al. realization of the data is not perfect. It is still pretty damn good by astronomical standards, and better than the Keplerian dotted line. Cosmologists would be wetting themselves with excitement if they could come this close to predicting anything. Heck, they’re known to do that even when they’re obviously wrong*.
If the difference between the outermost data and the blue line is correct, then all it means is that we have to tweak the model to have a bit less mass than assumed in the extrapolation. I call it a tweak because it would be exactly that: a small change to an assumption I was obliged to make in order to do the calculation. I could have assumed something else, and almost did: there is discussion in the literature that the disk of the Milky Way is truncated at 20 kpc. I considered using a mass model with such a feature, but one can’t make it a sharp edge as that introduces numerical artifacts when solving the Poisson equation numerically, as this procedure depends on derivatives that blow up when they encounter sharp features. Presumably the physical truncation isn’t unphysically sharp anyway, rather being a transition to a steeper exponential decline as we sometimes see in other galaxies. However, despite indications of such an effect, there wasn’t enough data to constrain it in a way useful for my model. So rather than introduce a bunch of extra, unconstrained freedom into the model, I made a straight extrapolation from what I had all the way to infinity in the full knowledge that this had to be wrong at some level. Perhaps we’ve found that level.
That said, I’m happy with the agreement of the data with the model as is. The data become very sparse where there is even a hint of disagreement. Where there are thousands of stars per bin in the well-fit portion of the rotation curve, there are only tens per bin outside 20 kpc. When the numbers get that small, one has to start to worry that there are not enough independent samples of phase space. A sizeable fraction of those tens of stars could be part of the same stellar stream, which would bias the results to that particular unrepresentative orbit. I don’t know if that’s the case, which is the point: it is just one of the many potential systematic uncertainties that are not represented in the formal error bars. Missing those last five points by two sigma is as likely to be an indication that the error bars have been underestimated as it is to be an indication that the model is inadequate. Trying to account for this sort of thing is why the error bars of Jiao et al. are so much bigger than the formal uncertainties in the three realization papers.
That’s the outer regions. The place where the RAR model disagrees the most with the Gaia data is from 5 < R < 8 kpc, which is in the range where it was fit! So what’s going on there?
Again, the data disagree with the data. The stellar data from Gaia disagree with the terminal velocity data from interstellar gas at high significance. The RAR model was fit to the latter, so it must per force disagree with the former. It is tempting to dismiss one or the other as wrong, but do they really disagree?
Adapted from Fig. 4 of McGaugh (2019). Grey points are the first and fourth quadrant terminal velocity data to which the model (blue line) was matched. The red squares are the stellar rotation curve estimated with Gaia DR2 (DR3 is indistinguishable). The black squares are the stellar rotation curve after adjustment to be consistent with a mass profile that includes spiral arms. This adjustment for self-consistency remedies the apparent discrepancy between gas and stellar data.
In order to build the model depicted above, I chose to split the difference between the first and fourth quadrant terminal velocity data. I fit them separately in McGaugh (2016) where I made the additional point that the apparent difference between the two quadrants is what we expect from an m=2 mode – i.e., a galaxy with spiral arms. That means these velocities are not exactly circular as commonly assumed, and as I must per force assume to build the model. So I split the difference above in the full knowledge that this is not the exact circular velocity curve of the Galaxy, it’s just the best I can do at present. This is another example of the systematic uncertainties we encounter: the difference between the first and fourth quadrant is real and is telling us that the galaxy is not azimuthally symmetric – as anyone can tell by looking at any spiral galaxy, but is a detail we’d like to ignore so we can talk about disk+dark matter halo models in the convenient limit of axisymmetry.
Though not perfect – no model is – the RAR model Milky Way is a lot better than models that ignore spiral structure entirely, which is basically all of them. The standard procedure assumes an exponential disk and some form of dark matter halo. Allowance is usually made for a central bulge component, but it is relatively rare to bother to include the interstellar gas, much less consider deviations from a pure exponential disk. Having adopted the approximation of an exponential disk, one inevitably get a smooth rotation curve like the dashed line below:
Fig. 1 from McGaugh (2019). Red points are the binned fourth quadrant molecular hydrogen terminal velocities to which the model (blue line) has been fit. The dotted lines shows the corresponding Newtonian rotation curve of the baryons. The dashed line is the model of Bovy & Rix (2013) built assuming an exponential disk. The inset shows residuals of the models from the data. The exponential model does not and cannot fit these data.
The common assumption of exponential disk precludes the possibility of fitting the bumps and wiggles observed in the terminal velocities. These occur because of deviations from a pure exponential profile caused by features like spiral arms. By making this assumption, the variations in mass due to spiral arms is artificially smoothed over. They are not there by assumption, and there is no way to recover them in a dark matter fit that doesn’t know about the RAR.
Depending on what one is trying to accomplish, an exponential model may suffice. The Bovy & Rix model shown above is perfectly reasonable for what they were trying to do, which involved the vertical motions of stars, not the bumps and wiggles in the rotation curve. I would say that the result they obtain is in reasonable agreement with the rotation curve, given what they were doing and in full knowledge that we can’t expect to hit every error bar of every datum of every sort. But for the benefit of the chi-square enthusiasts who are concerned about missing a few data points at large radii, the reduced chi-squared of the Bovy & Rix model is 14.35 while that of the RAR model is 0.6. A good fit is around 1, so the RAR model is a good fit while the smooth exponential is terrible – as one can see by eye in the residual inset: the smooth exponential model gets the overall amplitude about right, but hits none of the data. That’s the starting point for every dark matter model that assumes an exponential disk; even if they do a marginally better job of fitting the alleged Keplerian downturn, they’re still a lot worse if we consider the terminal velocity data, the details of which are usually ignored.
If instead we pay attention the details of the terminal velocity data, we discover that the broad features seen there in are pretty much what we expect for the kinematic signatures of photometrically known spiral arms. That is, the mass density variations inferred by fitting the RAR correspond to spiral arms that are independently known from star counts. We’ve discussed this before.
Spiral structure in the Milky Way (left) as traced by HII regions and Giant Molecular Clouds (GMCs). These correspond to bumps in the surface density profile inferred from kinematics with the RAR (right).
If we accept that the bumps and wiggles in the terminal velocities are tracers of bumps and wiggles in the stellar mass profiles, as seen in external galaxies, then we can return to examining the apparent discrepancy between them and the stellar rotation curve from Gaia. The latter follow from an application of the Jeans equation, which helps us sort out the circular motion from the mildly eccentric orbits of many stars. It includes a term that depends on the gradient of the density profile of the stars that trace the gravitational potential. If we assume an exponential disk, then that term is easily calculated. It is slowly and smoothly varying, and has little impact on the outcome. One can explore variations of the assumed scale length of the disk, and these likewise have little impact, leading us to infer that we don’t need to worry about it. The trouble with this inference is that it is predicated on the assumption of a smooth exponential disk. We are implicitly assuming that there are no bumps and wiggles.
The bumps and wiggles are explicitly part of the RAR model. Consequently, the gradient term in the Jeans equation has a modest but important impact on the result. Applying it to the Gaia data, I get the black points:
The red squares are the Gaia DR2 data. The black squares are the same data after including in the Jeans equation the effect of variations in the tracer gradient. This term dominates the uncertainties.
The velocities of the Gaia data in the range illustrated all go up. This systematic effect reconciles the apparent discrepancy between the stellar and gas rotation curves. The red points are highly discrepant from the gray points, but the black points are not. All it took was to drop the assumption of a smooth exponential profile and calculate the density gradient numerically from the data. This difference has a more pronounced impact on rotation curve fits than any of the differences between the various realizations of the Gaia DR3 data – hence my cavalier attitude towards their error bars. Those are not the important uncertainties.
Indeed, I caution that we still don’t know what the effective circular velocity of the potential is. I’ve made my best guess by splitting the difference between the first and fourth quadrant terminal velocity data, but I’ve surely not got it perfectly right. One might view the difference between the quadrants as the level at which the perfect quantity is practically unknowable. I don’t think it is quite that bad, but I hope I have at least given the reader some flavor for some of the hidden systematic uncertainties that we struggle with in astronomy.
It gets worse! At small radii, there is good reason to be wary of the extent to which terminal velocities represent circular motion. Our Galaxy hosts a strong bar, as artistically depicted here:
Artist’s rendition of the Milky Way. Image credit: NASA/JPL-Caltech.
Bars are a rich topic in their own right. They are supported by non-circular orbits that maintain their pattern. Consequently, one does not expect gas in the region where the bar is to be on circular orbits. It is not entirely clear how long the bar in our Galaxy is, but it is at least 3 kpc – which is why I have not attempted to fit data interior to that. I do, however, have to account for the mass in that region. So I built a model based on the observed light distribution. It’s a nifty bit of math to work out the equivalent circular velocity corresponding to a triaxial bar structure, so having done it once I’ve not been keen to do it again. This fixes the shape of the rotation curve in the inner region, though the amplitude may shift up and down with the mass-to-light ratio of the stars, which dominate the gravitational potential at small radii. This deserves its own close up:
Colored points are terminal velocities from Marasco et al. (2017), from both molecular (red) and atomic (green) gas. Light gray circles are from Sofue (2020). These are plotted assuming they represent circular motions, which they do not. Dark grey squares are the equivalent circular velocity inferred from stars in the VVV survey. The black line is the Newtonian mass model for the central bar and disk, and the blue line is the corresponding RAR model as seen above.
Here is another place where the terminal velocities disagree with the stellar data. This time, it is because the terminal velocities do not trace circular motion. If we assume they do, then we get what is depicted above, and for many years, that was thought to be the Galactic rotation curve, complete with a pronounced classical bulge. Many decades later, we know the center of the Galaxy is not dominated by a bulge but rather a bar, with concominant non-circular motions – motions that have been observed in the stars and carefully used to reconstruct the equivalent circular velocity curve by Portail et al. (2017). This is exactly what we need to compare to the RAR model.
Note that 2008, when the bar model was constructed, predates 2017 (or the 2016 appearance of the preprint). While it would have been fair to tweak the model as the data improved, this did not prove necessary. The RAR model effectively predicted the inner rotation curve a priori. That’s a considerably more impressive feat than getting the outer slope right, but the model manages both sans effort.
No dark matter model can make an equivalent boast. Indeed, it is not obvious how to do this at all; usually people just make a crude assumption with some convenient approximation like the Hernquist potential and call it a day without bothering to fit the inner data. The obvious prediction for a dark matter model overshoots the inner rotation curve, as there is no room for the cusp predicted in cold dark matter halos – stars dominate the central potential. One can of course invoke feedback to fix this, but it is a post hoc kludge rather than a prediction, and one that isn’t supposed to apply in galaxies as massive as the Milky Way. Unless it needs to, of course.
So, lets’s see – the RAR model Milky Way reconciles the tension between stellar and interstellar velocity data, indicates density bumps that are in the right location to correspond to actual spiral arms, matches the effective circular velocity curve determined for stars in the Galactic bar, correctly predicted the slope of the rotation curve outside the solar circle out to at least 19 kpc, and is consistent with the bulk of the data at much larger radii. That’s a pretty successful model. Some realizations of the Gaia DR3 data are a bit lower than predicted, but others are not. Hopefully our knowledge of the outer rotation curve will continue to improve. Maybe the day will come when the data have improved to the point where the model needs to be tweaked a little bit, but it is not this day.
*To give one example, the BICEP II experiment infamously claimed in March of 2014 to have detected the Inflationary signal of primordial gravitational waves in their polarization data. They held a huge press conference to announce the result in clear anticipation of earning a Nobel prize. They did this before releasing the science paper, much less hearing back from a referee. When they did release the science paper, it was immediately obvious on inspection that they had incorrectly estimated the dust foreground. Their signal was just that – excess foreground emission. I could see that in a quick glance at the relevant figure as soon as the paper was made available. Literally – I picked it up, scanned through it, saw the relevant figure, and could immediately spot where they had gone wrong. And yet this huge group of scientists all signed their name to the submitted paper and hyped it as the cosmic “discovery of the century”. Pfft.
Continuing from last time, let’s compare recent rotation curve determinations from GaiaDR3:
Fig. 1 from Jiao et al. comparing three different realizations of the Galactic rotation curve from Gaia DR3. The vertical lines* mark the range of the Ou et al. data considered by Chan & Chung Law (2023).
These are different analyses of the same dataset. The Gaia data release is immense, with billions of stars. There are gazillions of ways to parse these data. So it is reasonable to have multiple realizations, and we shouldn’t expect them to necessarily agree perfectly: do we look exclusively at K giants? A stars? Only stars with proper motion and/or parallax data more accurate than some limit? etc. Of course we want to understand any differences, but that’s not going to happen here.
My first observation is that the various analyses are broadly consistent. They all show a steady decline over a large range of radii. Nothing shocking there; it is fairly typical for bright, compact galaxies like the Milky Way to have somewhat declining rotation curves. The issue here, of course, is how much, and what does it mean?
Looking more closely, not all of the data agree with each other, or even with themselves. There are offsets between the three at radii around the sun (we live just outside R = 8 kpc) where you’d naively think they would agree the best. They’re very consistent from 13 < R < 17 kpc, then they start to diverge a little. The Ou data have a curious uptick right around R = 17 kpc, which I wouldn’t put much stock in; weird kinks like that sometimes happen in astronomical data. But it can’t be consistent with a continuous mass distribution, and will come up again for other reasons.
As an astronomer, I’m happy with the level of agreement I see here. It is not perfect, in the sense that there are some points from one data set whose error bars do not overlap with those of other data sets in places. That’s normal in astronomy, and one of the reasons that we can never entirely trust the stated uncertainties. Jiao et al. make a thorough and yet still incomplete assessment of the systematic uncertainties, winding up with larger error bars on the Wang et al. realization of the data.
For example, one – just one of the issues we have to contend with – is the distance to each star in the sample. Distances to individual objects are hard, and subject to systematic uncertainties. The reason to choose A stars or K giants is because you think you know their luminosity, so can estimate their distance. That works, but aren’t necessarily consistent (let alone correct) among the different groups. That by itself could be the source of the modest difference we see between data sets.
Chan & Chung Law use the Ou et al. realization of the data to make some strong claims. One is that the gradient of the rotation curve is -5 km/s/kpc, and this excludes MOND at high confidence. Here is their plot.
You will notice that, as they say, these are the data of Ou et al, being identical to the same points in the plot from Jiao et al. above – provided you only look in the range between the lines, 17 < R < 23 kpc. This is where the kink at R = 17 kpc comes in. They appear to have truncated the data right where it needs to be truncated to ignore the point with a noticeably lower velocity, which would surely affect the determination of the slope and reduce its confidence level. They also exclude the point with a really big error bar that nominally is within their radial range. That’s OK, as it has little significance: it’s large error bar means it contributes little to the constraint. That is not the case for the datum just inside of R = 17 kpc, or the rest of the data at smaller radii for that matter. These have a manifestly shallower slope. Looking at the line boundaries added to Jiao’s plot, it appears that they selected the range of the data with the steepest gradient. This is called cherry-picking.
It is a strange form of cherry-picking, as there is no physical reason to expect a linear fit to be appropriate. A Keplerian downturn has velocity decline as the inverse square root of radius (see the dotted line above.) These data, over this limited range, may be consistent with a Keplerian downturn, but certainly do not establish that it is required.
Contrast the statements of Chan & Chung Law with the more measured statement from the paper where the data analysis is actually performed:
… a low mass for the Galaxy is driven by the functional forms tested, given that it probes beyond our measurements. It is found to be in tension with mass measurements from globular clusters, dwarf satellites, and streams.
What this means is that the data do not go far enough out to measure the total mass. The low mass that is inferred from the data is a result of fitting some specific choice of halo form to it. They note that the result disagrees with other data, as I discussed last time.
Rather than cherry pick the data, we should look at all of it. Let’s see, I’ve done thatbefore. We looked at the Wang et al. (2023) data via Jiao et al.previously, and just discussed the Ou et al. data. That leaves the new Zhao et al. data, so let’s look at those:
Milky Way rotation curve with RAR model (blue line from 2018) and the Gaia DR3 data as realized by Zhou et al. (2023: purple triangles). The dashed line shows the number of stars (right axis) informing each datum.
These data were the last of the current crop that I looked at. They look… pretty good in comparison with the pre-existing RAR model. Not exactly the falsification I had been led to expect.
So – the three different realizations of the Gaia DR3 data are largely consistent, yet one is being portrayed as a falsification of MOND while another is in good agreement with its prediction.
This is why you have to take astronomical error bars with a grain of salt. Three different groups are using data from the same source to obtain very nearly the same result. It isn’t quite the same result, as some of the data disagree at the formal limits of their uncertainty. No big deal – that’s what happens in astronomy. The number of stars per bin helps illustrate one reason why: we go from thousands of stars per bin near the sun to tens of stars in wider bins at R > 20 kpc. That’s not necessarily problematic, but it is emblematic of what we’re dealing with: great gobs of data up close, but only scarce scratches of it far away where systematic effects are more pernicious.
In the meantime, one realization of these data are being portrayed as a death knell for a theory that successfully predicts another realization of the same data. Well, which is it?
*Thanks to Moti Milgrom for pointing out the restricted range of radii considered by Chan & Chung Law and adding the vertical lines to this figure.
Recent results from the third data release (DR3) from Gaia has led to a flurry of papers. Some are good, some are great, some are neither of those. It is apparent from the comments last time that while I’ve kept my pledge to never dumb it down, I have perhaps been assuming more background knowledge on the part of readers than is adequate. I can’t cram a graduate education in astronomy into one web page, but will try to provide a little relevant context.
Galactic Astronomy is an ancient field, dating back at least to the Herschels. There is a lot that is known in the field. There have also been a lot of misleading observations, going back just as far to the Herschel’s map of the Milky Way, which was severely limited by extinction from interstellar dust. That’s easy to say now, but Herschel’s map was the standard for over a century – longer than our modern map has persisted.
So a lot has changed, including a lot that seemed certain, so I try to keep an open mind. The astronomers working with the Gaia data – the ones deriving the rotation curve – are simply following where those data take them, as they should. There are others using their analyses to less credible ends. A lot of context is required to distinguish the two.
The total mass of the Milky Way
There are a lot of constraints on the mass of the Milky Way that predate Gaia; it’s not like these are the first data that address the issue. Indeed, there are lots and lots and lots of other applicable data acquired using different methods over the course of many decades. Here is a summary plot of determinations of the mass of the Milky Way compiled by Wang et al. (2019).
This is an admirable compilation, and yet no such compilation can be complete. There are just so many determinations by lots of independent authors. Still, this is nice for listing multiple results from many distinct methodologies. They all consistently give numbers around 1012 solar masses. (Cast in these terms, my own estimate is 1.4 x 1012 albeit with a substantial systematic uncertainty.) I’ve added a point for the total mass according to the alleged Keplerian downturn seen in the Gaia data, 2 x 1011 solar masses. One of these things is not like the others.
The difference from the bulk of the data has nearly every astronomer rolling our collective eyes. Most of us straight up don’t believe it. That’s not to say the Gaia data are wrong, but the interpretation of those data as indicative of such a small, finite total mass seems unlikely in the light of all other results.
As I discussed briefly last time, it is conceivable that previous results are wrong or misleading due to some systematic effect or bad assumption. For example, mass estimates based on “satellite phenomenon” require the assumption that the satellite galaxies are indeed satellites of the Milky Way on bound orbits. That seems like a really good assumption, as without it, their presence is an instantaneous coincidence particular to the most recent few percent of a Hubble time: they wouldn’t have been nearby more than a billion years ago, and won’t be around another for even a few hundred million more. That sounds like a long time to you and me, but it is not that long on a cosmic scale. Maybe they’re raining down all the time to give the appearance of a steady state? Where have I heard that before?
Even if we’re willing to dismiss satellite constraints, that doesn’t suffice. It isn’t good enough to find flaw with one set of determinations; one must question all distinct methods. I could probably do that; there’s always a systematic uncertainty that might be bigger than expected or an assumption that could go badly wrong. But it is asking a lot for all of them to conspire to be wrong at the same time by the same amount. (The assumption of Newtonian gravity is a catch-all.)
Some constraints are more difficult to dodge than others. For example, the escape velocity method merely notes that there are fast moving stars in the solar neighborhood. Those stars are many billions of years old, and wouldn’t be here if the gravitational potential couldn’t contain them. The mass implied by the Gaia quasi-Keplerian downturn doesn’t suffice.
That said, the total mass of the Milky Way as expressed above is a rather notional quantity. M200 occurs roughly 200 kpc out for the Milky Way, give or take a lot. And the “200” in the subscript has nothing to do with that radius being 200 kpc for reasons too technical and silly to delve into. So my biggest concern about the compilation above is not that the data are wrong so much as they are being extrapolated to an idealized radius that we don’t directly observe. This extrapolation is usually done by assuming the potential of an NFW halo, which makes perfect sense in terms of LCDM but none whatsoever empirically, since NFW predicts the wrong density profile at small, intermediate, and large radii: where the density profile ฯ โ r-ฮฑ is predicted to have ฮฑ = (1,2,3), it is persistently observed to be more like (0,1,2). While the latter profile is empirically more realistic, it also fails to converge to a finite total mass, rendering the concept meaningless.
Rather than indulge yet again in a discussion of the virtues and vices of different dark matter halo profiles, let’s look at an observationally more robust quantity: the enclosed mass. Wang et al. also provide a tabulation of this quantity from many sources, as depicted here:
Rotation curve constraints implied by the enclosed mass measurements tabulated by Wang et al. (2019) combined with the halo stars and globular clusterspreviously discussed. The location of the Large Magellanic Cloud is also indicated; data beyond this radius (and perhaps even within it) are subject to perturbation by the passage of the LMC. The RAR-based model is shown as the blue line; the light blue line includes a very uncertain estimate of the effect of the coronal gas. This is very diffuse and extended, and only becomes significant at very large radii. The dotted line is the Keplerian curve for a mass of 2 x 1011 Mโ.
Not all of the enclosed mass data are consistent with one another. The bulk of them are consistent with the RAR model Milky Way (blue line). None of them are consistent with the small mass indicated by recent Gaia analyses (dotted line). Hence the collective unwillingness of most astronomers to accept the low-mass interpretation.
An important thing to note when considering data at large radii, especially those beyond 50 kpc, is that 50 kpc is the current Galactocentric radius of the Large Magellanic Cloud. The LMC brings with it its own dark matter halo, which perturbs the outer regions of the Milky Way. This effect is surprisingly strong*, and leads to the inference that the mass ratio of the two is only 4 or 5:1 even though the luminosity ratio is more like 20:1. This makes the interpretation of the data beyond 50 kpc problematic. If we use that as a pretext to ignore it, then we infer that our low mass Milky Way is no more massive then the LMC – an apparently absurd situation.
There are many rabbit holes we could dig down here, but the basic message is that a small Milky Way mass violates a gazillion well-established constraints. That doesn’t mean the Gaia data are wrong, but it does call into question their interpretation. So next time we’ll look more closely at the data.
*This is not surprising in MOND. The LMC is in the right place at the right time to cause the Galactic warp. The LMC as a candidate perturber to excite the Galactic warp was recognized early, but the conventional mass was thought to be much too small to do the job. The small baryonic mass of the LMC in MOND is not a problem as the long range nature of the force law makes tidal effects more pronounced: it works out about right.
Yes, some. That much is a step forward from a decade ago, when a common assumption was that the Milky Way’s rotation curve remained flat at the speed at which the sun orbited. This was a good guess based on empirical experience with other galaxies, but not all galaxies have rotation curves that are completely flat, nor can we be sure the sun is located where that is the case.
A bigger question whether the Milky Way’s rotation curve is declining in a Keplerian fashion. This would indicate that the total mass has been enclosed. That would be a remarkable result. If true, it would be the first time that the total mass of an individual galaxy has been measured. There have been claims to this effect before that have not panned out when the data have been extended to larger radii, so one might be inclined to be skeptical.
There are several claims now to see a distinctly declining rotation curve based on the third data release (DR3) from Gaia. The most recent, Jiao et al., has gained some note by virtue of putting “Keplerian decline” in the title, but very similar results have also been reported by Ou et al., Wang et al. and Sylos Labini et al. They all obtain basically the same answer using the same data, with minor differences in the error assessment and other details. There are also differences in interpretation*, which is always possible even when everyone agrees about what the data say.
Jiao et al. measure a total mass for the Milky Way of about 2 x 1011 Mโ. Before looking at the data, let’s take a moment to think about that number. Most mass determinations – and there are lots, see Fig. 2 of Wang et al. – for the Milky Way have been in the neighborhood of 1012 Mโ. Indeed, for most of my career, it was traditionally Known to be 2 x 1012 Mโ. The new measurement is an order of magnitude smaller. That’s a lot to be off by, even in extragalactic astronomy. The difference, as we’ll see, has to do with what data we use.
The mass of stars and gas in the Milky Way is about 6 x 1010 Mโ, give or take ten billion. That means that nearly a third of the total mass is normal baryonic matter that we can readily see. So the ratio of dark-to-baryonic mass is only 2.3:1, well short of the cosmic ratio of about 6:1. That’s embarrassing – especially since much of the effort in galaxy formation theory has been to explain why the baryon fraction is much less than the cosmic fraction, not much more. And here our Galaxy is an outlier, having much less dark matter for its stellar mass than everything else. It is always a bad sign when the Galaxy appears to violate the Copernican Principle.
Nonetheless, this is what we find if we look at the Gaia DR3 data. Here is a model I’ve shown before, extrapolated to larger radii with some new data added. The orange circles are the Gaia DR3 rotation curve as given by Jiao et al. For radii greater than 18 kpc, they show a clear decline consistent with a Keplerian curve for a 1.95 x 1011 Mโ point mass (dotted line), as per Fig. 9 of Jiao et al.
This is the first time we’ve been able to trace the rotation curve so far out with stars in the disk of the Milky Way, and the Keplerian line is a good match. If that’s all we know, then a total mass of only 2 x 1011 Mโ is a reasonable inference. That’s not all we know.
As I alluded above, a halo mass this small makes no sense in the context of cosmology. Not only is 2 x 1011 Mโ too small, the more commonly inferred dynamical mass of 1012 Mโ is also too small. According to abundance matching, which has become an important aspect of LCDM, the Milky Way should reside in a 3 or 4 x 1012 Mโ halo. So the new mass makes a factor of 2 or 3 problem into a factor a ten problem. That is too large to attribute to scatter in the stellar mass-halo mass relation. Worse, there is no evidence that the Milky Way is an outlier from scaling relations like Tully-Fisher. We can’t have it one way and not the other.
The traditional mass estimates that obtain ~1012 Mโ rely on dwarf satellite galaxies as tracers of the gravitational potential of the Milky Way. Maybe they’re not fair tracers? We have to make assumptions about their orbits to use them to infer a mass; perhaps these assumptions are wrong? It is conceivable that many of our satellites are on first infall rather than in well-established orbits. Indeed, the consensus is that our largest satellites, the Magellanic Clouds, are on first infall, and that they cause a substantial perturbation to the halo of the Milky Way. This was an absurd thought 15 years ago – the Magellanic clouds must have been here forever, and were far too small to do damage – but now this is standard lore.
There are tracers at large radii besides dwarf satellite galaxies. The figure above shows three: globular clusters (pink triangles) and two types of stars in the halo: blue horizontal branch stars (green squares) and K giants (red squares). These are well-known parts of the Milky Way that have been with us for many billions of years, so they’ve had plenty of time to become equilibrium tracers of the gravitational potential. They clearly indicate a larger enclosed mass than predicted by the Keplerian decline traced by the Gaia rotation curve, and are consistent with traditional satellite analyses. Perhaps these data are somehow misleading, but it is hard to see how.
Gaia is great, but has its limits. It is really optimized for nearby stars (within a few kpc). Outside of that, the statistics… leave something to be desired. Is it safe to push out beyond 20 kpc? I don’t know, but I did notice this panel from Fig. 8 of Wang et al.:
Radial velocities of stars at different heights above the Galactic plane.
The radial velocity is a minor component of disk motion, where azimuthal motion dominates. However, one does need to know it to solve the Jeans equation. Having it wrong will cause a perceptible systematic error. You notice the bifurcation in the data for R > 22 kpc? That, in technical terms, is Messed Up. I don’t know what goes awry there, but I’ve done this exercise enough times for the sight of this to scare the bejeepers out of me. No way I trust any of these data at R > 22 kpc, and I hope having seen this doesn’t give me nightmares tonight.
Perhaps the uncertainty caused by this is adequately reflected in the large error bars on the orange points above. Those with R > 22 kpc are nicely Keplerian, but also consistent with a lot of things, including the blue line that successfully predicts the halo stars and globular clusters. That’s not true for the data around R = 20 kpc where the error bars are much smaller: there the discrepancy with the blue line I take seriously. But that is a much more limited affair that might indicate the presence of a ring of mass – that’s what gives the bumps and wiggles at smaller radii – and certainly isn’t enough to imply the entire mass of the Milky Way has been enclosed.
But who knows? Perhaps fifteen years hence it will be the standard lore that all galaxies reside in dark matter halos that are only twice the mass of their luminous disks. At that mass ratio, all the galactic dark matter could be baryonic. I wouldn’t bet on it, but stranger things have happened before, and will happen again.
*A difference in interpretation is largely what the debate about dark matter and MOND boils down to. There is no doubt that there are acceleration discrepancies in extragalactic objects that require something beyond what you see is what you get with normal gravity. Whether we should blame what we can’t see or the assumption of normal gravity is open to interpretation. I would hope this is obvious, but this elementary point seems to be lost on many.
The community seems to react to problems with the dark matter interpretation in one of several ways. Physicists often seem to simply ignore them, presuming that any problems are mere astronomical details that aren’t relevant to fundamental physics. Among more serious scientists, there is a tendency to bicker over solutions, settle on something (satisfactory or not), then forget that there was ever a problem.
Benoit Famaey and I wrote a long review for Living Reviews in Relativity about a decade ago. In it, we listed some of the problems that afflicted LCDM. It is instructive to review what those were, and examine what progress has been made. The following is based on section 4 of the review. I will skip over the discussion of coincidences, which remain an issue, to focus on specific astronomical problems.
Unobserved predictions
A problem for LCDM, and indeed, any theory, is when it makes predictions that are not confirmed. Here are a list of challenges stemming from observational reality deviating from the expectations or LCDM that we identified in our review, together with an assessment of whether they remain a concern. The bulk flow challenge Peculiar velocities of galaxy clusters are predicted to be on the order of 200 km/s in the ฮCDM model: as massive, recently formed objects, they should be nearly at rest with respect to the frame of the cosmic microwave background. Instead, they are observed to have bulk flows of order 1000 km/s.
This appears to remain a problem, and is related to the high collision speeds of objects like the bullet cluster, which basically shouldn’t exist.
The high-z clusters challenge Structure formation is reputed to be one of the greatest strengths of LCDM, but the observers’ experience has consistently been to find more structure in place earlier than expected. This goes back at least to the 1987 CfA redshift survey stick man figure, which may seem normal now but surprised the bejeepers out of us at the time. It also includes clusters of galaxies, which appear at higher redshift than they should. At the time, we pointed out XMMU J2235.3-2557 with a mass of of โผ 4 ร 1014 Mโ at z = 1.4 as being very surprising.
More recently we have El Gordo, so this remains a problem.
The Local Void challenge Peebles has been pointing out for a long time that voids are more empty than they should be, and do not contain the population of galaxies expected in LCDM. They’re too normal, too big, and gee it would help if structure formed faster. In our review, we pointed out that the โLocal Voidโ hosts only 3 galaxies, which is much less than the expected โผ 20 for a typical similar void in ฮCDM.
I am not seeing much in the literature in the way of updates, so I guess this one has been forgotten and remains a problem.
The missing satellites challenge LCDM predicts that there are many subhalos in every galactic halo, and one would naturally expect each of these to host a dwarf satellite galaxy. While galaxies like the Milky Way do have dwarf satellites, they number in the dozens when there should be thousands of subhalos. This is manifestly not the case.
The trick with this test is mapping the predicted number of halos to the corresponding galaxies that inhabit them. If there is a nonlinear relation between mass and light, then there can be fewer (or more) dwarf galaxies than halos. People seem to have decided that this problem has been solved.
It is not clear to me how the solutions map to the (contemporaneous with our review) Too Big To Fail problem in which the most massive predicted subhaloes are incompatible with hosting any of the known Milky Way satellites. It isn’t a simple nonlinearity in mass-to-light; some biggish subhalos simply don’t host galaxies, apparently, while many smaller ones do. That doesn’t make sense in terms of the many mass-dependent mechanisms that are invoked to suppress dwarf galaxy formation. Nevertheless, we are assured that it all works out.
The satellites phase-space correlation challenge This is also known as the planes of satellites problem. At the time of our review, it had recently been recognized that the satellite galaxies of the Milky Way are observed to correlate in phase-space, lying in a seemingly rotation-supported disk. This is pretty much the opposite of what one expects in LCDM, in which subhalos are on randomly oriented, radial orbits.
The problem has gotten worse with more planes now being known around Andromeda and Centaurus A and other galaxies. There have been a steady stream of papers asserting that this is not a problem, but the “solution” seems to be to declare planes to be “common” if their incidence in simulations is a few percent. That is, they seem to agree with the observers who point out that this is a problem, and simply declare it not to be a problem.
The cusp-core challenge The cusp-core problem is that cold dark matter halos are predicted to have cuspy central regions in which the density of dark matter rises continuously towards their centers, while fitting a dark matter mass distribution to observed galaxies prefers cored halos with a rougly constant density within some finite radius. This has a long history. Observers traditionally used the pseudoisothermal halo profile (with a constant density core) to fit rotation curve data. This was the standard model for a decade before CDM simulations predicted the presence of a central cusp. The pseudoisothermal halo continues to provide a better description of the data. The initial reaction of the theoretical community was to blame the data for not conforming to their predictions: they came up with a series of lame excuses (beam smearing, slit misplacement) for why the data were wrong. Serial improvements in the quality of data showed that these ideas were wrong, and effort switched from reality denial to model modification.
People generally seem to think this problem is solved through the use of baryon feedback to erase the cusps from galaxy halos. I do not find these explanations satisfactory, as they require a just-so fine-tuning to get things right. More generally, this is just one aspect of the challenge presented by galaxy kinematic data. This is what happens if you insist on fitting dark matter halos to data the looks like what MOND predicts. Lots of people seem to think that explaining the cusp-cpore problem solves everything, but this is just one piece of a more general problem, which is not restricted to the central regions. Ultimately, the question remains why MOND works at all in a universe run by dark matter.
I mention all this because it is the prototypical example of why one should take the claims of theorists to have solved a problem with a huge grain of salt. Here, the problem has been redefined into something more limited, then the limited problem has been solved in a seemingly-plausible yet unconvincing way, victory is declared, and the original, more difficult problem (MOND works when it should not) is forgotten or considered to be solved by extension.
The angular momentum challenge During galaxy formation, the baryons sink to the centers of their dark matter halos. A persistent idea is that they spin up as they do so (like a figure skater pulling her arms in), ultimately establishing a rotationally supported equilibrium in which the galaxy disk is around ten or twenty times smaller than the dark matter halo that birthed it, depending on the initial spin of the halo. This is a seductively simple picture that still has many adherents despite never having really worked. In live simulations, in which baryonic and dark matter particles interact, there is a net transfer of angular momentum from the baryonic disk to the dark halo. This results in simulated disks being much too small.
This problem is solved by invoking just-so feedback again. Whether the feedback one needs to solve this problem is consistent with the feedback one needs to solve the cusp-core problem is unclear, in large part because different groups have different implementations of feedback that all do different things. At most one of them can be right. Given familiarity with the approximations involved, a more likely number is Zero.
The pure disk challenge Structure forms hierarchically in CDM: small galaxies merge into larger ones. This process is hostile to the existence of dynamically cold, rotating disks, preferring instead to construct dynamically hot, spheroidal galaxies. All the merging destroys disks. Yet spiral galaxies are ubiquitous, and many late type galaxies have no central bulge component at all. At some point it was recognized that the existence of quiescent disks didn’t make a whole lot of sense in LCDM. To form such things, one needs to let gas dissipate and settle into a plane without getting torqued and bombarded by lots of lumps falling onto it from random directions. Indeed, it proved difficult to form large, bulgeless, thin disk galaxies in simulations.
The solution seems to be just-so feedback again, though I don’t see how that can preclude the dynamical chaos caused by merging dark matter halos regardless of what the baryons do.
The stability challenge One of the early indications of the need for spiral galaxies to be embedded in dark matter halos was the stability of disks. Thin, dynamically cold spiral disks are everywhere around us, yet Newton can’t hold them together by himself: simulated spirals self destruct on a short timescale (a few orbits). A dark matter halo precludes this from happening by counterbalancing the self-gravity of the disk. This is a somewhat fine-tuned situation: too little halo, and a disk goes unstable; too much and disk self-gravity is suppressed – and spiral arms and bars along with it.
I recognized this as a potential test early on. Dark matter halos tend to over-stabilize low surface density disks against the formation of bars and spirals. You need a lot of dark matter to explain the rotation curve, but not too much so as to allow for spiral structure. These tensions can be contradictory, and the tension I anticipated long ago has been realized in subsequent analyses.
The low surface brightness spiral F568-1 (left) and its rotation curve (right). The heavy line indicates the stellar disk mass required to sustain the observed spiral arms; the light line shows what is reasonable for a normal stellar population for which the galaxy consistent with the BTFR and RAR. We can’t have it both ways; this is the predicted contradiction to invoking dark matter to explain both disk stability and kinematics.
I’m not aware of this problem being addressed in the context of cold dark matter models, much less solved. The problem is very much present in modern hydrodynamical simulations, as illustrated by this figure from the enormous review by Banik & Zhao:
The pattern speeds of bars as observed and simulated. Real bars are fast (R = 1) while simulated bars are slow (R > 2) due to the excessive dynamical friction from cuspy dark matter halos. (Fig. 21 from Banik & Zhao 2022).
The missing baryons challenge The cosmic fraction of baryons – the ratio of normal matter to dark matter – is well known (16 ยฑ 1%). One might reasonably expect individual CDM halos to be in in possession of this universal baryon fraction: the sum of the stars and gas in a galaxy should be 16% of the total, mostly dark mass. However, most objects fall well short of this mark, with the only exception being the most massive clusters of galaxies. So where are all the baryons?
The answer seems to be that we don’t have to answer that. Initially, the poroblem was overcooling: low mass galaxies should turn more of their baryons into stars than is observed. Feedback was invoked to prevent that, and it seems to be widely accepted that feedback from those stars that do form heat much of the surrounding gas so it remains mixed in with the halo in some conveniently unobservable form, or that the feedback is so vigorous that it expells the excess baryons entirely. That the observed baryon fraction declines with declining mass is attributed to the lesser potential wells of smaller galaxies not being able to hang on to their baryons as well – they are more readily expelled. That sounds reasonable at a hand-waving level, but getting it right quantitatively presents a fine-tuning problem: the observed baryon fraction correlates strongly with mass with practically no scatter. One would expect feedback to be rather stochastic and result in a lot of scatter, but if it did it would propagate straight into the Tully-Fisher relation, which has practically no scatter. This fine-tuning problem is addressed by ignoring it.
The more things change
So those are the things that concerned us a decade ago. Looking back on them, there has been some progress on some items and less on others. Being generous, I would say there has at least been progress on the missing satellite problem, cusp-core, angular momentum, and pure disks. There has been no perceptible progress on the other problems, some of which (high-z clusters, disk stability) have gotten worse.
This is all written in the context of dark matter, with only passing reference to MOND. How does MOND fare for these same issues? MOND is good at making things move fast; it naturally predicts the scale of the bulk flows. It also predicted early structure formation, and is good at sweeping the voids clean. It has nothing to say about missing satellites. There are no subhalos that might be populated with dwarfs in MOND, so the question doesn’t arise. It might provide an explanation for the planes of satellites, but I am underwhelmed by this idea (or any others that I’ve heard for this particular problem). MOND is the underlying cause of the cusp-core problem, which arises entirely from trying to fit dark matter halos to galaxies that obey MOND. MOND suffers no angular momentum problem; what you see is what you get. It is noteworthy that angular momentum is not an additonal free parameter as there is no dark component with an unspecified quantity of it; it is specified entirely by the observed distribution of baryons and their motions. Similarly, making pure disks is not a problem for MOND. One can have hierarchical structure formation, but it is not required to the degree that it wipes out nascent disks in the way it did in LCDM simulations before steps were taken to make them stop doing that. Disk stability in MOND stems from the longer range of the force law rather than piling on dark matter; it is comparable for high surface brightness galaxies in both theories, but readily distinguishable for low surface brightness galaxies. This test clearly prefers MOND. Finally, the missing baryon problem doesn’t really pertain in MOND. Objects just have the baryons they have; only in rich clusters of galaxies is there a residual missing baryon problem (albeit a serious one!)
At a conservative count, that is four distinct items that have nothing to do with rotation curves where MOND performs better than LCDM. But go ahead, tell me again how MOND only explains rotation curves and nothing else.
This was basically just section 4.2 of the review. Section 4.3 was about unexpected observations – observations that were surprising in the context of LCDM. I think this post is been long enough, so I won’t go there except to say that these unexpected things were either predicted a priori by MOND, or follow so naturally from it that they could have been if the question had been posed. So it’s not just that MOND explains some things better than dark matter, it’s that it correctly predicted in advance things that were not predicted by dark matter, and that are often not well-explained by it.
I’m back from the meeting in St. Andrews, and am mostly recovered from the jet lag and the hiking (it was hot and sunny, we did not pack for that!) and the driving on single-track roads like Mr. Toad. The A835 north from Ullapool provides some spectacular mountain views, but the A837 through Rosehall is more perilous carnival attraction than well-planned means of conveyance.
As expected, the most contentious issue was that of wide binaries. The divide was stark: there were two talks finding nary a hint of MONDian signal, just old Newton, and two talks claiming a clear MONDian signal. Nothing was resolved in the sense of one side convincing the other it was right, but there was progress in terms of [mostly] amicable discussion, with some sensible suggestions for how to proceed. One suggestion was that a neutral party should provide all the groups with several sets of mock data, one Newtonian, one MONDian, and one something else, to see if they all recovered the right answers. That’s a good test in principle, but it is a hassle to do in practice, as it is highly nontrivial to produce realistic mock Gaia data, so no one was leaping at the opportunity to stick their hand in this particular bear trap.
Xavier Hernandez made the excellent point that one should check that one’s method recovers Newtonian behavior for close binaries before making any claims to require/exclude such behavior for wide binaries. Neither MOND nor dark matter predicts any deviation from Newtonian behavior where stars are orbiting each other well in excess of a0, of which there are copious examples, so they provide a touchstone on which all should agree. He also convinced me that it was a Good Idea to have radial velocities as well as proper motions. This limits the sample size, but it helps immensely to insure that sample binaries are indeed bound pairs of binary stars. Doing this, he finds MOND-like behavior.
Previously, I linked to a talk by Indranil Banik, who found Newtonian behavior. This led to an exchange with Kyu-Hyun Chae, who has now posted an update to his own analysis in which he finds MONDian behavior. It is a clear signal, and if correct, could be the smoking gun for MOND. It wouldn’t be the first one; that honor probably goes to NGC 1560, and there have been plenty of other smokinggunssincethen. The trick seems to be finding something than cannot be explained with dark matter, and this could play that role since dark matter shouldn’t be relevant to binary stars. But dark matter is pretty much the ultimate Rube Goldberg machine of science, so we’ll see explanation people come up with, should they need to do so.
At present, the facts of the matter are still in dispute, so that’s the first thing to get straight.
Thanks to everyone I met at the conference who told me how useful this blog is. That’s good to know. Communication is inefficient at best, counterproductive at worst, and most often practically nonexistent. So it is good to hear that this does some small good.
My last post about the Milky Way was intended to be a brief introduction to our home galaxy in order to motivate the topic of binary stars. There’s too much interesting to say about the Milky Way as a galaxy, so I never got past that. Even now I feel the urge to say more, like with this extended rotation curve that I included in my contribution to the proceedings of IAU 379.
The RAR-based model rotation curve of the Milky Way extrapolated to large radii (note the switch to a logarithmic scale at 20 kpc!) for comparison to the halo stars of Bird et al (2022) and the globular clusters of Watkins et al (2019). The location of the solar system is noted by the red circle.
But instead I want to talk about data for binary stars from the Gaia mission. Gaia has been mapping the positions and proper motions of stars in the local neighborhood with unprecedented accuracy. These can be used to measure distances via trigonometric parallax, and speeds along the sky. The latter once seemed impossible to obtain in numbers with much precision; thanks to Gaia such data now outnumber radial (line of sight) velocities of comparable accuracy from spectra. That is a mind-boggling statement to anyone who has worked in the field; for all of my career (and that of any living astronomer), radial velocities have vastly outnumbered comparably well-measured proper motions. Gaia has flipped that forever-reality upside down in a few short years. It’s third data release was in June of 2022; this provides enough information to identify binary stars, and we’ve had enough time to start (and I do mean start) sorting through the data.
OK, why are binary stars interesting to the missing mass (really the acceleration discrepancy) problem? In principle, they allow us to distinguish between dark matter and modified gravity theories like MOND. If galactic mass discrepancies are caused by a diffuse distribution of dark matter, gravity is normal, and binary stars should orbit each other as Newton predicts, no matter their separation: the dark matter is too diffuse to have an impact on such comparatively tiny scales. If instead the force law changes at some critical scale, then the orbital speeds of widely separated binary pairs that exceed this scale should get a boost relative to the Newtonian case.
The test is easy to visualize for a single binary system. Imagine two stars orbiting one another. When they’re close, they orbits as Newton predicts. This is, after all, how we got Newtonian gravity – as an explanation for Kepler’s Laws or planetary motion. Ours is a lonely star, not a binary, but that makes no difference to gravity: Jupiter (or any other planet) is an adequate stand-in. Newton’s universal law of gravity (with tiny tweaks from Einstein) is valid as far out in the solar system as we’ve been able to probe. For scale, Pluto is about 40 AU out (where Earth, by definition, is 1 AU from the sun).
Let’s start with a pair of stars orbiting at a distance that is comfortably in the Newtonian regime, say with a separation of 40 AU. If we know the mass of the stars, we can calculate what their orbital speed will be. Now imagine gradually separating the stars so they are farther and farther apart. For any new separation s, we can predict what the new orbital speed will be. According to Newton, this will decline in a Keplerian fashion, v ~ 1/โs. This will continue indefinitely if Newton remains forever the law of the land. If instead the force law changes at some critical scale sc, then we would expect to see a change when the separation exceeds that scale. Same binary pair, same mass, but relatively faster speed – a faster speed that on galaxy scales leads to the inference of dark matter. In essence, we want to check if binary stars also have flat rotation curves if we look far enough out.
We have long known that simply changing the force law at some length scale sc does not work. In MOND, the critical scale is an acceleration, a0. This will map to a different sc for binary stars of different masses. For the sun, the critical acceleration scale is reached at sc โ 7000 AU โ 0.034 parsecs (pc), about a tenth of a light-year. That’s a lot bigger than the solar system (40 AU) but rather smaller than the distance to the next star (1.3 pc = 4.25 light-years). So it is conceivable that there are wide binaries in the solar neighborhood for which this test can be made – pairs of stars with separations large enough to probe the MOND regime without being so far apart that they inevitably get broken up by random interactions with unrelated stars.
Gaia is great for identifying binaries, and space is big. There are thousands of wide binaries within 200 pc of the sun where Gaia can obtain excellent measurements. That’s not a big piece of the galaxy – it is a patch roughly the size of the red circle in the rotation curve plot above – but it is still a heck of a lot of stars. A signal should emerge, and a number of papers have now appeared that attempt this exercise. And ooooo-buddy, am I confused. Frequent readers will have noticed that it has been a long time between posts. There are lots of reasons for this, but a big one is that every time I think I understand what is going on here, another paper appears with a different result.
OK, first, what do we expect? Conventionally, binaries should show Keplerian behavior whatever their separation. Dark matter is not dense enough locally to have any perceptible impact. In MOND, one might expect an effect analogous to the flattening of rotation curves, hence higher velocities than predicted by Newton. And that’s correct, but it isn’t quite that simple.
In MOND, there is the External Field Effect (EFE) in which the acceleration from distant sources can matter to the motion of a local system. This violates the strong but not the weak Equivalence Principle. In MOND, all accelerative tugs matter, whereas conventionally only local effects matter.
This is important here, as we live in a relatively high acceleration neighborhood that is close to a0. The acceleration the sun feels towards the Galactic center is about 1.8 a0. This applies to all the stars in the solar neighborhood, so even if one finds a binary pair that is widely separated enough for the force of one star on another to be less than a0, they both feel the 1.8 a0 of the greater Galaxy. A lot of math intervenes, with the net effect being that the predicted boost over Newton is less than it would have been in the absence of this effect. There is still a boost, but its predicted amplitude is less than one might naively hope.
The location of the solar system along the radial acceleration relation is roughly (gbar, gobs) = (1.2, 1.8) a0. At this acceleration, the effects of MOND are just beginning to appear, and the external field of the Galaxy can affect local binary stars.
One of the first papers to address this is Hernandez et al (2022). They found a boost in speed that looks like MOND but is not MOND. Rather, it is consistent with the larger speed that is predicted by MOND in the absence of the EFE. This implies that the radial acceleration relation depicted above is absolute, and somehow more fundamental than MOND. This would require a new theory that is very similar to MOND but lacks the EFE, which seems necessary in othersituations. Weird.
A thorough study has independently been made by Pittordis & Sutherland (2023). I heard a talk by them over Zoom that motivated the previous post to set the stage for this one. They identify a huge sample of over 73,000 wide binaries within 300 pc of the sun. Contrary to Hernandez et al., they find no boost at all. The motions of binaries appear to remain perfectly Keplerian. There is no hint of MOND-like effects. Different.
OK, so that is pretty strong evidence against MOND, as Indranil Banik was describing to me at the IAU meeting in Potsdam, which is why I knew to tune in for the talk by Pittordis. But before I could write this post, yet another paper appeared. This preprint by Kyu-Hyun Chae splits the difference. It finds a clear excess over the Newtonian expectation that is formally highly significant. It is also about right for what is expected in MOND with the EFE, in particular with the AQUAL flavor of MOND developed by Bekenstein & Milgrom (1984).
So we have one estimate that is MOND-like but too much for MOND, one estimate that is straight-laced Newton, and one estimate that is so MOND that it can start to discern flavors of MOND.
I really don’t know what to make of all this. The test is clearly a lot more complicated than I made it sound. One does not get to play God with a single binary pair; one instead has to infer from populations of binaries of different mass stars whether a statistical excess in orbital velocity occurs at wide separations. This is challenging for lots of reasons.
For example, we need to know the mass of each star in each binary. This can be gauged by the mass-luminosity relation – how bright a main sequence star is depends on its mass – but this must be calibrated by binary stars. OK, so, it should be safe to use close binaries that are nowhere near the MOND limit, but it can still be challenging to get this right for completely mundane, traditional astronomical reasons. It remains challenging to confidently infer the properties of impossibly distant physical objects that we can never hope to visit, much less subject to laboratory scrutiny.
Another complication is the orientation and eccentricity of orbits. The plane of the orbit of each binary pair will be inclined to our line of sight so that the velocity we measure is only a portion of the full velocity. We do not have any way to know what the inclination of any one wide binary is; it is hard enough to identify them and get a relative velocity on the plane of the sky. So we have to resort to statistical estimates. The same goes for the eccentricities of the orbits: not all orbits are circles; indeed, most are not. The orbital speed depends on where an object is along its elliptical orbit, as Kepler taught us. So yet again we must make some statistical inference about the distribution of eccentricities. These kinds of estimates are both doable and subject to going badly wrong.
The net effect is that we wind up looking at distributions of relative velocities, and trying to perceive whether there is an excess high-velocity tail over and above the Newtonian expectation. This is far enough from my expertise that I do not feel qualified to judge between the works cited above. It takes time to sort these things out, and hopefully we can all come to agreement on what it is that we’re seeing. Right now, we’re not all seeing eye-to-eye.
There is a whole session devoted to this topic at the upcoming meeting on MOND. The primary protagonists will be there, so hopefully some progress can be made. At least it should be entertaining.
I recently traveled to my first international meeting since the Covid pandemic began. It was good to be out in the world again. It also served as an excellent reminder of the importance of in-person interactions. On-line interactions are not an adequate substitute. I’d like to be able to recount all that I learned there, but it is too much. This post will touch on one of the much-discussed topics, our own Milky Way Galaxy.
The Milky Way is a normal galaxy – it is part of the forest. It is easy to get lost in the leaves when one has access to data for millions going on billions of individual stars. These add up to a normal spiral galaxy, and we know a lot about external spirals that can help inform our picture of our own home.
For example, by assuming that the Milky Way falls along the radial acceleration relation defined by other spiral galaxies, I was able to build a mass model of its surface density profile. The resulting mass distribution is considerably more detailed than the usual approach of assuming a smooth exponential disk, which would be a straight line in the right-hand plot below. With the level of detail becoming available from missions like the Gaia satellite, it is necessary to move beyond such approximations.
Left: Spiral structure in the Milky Way traced by regions of gas ionized by young stars (HII regions, in red) and by the birthplaces of giant molecular clouds (GMCs, in blue). Right: the azimuthally-averaged surface density profile of stars inferred from the rotation curve of the Milky Way using the Radial Acceleration Relation. The features inferred kinematically correspond to the spiral arms known from star counts, providing a local example of Renzo’s Rule.
This model was built before Gaia data became available, and is not informed by it. Rather, I took the terminal velocities measured by McClure-Griffiths & Dickey, which provide the estimate of the Milky Way rotation curve that is most directly comparable to what we measure in external spirals, and worked out the surface density profile using the radial acceleration relation. The resulting model possesses bumps and wiggles like those we see corresponding to spiral arms in external galaxies. And indeed, it turns out that the locations of these features correspond with known spiral arms. Those are independent observations: one is from the kinematics of interstellar gas, the other from traditional star counts.
The model turns out to have a few further virtues. It matches the enclosed mass profile of the inner bulge/bar region of the Galaxy without any attempt at a specific fit. It reconciles the rotation curve measured with stars using Gaia data with that measured using gas in the interstellar medium – a subtle difference that was nevertheless highly significant. It successfully predicts that the rotation curve beyond the solar radius would not be perfectly flat, but rather decline at a specific rate – and exactly that rate was subsequently measured using Gaia. These are the sort of results that inclines one to believe that the underlying physics has to be MOND. Inferring maps of the mass distribution with this level of detail is simply not possible using a dark matter model.
The rotation curve of the Milky as observed in interstellar gas (light grey) and as fit to the radial acceleration relation (blue line). Only the region from 3 to 8 kpc has been fit; the rest follows. This matches well stellar observations from the inner, barred region of the Milky Way (dark grey squares: Portail et al. 2017) and the gradual decline of the outer rotation curve (black squares: Eilers et al. 2019) once corrected for the presence of bumps and wiggles due to spiral arms. These require taking numerical derivatives for use in the Jeans equation; the red squares show the conventional result obtained when neglecting this effect by assuming a smooth exponential surface density profile. See McGaugh (2008 [when the method was introduced and the bulge/bar model for the inner region was built], 2016 [the main fitting paper], 2018 [an update to the distance to the Galactic center], 2019 [including bumps & wiggles in the Gaia analysis]).
Great, right? It is. It also makes a further prediction: we can use the mass model to predict the vertical motions of stars perpendicular to the Milky Way’s disk.
Most of the kinetic energy of stars orbiting in the solar neighborhood is invested in circular motion: the vast majority of stars are orbiting in the same direction in the same plane at nearly the same speed. There is some scatter, of course, but radial motions due to orbital eccentricities represent a small portion of the kinetic energy budget. As stars go round and round, the also bob up and down, oscillating perpendicular to the plane of the disk. The energy invested in these vertical motions is also small, which is why the disk of the Milky Way is thin.
View of the Milky Way in the infrared provided by the COBE satellite. The dust lanes that afflict optical light are less severe at these wavelengths, revealing that the stellar disk of the Milky Way is thin but for the peanut-shaped bulge/bar at the center.
Knowing the surface density profile of the Milky Way disk, we can predict the vertical motions. In the context of dark matter, most of the restoring force that keeps stars near the central plane is provided by the stars themselves – the dark matter halo is quasi-spherical, and doesn’t contribute much to the restoring force of the disk. In MOND, the stars and gas are all there is. So the prediction is straightforward (if technically fraught) in both paradigms. Here is a comparison of both predictions with data from Bovy & Rix (2013).
The dynamical surface density implied by vertical motions (data from Bovy & Rix 2013). The dark blue line is the prediction of the model surface density described above – assuming Newtonian gravity. The light blue line is the naive prediction of MOND.
Looks great again, right? The dark blue line goes right through the data with zero fitting. The only exception is in the radial range 5.5 to 6.4 kpc, which turns out to be where the stars probing the vertical motion are maximally different from the gas informing the prediction: we’re looking at different Galactic longitudes, right where there is or is not a spiral arm, so perhaps we should get a different answer in this range. Theory gives us the right answer, no muss, no fuss.
Except, hang on – the line that fits is the Newtonian prediction. The prediction of MOND overshoots the data. It gets the shape right, but the naive MOND prediction is for more vertical motion than we see.
By the “naive” MOND prediction, I mean that we assume that MOND gives the same boost in the vertical direction as it does in the radial direction. This is the obvious first thing to try, but it is not necessarily what happens in all possible MOND theories. Indeed, there are some flavors of modified inertia in which it should not. However, one would expect some boost, and in these data there appears to be none. We get the right answer with just Newton and stars. There’s not even room for much dark matter.
I hope Gaia helps us sort this out. I worry that it will provide so much information that we risk missing the big picture for all the leaves.
This leaves us in a weird predicament. The radial force is extraordinarily well-described by MOND, which reveals details that we could never hope to access if all we know about is dark matter. But if we spot Newtonian gravity this non-Newtonian information from the radial motion, it predicts the correct vertical motion. It’s like we have MOND in one direction and Newton in another.
This makes no sense, so is one of the things that worries me most about MOND. It is not encouraging for dark matter either – we don’t get to spot ourselves MOND in the radial direction then pretend that dark matter did it. At present, it feels like we are up the proverbial creek without a paddle.
I had written most of the post below the line before an exchange with a senior colleague who accused me of asking us to abandon General Relativity (GR). Anyone who read the last post knows that this is the opposite of true. So how does this happen?
Much of the field is mired in bad ideas that seemed like good ideas in the 1980s. There has been some progress, but the idea that MOND is an abandonment of GR I recognize as a misconception from that time. It arose because the initial MOND hypothesis suggested modifying the law of inertia without showing a clear path to how this might be consistent with GR. GR was built on the Equivalence Principle (EP), the equivalence1 of gravitational charge with inertial mass. The original MOND hypothesis directly contradicted that, so it was a fair concern in 1983. It was not by 19842. I was still an undergraduate then, so I don’t know the sociology, but I get the impression that most of the community wrote MOND off at this point and never gave it further thought.
I guess this is why I still encounter people with this attitude, that someone is trying to rob them of GR. It’s feels like we’re always starting at square one, like there has been zero progress in forty years. I hope it isn’t that bad, but I admit my patience is wearing thin.
I’m trying to help you. Don’t waste you’re entire career chasing phantoms.
What MOND does ask us to abandon is the Strong Equivalence Principle. Not the Weak EP, nor even the Einstein EP. Just the Strong EP. That’s a much more limited ask that abandoning all of GR. Indeed, all flavors of EP are subject to experimental test. The Weak EP has been repeatedly validated, but there is nothing about MOND that implies platinum would fall differently from titanium. Experimental tests of the Strong EP are less favorable.
I understand that MOND seems impossible. It also keeps having its predictions come true. This combination is what makes it important. The history of science is chock full of ideas that were initially rejected as impossible or absurd, going all the way back to heliocentrism. The greater the cognitive dissonance, the more important the result.
Continuing the previous discussion of UT, where do we go from here? If we accept that maybe we have all these problems in cosmology because we’re piling on auxiliary hypotheses to continue to be able to approximate UT with FLRW, what now?
I don’t know.
It’s hard to accept that we don’t understand something we thought we understood. Scientists hate revisiting issues that seem settled. Feels like a waste of time. It also feels like a waste of time continuing to add epicycles to a zombie theory, be it LCDM or MOND or the phoenix universe or tired light or whatever fantasy reality you favor. So, painful as it may be, one has find a little humility to step back and take account of what we know empirically independent of the interpretive veneer of theory.
Still, to give one pertinent example, BBN only works if the expansion rate is as expected during the epoch of radiation domination. So whatever is going on has to converge to that early on. This is hardly surprising for UT since it was stipulated to contain GR in the relevant limit, but we don’t actually know how it does so until we work out what UT is – a tall order that we can’t expect to accomplish overnight, or even over the course of many decades without a critical mass of scientists thinking about it (and not being vilified by other scientists for doing so).
Another example is that the cosmological principle – that the universe is homogeneous and isotropic – is observed to be true in the CMB. The temperature is the same all over the sky to one part in 100,000. That’s isotropy. The temperature is tightly coupled to the density, so if the temperature is the same everywhere, so is the density. That’s homogeneity. So both of the assumptions made by the cosmological principle are corroborated by observations of the CMB.
The cosmological principle is extremely useful for solving the equations of GR as applied to the whole universe. If the universe has a uniform density on average, then the solution is straightforward (though it is rather tedious to work through to the Friedmann equation). If the universe is not homogeneous and isotropic, then it becomes a nightmare to solve the equations. One needs to know where everything was for all of time.
Starting from the uniform condition of the CMB, it is straightforward to show that the assumption of homogeneity and isotropy should persist on large scales up to the present day. “Small” things like galaxies go nonlinear and collapse, but huge volumes containing billions of galaxies should remain in the linear regime and these small-scale variations average out. One cubic Gigaparsec will have the same average density as the next as the next, so the cosmological principle continues to hold today.
Anyone spot the rub? I said homogeneity and isotropy should persist. This statement assumes GR. Perhaps it doesn’t hold in UT?
This aspect of cosmology is so deeply embedded in everything that we do in the field that it was only recently that I realized it might not hold absolutely – and I’ve been actively contemplating such a possibility for a long time. Shouldn’t have taken me so long. Felten (1984) realized right away that a MONDian universe would depart from isotropy by late times. I read that paper long ago but didn’t grasp the significance of that statement. I did absorb that in the absence of a cosmological constant (which no one believed in at the time), the universe would inevitably recollapse, regardless of what the density was. This seems like an elegant solution to the flatness/coincidence problem that obsessed cosmologists at the time. There is no special value of the mass density that provides an over/under line demarcating eternal expansion from eventual recollapse, so there is no coincidence problem. All naive MOND cosmologies share the same ultimate fate, so it doesn’t matter what we observe for the mass density.
MOND departs from isotropy for the same reason it forms structure fast: it is inherently non-linear. As well as predicting that big galaxies would form by z=10, Sanders (1998) correctly anticipated the size of the largest structures collapsing today (things like the local supercluster Laniakea) and the scale of homogeneity (a few hundred Mpc if there is a cosmological constant). Pretty much everyone who looked into it came to similar conclusions.
But MOND and cosmology, as we know it in the absence of UT, are incompatible. Where LCDM encompasses both cosmology and the dynamics of bound systems (dark matter halos3), MOND addresses the dynamics of low acceleration systems (the most common examples being individual galaxies) but says nothing about cosmology. So how do we proceed?
For starters, we have to admit our ignorance. From there, one has to assume some expanding background – that much is well established – and ask what happens to particles responding to a MONDian force-law in this background, starting from the very nearly uniform initial condition indicated by the CMB. From that simple starting point, it turns out one can get a long way without knowing the details of the cosmic expansion history or the metric that so obsess cosmologists. These are interesting things, to be sure, but they are aspects of UT we don’t know and can manage without to some finite extent.
For one, the thermal history of the universe is pretty much the same with or without dark matter, with or without a cosmological constant. Without dark matter, structure can’t get going until after thermal decoupling (when the matter is free to diverge thermally from the temperature of the background radiation). After that happens, around z = 200, the baryons suddenly find themselves in the low acceleration regime, newly free to respond to the nonlinear force of MOND, and structure starts forming fast, with the consequences previously elaborated.
But what about the expansion history? The geometry? The big questions of cosmology?
Again, I don’t know. MOND is a dynamical theory that extends Newton. It doesn’t address these questions. Hence the need for UT.
I’ve encountered people who refuse to acknowledge4 that MOND gets predictions like z=10 galaxies right without a proper theory for cosmology. That attitude puts the cart before the horse. One doesn’t look for UT unless well motivated. That one is able to correctly predict 25 years in advance something that comes as a huge surprise to cosmologists today is the motivation. Indeed, the degree of surprise and the longevity of the prediction amplify the motivation: if this doesn’t get your attention, what possibly could?
There is no guarantee that our first attempt at UT (or our second or third or fourth) will work out. It is possible that in the search for UT, one comes up with a theory that fails to do what was successfully predicted by the more primitive theory. That just lets you know you’ve taken a wrong turn. It does not mean that a correct UT doesn’t exist, or that the initial prediction was some impossible fluke.
One candidate theory for UT is bimetric MOND. This appears to justify the assumptions made by Sanders’s early work, and provide a basis for a relativistic theory that leads to rapid structure formation. Whether it can also fit the acoustic power spectrum of the CMB as well as LCDM and AeST has yet to be seen. These things take time and effort. What they really need is a critical mass of people working on the problem – a community that enjoys the support of other scientists and funding institutions like NSF. Until we have that5, progress will remain grudgingly slow.
1The equivalence of gravitational charge and inertial mass means that the m in F=GMm/d2 is identically the same as the m in F=ma. Modified gravity changes the former; modified inertia the latter.
2Bekenstein & Milgrom (1984) showed how a modification of Newtonian gravity could avoid the non-conservation issues suffered by the original hypothesis of modified inertia. They also outlined a path towards a generally covariant theory that Bekenstein pursued for the rest of his life. That he never managed to obtain a completely satisfactory version is often cited as evidence that it can’t be done, since he was widely acknowledged as one of the smartest people in the field. One wonders why he persisted if, as these detractors would have us believe, the smart thing to do was not even try.
4I have entirely lost patience with this attitude. If a phenomena is correctly predicted in advance in the literature, we are obliged as scientists to take it seriously+. Pretending that it is not meaningful in the absence of UT is just an avoidance strategy: an excuse to ignore inconvenient facts.
+I’ve heard eminent scientists describe MOND’s predictive ability as “magic.” This also seems like an avoidance strategy. I, for one, do not believe in magic. That it works as well as it does – that it works at all – must be telling us something about the natural world, not the supernatural.
5There does exist a large and active community of astroparticle physicists trying to come up with theories for what the dark matter could be. That’s good: that’s what needs to happen, and we should exhaust all possibilities. We should do the same for new dynamical theories.