[astro-ph/0409768] Constraining Neutrino Masses by CMB Experiments Alone

Authors:  Kazuhide Ichikawa, Masataka Fukugita, Masahiro Kawasaki
Abstract:  It is shown that a subelectronvolt upper limit can be derived on the neutrino mass from the CMB data alone in the Lambda CDM model with the power-law adiabatic perturbations, without the aid of any other cosmological data. Assuming the flatness of the universe, the constraint we can derive from the current WMAP observations is \sum m_nu < 2.2 eV at the 95% confidence level for the sum over three species of neutrinos (m_nu < 0.75 eV for the degenerate neutrinos). This constraint may be loosened if we abandon the flatness assumption but only by up to 4% for Omega_tot=1.02, the WMAP limit on the spatial curvature. We argue that it would be difficult to improve the limit much beyond \sum m_nu \lesssim 1.5 eV using only the CMB data, even if their statistics are substantially improved. However, a significant improvement of the limit is possible if an external input is introduced that constrains the Hubble constant from below. The parameter correlation and the mechanism of CMB perturbations that give rise to the limit on the neutrino mass are also elucidated.
[PDF]  [PS]  [BibTex]  [Bookmark]

Discussion related to specific recent arXiv papers
Post Reply
Anze Slosar
Posts: 183
Joined: September 24 2004
Affiliation: Brookhaven National Laboratory
Contact:

[astro-ph/0409768] Constraining Neutrino Masses by CMB Exper

Post by Anze Slosar » October 01 2004

The article claims to get a decent constraint on neutrino mass from CMB data alon. I always though you cannot constrain neutrino mass from CMB alone and that this is the main reason why WMAP didn't put constraint on neutrino mass from its data alone, but combined with LSS (basically you get other parameters from CMB and LSS then constrains omega_nu in my view)

Authors of this article use a very complicated data reduction procedure in which they find global minimum in chi^2 by comparing lots of CMBFAST models with data and then they constrain position and ration of peak height which they in turn use to constrain omega_nu. What is the advantage of doing this instead of straight MCMC and how can you reliably estimate errors in this manner? Is there something fundamentally new that I overlooked?

Alessandro Melchiorri
Posts: 129
Joined: September 24 2004
Affiliation: University of Rome
Contact:

Post by Alessandro Melchiorri » October 01 2004

I made a similar analysis and I have'nt found any interesting constraint from WMAP on neutrino absolute mass (so I agree with Anze).
The thing you can due is to rule out at few sigmas flat mixed dark matter model with no lambda.

Sarah Bridle
Posts: 144
Joined: September 24 2004
Affiliation: University College London (UCL)
Contact:

delta chisq and confidence limits

Post by Sarah Bridle » October 02 2004

Hi, yes this result is quite surprising.

I think looking at the best fit point for fixed values of [tex]\omega_{\nu}[/tex] is quite interesting in itself, even though the interpretation in terms of probability as a function of [tex]\omega_{\nu}[/tex] is not so simple.
It is surprising to me that you get such a big [tex]\Delta \chi^2[/tex] for such a small change in the neutrino mass.
It also seems like a nice idea to me to look at how the features of the CMB power spectrum depend on [tex]\omega_{\nu}[/tex].

Not sure how significant this point is, but re interpretation of [tex]\Delta \chi^2[/tex] in terms of [tex]{\rm Pr}(\omega_{\nu})[/tex]:

The authors' main result comes from the observation that [tex]\Delta \chi^2=4[/tex] for [tex]\omega_{\nu}=0.024[/tex] which can be converted into [tex]\sum m_{\nu} = 2.25[/tex] as quoted in the abstract.
They interpret this as a 95 per cent detection, since [tex]\Delta \chi^2=4[/tex].
However this conversion is only possible for a 1d Gaussian.

Whereas I think that if they are going to extrapolate the results from a Gaussian then a 7d Gaussian would be more appropriate, since they are looking at a minimum in 7d space.
The [tex]\Delta \chi^2[/tex] for a 95 per cent detection is much higher for higher dimensions.
The table in Numerical Recipes only goes up to 6d, but for 95 per cent confidence in 6d you need [tex]\Delta \chi^2=12.8[/tex]. For a 6d Gaussian [tex]\Delta \chi^2=7[/tex] corresponds to 68 per cent.
So maybe for a 7d Gaussian [tex]\Delta \chi^2=4[/tex] is about a half sigma detection?

Also, if you have a 1d Gaussian peaked at [tex]x=0[/tex] and you say that x>0, then you would need [tex]\Delta \chi^2[/tex] greater than 4 for a 95 per cent confidence detection.
Since [tex]\omega_{\nu}[/tex]>0 then this would also need to be taken into account too which would also weaken the detection.

Anze Slosar
Posts: 183
Joined: September 24 2004
Affiliation: Brookhaven National Laboratory
Contact:

Post by Anze Slosar » October 04 2004

I see what do you mean, but my main objection is still against their method: basically, by reducing the entire power spectrum to 6 numbers representing the position of the peak, they are throwing away a lot of data... The correct thing would be to run MCMC and then simply look at the marginalised probability dist. for [tex]\omega_\nu[/tex]: and then they will probably get a half sigma detection or something as you say.

While analyzing latest VSA data batch there was some uproad re neutrino mass, but I think it turned out to be just sampling artifact... Have a look at [tex]f_\nu[/tex] plot on Tegmark's site (http://www.hep.upenn.edu/~max/), even for [tex]f_\nu =1[/tex] changes are fairly small, while their detection is for [tex]f_\nu[/tex] <<1.

Sarah Bridle
Posts: 144
Joined: September 24 2004
Affiliation: University College London (UCL)
Contact:

checking delta chisq

Post by Sarah Bridle » October 04 2004

I got the impression they only used the additional fitting parameters later in the paper, and that the main result does not use this.

I guess it would be very easy to check the [tex]\chi^2[/tex] values for the two sets of parameters in rows 1 and 5 of Table 1..

If these are right (and assuming they really do find the minimum in 6d, which seems likely given that they have found a lower minimum than those in Spergel et al and Tegmark et al) then it is surely just a question of how one interprets a [tex]\Delta \chi^2[/tex] of 4 ?

Anze Slosar
Posts: 183
Joined: September 24 2004
Affiliation: Brookhaven National Laboratory
Contact:

Re: checking delta chisq

Post by Anze Slosar » October 04 2004

Sarah Bridle wrote: If these are right (and assuming they really do find the minimum in 6d, which seems likely given that they have found a lower minimum than those in Spergel et al and Tegmark et al) then it is surely just a question of how one interprets a \Delta \chi^2 of 4 ?

Ahh, sorry, I misread the paper, now I see what they are doing... Now it sounds somewhat more sensible, although I am not sure either how should one interpret this [tex]\Delta \chi^2[/tex] of 4... Need to think about...

Alessandro Melchiorri
Posts: 129
Joined: September 24 2004
Affiliation: University of Rome
Contact:

Post by Alessandro Melchiorri » October 05 2004

Uhm, no I don't believe this result. 7-10 eV would be a more reasonable limit.
BTW a more general point.
I always feel a bit confused in interpreting a \delta chi2=4 as 95% c.l. when the best fit is 1428 and the degrees of freedom are 1341. You should also consider that many of the degrees of freedom quoted come from datapoints at very small scales (or from TE) with big error bars so their contribution to the overall chi2 is small and considering them IMHO is like cheating a bit. What do you think ? :-)

conservative Al.

Anze Slosar
Posts: 183
Joined: September 24 2004
Affiliation: Brookhaven National Laboratory
Contact:

Post by Anze Slosar » October 05 2004

Alessandro Melchiorri wrote:Uhm, no I don't believe this result. 7-10 eV would be a more reasonable limit.
BTW a more general point.
I always feel a bit confused in interpreting a \delta chi2=4 as 95% c.l. when the best fit is 1428 and the degrees of freedom are 1341. You should also consider that many of the degrees of freedom quoted come from datapoints at very small scales (or from TE) with big error bars so their contribution to the overall chi2 is small and considering them IMHO is like cheating a bit. What do you think ? :-)
Yeah, I tend to agree. There is a well prescribed procedure to calculate the marginalised probability distribution for [tex]\omega_\nu[/tex] and this method gives a few eV as you say. Since they are not putting in extra physics they cannot obtain stronger constraints unless cheating. :)

Sarah Bridle
Posts: 144
Joined: September 24 2004
Affiliation: University College London (UCL)
Contact:

Delta chisq and dof

Post by Sarah Bridle » October 05 2004

I agree with Alessandro's criticism that looking at the [tex]\chi^2[/tex] per dof is misleading when you have lots of noisy data. But I think this is a separate issue from the issue of interpreting [tex]\Delta \chi^2[/tex].

If the probability distribution in 7d were a 7d Gaussian then it would be possible to look at the [tex]\Delta \chi^2[/tex] and convert into confidence levels.

If I had a 2d Gaussian in x and y I could do something analogous to what they did with the neutrino mass:
* fix x
* find the best fitting y
* note down the chisq at this best fit y and fixed x
* repeat for all values of fixed x
* plot [tex]\chi^2[/tex] at the best fit y as a function of fixed x
* find the minimum wrt x
* find the x value where the [tex]\chi^2[/tex] is increased by 4 wrt this minimum

The question is, how does this x value relate to the 95 per cent marginalised limit on x? I am saying that it does not have any relation for an nd Gaussian where n \ne 1.

I am saying that when they say 95 per cent confidence they should have put about 10 or 20 per cent confidence (and that's assuming its all Gaussian..)!

Please could someone check I'm not going mad about this point? Thanks very much indeed.

Anze Slosar
Posts: 183
Joined: September 24 2004
Affiliation: Brookhaven National Laboratory
Contact:

Re: Delta chisq and dof

Post by Anze Slosar » October 05 2004

Sarah Bridle wrote:
I am saying that when they say 95 per cent confidence they should have put about 10 or 20 per cent confidence (and that's assuming its all Gaussian..)!

Please could someone check I'm not going mad about this point? Thanks very much indeed.
Ok,

have a loot at

http://www.fiz.uni-lj.si/cosmo/aslosar/dstat.mws

I did exactly what you advocated. For 2D gaussians it seems that the
advocated procedure is exactly the same as full marginalisation, i.e.
the probablity of a funcion at its maximum given fixed x is proportional
to its marginalised values... However, I don't think it works in general...

Jochen Weller
Posts: 45
Joined: September 24 2004
Affiliation: Ludwig-Maximilians-University Munich
Contact:

Don't go mad

Post by Jochen Weller » October 05 2004

Hi Sarah

I agree entirely with you. The [tex]\Delta\Chi^2 = 4 \equiv 95\%[/tex] only applies for a 1-d likelihhod. But they have a 7-d likelihood. It would be like having a prior with NO errorbars
on all the other parameters fixed at the best fit value. Which I would say is wrong.

Sarah Bridle
Posts: 144
Joined: September 24 2004
Affiliation: University College London (UCL)
Contact:

my comments were barking up the wrong tree

Post by Sarah Bridle » October 05 2004

DOH! Thanks very much for your post Anze, this brings me back into the real world at least..
Somehow I had got sidetracked with an irrelevant issue: where the 95 per cent contour would be in 7d space.. so please ignore my postings above!

So I agree with what you say:
* what they do is the old "maximising" over other parameters trick, which the majority of people used to do before MCMC to save time;
* this is the same as marginalising properly if the pdf is a Gaussian;
* one possible explanation for their result cf others is that the pdf is not very Gaussian and so this maximisation gives a misleading answer

Sarah Bridle
Posts: 144
Joined: September 24 2004
Affiliation: University College London (UCL)
Contact:

Seems like they may be right...?

Post by Sarah Bridle » November 06 2004

We concluded that their result might not hold when you do the full MCMC. So I ran a few chains and it looks like they might be right.. I put the chains at www.star.ucl.ac.uk/~sarah/chains

I know the chains are extremely ropey, but I'm posting this anyway because I realised I am not finding time to improve it and it seemed better to post this than not say anything.

I started all the chains from fnu=0.5 and I did a couple of CosmoMC runs to improve the covariance matrix. Cutting a measly 500 samples for burn-in (after which I see no trend in fnu or likelihood changing) gives www.star.ucl.ac.uk/~sarah/chains/fnu.png
You can see just how dodgy the chains are by looking at how each chain contributes to this result www.star.ucl.ac.uk/~sarah/chains/fnu_eachchain.png
(The vertical lines are two-tailed limits wheras we want an upper limit on fnu).
Seems to agree with their result pretty well.

This has made me think that they may not be wrong. One thing to note is that the dashed lines (mean likelihood) are v similar to the solid lines (marginalised likelihood) which happens when the pdf is quite Gaussian. Their method would give the right answer if the pdf were Gaussian. So that hangs together.

Anyway, obviously these chains are awful but I thought better to share it with you than to keep it on my disk. Does anyone have anything more reliable?

Kazuhide Ichikawa
Posts: 3
Joined: March 25 2006
Affiliation: Institute for Cosmic Ray Research, University of Tokyo

[astro-ph/0409768] Constraining Neutrino Masses by CMB Exper

Post by Kazuhide Ichikawa » March 25 2006

Recent analyses (using WMAP 1st yr only) are consistent with ours:
astro-ph/0507503, MacTavish et al., Fig. 8,
hep-ph/0602058, Hannestad, sec. 4.3.2,
astro-ph/0603494, Lesgourgues & Pastor, sec. 5.2.

Also, WMAP 3rd yr result is consistent.
astro-ph/0603449, Spergel et al., sec. 7.2.1.

Post Reply