Just a link April 5, 2009Posted by dorigo in Blogroll, news, physics, science.
Tags: Higgs boson, science reporting, Tevatron
I read with amusement (and some effort) a spanish account by Francis (th)E mule of Michael Dittmar’s controversial seminar of last March 19th. I paste the link here for several reasons: since I believe it might be of interest to some of you, to have a place to store it, and because I am not insensitive to flattery:
“Entre el público se encontraba Tomasso Dorigo [...] (r)esponsable del mejor blog sobre física de partículas elementales del mundo”
Muchas gracias, Francis -but please note: my name spells with two m’s and one s!
Tags: CDF, DZERO, electroweak fits, Gfitter, Higgs boson, LEP, SLD, standard model, Tevatron, top quark, W boson
A recent discussion in this blog between well-known theorists and phenomenologists, centered on the real meaning of the experimental measurements of top quark and W boson masses, Higgs boson cross-section limits, and other SM observables, convinces me that some clarification is needed.
The work has been done for us: there are groups that do exactly that, i.e. updating their global fits to express the internal consistency of all those measurements, and the implications for the search of the Higgs boson. So let me go through the most important graphs below, after mentioning that most of the material comes from the LEP electroweak working group web site.
First of all, what goes in the soup ? Many things, but most notably, the LEP I/SLD measurements at the Z pole, the top quark mass measurements by CDF and DZERO, and the W mass measurements by CDF, DZERO, and LEP II. Let us give a look at the mass measurements, which have recently been updated.
For the top mass, the situation is the one pictured in the graph shown below. As you can clearly see, the CDF and DZERO measurements have reached a combined precision of 0.75% on this quantity.
The world average is now at . I am amazed to see that the first estimate of the top mass, made by a handful of events published by CDF in 1994 (a set which did not even provide a conclusive “observation-level” significance at the time) was so dead-on: the measurement back then was ! (for comparison, the DZERO measurement of 1995, in their “observation” paper, was ).
As far as global fits are concerned, there is one additional point to make for the top quark: knowing the top mass any better than this has become, by now, useless. You can see it by comparing the constraints on coming from the indirect measurements and W mass measurements (shown by the blue bars at the bottom of the graph above) with the direct measurements at the Tevatron (shown with the green band). The green band is already too narrow: the width of the blue error bars compared to the narrow green band tells us that the SM does not care much where exactly the top mass is, by now.
Then, let us look at the W mass determinations. Note, the graph below shows the situation BEFORE the latest DZERO result;, obtained with 1/fb of data, and which finds ; its inclusion would not change much of the discussion below, but it is important to stress it.
Here the situation is different: a better measurement would still increase the precision of our comparisons with indirect information from electroweak measurements at the Z. This is apparent by observing that the blue bars have width still smaller than the world average of direct measurements (again in green). Narrow the green band, and you can still collect interesting information on its consistency with the blue points.
Finally, let us look at the global fit: the electroweak working group at LEP displays in the by now famous “blue band plot”, shown below for March 2009 conferences. It shows the constraints on the Higgs boson mass coming from all experimental inputs combined, assuming that the Standard Model holds.
I will not discuss this graph in details, since I have done it repeatedly in the past. I will just mention that the yellow regions have been excluded by direct searches of the Higgs boson at LEP II (on the left, the wide yellow area) and the Tevatron ( the narrow strip on the right). From the plot you should just gather that a light Higgs mass is preferred (the central value being 90 GeV, with +36 and -27 GeV one-sigma error bars). Also, a 95% confidence-level exclusion of masses above 163 GeV is implied by the variation of the global fit with Higgs mass.
I have started to be a bit bored by this plot, because it does not do the best job for me. For one thing, the LEP II limit and the Tevatron limit on the Higgs mass are treated as if they were equivalent in their strength, something which could not be possibly farther from the truth. The truth is, the LEP II limit is a very strong one -the probability that the Higgs has a mass below 112 GeV, say, is one in a billion or so-, while the limit obtained recently by the Tevatron is just an “indication”, because the excluded region (160 to 170 GeV) is not excluded strongly: there still is a one-in-twenty chance or so that the real Higgs boson mass indeed lies there.
Another thing I do not particularly like in the graph is that it attempts to pack too much information: variations of , inclusion of low-Q^2 data, etcetera. A much better graph to look at is the one produced by the GFitter group instead. It is shown below.
In this plot, the direct search results are introduced with their actual measured probability of exclusion as a function of Higgs mass, and not just in a digital manner, yes/no, as the yellow regions in the blue band plot. And in fact, you can see that the LEP II limit is a brick wall, while the Tevatron exclusion acts like a smooth increase in the global of the fit.
From the black curve in the graph you can get a lot of information. For instance, the most likely values, those that globally have a 1-sigma probability of being one day proven correct, are masses contained in the interval 114-132 GeV. At two-sigma, the Higgs mass is instead within the interval 114-152 GeV, and at three sigma, it extends into the Tevatron-excluded band a little, 114-163 GeV, with a second region allowed between 181 and 224 GeV.
In conclusion, I would like you to take away the following few points:
- Future indirect constraints on the Higgs boson mass will only come from increased precision measurements of the W boson mass, while the top quark has exhausted its discrimination power;
- Global SM fits show an overall very good consistency: there does not seem to be much tension between fits and experimental constraints;
- The Higgs boson is most likely in the 114-132 GeV range (1-sigma bounds from global fits).
Zooming in on the Higgs March 24, 2009Posted by dorigo in news, physics, science.
Tags: CDF, DZERO, Higgs boson, LEP, MSSM, standard model, supersymmetry, Tevatron, top quark, W boson
Yesterday Sven Heinemeyer kindly provided me with an updated version of a plot which best describes the experimental constraints on the Higgs boson mass, coming from electroweak observables measured at LEP and SLD, and from the most recent measurements of W boson and top quark masses. It is shown on the right (click to get the full-sized version).
The graph is a quite busy one, but I will try below to explain everything one bit at a time, hoping I keep things simple enough that a non-physicist can understand it.
The axes show suitable ranges of values of the top quark mass (varying on the horizontal axis) and of the W boson masses (on the vertical axis). The value of these quantities is functionally dependent (because of quantum effects connected to the propagation of the particles and their interaction with the Higgs field) on the Higgs boson mass.
The dependence, however, is really “soft”: if you were to double the Higgs mass by a factor of two from its true value, the effect on top and W masses would be only of the order of 1% or less. Because of that, only recently have the determinations of top quark and W boson masses started to provide meaningful inputs for a guess of the mass of the Higgs.
Top mass and W mass measurements are plotted in the graphs in the form of ellipses encompassing the most likely values: their size is such that the true masses should lie within their boundaries, 68% of the time. The red ellipse shows CDF results, and the blue one shows DZERO results.
There is a third measurement of the W mass shown in the plot: it is displayed as a horizontal band limited by two black lines, and it comes from the LEP II measurements. The band also encompasses the 68% most likely W masses, as ellipses do.
In addition to W and top masses, other experimental results constrain the mass of top, W, and Higgs boson. The most stringent of these results are those coming from the LEP experiment at CERN, from detailed analysis of electroweak interactions studied in the production of Z bosons. A wide band crossing the graph from left to right, with a small tilt, encompasses the most likely region for top and W masses.
So far we have described measurements. Then, there are two different physical models one should consider in order to link those measurements to the Higgs mass. The first one is the Standard Model: it dictates precisely the inter-dependence of all the parameters mentioned above. Because of the precise SM predictions, for any choice of the Higgs boson mass one can draw a curve in the top mass versus W mass plane. However, in the graph a full band is hatched instead. This correspond to allowing the Higgs boson mass to vary from a minimum of 114 GeV to 400 GeV. 114 GeV is the lower limit on the Higgs boson mass found by the LEP II experiments in their direct searches, using electron-positron collisions; while 400 GeV is just a reference value.
The boundaries of the red region show the functional dependence of Higgs mass on top and W masses: an increase of top mass, for fixed W mass, results in an increase of the Higgs mass, as is clear by starting from the 114 GeV upper boundary of the red region, since one then would move into the region, to higher Higgs masses. On the contrary, for a fixed top mass, an increase in W boson mass results in a decrease of the Higgs mass predicted by the Standard Model. Also note that the red region includes a narrow band which has been left white: it is the region corresponding to Higgs masses varying between 160 and 170 GeV, the masses that direct searches at the Tevatron have excluded at 95% confidence level.
The second area, hatched in green, is not showing a single model predictions, but rather a range of values allowed by varying arbitrarily many of the parameters describing the supersymmetric extension of the SM called “MSSM”, its “minimal” extension. Even in the minimal extension there are about a hundred additional parameters introduced in the theory, and the values of a few of those modify the interconnection between top mass and W mass in a way that makes direct functional dependencies in the graph impossible to draw. Still, the hatched green region shows a “possible range of values” of the top quark and W boson masses. The arrow pointing down only describes what is expected for W and top masses if the mass of supersymmetric particles is increased from values barely above present exclusion limits to very high values.
So, to summarize, what to get from the plot ? I think the graph describes many things in one single package, and it is not easy to get the right message from it alone. Here is a short commentary, in bits.
- All experimental results are consistent with each other (but here, I should add, a result from NuTeV which finds indirectly the W mass from the measured ratio of neutral current and charged current neutrino interactions is not shown);
- Results point to a small patch of the plane, consistent with a light Higgs boson if the Standard Model holds
- The lower part of the MSSM allowed region is favored, pointing to heavy supersymmetric particles if that theory holds
- Among experimental determinations, the most constraining are those of the top mass; but once the top mass is known to within a few GeV, it is the W mass the one which tells us more about the unknown mass of the Higgs boson
- One point to note when comparing measurements from LEP II and the Tevatron experiments: when one draws a 2-D ellipse of 68% contour, this compares unfavourably to a band, which encompasses the same probability in a 1-D distribution. This is clear if one compares the actual measurements: CDF (with 200/pb of data), DZERO (with five times more statistics), LEP II (average of four experiments). The ellipses look like they are half as precise as the black band, while they are actually only 30-40% worse. If the above is obscure to you, a simple graphical explanation is provided here.
- When averaged, CDF and DZERO will actually beat the LEP II precision measurement -and they are sitting on 25 times more data (CDF) or 5 times more (DZERO).
A seminar against the Tevatron! March 20, 2009Posted by dorigo in news, physics, science.
Tags: CDF, DZERO, Higgs boson, LHC, Tevatron
I spent this week at CERN to attend the meetings of the CMS week – an event which takes place four times a year, when collaborators of the CMS experiment, coming from all parts of the world, get together at CERN to discuss detector commissioning, analysis plans, and recent results. It was a very busy and eventful week, and only now, sitting on a train that brings me back from Geneva to Venice, can I find the time to report with the due dedication on some things you might be interested to know about.
One thing to report on is certainly the seminar I eagerly attended on Thursday morning, by Michael Dittmar (ETH-Zurich). Dittmar is a CMS collaborator, and he talked at the CERN theory division on a tickling subject:”Why I never believed in the Tevatron Higgs sensitivity claims for Run 2ab”. The title did promise a controversial discussion, but I was really startled by its level, as much as by the defamation of which I felt personally to be a target. I will explain this below.
I have also to mention that by Thursday I had already attended to a reduced version of his talk, since he had given it on the previous day in another venue. Both I and John Conway had corrected him on a few plainly wrong statements back then, but I was puzzled to see he reiterated those false statements in the longer seminar! More on that below.
Dittmar’s obnoxious seminar
Dittmar started by saying he was infuriated by the recent BBC article where “a statement from the director of a famous laboratory” claimed that the Tevatron had 50% odds of finding a Higgs boson, in a certain mass range. This prompted him to prepare a seminar to express his scepticism. However, it turned out that his scepticism was not directed solely at the optimistic statement he had read, but at every single result on Higgs searches that CDF and DZERO had produced since Run I.
In order to discuss sensitivity and significances, the speaker made a un-illuminating digression on how counting experiments can or cannot obtain observation-level significances with their data depending on the level of background of their searches and the associated systematical uncertainties. His statements were very basic and totally uncontroversial on this issue, but he failed to focus on the fact that nowadays, nobody does counting experiments any more when searching for evidence of a specific model: our confidence in advanced analysis methods involving neural networks, shape analysis, and likelihood discriminants; the tuning of Monte Carlo simulations; and the accurate analytical calculations of high-order diagrams for Standard Model processes, have all grown tremendously with years of practice and studies, and these methods and tools overcome the problems of searches for small signals immersed in large backgrounds. One can be sceptical, but one cannot ignore the facts, as the speaker seemed inclined to.
Then Dittmar said that in order to judge the value of sensitivity claims for the future, one may turn to past studies and verify their agreement with the actual results. So he turned to the Tevatron Higgs Sensitivity studies of 2000 and 2003, two endeavours to which I had participated with enthusiasm.
He produced a plot showing the small signal of decays that the Tevatron 2000 study believed the two experiments could achieve with 10 inverse femtobarns of data, expressing his doubts that the “tiny excess” could mean an evidence for Higgs production. On the side of that graph, he had for comparison placed a result of CDF on real Run I data, where a signal of WH or ZH decays to four jets had been searched in the dijet invariant mass distribution of the two b-jets.
He commented that figure by saying half-mockingly that the data could have been used to exclude the standard model process of associated production, since the contribution from Z decays to b-quark pairs was sitting at a mass where one bin had fluctuated down by two standard deviations with respect to the sum of background processes. This ridiculous claim was utterly unsupported by the plot -which had an overall very good agreement between data and MC sources- and by the fact that the bins adjacent to the downward-fluctuating one were higher than the prediction. I found this claim really disturbing, because it tried to denigrate my experiment with a futile and incorrect argument. But I was about to get more upset for his next statement.
In fact, he went on to discuss the global expectation of the Tevatron on Higgs searches, a graph (see below) produced in 2000 after a big effort from several tens of people in CDF and DZERO.
He started by saying that the graph was confusing, and that it was not clear in the documentation how it had been produced, nor that it was the combination of CDF and DZERO sensitivity. This was very amusing, since sitting from the far back John Conway, a CDF colleague, shouted: “It says it in print on top of it: combined thresholds!”, then adding, with a pacate voice “…In case you’re wondering, I made that plot.” John had in fact been the leader of the Tevatron Higgs sensitivity study, not to mention the author of many of the most interesting searches for the higgs boson in CDF since then.
Dittmar continued his surreal talk with an overbid, by claiming that the plot had been produced “by assuming a 30% improvement in the mass resolution of pairs of b-jets, when nobody had not even the least idea on how such improvement could be achieved”.
I could not have put together a more personal, direct attack to years of my own work myself! It is no mystery that I worked on dijet resonances since 1992, but of course I am a rather unknown soldier in this big game; however, I felt the need to interrupt the speaker at this point -exactly as I had done at the shorter talk the day before.
I remarked that in 1998, one year before the Tevatron sensitivity study, I had produced a PhD thesis and public documents showing the observation of a signal of decays in CDF Run I data, and had demonstrated on that very signal how the use of ingenuous algorithms could reduce by at least 30% the dijet mass resolution, making the signal more prominent. The relevant plots are below, directly from my PhD thesis: judge for yourself.
In the plots, you can see how the excess over background predictions moves to the right as more and more refined jet energy corrections are applied, starting from the result of generic jet energy corrections (top) to optimized corrections (bottom) until the signal becomes narrower and centered at the true value. The plots on the left show the data and the background prediction, those on the right show the difference, which is due to Z decays to b-quark jet pairs. Needless to say, the optimization is done on Monte Carlo Z events, and only then checked on the data.
So I said that Dittmar’s statement was utterly false: we had an idea of how to do it, we had proven we could do it, and besides, the plots showing what we had done had been indeed included in the Tevatron 2000 report. Had he overlooked them ?
Dittmar seemed unbothered by my remark, and he responded that that small signal had not been confirmed in Run II data. His statement constituted an even more direct attack to four more years of my research time, spent on that very topic. I kept my cool, because when your opponent offers you on a silver plate the chance to verbally sodomize him, you cannot be too angry with him.
I remarked that a signal had indeed been found in Run II, amounting to about 6000 events after all selection cuts; it confirmed the past results. Dittmar then said that “to the best of his knowledge” this had not been published, so it did not really count. I then explained it was a 2008 NIM publication, and would he please document himself before making such unsubstantiated allegations? He shrugged his shoulders, said he would look more carefully for the paper, and went back to his talk.
His points about the Tevatron sensitivity studies were laid down: for a low-mass Higgs boson, the signal is just too small and backgrounds are too large, and the sensitivity of real searches is below expectations by a large factor. To stress this point, he produced a slide containing a plot he had taken from this blog! The plot (see on the left), which is my own concoction and not Tevatron-approved material, shows the ratio between observed limit to Higgs production and the expectations of the 2000 study. He pointed at the two points for 100-140 GeV Higgs boson masses, trying to prove his claim: The Tevatron is now doing three times worse than expected. He even uttered “It is time to confess: the sensitivity study was wrong by a large factor!”.
I could not help interrupting again: I had to stress that the plot was not approved material and was just a private interpretation of Tevatron results, but I did not deny its contents. The plot was indeed showing that low-mass searches were below par, but it was also showing that high-mass ones were amazingly in agreement with expectations worked at 10 years before. Then John Conway explained the low-mass discrepancy for the benefit of the audience, as he had done one day before for no apparent benefit of the speaker.
Conway explained that the study had been done under the hypothesis that an upgrade of our silicon detector would be financed by the DoE: it was in fact meant to prove the usefulness of funding an upgrade. A larger acceptance of inner silicon tracking boosts the sensitivity to identify b-quark jets from Higgs decays by a large factor, because any acceptance increase gets squared when computing the over-efficiency. So Dittmar could not really blame the Tevatron experiments for predicting something that would not materialize in a corresponding result, given that the DoE had denied the funding to build the upgraded detector!
I then felt compelled to add that by using my plot Dittmar was proving the opposite thesis of what he wanted to demonstrate: low-mass Tevatron searches were shown to underperform because of funding issues, rather than because of a wrong estimate of sensitivity; and high-mass searches, almost unhindered by the lack of an upgraded silicon, were in excellent agreement with expectations!
The speaker said that no, the high-mass searches were not in agreement, because their results could not be believed, and moved on to discuss those by taking real-data results by the Tevatron.
He said that the is a great channel at the LHC.
“Possible at the Tevatron ? I believe that the WW continuum background is much larger at a ppbar collider than at a pp collider, so my personal conclusion is that if the Tevatron people want to waste their time on it, good luck to them.”
Now, come on. I cannot imagine how a respectable particle physicist could drive himself into making such statements in front of a distinguished audience (which, have I mentioned it, included several theorists of the highest caliber, including none less than Edward Witten). Waste their time ? I felt I was wasting my time listening to him, but my determination of reporting his talk here kept me anchored to my chair, taking notes.
So this second part of the talk was not less unpleasant than the first part: Dittmar criticized the Tevatron high-mass Higgs results in the most incorrect, and scientifically dishonest, way that I could think of. Here is just a summary:
- He picked up a distribution of one particular sub-channel from one experiment, noting that it seemed to have the most signal-rich region showing a deficit of events. He then showed the global CDF+DZERO limit, which did not show a departure between expected and observed limit on Higgs cross section, and concluded that there was something fishy in the way the limit had been evaluated. But the limit is extracted from literally several dozens of those distributions -something he failed to mention despite having been warned of that very issue in advance.
- He picked up two neural-network output distributions for a search of Higgs at 160 and 165 GeV, and declared they could not be correct since they were very different in shape! John, from the back, replied “You have never worked with neural networks, have you ?” No, he had not. Had he, he would probably have understood that different mass points, optimized differently, can provide very different NN outputs.
- He showed another Neural Network output based on 3/fb of data, which had a pair of data points lying one standard deviation above the background predictions, and the corresponding plot for a search performed with improved statistics, which had instead a underfluctuation. He said he was puzzled by the effect. Again, some intervention from the audience was necessary, explaining that the methods are constantly reoptimized, and there is no wonder that adding more data can result in a different outcome. This produced a discussion when somebody from the audience tried to speculate that searches were maybe performed by looking at the data before choosing which method to use for a limit extraction! On the contrary of course, all Tevatron searches of the Higgs are blind analyses, where the optimization is performed on expected limits, using control samples, and Monte Carlo, and the data is only looked at afterwards.
- He showed that the Tevatron 2000 report had estimated a maximum Signal/Noise ratio for the H–>WW search of 0.34, and he picked up one random plot from the many searches of that channel by CDF and DZERO, showing that the signal to noise there was never larger than 0.15 or so. Explaining to him that the S/N of searches based on neural networks and combined discriminants is not a fixed value, and that many improvements have occurred in data analysis techniques in 10 years was useless.
Dittmar concluded his talk by saying that:
“Optimistic expectations might help to get funding! This is true, but it is also true that this approach eventually destroys some remaining confidence in science of the public.”.
His last slide even contained the sentence he had previously brought himself to uttering:
“It is the time to confess and admit that the sensitivity predictions were wrong”.
Finally, he encouraged LHC experiments to looking for the Higgs where the Tevatron had excluded it -between 160 and 170 GeV- because Tevatron results cannot be believed. I was disgusted: he most definitely places a strong claim on the prize of the most obnoxious talk of the year. Unfortunately for all, it was just as much an incorrect, scientifically dishonest, and dilettantesque lamentation, plus a defamation of a community of 1300 respected physicists.
In the end, I am really wondering what really moved Dittmar to such a disastrous performance. I think I know the answer, at least in part: he has been an advocate of the signature since 1998, and he must now feel bad for that beautiful process being proven hard to see, by his “enemies”. Add to that the frustration of seeing the Tevatron producing brilliant results and excellent performances, while CMS and Atlas are sitting idly in their caverns, and you might figure out there is some human factor to take into account. But nothing, in my opinion, can justify the mix he put together: false allegations, disregard of published material, manipulation of plots, public defamation of respected colleagues. I am sorry to say it, but even though I have nothing personal against Michael Dittmar -I do not know him, and in private he might even be a pleasant person-, it will be very difficult for me to collaborate with him for the benefit of the CMS experiment in the future.
Streaming video for Y(4140) discovery March 17, 2009Posted by dorigo in news, physics, science.
Tags: B physics, CDF, discoveries, QCD, standard model, Tevatron
The CDF collaboration will present at a public venue (Fermilab’s Wilson Hall) its discovery of the new Y(4140) hadron, a mysterious particle created in B meson decays, and observed to decay strongly into a state, a pair of vector mesons. I have described that exciting discovery in a recent post.
From this site you can connect to streaming video (starting at 4.00PM CDT, or 9.00PM GMT – should last about 1.30 hours).
DZERO refutes CDF’s multimuon signal… Or does it ? March 17, 2009Posted by dorigo in news, physics, science.
Tags: anomalous muons, CDF, DZERO, new physics, standard model, Tevatron
Hot off the press: Mark Williams, a DZERO member speaking at Moriond QCD 2009 -a yearly international conference in particle physics, where HEP experimentalists regularly present their hottest results- has shown today the preliminary results of their analysis of dimuon events, based on 900 inverse picobarns of proton-antiproton collision data. And the conclusion is…
DZERO searched for an excess of muons with large impact parameter by applying a data selection very similar, and when possible totally equivalent, to the one used by CDF in its recent study. Of course, the two detectors have entirely different hardware, software algorithms, and triggers, so there are certain limits to how closely one analysis can be replicated by the other experiment. However, the main machinery is quite similar: they count how many events have two muons produced within the first layer of silicon detector, and extrapolate to determine how many they expect to see which fail to yield a hit in that first layer, comparing to the actual number. They find no excess of large impact parameter muons.
Impact parameter, for those of you who have not followed this closely in the last few months, is the smallest distance between a track and the proton-antiproton collision vertex, in the plane transverse to the beam direction. A large impact parameter indicates that a particle has been produced in the decay of a parent body which was able to travel away from the interaction point before disintegrating. More information about the whole issue can be found in this series of posts, or by just clicking the “anomalous muons” tab in the column on the right of this text.
There are many things to say, but I will not say them all here now, because I am still digesting the presentation, the accompanying document produced by DZERO (not ready for public consumption yet), and the implications and subtleties involved. However, let me flash a few of the questions I am going to try and give an answer to with my readings:
- The paper does not address the most important question – what is DZERO’s track reconstruction efficiency as a function of track impact parameter ? They do discuss with some detail the complicated mixture of their data, which results from triggers which enforce that tracks have very small impact parameter -effectively cutting away all tracks with an impact parameter larger than 0.5cm- and a dedicated trigger which does not enforce an IP requirement; they also discuss their offline track reconstruction algorithms. But at a first sight it did not seem clear to me that they can actually reconstruct effectively tracks with impact parameters up to 2.5 cm as they claim. I would have inserted in the documents an efficiency graph for the reconstruction efficiency as a function of impact parameter, had I authored it.
- The paper shows a distribution of the decay radius of neutral K mesons, reconstructed from their decay into pair of charged pions. From the plot, the efficiency of reconstructing those pions is excessively small -some three times smaller than what it is in CMS, for instance. I need to read another paper by DZERO to figure out what drives their K-zero reconstruction efficiency to be so small, and whether this is in fact due to the decrease of effectiveness with track displacement.
- What really puzzles me, however, is the fact that they do not see *any* excess, while we know there must be in any case a significant one: decays in flight of charged kaons and pions. Why is it that CDF is riddled with those, while DZERO appears free of them ? To explain this point: charged kaons and pions yield muons, which get reconstructed as real muons with large impact parameter. If the decay occurs within the tracking volume, the track is partly reconstructed with the muon hits and partly with the kaon or pion hits. Now, while pions have a mass similar to that of muons, and thus the muon practically follows the pion trajectory faithfully, for kaons there must be a significant kink in the track trajectory. One expects that the track reconstruction algorithm will fail to associate inner hits to a good fraction of those tracks, and the resulting muons will belong to the “loose” category, without a correspondence in the “tight” muon category which has muons containing a silicon hit in the innermost layer of the silicon detector. This creates an excess of muons with large impact parameter. CDF does estimate that contribution, and it is quite large, of the order of tens of thousands of events in 743 inverse picobarns of data! Now where are those events in the DZERO dataset, then ?
Of course, you should not expect that my limited intellectual capabilities and my slow reading of a paper I have had in my hands for no longer than two hours can produce foulproof arguments. So the above is just a first pass, sort of a quick and dirty evaluation. I imagine I will be able to give an answer to those puzzles myself, at least in part, with a deeper look at the documentation. But, for the time being, this is what I have to say about the DZERO analysis.
Or rather, I should add something. By reading the above, you might get the impression that I am only criticizing DZERO out of bitterness for the failed discovery of the century by CDF… No, it is not the case: I have always thought, and I continue to think, that the multi-muon signal by CDF is some unaccounted-for background. And I do salute with relief and interest the new effort by DZERO on this issue. I actually thank them for providing their input on this mystery. However, I still retain some scepticism with respect to the findings of their study. I hope that scepticism can be wiped off by some input – maybe some reader belonging to DZERO wants to shed some light on the issues I mentioned above ? You are most welcome to do so!
UPDATE: Lubos pitches in, and guess what, he blames CDF… But Lubos the experimentalist is not better than Lubos the diplomat, if you know what I mean…
Other reactions will be collected below – if you have any to point to, please do so.
Tevatron excludes chunk of Higgs masses! March 13, 2009Posted by dorigo in news, physics, science.
Tags: CDF, DZERO, Higgs boson, Tevatron
This just in – the Fermilab site has the news on the new exclusion in a range of Higgs masses. At 95% C.L., the Higgs boson cannot have a mass in the 160-170 GeV range, as shown in the graph below. The new limit is shown by the orange band.
This is the first real exclusion range on the Higgs boson mass from CDF and DZERO. I will have more to say about this great new result during the weekend.
UPDATE: maybe the most interesting thing is not the limit shown above, but the information contained in the graph shown below. It shows how the combination of CDF and DZERO searches for the Higgs bosons end up agreeing with the background-only hypothesis (black hatched curve) or the background plus signal hypothesis (red curve), as a function of the unknown value of the Higgs boson mass. The full black line seems to favor the signal plus background hypothesis, although only marginally and at just the 1-sigma level, at around 130 GeV of mass:
However, they say that if you like sausages and if you follow laws, you should not ask how these things are made. The same goes with global limits, to some extent. In this case it is not a criticism of the limit by itself, but rather of the interpretation that one might be led to give to it. In fact, the width of the green band should put you en garde against wild speculations: It would be extremely suspicious if the black line did not venture outside of the green band somewhere, even in case the Higgs boson does not exist!
That is because the band shows the expected range of 1-sigma fluctuations -due to statistical effects, and not to systematic ones such as the real presence of a signal!- and since the black curve is extracted from the data by combining many datasets and each individual point of the line (in, say, 5-GeV intervals) has little correlation with the others, it is entirely appropriate for the curve to not be fully contained in the green area! So, the fact that the black curve overlaps with the signal plus background hypothesis at 130 GeV really -really!- means very, very little.
What does mean something is that the hatched black and red curves appear separated by about one-sigma (the width of the green band surrounding the background-only black hatched curve) over a wide range of Higgs masses. This says that the two Tevatron experiments have by now reached a sensitivity of about 1-sigma to the signal with the data they have analyzed so far. Beware: they are already sitting on about twice as much data (most analyses rely on about 2.5/fb of collisions, but the Tevatron has already delivered to the experiments over 5/fb). So they expect new results, significantly improved, by this summer.
It does seem that at last, the game of Higgs hunting is starting to get exciting again, after a hiatus of about 7 years following the tentative signal seen by the LEP II experiments!
CDF discovers a new hadron! March 13, 2009Posted by dorigo in news, physics, science.
Tags: CDF, discoveries, QCD, standard model, Tevatron
This morning CDF released the results of a search for narrow resonances produced in B meson decays, and in turn decaying into a pair of vector mesons: namely, . This Y state is a new particle whose exact composition is as of yet unknown, except that CDF has measured its mass (4144 MeV) and established that its decay appears to be mediated by strong interactions, given that the natural width of the state is in the range of a few MeV. I describe succintly the analysis below, but first let me make a few points on the relevance of area of investigation.
Heavy meson spectroscopy appears to be a really entertaining research field these days. While all eyes are pointed at the searches for the Higgs boson and supersymmetric particles, if not at even more exotic high-mass objects, and while careers are made and unmade on those uneventful searches, it is elsewhere that action develops. Just think about it: the baryon, the , those mysterious X and Y states which are still unknown in their quark composition. Such discoveries tell the tale of a very prolific research field: one where there is really a lot to understand.
Low-energy QCD is still poorly known and not easily calculable. In frontier High-Energy Physics we bypassed the problem for the sake of studying high-energy phenomena by tuning our simulations such that their output well resembles the result of low-energy QCD processes in all cases where we need them -such as the details of parton fragmentation, or jet production, or transverse momentum effects in the production of massive bodies. However, we have not learnt much with our parametrizations: those describe well what we already know, but they do not even come close to guessing whatever we do not know. Our understanding of low-energy QCD is starting to be a limiting factor in cosmological studies, such as in baryogenesis predictions. So by all means, let us pursue low-energy QCD in all the dirty corners of our produced datasets at hadron colliders!
CDF is actively pursuing this task. The outstanding spectroscopic capabilities of the detector, combined with the huge size of the dataset collected since 2002, allow searches for decays in the one-in-a-million range of branching ratios. The new discovery I am discussing today has indeed been made possible by pushing to the limit our search range.
The full decay chain which has been observed is the following: . That mesons decay to muon pairs is not a surprise, as is the decay to two charged kaons of the vector meson. Also the original decay of the B hadron into the final state is not new: it had been in fact observed previously. What had not been realized yet, because of the insufficient statistics and mass resolution, is that the and mesons produced in that reaction often “resonate” at a very definite mass value, indicating that in those instances the decay actually takes place in two steps as the chain of two two-body decays: and .
The new analysis by CDF is a pleasure to examine, because the already excellent momentum resolution of the charged particle tracking system gets boosted when constraints are placed on the combined mass of multi-body systems. Take the B meson, reconstructed with two muons and three charged tracks, each assumed to be a kaon: if you did not know that the muon pair comes from a nor that two of the kaons come from a , the mass resolution of the system would be in the few tens of MeV range. Instead, by forcing the momenta of the two muons to be consistent with the World average mass of the , , and by imposing that the two kaons make exactly the extremely well-known mass (), much of the uncertainty on the daughter particle momenta disappears, and the B meson becomes an extremely narrow signal: its mass resolution is just 5.9 MeV, a per-mille measurement event-by-event!
The selection of signal events requires several cleanup cuts, including mass window cuts around the J/Psi and phi masses, a decay length of the reconstructed B+ meson longer than 500 microns, and a cut on the log-likelihood ratio fed with dE/dx and time-of-flight information capable of discriminating kaon tracks from other hadrons. After those cuts, the B+ signal really stands above the flat background. There is a total of 78+-10 events in the sample after these cuts, and this is the largest sample of such decays ever isolated. It is shown above (left), together with the corresponding distribution in the candidate mass (right).
A Dalitz plot of the reconstructed decay candidates is shown in the figure on the right. A Dalitz plot is a scatterplot of the squared invariant mass of a subset of the particles emitted in the decay, versus the squared invariant mass of another subset. If the decay proceeds via the creation of an intermediate state, one may observe a horizontal or vertical cluster of events. Judge by yourself: do the points appear to spread evenly in the allowed phase space of the B+ decays ?
The answer is no: a significant structure is seen corresponding to a definite mass of the system. A histogram of the difference between the reconstructed mass of the system and the mass is shown in the plot below: a near-threshold structure appears at just above 1 GeV energy. An unbinned fit to a relativistic Breit-Wigner signal shape on top of the expected background shape shows a signal at a mass difference of , with a width of 11.7+-5.7 MeV.
The significance of the signal is, after taking account of trial factors, equal to 3.8 standard deviations. For the non-zero width hypothesis, the significance is of 3.4 standard deviations, implying that the newfound structure has strong decay. The mass of the new state is thus of 4143+-2.9 MeV.
The new state is above the threshold for decay to pair of charmed hadrons. The decay of the state appears to occur to a pair of vector mesons, , in close similarity to a previous state found at 3930 MeV, the Y(3930), which also decays to two vector mesons in . Therefore, the new state can be also called a Y(4140).
Although the significance of this new signal has not reached the coveted threshold of 5 standard deviations, there are few doubts about its nature. Being a die-hard sceptic, I did doubt about the reality of the signal shown above for a while when I first saw it, but I must admit that the analysis was really done with a lot of care. Besides, CDF now has tens of thousands of fully reconstructed B meson decays available, with which it is possible to study and understand even the most insignificant nuances to every effect, including reconstruction problems, fit method, track characteristics, kinematical biases, you name it. So I am bound to congratulate with the authors of this nice new analysis, which shows once more how the CDF experiment is producing star new results not just in the high-energy frontier, but as well as in low-energy spectroscopy. Well done, CDF!
Live video streaming of single top observation NOW March 10, 2009Posted by dorigo in news, physics, science.
Tags: CDF, Tevatron, top quark
You can follow it at this link.
Who discovered single top production ? March 5, 2009Posted by dorigo in news, physics, science.
Tags: CDF, DZERO, Tevatron, top quark
Both CDF and DZERO have announced yesterday the first observation of electroweak production of single top quarks in proton-antiproton collisions. Both papers (this one from CDF, and this one from DZERO) claim theirs is the first observation of the long sought-after subatomic reaction. Who is right ? Who has more merit in this advancement in human knowledge of fundamental interactions ? Whose analysis is more credible ? Which of the two results has fewer blemishes ?
To me, it is always a matter of which one is the most relevant question. And to me, the most relevant question is, Who cares who did it ? ... with the easy-to-guess answer: not me. As I have had other occasions to say, I am for the advancement of Science, much less for the advancement of scientific careers, leave alone to which experiments those careers belong.
The top quark is interesting, but so far the Tevatron experiments had only studied it when produced in pairs with its antiparticle, through strong interactions. Electroweak production of the top quark is also possible in proton-antiproton collisions, at half the rate. It is one of those rare instances when the electroweak force competes with the strong one, and it is due to the large mass of the top quark: producing two is much more demanding than producing only one, due to the limited energy budget of the collisions. The reactions capable of producing a single top quark are described by the diagrams shown above. In a), a b-quark from one of the projectiles becomes a top by intervention of a weak vector boson; in b), a gluon “fuses” with a W boson and a top quark is created; in c), a W boson is produced off-mass-shell, and it possesses enough energy to decay into a top-bottom pair.
Since 1995, when CDF and DZERO published jointly the observation of the top quark, nobody has ever doubted that electroweak processes would produce single tops as well. Not even one article, to my knowledge, tried to speculate that the top might be so special to have no weak couplings. The very few early attempts at casting doubt on the real nature of what the Tevatron experiments were producing died quickly as statistics improved and the characterization of the newfound quark was furthered. So what is the fuss about finding out that the reaction resulting from the Feynman diagrams shown above can indeed be directly observed ?
There are different facets in a thorough answer to the above question. First of all, competition between CDF and DZERO: each collaboration badly wanted to get there first, especially since this was correctly predicted from the outset to be a tough nut to crack. Second, because seeing single top production implies having direct access to one element of the Cabibbo-Kobayashi-Maskawa mixing matrix, the element , which is after all a fundamental parameter in the standard model (well, to be precise it is a function of some of the latter, namely of the CKM matrix parameters, but let’s not split hairs here). Third, you cannot really see a low-mass Higgs at the Tevatron if you did not measure single top production first, because single top is a background in Higgs boson searches, and one cannot really discover something by assuming something else is there, if one has not proven that beforehand.
So, single top observation is important after all. I am a member of the CDF collaboration, and I am really proud I belong to it, so my judgement on the whole issue might be biased. But if I have to answer the question that gave the title to this post, I will first give you a very short summary of the results of the two analyses, deferring to a better day a more detailed discussion. This will allow me to drive home a few points.
The two analyses: a face-to-face summary
- Significance: both experiments claim that the signal they observe has a statistical significance of 5.0 standard deviations.
- CDF uses 3.2 inverse femtobarns, and finds a 5.0-sigma-significance signal of single top production. The sensitivity of the analysis is better measured by the expected significance, which is quoted at 5.9-sigma.
- DZERO uses 2.4 inverse femtobarns, and finds a 5.0-sigma-significance of single top production. The sensitivity of the DZERO analysis is quoted at 4.5-sigma.
- Cross-section: both experiments measure a cross section in agreement with standard model expectations.
- CDF measures , a relative uncertainty of about 24%.
- DZERO measures , a relative uncertainty of about 23%.
- Measurements of the CKM matrix element: both experiments quote a direct determination of that quantity, which is very close to 1.0 in the SM, but cannot exceed unity.
- CDF finds , a 12% accuracy.
- DZERO finds , a 11% accuracy.
- Data distributions: both experiments have a super-discriminant which combines the information from different searches. This is a graphical display of the power of the analysis, and should be examined with care.
1. CDF in its paper shows the distribution below, as well as the five inputs that were used to obtain it. The distribution shows the single-top contribution in red, stacked over the concurring backgrounds. At high values of the discriminant, the single top signal does stick out, and the black points -the data- follow the sum of all processes nicely.
2.DZERO in its paper has only the distribution shown below. I was underwhelmed when I saw it. Again, backgrounds are stacked one on top of the other, the top distribution is the one from single top (this time shown in blue), and the data is shown by black dots. It does not look like the data prefer the hypothesis of backgrounds+single top over the background-only one all that much!
Maybe I am too partisan to really make a credible point here, and since I did not follow in detail the development of these analyses -from their first publications as evidence for single top, to updates, until yesterday’s papers- I may very well be proven wrong; however, by looking at the two plots above, and by knowing that they both appear to provide a 5.0-sigma significance, I am drawn to the conclusion that DZERO believes their background shapes and normalization much better than CDF does!
Now, believing something is a good thing in almost all human activities except Science. And if two scientific collaborations have a very different way of looking at how well their backgrounds are modeled by Monte Carlo simulations (which, at least as far as the generation of subatomic processes is concerned, are -or can be- the same), which one is to praise more: the one which believes the simulations more to extract their signal, or the one which relies less on them?
The above question is rethorical, and you should have already agreed that you value more a result which is less based on simulations. So let us look into this issue a bit further. CDF bases its result on a total sample of 4780 events, where the total uncertainty is estimated at +-533 events. DZERO bases its own on a sample of 4651 events, with a total uncertainty estimated at +-234 events! What drives such a large difference in the precision of these predictions ?
The culprit is one of the backgrounds, the production of W bosons in association with heavy flavor quarks – an annoying process, which enters all selection of top quarks and Higgs bosons at the Tevatron. CDF has it at 1855 events, with an uncertainty of 486 -or 26.2%; it is shown in green in the CDF plot above. DZERO has it at 2646 events, with an uncertainty of 173, or 6.5%; it is also shown in green in the DZERO plot. Do not be distracted by the different size of the contribution of W+heavy flavor in the two datasets: different selection strategies drive the numbers to differ, and besides, it is rather the total number of events of the two analyses which is similar by pure chance. The point here is the uncertainty.
Luckily, the DZERO analysis does not appear to rely too much on the background normalization -this is not a simple counting experiment, where the better you know the size of expected backgrounds, the smaller your uncertainty on the signal; rather, the shapes of backgrounds are important, and the graphs above show that the data appears indeed well-described by the discriminant shape. And of course, background shapes are checked in control samples, so both experiments have many tools to ensure that the different contributions are well understood. However, the issue remains: how much do the different estimates of the W plus heavy flavor uncertainty impacts the significance of the measurements ? The DZERO paper mentions that one of their largest uncertainties arises from the modeling of the heavy flavor composition of W+jet events, but it does not provide further details.
I would be happy to receive an informed answer in the comments thread about the points I mention above…