Journal Club – This new Chi b (3P) thingy

Today's Journal Club is about a new addition to the Standard Model of fundamental particles.

[/caption]

According to Wikipedia, a Journal Club is a group of individuals who meet regularly to critically evaluate recent articles in the scientific literature. Since this is Universe Today if we occasionally stray into critically evaluating each other’s critical evaluations, that’s OK too.

And of course, the first rule of Journal Club is… don’t talk about Journal Club. So, without further ado – today’s journal article is about a new addition to the Standard Model of fundamental particles.

The good folk at the CERN Large Hadron Collider finished off 2011 with some vague murmurings about the Higgs Boson – which might have been kind-of sort-of discovered in the data already, but due to the degree of statistical noise around it, no-one’s willing to call it really found yet.

Since there is probably a Nobel prize in it – this seems like a good decision. It is likely that a one-way-or-the-other conclusion will be possible around this time next year – either because collisions to be run over 2012 reveal some critical new data, or because someone sifting through the mountain of data already produced will finally nail it.

But in the meantime, they did find something in 2011. There is a confirmed Observation of a new chi_b state in radiative transitions to Upsilon(1S) and Upsilon(2S) at the ATLAS experiment – or, in a nutshell… we hit Bottomonium.

In the lexicon of sub-atomic particle physics, the term Quarkonium is used to describe a particle whose constituents comprise a quark and its own anti-quark. So for example you can have Charmonium (a charm quark and a charm anti-quark) and you can have Bottomonium (a bottom quark and a bottom anti-quark).

The new Chi b (3P) particle has been reported as a boson – which is technically correct, since it has integer spin, while fermions (hadrons and leptons) have half spins. But it’s not an elementary boson like photons, gluons or the (theoretical) Higgs – it’s a composite boson composed of quarks. So, it is perhaps less confusing to consider it a meson (which is a bosonic hadron). Like other mesons, Chi b (3P) is a hadron that would not be commonly found in nature. It just appears briefly in particle accelerator collisions before it decays.

So comments? Has the significance of this new finding been muted because the discoverers thought it would just prompt a lot of bottom jokes? Is Chi_b (3P) the ‘Claytons Higgs’ (the boson you have when you’re not having a Higgs?). Want to suggest an article for the next edition of Journal Club?

Otherwise, have a great 2012.

Today’s article:
The ATLAS collaboration Observation of a new chi_b state in radiative transitions to Upsilon(1S) and Upsilon(2S) at the ATLAS experiment.

Slower than Light Neutrinos

The first annotated neutrino event. Image credit:

[/caption]

Earlier this year, an international team of scientists announced they had found neutrinos — tiny particles with an equally tiny but non-zero mass — traveling faster than the speed of light. Unable to find a flaw themselves, the team put out a call for physicists worldwide to check their experiment. One physicist who answered the call was Dr. Ramanath Cowsik. He found a potentially fatal flaw in the experiment that challenged the existence of faster than light neutrinos. 

Superluminal (faster than light) neutrinos were the result of the OPERA experiment, a collaboration between the CERN physics laboratory in Geneva, Switzerland, and the Laboratori Nazionali del Gran Sasso in Gran Sasso, Italy.

Scientists at CERN managed to repeat their result of neutrinos travelling faster than the speed of light. Image credit: Cern/Science Photo Library

The experiment timed neutrinos as they traveled 730 kilometres (about 450 miles) through Earth from their origin point at CERN to a detector in Gran Sasso. The team was shocked to find that the neutrinos arrived at Gran Sasso 60 nanoseconds sooner than they would have if they were traveling at the speed of light in a vacuum. In short, they appeared to be superluminal.

This result created either a problem for physics or a breakthrough. According to Einstein’s theory of special relativity, any particle with mass can come close to the speed of light but can’t reach it. Since neutrinos have mass, superluminal neutrinos shouldn’t exist. But, somehow, they did.

But Cowsik questioned the neutrinos’ genesis. The OPERA experiments generated neutrinos by slamming protons into a stationary target. This produced a pulse of pions, unstable particles that were magnetically focused into a tunnel where they decayed into neutrinos and muons (another tiny elementary particle). The muons never went further than the tunnel, but the neutrinos, which can slip through matter like a ghost passes through a wall, kept going towards Gran Sasso.

The creation of a neutrino and a muon. Image credit: J. Sonier

Cowsik’s and his team looked closely at this first step of the OPERA experiment. They investigated whether “pion decays would produce superluminal neutrinos, assuming energy and momentum are conserved,” he said. The OPERA neutrinos had a lot of energy but very little mass, so the question was whether they could really move faster than light.

What Cowsik and his team found was that if neutrinos produced from a pion decay were traveling faster than light, the pion lifetime would get longer and each neutrino would carry a smaller fraction of the energy it shares with the muon. Within the present framework of physics, superluminal neutrinos would be very difficult to produce. “What’s more,”Cowsik explains, “these difficulties would only increase as the pion energy increases.

There is an experimental check of Cowsik’s theoretical conclusion. CERN’s method of producing neutrinos is duplicated naturally when cosmic rays hit Earth’s atmosphere. An observatory called IceCube is set up to observe these naturally occurring neutrinos in Antarctica; as neutrinos collide with other particles, they generate muons that leave trails of light flashes as they pass through a nearly 2.5 kilometre (1.5 mile) thick block of clear ice.

A schematic image of IceCube. ICE.WISC.EDU / PETE GUEST

IceCube has detected neutrinos with energy 10,000 times higher than any generated as part of the OPERA experiment, leading Cowsik to conclude that their parent pions must have correspondingly high energy levels. His team’s calculations based on laws of the conservation of energy and momentum revealed that the lifetimes of those pions should be too long for them to decay into superluminal neutrinos.

As Cowsik explains, IceCube’s detection of high-energy neutrinos is indicative that pions do decay according to standard ideas of physics, but the neutrinos will only approach the speed of light; they will never exceed it.

Source: Pions Don’t Want to Decay into Faster the Light Neutrinos

Faster than Light Neutrinos (maybe): Field Trip!

Underwater Neutrino Detector Will Be Second-Largest Structure Ever Built

Artist's rendering of the KM3NeT array. (Marco Kraan/Property KM3NeT Consortium)

[/caption]

The hunt for elusive neutrinos will soon get its largest and most powerful tool yet: the enormous KM3NeT telescope, currently under development by a consortium of 40 institutions from ten European countries. Once completed KM3NeT will be the second-largest structure ever made by humans, after the Great Wall of China, and taller than the Burj Khalifa in Dubai… but submerged beneath 3,200 feet of ocean!

KM3NeT – so named because it will encompass an area of several cubic kilometers – will be composed of lengths of cable holding optical modules on the ends of long arms. These modules will stare at the sea floor beneath the Mediterranean in an attempt to detect the impacts of neutrinos traveling down from deep space.

Successfully spotting neutrinos – subatomic particles that don’t interact with “normal” matter very much at all, nor have magnetic charges – will help researchers to determine which direction they originated from. That in turn will help them pinpoint distant sources of powerful radiation, like quasars and gamma-ray bursts. Only neutrinos could make it this far and this long after such events since they can pass basically unimpeded across vast cosmic distances.

“The only high energy particles that can come from very distant sources are neutrinos,” said Giorgio Riccobene, a physicist and staff researcher at the National Institute for Nuclear Physics. “So by looking at them, we can probe the far and violent universe.”

Each Digital Optical Module (DOM) is a standalone sensor module with 31 3-inch PMTs in a 17-inch glass sphere.

In effect, by looking down beneath the sea KM3NeT will allow scientists to peer outward into the Universe, deep into space as well as far back in time.

The optical modules dispersed along the KM3NeT array will be able to identify the light given off by muons when neutrinos pass into the sea floor. The entire structure would have thousands of the modules (which resemble large versions of the hovering training spheres used by Luke Skywalker in Star Wars.)

In addition to searching for neutrinos passing through Earth, KM3NeT will also look toward the galactic center and search for the presence of neutrinos there, which would help confirm the purported existence of dark matter.

Read more about the KM3NeT project here, and check out a detailed article on the telescope and neutrinos on Popsci.com.

Height of the KM3NeT telescope structure compared to well-known buildings

Images property of KM3NeT Consortium 

Astronomy Without A Telescope – Special Relativity From First Principles

CAPTION

[/caption]

Einstein’s explanation of special relativity, delivered in his 1905 paper On the Electrodynamics of Moving Bodies focuses on demolishing the idea of ‘absolute rest’, exemplified by the theoretical luminiferous aether. He achieved this very successfully, but many hearing that argument today are left puzzled as to why everything seems to depend upon the speed of light in a vacuum.

Since few people in the 21st century need convincing that the luminiferous aether does not exist, it is possible to come at the concept of special relativity in a different way and just through an exercise of logic deduce that the universe must have an absolute speed – and from there deduce special relativity as a logical consequence.

The argument goes like this:

1) There must be an absolute speed in any universe since speed is a measure of distance moved over time. Increasing your speed means you reduce your travel time between a distance A to B. A kilometre walk to the shops might take 25 minutes, but if you run it might take only 15 minutes – and if you take the car, only 2 minutes. At least theoretically you should be able to increase your speed up to the point where that travel time reaches zero – and whatever speed you are at when that happens will represent the universe’s absolute speed.

2) Now consider the principle of relativity. Einstein talked about trains and platforms to describe different inertial frame of references. So for example, you can measure someone throwing a ball forward at 10 km/hr on the platform. But put that someone on the train which is travelling at 60 km/hr and then the ball measurably moves forward at nearly 70 km/hr (relative to the platform).

3) Point 2 is a big problem for a universe that has an absolute speed (see Point 1). For example, if you had an instrument that projected something forward at the absolute speed of the universe and then put that instrument on the train – you would expect to be able to measure something moving at the absolute speed + 60 km/hr.

4) Einstein deduced that when you observe something moving in a different frame of reference to your own, the components of speed (i.e. distance and time), must change in that other frame of reference to ensure that anything that moves can never be measured moving at a speed greater than the absolute speed.

Thus on the train, distances should contract and time should dilate (since time is the denominator of distance over time).

The effect of relative motion. Measurable time dilation is negligible on a train moving past a platform at 60 km/hr, but increases dramatically if that train acquires the capacity to approach the speed of light. Time (and distance) will change to ensure that light speed is always light speed, not light speed + the speed of the train.

And that’s it really. From there one can just look to the universe for examples of something that always moves at the same speed regardless of frame of reference. When you find that something, you will know that it must be moving at the absolute speed.

Einstein offers two examples in the opening paragraphs of On the Electrodynamics of Moving Bodies:

  • the electromagnetic output produced by the relative motion of a magnet and an induction coil is the same whether the magnet is moved or whether the coil is moved (a finding of James Clerk Maxwell‘s electromagnetic theory) and;
  • the failure to demonstrate that the motion of the Earth adds any additional speed to a light beam moving ahead of the Earth’s orbital trajectory (presumably an oblique reference to the 1887 Michelson-Morley experiment).

In other words, electromagnetic radiation (i.e. light) demonstrated the very property that would be expected of something which moved at the absolute speed that it is possible to move in our universe.

The fact that light happens to move at the absolute speed of the universe is useful to know – since we can measure the speed of light and hence we can then assign a numerical value to the universe’s absolute speed (i.e. 300,000 km/sec), rather than just calling it c.

Further reading:
None! That was AWAT #100 – more than enough for anyone. Thanks for reading, even if it was just today. SN.

Looking at Early Black Holes with a ‘Time Machine’

The large scale cosmological mass distribution in the simulation volume of the MassiveBlack. The projected gas density over the whole volume ('unwrapped' into 2D) is shown in the large scale (background) image. The two images on top show two zoom-in of increasing factor of 10, of the regions where the most massive black hole - the first quasars - is formed. The black hole is at the center of the image and is being fed by cold gas streams. Image Courtesy of Yu Feng.

[/caption]

What fed early black holes enabling their very rapid growth? A new discovery made by researchers at Carnegie Mellon University using a combination of supercomputer simulations and GigaPan Time Machine technology shows that a diet of cosmic “fast food” (thin streams of cold gas) flowed uncontrollably into the center of the first black holes, causing them to be “supersized” and grow faster than anything else in the Universe.

When our Universe was young, less than a billion years after the Big Bang, galaxies were just beginning to form and grow. According to prior theories, black holes at that time should have been equally small. Data from the Sloan Digital Sky Survey has shown evidence to the contrary – supermassive black holes were in existence as early as 700 million years after the Big Bang.

“The Sloan Digital Sky Survey found supermassive black holes at less than 1 billion years. They were the same size as today’s most massive black holes, which are 13.6 billion years old,” said Tiziana Di Matteo, associate professor of physics (Carnegie Mellon University). “It was a puzzle. Why do some black holes form so early when it takes the whole age of the Universe for others to reach the same mass?”

Supermassive black holes are the largest black holes in existence – weighing in with masses billions of times that of the Sun. Most “normal” black holes are only about 30 times more massive than the Sun. The currently accepted mechanism for the formation of supermassive black holes is through galactic mergers. One problem with this theory and how it applies to early supermassive black holes is that in early Universe, there weren’t many galaxies, and they were too distant from each other to merge.

Rupert Croft, associate professor of physics (Carnegie Mellon University) remarked, “If you write the equations for how galaxies and black holes form, it doesn’t seem possible that these huge masses could form that early, But we look to the sky and there they are.”

In an effort to understand the processes that formed the early supermassive black holes, Di Matteo, Croft and Khandai created MassiveBlack – the largest cosmological simulation to date. The purpose of MassiveBlack is to accurately simulate the first billion years of our universe. Describing MassiveBlack, Di Matteo remarked, “This simulation is truly gigantic. It’s the largest in terms of the level of physics and the actual volume. We did that because we were interested in looking at rare things in the universe, like the first black holes. Because they are so rare, you need to search over a large volume of space”.

Croft and the team started the simulations using known models of cosmology based on theories and laws of modern day physics. “We didn’t put anything crazy in. There’s no magic physics, no extra stuff. It’s the same physics that forms galaxies in simulations of the later universe,” said Croft. “But magically, these early quasars, just as had been observed, appear. We didn’t know they were going to show up. It was amazing to measure their masses and go ‘Wow! These are the exact right size and show up exactly at the right point in time.’ It’s a success story for the modern theory of cosmology.”

The data from MassiveBlack was added to the GigaPan Time Machine project. By combining the MassiveBlack data with the GigaPan Time Machine project, researchers were able to view the simulation as if it was a movie – easily panning across the simulated universe as it formed. When the team noticed events which appeared interesting, they were also able to zoom in to view the events in greater detail than what they could see in our own universe with ground or space-based telescopes.

When the team zoomed in on the creation of the first supermassive black holes, they saw something unexpected. Normal observations show that when cold gas flows toward a black hole it is heated from collisions with other nearby gas molecules, then cools down before entering the black hole. Known as ‘shock heating’, the process should have stopped early black holes from reaching the masses observed. Instead, the team observed thin streams of cold dense gas flowing along ‘filaments’ seen in large-scale surveys that reveal the structure of our universe. The filaments allowed the gas to flow directly into the center of the black holes at incredible speed, providing them with cold, fast food. The steady, but uncontrolled consumption provided a mechanism for the black holes to grow at a much faster rate than their host galaxies.

The findings will be published in the Astrophysical Journal Letters.

If you’d like to read more, check out the papers below ( via Physics arXiv ):
Terapixel Imaging of Cosmological Simulations
The Formation of Galaxies Hosting z~6 Quasars
Early Black Holes in Cosmological Simulations
Cold Flows and the First Quasars

Learn more about Gigapan and MassiveBlack at: http://gigapan.org/gigapans/76215/ and http://www.psc.edu/science/2011/supermassive/

Source: Carnegie Mellon University Press Release

Particle Physicists Put the Squeeze on the Higgs Boson; Look for Conclusive Results in 2012

Scientists gather as the ATLAS and CMS experiments present the status of their searches for the Standard Model Higgs boson. Credit: CERN

[/caption]

With “freshly squeezed” plots from the latest data garnered by two particle physics experiments, teams of scientists from the Large Hadron Collider at CERN, the European Center for Nuclear Research, said Tuesday they had recorded “tantalizing hints” of the elusive subatomic particle known as the Higgs Boson, but cannot conclusively say it exists … yet. However, they predict that 2012 collider runs should bring enough data to make the determination.

“The very fact that we are able to show the results of very sophisticated analysis just one month after the last bit of data we used has been recorded is very reassuring,” Dr. Greg Landsberg, physics coordinator for the Compact Muon Solenoid (CMS) detector at the LHC told Universe Today. “It tells you how quick the turnaround time is. This is truly unprecedented in the history of particle physics, with such large and complex experiments producing so much data, and it’s very exciting.”

For now, the main conclusion of over 6,000 scientists on the combined teams from CMS and the ATLAS particle detectors is that they were able to constrain the mass range of the Standard Model Higgs boson — if it exists — to be in the range of 116-130 GeV by the ATLAS experiment, and 115-127 GeV by CMS.

The Standard Model is the theory that explains the interactions of subatomic particles – which describes ordinary matter that the Universe is made of — and on the whole works very well. But it doesn’t explain why some particles have mass and others don’t, and it also doesn’t describe the 96% of the Universe that is invisible.

In 1964, physicist Peter Higgs and colleagues proposed the existence of a mysterious energy field that interacts with some subatomic particles more than others, resulting in varying values for particle mass. That field is known as the Higgs field, and the Higgs Boson is the smallest particle of the Higgs field. But the Higgs Boson hasn’t been discovered yet, and one of the main reasons the LHC was built was to try to find it.

To look for these tiny particles, the LHC smashes high-energy protons together, converting some energy to mass. This produces a spray of particles which are picked up by the detectors. However, the discovery of the Higgs relies on observing the particles these protons decay into rather than the Higgs itself. If they do exist, they are very short lived and can decay in many different ways. The problem is that many other processes can also produce the same results.

How can scientists tell the difference? A short answer is that if they can figure out all the other things that can produce a Higgs-like signal and the typical frequency at which they will occur, then if they see more of these signals than current theories suggest, that gives them a place to look for the Higgs.

The experiments have seen excesses in similar ranges. And as the CERN press release noted, “Taken individually, none of these excesses is any more statistically significant than rolling a die and coming up with two sixes in a row. What is interesting is that there are multiple independent measurements pointing to the region of 124 to 126 GeV.”

“This is very promising,” said Landsberg, who is also a professor at Brown University. “This shows that both experiments understand what is going on with their detectors very, very well. Both calibrations saw excesses at low masses. But unfortunately the nature of our process is statistical and statistics is known to play funny tricks once in a while. So we don’t really know — we don’t have enough evidence to know — if what we saw is a glimpse of the Higgs Boson or these are just statistical fluctuations of the Standand Model process which mimic the same type of signatures as would come if the Higgs Boson is produced.”

Landsberg said the only way to cope with statistics is to get more data, and the scientists need to increase the size of the data samples considerably in order to definitely answer the question on whether the Higgs Boson exists at the mass of 125 GeV or any mass range which hasn’t been excluded yet.

The good news is that loads of data are coming in 2012.

“We hope to quadruple the data sample collected this year,” Landsberg said. “And that should give us enough statistical confidence to essentially solve this puzzle and tell the world whether we saw the first glimpses of the Higgs Boson. As the team showed today, we will keep increasing until we reach a level of statistical significance which is considered to be sufficient for discovery in our field.”

Landsberg said that within this small range, there is not much room for the Higgs to hide. “This is very exciting, and it tells you that we are almost there. We have enough sensitivity and beautiful detectors; we need just a little bit more time and a little more data. I am very hopeful we should be able to say something definitive by sometime next year.”

So the suspense is building and 2012 could be the year of the Higgs.

More info: CERN press release, ArsTechnica

Mapping The Milky Way’s Magnetic Fields – The Faraday Sky

Fig. 3: In this map of the sky, a correction for the effect of the galactic disk has been made in order to emphasize weaker magnetic field structures. The magnetic field directions above and below the disk seem to be diametrically opposed, as indicated by the positive (red) and negative (blue) values. An analogous change of direction takes place accross the vertical center line, which runs through the center of the Milky Way.

[/caption]

Kudos to the scientists at the Max Planck Institut and an international team of radio astronomers for an incredibly detailed new map of our galaxy’s magnetic fields! This unique all-sky map has surpassed its predecessors and is giving us insight into the magnetic field structure of the Milky Way beyond anything so far seen. What’s so special about this one? It’s showing us a quality known as Faraday depth – a concept which works along a specific line of sight. To construct the map, data was melded from 41,000 measurements collected from a new image reconstruction technique. We can now see not only the major structure of galactic fields, but less obvious features like turbulence in galactic gas.

So, exactly what does a new map of this kind mean? All galaxies possess magnetic fields, but their source is a mystery. As of now, we can only guess they occur due to dynamo processes… where mechanical energy is transformed into magnetic energy. This type of creation is perfectly normal and happens here on Earth, the Sun, and even on a smaller scale like a hand-crank powered radio – or a Faraday flashlight! By showing us where magnetic field structures occur in the Milky Way, we can get a better understanding of galactic dynamos.

Fig. 1: The sky map of the Faraday effect caused by the magnetic fields of the Milky Way. Red and blue colors indicate regions of the sky where the magnetic field points toward and away from the observer, respectively. The band of the Milky Way (the plane of the galactic disk) extends horizontally in this panoramic view. The center of the Milky Way lies in the middle of the image. The North celestial pole is at the top left and the South Pole is at the bottom right.
For the last century and a half, we’ve known about Faraday rotation and scientists use it to measure cosmic magnetic fields. This action happens when polarized light goes through a magnetized medium and the plane of polarization revolves. The amount of turn is dependent on the strength and direction of the magnetic field. By observation of the rotation we can further understand the properties of the intervening magnetic fields. Radio astronomers gather and examine the polarized light from distant radio sources passing through our galaxy on its way to us. The Faraday effect can then be judged by measuring the source polarization at various frequencies. However, these measurements can only tell us about the one path through the Milky Way. To see things as a whole, one needs to know how many sources are scattered over the visible sky. This is where the international group of radio astronomers played an important role. They proved data from 26 different projects which gave a grand total of 41,300 pinpoint sources – at an average of about one radio source per square degree of sky.

Although that sounds like a wealth of information, it’s still not really enough. There are huge areas, particularly in the southern sky, where only a few measurements exist. Because of this lack of data, we have to interpolate between existing data points and that creates its own problems. First, the accuracy varies and more precise measurements should help. Also, astronomers are not exactly sure of how reliable a single measurement can be – they just have to take their best guess based on what information they have. Still, other problems exist. There are measurement uncertainties due to the complex nature of the process. A small error can increase by tenfold and this could convolute the map if not corrected. To help fix these problems, scientists at MPA developed a new algorithm for image capture, named the “extended critical filter”. In its creation, the team utilizes tools provided by the new discipline known as information field theory – a powerful tool that blends logical and statistical methods to applied fields and stacks it up against inaccurate information. This new work is exciting because it can also be applied to other imaging and signal-processing venues in alternate scientific fields.

Fig. 2: The uncertainty in the Faraday map. Note that the range of values is significantly smaller than in the Faraday map (Fig. 1). In the area of the celestial south pole, the measurement uncertainties are particularly high because of the low density of data points.
“In addition to the detailed Faraday depth map (Fig. 1), the algorithm provides a map of the uncertainties (Fig. 2). Especially in the galactic disk and in the less well-observed region around the south celestial pole (bottom right quadrant), the uncertainties are significantly larger.” says the team. “To better emphasize the structures in the galactic magnetic field, in Figure 3 (above) the effect of the galactic disk has been removed so that weaker features above and below the galactic disk are more visible. This reveals not only the conspicuous horizontal band of the gas disk of our Milky Way in the middle of the picture, but also that the magnetic field directions seem to be opposite above and below the disk. An analogous change of direction also takes place between the left and right sides of the image, from one side of the center of the Milky Way to the other.”

The good news is the galactic dynamo theory seems to be spot on. It has predicted symmetrical structures and the new map reflects it. In this projection, the magnetic fields are lined up parallel to the plane of the galactic disc in a spiral. This direction is opposite above and below the disc and the observed symmetries in the Faraday map arise from our location within the galactic disc. Here we see both large and small structures tied in with the turbulent, dynamic Milky Way gas structures. This new map algorithm has a great side-line, too… it characterizes the size distribution of these structures. Larger ones are more definitive than smaller ones, which is normal for turbulent systems. This spectrum can then be stacked against computer models of dynamics – allowing for intricate testing of the galactic dynamo models.

This incredible new map is more than just another pretty face in astronomy. By providing information of extragalactic magnetic fields, we’re enabling radio telescope projects such as LOFAR, eVLA, ASKAP, Meerkat and the SKA to rise to new heights. With this will come even more updates to the Faraday Sky and reveal the mystery of the origin of galactic magnetic fields.

Original Story Source: Max Planck Institut for Astrophysics News Release. For Further Reading: An improved map of the galactic Faraday sky”. Download the map HERE.

Are Pulsars Giant Permanent Magnets?

The Vela Pulsar, a neutron star corpse left from a titanic stellar supernova explosion, shoots through space powered by a jet emitted from one of the neutron star's rotational poles. Now a counter jet in front of the neutron star has been imaged by the Chandra X-ray observatory. The Chandra image above shows the Vela Pulsar as a bright white spot in the middle of the picture, surrounded by hot gas shown in yellow and orange. The counter jet can be seen wiggling from the hot gas in the upper right. Chandra has been studying this jet so long that it's been able to create a movie of the jet's motion. The jet moves through space like a firehose, wiggling to the left and right and up and down, but staying collimated: the "hose" around the stream is, in this case, composed of a tightly bound magnetic field. Image Credit:

[/caption]
Some of the most bizarre phenomena in the universe are neutron stars. Very few things in our universe can rival the density in these remnants of supernova explosions. Neutron stars emit intense radiation from their magnetic poles, and when a neutron star is aligned such that these “beams” of radiation point in Earth’s direction, we can detect the pulses, and refer to said neutron star as a pulsar.

What has been a mystery so far, is how exactly the magnetic fields of pulsars form and behave. Researchers had believed that the magnetic fields form from the rotation of charged particles, and as such should align with the rotational axis of the neutron star. Based on observational data, researchers know this is not the case.

Seeking to unravel this mystery, Johan Hansson and Anna Ponga (Lulea University of Technology, Sweden) have written a paper which outlines a new theory on how the magnetic fields of neutron stars form. Hansson and Ponga theorize that not only can the movement of charged particles form a magnetic field, but also the alignment of the magnetic fields of components that make up the neutron star – similar to the process of forming ferromagnets.

Getting into the physics of Hansson and Ponga’s paper, they suggest that when a neutron star forms, neutron magnetic moments become aligned. The alignment is thought to occur due to it being the lowest energy configuration of the nuclear forces. Basically, once the alignment occurs, the magnetic field of a neutron star is locked in place. This phenomenon essentially makes a neutron star into a giant permanent magnet, something Hansson and Ponga call a “neutromagnet”.

Similar to its smaller permanent magnet cousins, a neutromagnet would be extremely stable. The magnetic field of a neutromagnet is thought to align with the original magnetic field of the “parent” star, which appears to act as a catalyst. What is even more interesting is that the original magnetic field isn’t required to be in the same direction as the spin axis.

One more interesting fact is that with all neutron stars having nearly the same mass, Hansson and Ponga can calculate the strength of the magnetic fields the neutromagnets should generate. Based on their calculations, the strength is about 1012 Tesla’s – almost exactly the observed value detected around the most intense magnetic fields around neutron stars. The team’s calculations appear to solve several unsolved problems regarding pulsars.

Hansson and Ponga’s theory is simple to test – since they state the magnetic field strength of neutron stars cannot exceed 1012 Tesla’s. If a neutron star were to be discovered with a stronger magnetic field than 1012 Tesla’s, the team’s theory would be proven wrong.

Due to the Pauli exclusion principle possibly excluding neutrons aligning in the manner outlined in Hansson and Ponga’s paper, there are some questions regarding the team’s theory. Hansson and Ponga point to experiments that have been performed which suggest that nuclear spins can become ordered, like ferromagnets, stating: “One should remember that the nuclear physics at these extreme circumstances and densities is not known a priori, so several unexpected properties might apply,”

While Hansson and Ponga readily agree their theories are purely speculative, they feel their theory is worth pursuing in more detail.

If you’d like to learn more, you can read the full scientific paper by Hansson & Pong at: http://arxiv.org/pdf/1111.3434v1

Source: Pulsars: Cosmic Permanent ‘Neutromagnets’ (Hansson & Pong)

Astronomy Without A Telescope – Mass Is Energy

The USS Enterprise in 1964 (pre Zephram Cochran era), which sailed around the world in 64 days without refuelling to demonstrate the capability of nuclear-powered surface ships. Credit: US Navy.

[/caption]

Some say that the reason you can’t travel faster than light is that your mass will increase as your speed approaches light speed – so, regardless of how much energy your star drive can generate, you reach a point where no amount of energy can further accelerate your spacecraft because its mass is approaching infinite.

This line of thinking is at best an incomplete description of what’s really going on and is not a particularly effective way of explaining why you can’t move faster than light (even though you really can’t). However, the story does offer useful insight into why mass is equivalent to energy, in accordance with the relationship e=mc2.

Firstly, here’s why the story isn’t complete. Although someone back on Earth might see your spacecraft’s mass increase as you move near light speed – you the pilot aren’t going notice your mass change at all. Within your spacecraft, you would still be able to climb stairs, jump rope – and if you had a set of bathroom scales along for the ride you would still weigh just the same as you did back on Earth (assuming your ship is equipped with the latest in artificial gravity technology that mimics conditions back on Earth’s surface).

The change perceived by an Earth observer is just relativistic mass. If you hit the brakes and returned to a more conventional velocity, all the relativistic mass would go away and an Earth observer would just see you retaining with same proper (or rest) mass that the spacecraft and you had before you left Earth.

The Earth observer would be more correct to consider your situation in terms of momentum energy, which is a product of your mass and your speed. So as you pump more energy in to your star drive system, someone on Earth really sees your momentum increase – but interprets it as a mass increase, since your speed doesn’t seem to increase much at all once it is up around 99% of the speed of light. Then when you slow down again, although you might seem to be losing mass you are really offloading energy – perhaps by converting your kinetic energy of motion into heat (assuming your spacecraft is equipped with the latest in relativistic braking technology).

As the ratio of your velocity to light speed approaches 1, the ratio of your relativistic mass to your rest mass grows asymptotically - i.e. it approaches infinite.

From the perspective of the Earth-based observer, you can formulate that the relativistic mass gain observed when travelling near light speed is the sum of the spacecraft’s rest mass/energy plus the kinetic energy of its motion – all divided by c2. From that you can (stepping around some moderately complex math) derive that e=mc2. This is a useful finding, but it has little to do with why the spacecraft’s speed cannot exceed light speed.

The phenomenon of relativistic mass follows a similar, though inverse, asymptotic relationship to your speed. So as you approach light speed, your relativistic time approaches zero (clocks slow), your relativistic spatial dimensions approach zero (lengths contract) – but your relativistic mass grows towards infinite.

But as we’ve covered already, on the spacecraft you do not experience your spacecraft gaining mass (nor does it seem to shrink, nor its clocks slow down). So you must interpret your increase in momentum energy as a genuine speed increase – at least with respect to a new understanding you have developed about speed.

For you, the pilot, when you approach light speed and keep pumping more energy into your drive system, what you find is that you keep reaching your destination faster – not so much because you are moving faster, but because the time you estimated it would take you to cross the distance from point A to Point B becomes perceivably much less, indeed the distance between point A to Point B also becomes perceivably much less. So you never break light speed because the distance over time parameters of your speed keep changing in a way that ensures that you can’t.

In any case, consideration of relativistic mass is probably the best way to derive the relationship e=mc2 since the relativistic mass is a direct result of the kinetic energy of motion. The relationship does not easily fall out of consideration of (say) a nuclear explosion – since much of the energy of the blast derives from the release of the binding energy which holds a heavy atom together. A nuclear blast is more about energy transformation than about matter converting to energy, although at a system level it still represents genuine mass to energy conversion.

Similarly you might consider that your cup of coffee is more massive when it’s hot – and gets measurably less massive when it cools down. Matter, in terms of protons, neutrons, electrons …and coffee, is largely conserved throughout this process. But, for a while, the heat energy really does add to the mass of the system – although since it’s a mass of m=e/c2, it is a very tiny amount of mass.

Neutrinos Still Breaking Speed Limits

Particle Collider
Today, CERN announced that the LHCb experiment had revealed the existence of two new baryon subatomic particles. Credit: CERN/LHC/GridPP

[/caption]

New test results are in from OPERA and it seems those darn neutrinos, they just can’t keep their speed down… to within the speed of light, that is!

report released in September by scientists working on the OPERA project (Oscillation Project with Emulsion-tracking Apparatus) at Italy’s Gran Sasso research lab claimed that neutrinos emitted from CERN 500 miles away in Geneva arrived at their detectors 60 nanoseconds earlier than expected, thus traveling faster than light. This caused no small amount of contention in the scientific community and made news headlines worldwide – and rightfully so, as it basically slaps one of the main tenets of modern physics across the face.

Of course the scientists at OPERA were well aware of this, and didn’t make such a proclamation lightly; over two years of repeated research was undergone to make sure that the numbers were accurate… as well as could be determined, at least. And they were more than open to having their tests replicated and the results reviewed by their peers. In all regards their methods were scientific yet skepticism was widespread… even within OPERA’s own ranks.

One of the concerns that arose regarding the discovery was in regards to the length of the neutrino beam itself, emitted from CERN and received by special detector plates at Gran Sasso. Researchers couldn’t say for sure that any neutrinos detected were closer to the beginning of the beam versus the end, a disparity (on a neutrino-sized scale anyway) of 10.5 microseconds… that’s 10.5 millionths of a second! And so in October, OPERA requested that proton pulses be resent – this time lasting only 3 nanoseconds each.

The OPERA Neutrino Detector

The results were the same. The neutrinos arrived at Gran Sasso 60 nanoseconds earlier than anticipated: faster than light.

The test was repeated – by different teams, no less – and so far 20 such events have been recorded. Each time, the same.

Faster. Than light.

What does this mean? Do we start tearing pages out of physics textbooks? Should we draw up plans for those neutrino-powered warp engines? Does Einstein’s theory of relativity become a quaint memento of what we used to believe?

Hardly. Or, at least, not anytime soon.

OPERA’s latest tests have managed to allay one uncertainty regarding the results, but plenty more remain. One in particular is the use of GPS to align the clocks at the beginning and end of the neutrino beam. Since the same clock alignment system was used in all the experiments, it stands that there may be some as-of-yet unknown factor concerning the GPS – especially since it hasn’t been extensively used in the field of high-energy particle physics.

In addition, some scientists would like to see more results using other parts of the neutrino detector array.

Of course, like any good science, replication of results is a key factor for peer acceptance. And thus Fermilab in Batavia, Illinois will attempt to perform the same experiment with its MINOS (Main Injector Neutrino Oscillation Search) facility, using a precision matching OPERA’s.

MINOS hopes to have its independent results as early as next year.

No tearing up any textbooks just yet…

 

Read more in the Nature.com news article by Eugenie Samuel Reich. The new result was released on the arXiv preprint server on November 17. (The original September 2011 OPERA team paper can be found here.)