NASA Twins Study Researchers Take Genetic Data To Next Level

NASA is beginning to integrate the results of its twin study on astronauts Mark and Scott Kelly. Image: NASA

People who plan and conduct space missions never tire of telling us how hard it is to do things in space.

Our next big goal is getting humans to Mars, and establishing a colony there. There are a multitude of technical and engineering hurdles to be overcome, but we think we can do it.

But the other side of the coin is the physiological hurdles to be overcome. Those may prove to be much more challenging to deal with. NASA’s twins study is poised to add an enormous amount of data to our growing body of knowledge on the effects of space travel on human beings.

NASA's astronauts twins, Scott Kelly (l) and Mark Kelly (r). Image: NASA
NASA’s astronaut twins, Scott Kelly (l) and Mark Kelly (r). Image: NASA

Astronaut twins Scott and Mark Kelly are the basis of NASA’s study. Scott spent a year in space, returning to Earth on March 1st 2016, after spending 340 days aboard the ISS. Mark, himself a retired astronaut, remained on Earth during Scott’s year in space, providing a baseline for studying the effects on the human body of such a prolonged period of time away from Earth.

In February of 2016, NASA released preliminary results of the study. Now, the team studying the results of the twins study has started integrating the data. The way they’re doing this sets it apart from other studies.

“No one has ever looked this deeply at a human subject and profiled them in this detail.” – Tejaswini Mishra, Ph.D., Stanford University School of Medicine.

Typically, individual studies are released to appropriate journals more or less one at a time. But in the twins study, the data will be integrated and summarized before individual papers are published on separate themes. The idea is that taken together, their impact on our understanding of prolonged time in space will be much greater.

“The beauty of this study is when integrating rich data sets of physiological, neurobehavioral and molecular information, one can draw correlations and see patterns,” said Tejaswini Mishra, Ph.D., research fellow at Stanford University School of Medicine, who is creating the integrated database, recording results and looking for correlations. “No one has ever looked this deeply at a human subject and profiled them in this detail. Most researchers combine maybe two to three types of data but this study is one of the few that is collecting many different types of data and an unprecedented amount of information.”

“Each investigation within the study complements the other.” – Brinda Rana, Ph.D., U of C, San Diego School of Medicine

Mike Snyder, Ph.D, is the head of a team of people at Stanford that will work to synthesize the data. There are roughly three steps in the overall process:

  1. Individual researchers in areas like cognition, biochemistry, and immunology will analyze and compile their data then share their results with the Stanford team.
  2. The Stanford team will then further integrate those results into larger data sets.
  3. Those larger data sets will then be reviewed and analyzed to confirm and modify the initial findings.

“There are a lot of firsts with this study and that makes it exciting,” said Brinda Rana, Ph.D., associate professor of psychiatry, University of California San Diego School of Medicine. “A comparative study with one twin in space and one on Earth has never been done before. Each investigation within the study complements the other.”

NASA compares the twins study, and the new integrated method of handling all the results, to conducting a symphony. Each study is like an instrument, and instead of each one playing a solo, they will be added into a greater whole. The team at Stanford is like the conductor. If you’ve ever listened to an orchestra, you know how powerful that can be.

“The human systems in the body are all intertwined which is why we should view the data in a holistic way,” said Scott M. Smith, Ph.D., NASA manager for nutritional biochemistry at the Johnson Space Center. He conducts biochemical profiles on astronauts and his research is targeted to specific metabolites, end products of various biological pathways and processes.

“It is a more comprehensive way to conduct research.” – Chris Mason, Ph.D., associate professor, Department of Physiology and Biophysics Weill Cornell Medicine

Chris Mason Ph.D., at Weill Cornell Medicine said, “Both the universe and the human body are complicated systems and we are studying something hard to see. It’s like having a new flashlight that illuminates the previously dark gears of molecular interactions. It is a more comprehensive way to conduct research.”

Scientists involved with the twins study are very clearly excited about this new approach. Having twin astronauts is an extraordinary opportunity, and will advance our understanding of spaceflight on human physiology enormously.

“There is no doubt, the learnings from integrating our data will be priceless,” said Emmanuel Mignot, M.D., Ph.D., director of Center for Sleep Science and Medicine, Stanford University School of Medicine. He studies the immune system and is enthusiastic to study specific immune cell populations because many of the other immune studies focus only on general factors.

A summary of the early results should be out by early 2018, or possible late 2017. Individual papers on more detailed themes will follow shortly.

Canada To Get Its Own Spaceport

8 Ukrainian-built Cyclone 4 rockets will be launched each year from Maritime Launch Services' planned spaceport in Nova Scotia, Canada. Image: Maritime Launch Services

Canada is getting its own rocket-launching facility. Maritime Launch Services (MLS) has confirmed its plans to build and operate a commercial launch facility in Nova Scotia, on Canada’s east coast. The new spaceport should begin construction in 1 year, and should be in operation by 2022.

The facility will be built near Canso, in the province of Nova Scotia. Maritime Launch Services hopes to launch 8 rockets per year to place satellites in orbit. The Ukrainian Cyclone 4M medium-class rockets that will lift-off from Canso will have a payload of up to 3,350 kg.

The red marker in the map above shows the location of the Maritime Launch Services spaceport. Image: Google

Spaceports have certain requirements that make some locations more desirable. They need to be near transportation infrastructure so that rockets, payloads, and other materials can be transported to the site. They need to be away from major population centres in case of accidents. And they need to provide trajectories that give them access to desirable orbits.

The Nova Scotia site isn’t the only location considered by MLS. They evaluated 14 sites in North America before settling on the Canso, NS site, including ones in Mexico and the US. But it appears that interest and support from local governments helped MLS settle on Canso.

The Ukrainian Cyclone M4 rockets have an excellent track record for safety. The company who builds it, Yuzhnoye, has been in operation for 62 years and has launched 875 vehicles and built and launched over 400 spacecraft. Cyclone rockets have launched successfully 221 times.

The Cyclone 4. The Cyclone family of rockets have over 200 successful launches to their credit. Image: Yuzhnoye Design Office
The Cyclone 4. The Cyclone family of rockets have over 200 successful launches to their credit. Image: Yuzhnoye Design Office

MLS is a group of American aerospace experts including people who have worked with NASA. They are working with the makers of the Cyclone 4 rocket, who have wanted to open up operations in North America for some time.

The Cyclone rocket family first started operating in 1969. The Cyclone 4 is the newest and most powerful rocket in the Cyclone family. It’s a 3-stage rocket that runs on UDMH fuel and uses nitrogen tetroxide for an oxidizer.

There have been other proposals for a Canadian spaceport. The Canadian Space Agency was interested in Cape Breton, also in Nova Scotia, as a launch site for small satellites in 2010. A Canadian-American consortium called PlanetSpace also looked at a Nova Scotia site for a launch facility, but they failed to get the necessary funding from NASA in 2008. Fort Churchill, in the Province of Manitoba, was the site of over 3,500 sub-orbital flights before being shut down in 1985.

The Canso launch facility is an entirely private business proposal. Neither the Canadian government nor the Canadian Space Agency are partners. It’s not clear if having a launch facility on Canadian soil will impact the CSA’s activities in any way.

But at least Canadians won’t have to leave home to watch rocket launches.

Are You Ready For The NanoSWARM?

CubeSats NODes 1 & 2 and STMSat-1 are deployed from the International Space Station during Expedition 47. Image: NASA

We’re accustomed to the ‘large craft’ approach to exploring our Solar System. Probes like the Voyagers, the Mariners, and the Pioneers have written their place in the history of space exploration. Missions like Cassini and Juno are carrying on that work. But advances in technology mean that Nanosats and Cubesats might write the next chapter in the exploration of our Solar System.

Nanosats and Cubesats are different than the probes of the past. They’re much smaller and cheaper, and they offer some flexibility in our approach to exploring the Solar System. A Nanosat is defined as a satellite with a mass between 1 and 10 kg. A CubeSat is made up of multiple cubes of roughly 10cm³ (10cm x 10cm x 11.35cm). Together, they hold the promise of rapidly expanding our understanding of the Solar System in a much more flexible way.

A cubesat structure, made by ClydeSpace, 1U in size. Credit: Wikipedia Commons/Svobodat

NASA has been working on smaller satellites for a few years, and the work is starting to bear some serious fruit. A group of scientists at JPL predicts that by 2020 there will be 10 deep space CubeSats exploring our Solar System, and by 2030 there will be 100 of them. NASA, as usual, is developing NanoSat and CubeSat technologies, but so are private companies like Scotland’s Clyde Space.

Clyde Space from Clyde Space on Vimeo.

INSPIRE and MarCO

NASA has built 2 Interplanetary NanoSpacecraft Pathfinder In Relevant Environment (INSPIRE) CubeSats to be launched in 2017. They will demonstrate what NASA calls the “revolutionary capability of deep space CubeSats.” They’ll be placed in earth-escape orbit to show that they can withstand the rigors of space, and can operate, navigate, and communicate effectively.

Following in INSPIRE’s footsteps will be the Mars Cube One (MarCO) CubeSats. MarCO will demonstrate one of the most attractive aspects of CubeSats and NanoSats: their ability to hitch a ride with larger missions and to augment the capabilities of those missions.

In 2018, NASA plans to send a stationary lander to Mars, called Interior Exploration using Seismic Investigations, Geodesy and Heat Transport (InSight). The MarCO CubeSats will be along for the ride, and will act as communications relays, though they aren’t needed for mission success. They will be the first CubeSats to be sent into deep space.

So what are some specific targets for this new class of small probes? The applications for NanoSats and CubeSats are abundant.

Other NanoSat and CubeSat Missions

NASA’s Europa Clipper Mission, planned for the 2020’s, will likely have CubeSats along for the ride as it scrutinizes Europa for conditions favorable for life. NASA has contracted 10 academic institutes to study CubeSats that would allow the mission to get closer to Europa’s frozen surface.

The ESA’s AIM asteroid probe will launch in 2020 to study a binary asteroid system called the Didymos system. AIM will consist of the main spacecraft, a small lander, and at least two CubeSats. The CubeSats will act as part of a deep space communications network.

ESA’s Asteroid Impact Mission is joined by two triple-unit CubeSats to observe the impact of the NASA-led Demonstration of Autonomous Rendezvous Technology (DART) probe with the secondary Didymos asteroid, planned for late 2022. Image: ESA

The challenging environment of Venus is also another world where CubeSats and NanoSats can play a prominent role. Many missions make use of a gravity assist from Venus as they head to their main objective. The small size of NanoSats means that one or more of them could be released at Venus. The thick atmosphere at Venus gives us a chance to demonstrate aerocapture and to place NanoSats in orbit around our neighbor planet. These NanoSats could take study the Venusian atmosphere and send the results back to Earth.

NanoSWARM

But the proposed NanoSWARM might be the most effective demonstration of the power of NanoSats yet. The NanoSWARM mission would have a fleet of small satellites sent to the Moon with a specific set of objectives. Unlike other missions, where NanoSats and CubeSats would be part of a mission centered around larger payloads, NanoSWARM would be only small satellites.

NanoSWARM is a forward thinking mission that is so far only a concept. It would be a fleet of CubeSats orbiting the Moon and addressing questions around planetary magnetism, surface water on airless bodies, space weathering, and the physics of small-scale magnetospheres. NanoSWARM would target features on the Moon called “swirls“, which are high-albedo features correlated with strong magnetic fields and low surficial water. NanoSWARM CubeSats will make the first near-surface measurements of solar wind flux and magnetic fields at swirls.

This is an image of the Reiner Gamma lunar swirl from NASA’s Lunar Reconnaissance Orbiter.
Credits: NASA LRO WAC science team

NanoSWARM would have a mission architecture referred to as “mother with many children.” The mother ship would release two sets of CubeSats. One set would be released with impact trajectories and would gather data on magnetism and proton fluxes right up until impact. A second set would orbit the Moon to measure neutron fluxes. NanoSWARM’s results would tell us a lot about the geophysics, volatile distribution, and plasma physics of other bodies, including terrestrial planets and asteroids.

Space enthusiasts know that the Voyager probes had less computing power than our mobile phones. It’s common knowledge that our electronics are getting smaller and smaller. We’re also getting better at all the other technologies necessary for CubeSats and NanoSats, like batteries, solar arrays, and electrospray thrusters. As this trend continues, expect nanosatellites and cubesats to play a larger and more prominent role in space exploration.

And get ready for the NanoSTORM.

Earth’s Twisted Sister: How Will We Reveal Venus’ Secrets?

A radar view of Venus taken by the Magellan spacecraft, with some gaps filled in by the Pioneer Venus orbiter. Credit: NASA/JPL

Venus is known as Earth’s Sister Planet. It’s roughly the same size and mass as Earth, it’s our closest planetary neighbor, and Venus and Earth grew up together.

When you grow up with something, and it’s always been there, you kind of take it for granted. As a species, we occasionally glance over at Venus and go “Huh. Look at Venus.” Mars, exotic exoplanets in distant solar systems, and the strange gas giants and their moons in our own Solar System attract much more of our attention.

If a distant civilization searched our Solar System for potentially habitable planets, using the same criteria we do, then Venus would be front page news for them. It’s on the edge of the habitable zone and it has an atmosphere. But we know better. Venus is a hellish world, hot enough to melt lead, with crushing atmospheric pressure and acid rain falling from the sky. Even so, Venus still holds secrets we need to reveal.

Chief among those secrets is, “Why did Venus develop so differently?

Conditions on Venus pose unique challenges. The history of Venus exploration is littered with melted Soviet Venera Landers. Orbital probes like Pioneer 12 and Magellan have had more success recently, but Venus’ dense atmosphere still limits their effectiveness. Advances in materials, and especially in electronic circuitry that can withstand Venus’ heat, have buoyed our hopes of exploring the surface of Venus in greater detail.

At the Planetary Science Vision 2050 Workshop 2017, put on by the Lunar and Planetary Institute (LPI) a team from the Southwest Research Institute (SWRI) examined the future of Venus exploration. The team was led by James Cutts from JPL.

The group acknowledged several over-arching questions we have about Venus:

  • How can we understand the atmospheric formation, evolution, and climate history?
  • How can we determine the evolution of the surface and interior?
  • How can we understand the nature of interior-surface-atmosphere interactions over time, including whether liquid water was ever present?

Since the Vision 2050 Workshop is all about the next 50 years, Cutts and his team looked at the challenges posed by Venus’ unique conditions, and how they could answer questions in the near-term, mid-term, and long-term.

Near Term Exploration (Present to 2019)

Near-Term goals for the exploration of Venus include improved remote-sensing from orbital probes. This will tell us more about the gravity and topography of Venus. Improved radar imaging and infrared imaging will fill in more blanks. The team also promoted the idea of a sustained aerial platform, a deep probe, and a short duration lander. Multiple probes/dropsondes are also part of the plan.

Dropsondes are small devices that are released into the atmosphere to measure winds, temperature, and humidity. They’re used on Earth to understand the weather, and extreme phenomena like hurricanes, and can fulfill the same purpose at Venus.

Dropsondes are released into the atmosphere, and their descent is slowed by a small parachute. As they descend, they gather data on temperature, wind, and humidity. Image By Staff Sgt. Randy Redman of the US Air Force

In the near-term, missions whose final destination is not Venus can also answer questions. Fly-bys by craft such as Bepi-Colombo, Solar Probe Plus, and the Solar Orbiter missions can give us good information on their way to Mercury and the Sun respectively. These missions will launch in 2018.

Bepi-Colombo, a joint mission of the ESA and JAXA, will perform two fly-bys of Venus on its way to Mercury. Image: ESA/JAXA

The ESO’s Venus Express and Japan’s Akatsuki, (Venus Climate Orbiter), have studied Venus’ climate in detail, especially its chemistry and the interactions between the atmosphere and the surface. Venus Express ended in 2015, while Akatsuki is still there.

Mid-Term Exploration (2020-2024)

The mid-term goals are more ambitious. They include a long-term lander to study Venus’ geophysical properties, a short-duration tessera lander, and two balloons.

The tesserae lander would land in a type of terrain found on Venus known as tesserae. We think that at one time, Venus had liquid water on it. The fundamental evidence for this may lie in the tesserae regions, but the terrain is extremely rough. A short duration lander that could land and operate in the tesserae regions would help us answer Venus’ liquid water question.

Thanks to the continued development of heat-hardy electronics, a long-term duration lander (months or more) is becoming more feasible in the mid-term. Ideally, any long-term mobile lander would be able to travel tens to hundreds of kilometers, in order to acquire a regional sample of Venus’ surface. This is the only way to take geochemistry and mineralogy measurements at multiple sites.

On Mars the landers are solar-powered. Venus’ thick atmosphere makes that impossible. But the same dense atmosphere that prohibits solar power might offer another solution: a sail-powered rover. Old-fashioned sail power might hold the key to moving around on the surface of Venus. Because the atmosphere is so dense, only a small sail would be necessary.

A simple sail-powered rover may solve the problem of mobility on the Venusian surface. Image: NASA

Long-Term Exploration (2025 and Beyond)

The long-term goals from Cutts and his team are where things get really interesting. A long-lived surface rover is still on the list, or possibly a near-surface craft like a balloon. Also on there is a long-lived seismic network.

A seismic network would really start to reveal the secrets behind Venus’ geophysical life. Whereas a lander would give us estimates of seismic activity, they would be crude compared to what a network of seismic sensors would reveal about Venus’ inner workings. A more thorough understanding of quake mechanisms and locations would really get the theorists buzzing. But it’s the final thing on the list that would be the end-goal. A sample-return mission.

We’re getting good at in situ measurements on other worlds. But for Venus, and for all the other worlds we have visited or want to visit, a sample return is the holy grail. The Apollo missions brought back hundreds of kilograms of lunar samples. Other sample-return missions have been sent to Phobos, which failed, and to asteroids, with varying degrees of success.

Subjecting a sample to the kind of deep analysis that can only be done on labs here on Earth is the end-game. We can keep analyzing samples as we develop new technologies to examine them with. Science is iterative, after all.

An artist’s image of Hayabusa leaving Earth. Hayabusa was a Japanese sample return mission to the asteroid 25143 Itokawa. The mission was a partial success. A sample mission to Earth’s sister planet is the holy-grail for the exploration of Venus. Image credit: JAXA

The 2003 Planetary Science Decadal Survey identified the importance of a sample return mission to Venus’ atmosphere. A balloon would float aloft in the clouds, and an ascending rocket would launch a collected sample back to Earth. According to Cutts and his team, this kind of sample-return mission could act as a stepping stone to a surface sample mission.

A surface sample would likely be the pinnacle of achievement when it comes to understanding Venus. But like most of the proposed goals for Venus, we’ll have to wait awhile.

The Changing Future

Cutts and the team acknowledge that the technology to enable exploration of Venus is in flux. No more missions to Venus are planned before 2020. There’ve been proposals for things like sail powered landers, but we’re not there yet. We’re developing heat-resistant electronics, but so far they’re very simple. There’s a lot of work to do.

On the other hand, some things may happen sooner. It may turn out that we can learn about Venusian seismic activity from balloon-borne or orbital sensors. The team says that “Due to strong mechanical coupling between the atmosphere and ground, seismic waves are launched into the atmosphere, where they may be detected by infrasound on a balloon or infrared or ultraviolet signatures from orbit.” That’s thanks to Venus’ dense atmosphere. That means that the far-term goal of seismic sensing of the interior of Venus could be shifted to the near-term or mid-term.

Japan’s Akatsuki orbiter captured this image of a gravity wave in Venus’ upper cloud layer. Could orbiter sensors remove the need for a network of seismic sensors on the surface? Image credit: JAXA

As work on nanosatellites and cubesats continues, they may play a larger role at Venus, and shift the timelines. NASA wants to include these small satellites on every launch where there is a few kilograms of excess capacity. A group of these nanosatellites could form a network of seismic sensors much more easily and much sooner than an established network of surface sensors. A network of nanosatellites could also serve as a communications relay for other missions.

Venus doesn’t generate a lot of buzz these days. The discovery of Earth-like worlds in distant solar systems generates headline after headline. And the always popular search for life is centered on Mars, and the icy/sub-surface moons of our Solar System’s gas giants. But Venus is still a tantalizing target, and understanding Venus’ evolution will help us understand what we’re seeing in distant solar systems.

Are Fast Radio Bursts Evidence Of Alien Activity?

An artist's illustration of a light-sail powered by a radio beam (red) generated on the surface of a planet. Could the part of the beam that misses the sail be our mysterious Fast Radio Bursts? Image Credit: M. Weiss/CfA

The extremely energetic events that we see out there in the Universe are usually caused by cataclysmic astrophysical events and activities of one sort or another. But what about Fast Radio Bursts? A pair of astrophysicists at Harvard say that the seldom seen phenomena could, maybe, possibly, be evidence of an advanced alien technology.

Fast radio bursts (FRBs) are short-lived radio pulses that last only a few milliseconds. It’s been assumed that they have some astrophysical cause. Fewer than 2 dozen of them have been detected since their discovery in 2007. They’re detected by our huge radio telescopes like the Arecibo Observatory in Puerto Rico, and the Parkes Observatory in Australia. They’re extremely energetic, and their source is a great distance from us.

The NSF’s Arecibo Observatory, which is located in Puerto Rico, is the world largest radio telescope. Arecibo detected 11 FRBs over the course of 2 months. Credit: NAIC

The two astrophysicists, Avi Loeb at the Harvard-Smithsonian Center for Astrophysics, and Manasvi Lingam at Harvard University, decided to investigate the possibility that FRBs have an alien technological origin.

“Fast radio bursts are exceedingly bright given their short duration and origin at great distances, and we haven’t identified a possible natural source with any confidence. An artificial origin is worth contemplating and checking.” – Avi Loeb, Harvard-Smithsonian Center for Astrophysics

I’ll Take ‘Alien Signals’ For $200 Alex

Loeb and Lingam began by calculating how much energy would be needed to send a signal that strong across such an enormous distance. They found that doing so with solar energy requires a solar array with an area twice the surface area of Earth. That would be enough energy, if the alien civilization was as close as we are to a star similar to our Sun.

Obviously, such a massive construction project is well beyond us. But however unlikely it sounds, it can’t be ruled out.

The pair also asked themselves questions about the viability of such a project. Would the heat and energy involved in such a solar array melt the structure itself? Their answer is that water-cooling would be sufficient to keep an array like this operating.

Their next question was, “Why build something like this in the first place?”

I’ll Take ‘Alien Spacecraft Propulsion Systems’ For $400 Alex”

The thinking behind their idea is based on an idea that we ourselves have had: Could we power a spacecraft by pushing on it with lasers? Or Microwaves? If we’ve thought of it, why wouldn’t other existing civilizations? If another civilization were doing it, what would the technology look like?

Their investigation shows that the engineering they’re talking about could power a spacecraft with a payload of a million tons. That would be about 20 times bigger than our largest cruise ship. According to Lingam, “That’s big enough to carry living passengers across interstellar or even intergalactic distances.”

If FRBs are indeed the result of an alien propulsion system, here’s how it would work: Earth is rotating and orbiting, which means the alien star and galaxy are moving relative to us. That’s why we would only see a brief flash. The beam sweeps across the sky and only hits us for a moment. The repeated appearance of the FRB could be a clue to its alien, technological origin.

The authors of the study outlining this thinking know that it’s speculative. But it’s their job to speculate within scientific constraints, which they have done. As they say in the conclusion of their paper, “Although the possibility that FRBs are produced by extragalactic civilizations is more speculative than an astrophysical origin, quantifying the requirements necessary for an artificial origin serves, at the very least, the important purpose of enabling astronomers to rule it out with future data.”

There are other interpretations when it comes to FRBs, of course. The others of another paper say that for at least one group of FRBs, known as FRB 121102, the source is likely astrophysical. According to them, FRBs likely come from “a young, highly magnetized, extragalactic neutron star.”

Lurking behind these papers are some intriguing questions that are also fun to ponder.

If the system required a solar array twice the size of Earth, where would the materials come from? If the system required water-cooling to avoid melting, where would all the water come from? It’s impossible to know, or to even begin speculating. But a civilization able to do something like this would have to be master engineers and resource exploiters. That goes without saying.

Why they might do it is another question. Probably the same reasons we would: curiosity and exploration, or maybe to escape a dying world.

Either that or they ran out of beer.

How Will NASA Find Life On Other Worlds?

Is Earth in the range of normal when it comes to habitable planets? Or is it an outlier, with both large land masses, and large oceans? Image: Reto Stöckli, Nazmi El Saleous, and Marit Jentoft-Nilsen, NASA GSFC
Is Earth in the range of normal when it comes to habitable planets? Or is it an outlier, with both large land masses, and large oceans? Image: Reto Stöckli, Nazmi El Saleous, and Marit Jentoft-Nilsen, NASA GSFC

For a long time, the idea of finding life on other worlds was just a science fiction dream. But in our modern times, the search for life is rapidly becoming a practical endeavour. Now, some minds at NASA are looking ahead to the search for life on other worlds, and figuring out how to search more effectively and efficiently. Their approach is centered around two things: nano-satellites and microfluidics.

Life is obvious on Earth. But it’s a different story for the other worlds in our Solar System. Mars is our main target right now, with the work that MSL Curiosity is doing. But Curiosity is investigating Mars to find out if conditions on that planet were ever favorable for life. A more exciting possibility is finding extant life on another world: that is, life that exists right now.

MSL Curiosity is busy investigating the surface of Mars, to see if that planet could have harbored life. Image: NASA/JPL/Cal-Tech
MSL Curiosity is busy investigating the surface of Mars, to see if that planet could have harbored life. Image: NASA/JPL/Cal-Tech

At the Planetary Science Vision 2050 Workshop, experts in Planetary Science and related disciplines gathered to present ideas about the next 50 years of exploration in the Solar System. A team led by Richard Quinn at the NASA Ames Research Center (ARC) presented their ideas on the search for extant life in the next few decades.

Their work is based on the decadal survey “Vision and Voyages for Planetary Science in the Decade 2013-2022.” That source confirms what most of us are already aware of: that our search for life should be focussed on Mars and the so-called “Ocean Worlds” of our Solar System like Enceladus and Europa. The question is, what will that search look like?

The North Polar Region of Saturn’s moon, Enceladus. Could there be an ocean world full of life under its frozen surface? Credit: NASA/JPL/Space Science Institute

Quinn and his team outlined two technologies that we could center our search around.

Nanosatellites

A nanosatellite is classified as something with a mass between 1-10 kg. They offer several advantages over larger designs.

Firstly, their small mass keeps the cost of launching them very low. In many cases, nanosatellites can be piggy-backed onto the launch of a larger payload, just to use up any excess capacity. Nanosatellites can be made cheaply, and multiples of them can be designed and built the same. This would allow a fleet of nanosatellites to be sent to the same destination.

Most of the discussion around the search for life centers around large craft or landers that land in one location, and have limited mobility. The Mars rovers are doing great work, but they can only investigate very specific locations. In a way, this creates kind of a sampling error. It’s difficult to generalize about the conditions for life on other worlds when we’ve only sampled a small handful of locations.

In 2010, NASA successfully deployed the nanosatellite NANO-Sail D from a larger, microsatellite. Image: NASA

On Earth, life is everywhere. But Earth is also the home to extremophiles, organisms that exist only in extreme, hard-to-reach locations. Think of thermal vents on the ocean floor, or deep dark caves. If that is the kind of life that exists on the target worlds in our Solar System, then there’s a strong possibility that we’ll need to sample many locations before we find them. That is something that is beyond the capabilities of our rovers. Nanosatellites could be part of the solution. A fleet of them investigating a world like Enceladus or Europa could speed up our search for extant life.

NASA has designed and built nanosatellites to perform a variety of tasks, like performing biology experiments, and testing advanced propulsion and communications technologies. In 2010 they successfully deployed a nanosatellite from a larger, microsatellite. If you expand on that idea, you can see how a small fleet of nanosatellites could be deployed at another world, after arriving there on another larger craft.

Microfluidics

Microfluidics deals with systems that manipulate very small amounts of fluid, usually on the sub-millimeter scale. The idea is to build microchips which handle very small sample sizes, and test them in-situ. NASA has done work with microfluidics to try to develop ways of monitoring astronauts’ health on long space voyages, where there is no access to a lab. Microfluidic chips can be manufactured which have only one or two functions, and produce only one or two results.

In terms of the search for extant life in our Solar System, microfluidics is a natural fit with nanosatellites. Replace the medical diagnostic capabilities of a microfluidic chip with a biomarker diagnostic, and you have a tiny device that can be mounted on a tiny satellite. Since functioning microfluidic chips can be as small as microprocessors, multiples of them could be mounted.

” Technical constraints will inevitably limit robotic missions that search for evidence of life to a few selected experiments.” – Richard.C.Quinn, et. al.

When combined with nanosatellites, microfluidics offers the possibility of the same few tests for life being repeated over and over in multiple locations. This is obviously very attractive when it comes to the search for life. The team behind the idea stresses that their approach would involve the search for simple building blocks, the complex biomolecules involved in basic biochemistry, and also the structures that cellular life requires in order to exist. Performing these tests in multiple locations would be a boon in the search.

Some of the technologies for the microfluidic search for life have already been developed. The team points out that several of them have already had successful demonstrations in micro-gravity missions like the GeneSat, the PharmaSat, and the SporeSat.

“The combination of microfluidic systems with chemical and biochemical sensors and sensor arrays offer some of the most promising approaches for extant life detection using small-payload platforms.” – Richard.C.Quinn, et. al.

Putting It All Together

We’re a ways away from a mission to Europa or Enceladus. But this paper was about the future vision of the search for extant life. It’s never too soon to start thinking about that.

There are some obvious obstacles to using nanosatellites to search for life on Enceladus or Europa. Those worlds are frozen, and it’s the oceans under those thick ice caps that we need to investigate. Somehow, our tiny nanosatellites would need to get through that ice.

Also, the nanosatellites we have now are just that: satellites. They are designed to be in orbit around a body. How could they be transformed into tiny, ocean-going submersible explorers?
There’s no doubt that somebody, somewhere at NASA, is already thinking about that.

The over-arching vision of a fleet of small craft, each with the ability to repeat basic experiments searching for life in multiple locations, is a sound one. As for how it actually turns out, we’ll have to wait and see.

Rise of the Super Telescopes: The James Webb Space Telescope

A full-scale model of the JWST went on a bit of a World Tour. Here it is in Munich, Germany. Image Credit: EADS Astrium

We humans have an insatiable hunger to understand the Universe. As Carl Sagan said, “Understanding is Ecstasy.” But to understand the Universe, we need better and better ways to observe it. And that means one thing: big, huge, enormous telescopes.
In this series we’ll look at 6 of the world’s Super Telescopes:

The James Webb Space Telescope

The James Webb Space Telescope“>James Webb Space Telescope (JWST, or the Webb) may be the most eagerly anticipated of the Super Telescopes. Maybe because it has endured a tortured path on its way to being built. Or maybe because it’s different than the other Super Telescopes, what with it being 1.5 million km (1 million miles) away from Earth once it’s operating.

The JWST will do its observing while in what’s called a halo orbit at L2, a sort of gravitationally neutral point 1.5 million km from Earth. Image: NASA/JWST

If you’ve been following the drama behind the Webb, you’ll know that cost overruns almost caused it to be cancelled. That would’ve been a real shame.

The JWST has been brewing since 1996, but has suffered some bumps along the road. That road and its bumps have been discussed elsewhere, so what follows is a brief rundown.

Initial estimates for the JWST were a $1.6 billion price tag and a launch date of 2011. But the costs ballooned, and there were other problems. This caused the House of Representatives in the US to move to cancel the project in 2011. However, later that same year, US Congress reversed the cancellation. Eventually, the final cost of the Webb came to $8.8 billion, with a launch date set for October, 2018. That means the JWST’s first light will be much sooner than the other Super Telescopes.

The business end of the James Webb Space Telescope is its 18-segment primary mirror. The gleaming, gold-coated beryllium mirror has a collecting area of 25 square meters. Image: NASA/Chris Gunn

The Webb was envisioned as a successor to the Hubble Space Telescope, which has been in operation since 1990. But the Hubble is in Low Earth Orbit, and has a primary mirror of 2.4 meters. The JWST will be located in orbit at the LaGrange 2 point, and its primary mirror will be 6.5 meters. The Hubble observes in the near ultraviolet, visible, and near infrared spectra, while the Webb will observe in long-wavelength (orange-red) visible light, through near-infrared to the mid-infrared. This has some important implications for the science yielded by the Webb.

The Webb’s Instruments

The James Webb is built around four instruments:

  • The Near-Infrared Camera (NIRCam)
  • The Near-Infrared Spectrograph (NIRSpec)
  • The Mid-Infrared Instrument(MIRI)
  • The Fine Guidance Sensor/ Near InfraRed Imager and Slitless Spectrograph (FGS/NIRISS)
This image shows the wavelengths of the infrared spectrum that Webb’s instruments can observe. Image: NASA/JWST

The NIRCam is Webb’s primary imager. It will observe the formation of the earliest stars and galaxies, the population of stars in nearby galaxies, Kuiper Belt Objects, and young stars in the Milky Way. NIRCam is equipped with coronagraphs, which block out the light from bright objects in order to observe dimmer objects nearby.

NIRSpec will operate in a range from 0 to 5 microns. Its spectrograph will split the light into a spectrum. The resulting spectrum tells us about an objects, temperature, mass, and chemical composition. NIRSpec will observe 100 objects at once.

MIRI is a camera and a spectrograph. It will see the redshifted light of distant galaxies, newly forming stars, objects in the Kuiper Belt, and faint comets. MIRI’s camera will provide wide-field, broadband imaging that will rank up there with the astonishing images that Hubble has given us a steady diet of. The spectrograph will provide physical details of the distant objects it will observe.

The Fine Guidance Sensor part of FGS/NIRISS will give the Webb the precision required to yield high-quality images. NIRISS is a specialized instrument operating in three modes. It will investigate first light detection, exoplanet detection and characterization, and exoplanet transit spectroscopy.

The Science

The over-arching goal of the JWST, along with many other telescopes, is to understand the Universe and our origins. The Webb will investigate four broad themes:

  • First Light and Re-ionization: In the early stages of the Universe, there was no light. The Universe was opaque. Eventually, as it cooled, photons were able to travel more freely. Then, probably hundreds of millions of years after the Big Bang, the first light sources formed: stars. But we don’t know when, or what types of stars.
  • How Galaxies Assemble: We’re accustomed to seeing stunning images of the grand spiral galaxies that exist in the Universe today. But galaxies weren’t always like that. Early galaxies were often small and clumpy. How did they form into the shapes we see today?
  • The Birth of Stars and Protoplanetary Systems: The Webb’s keen eye will peer straight through clouds of dust that ‘scopes like the Hubble can’t see through. Those clouds of dust are where stars are forming, and their protoplanetary systems. What we see there will tell us a lot about the formation of our own Solar System, as well as shedding light on many other questions.
  • Planets and the Origins of Life: We now know that exoplanets are common. We’ve found thousands of them orbiting all types of stars. But we still know very little about them, like how common atmospheres are, and if the building blocks of life are common.

These are all obviously fascinating topics. But in our current times, one of them stands out among the others: Planets and the Origins of Life.

The recent discovery the TRAPPIST 1 system has people excited about possibly discovering life in another solar system. TRAPPIST 1 has 7 terrestrial planets, and 3 of them are in the habitable zone. It was huge news in February 2017. The buzz is still palpable, and people are eagerly awaiting more news about the system. That’s where the JWST comes in.

One big question around the TRAPPIST system is “Do the planets have atmospheres?” The Webb can help us answer this.

The NIRSpec instrument on JWST will be able to detect any atmospheres around the planets. Maybe more importantly, it will be able to investigate the atmospheres, and tell us about their composition. We will know if the atmospheres, if they exist, contain greenhouse gases. The Webb may also detect chemicals like ozone and methane, which are biosignatures and can tell us if life might be present on those planets.

You could say that if the James Webb were able to detect atmospheres on the TRAPPIST 1 planets, and confirm the existence of biosignature chemicals there, it will have done its job already. Even if it stopped working after that. That’s probably far-fetched. But still, the possibility is there.

Launch and Deployment

The science that the JWST will provide is extremely intriguing. But we’re not there yet. There’s still the matter of JWST’s launch, and it’s tricky deployment.

The JWST’s primary mirror is much larger than the Hubble’s. It’s 6.5 meters in diameter, versus 2.4 meters for the Hubble. The Hubble was no problem launching, despite being as large as a school bus. It was placed inside a space shuttle, and deployed by the Canadarm in low earth orbit. That won’t work for the James Webb.

This image shows the Hubble Space Telescope being held above the shuttle’s cargo bay by the Canadian-built Remote Manipulator System (RMS) arm, or Canadarm. A complex operation, but not as complex as JWST’s deployment. Image: NASA

The Webb has to be launched aboard a rocket to be sent on its way to L2, it’s eventual home. And in order to be launched aboard its rocket, it has to fit into a cargo space in the rocket’s nose. That means it has to be folded up.

The mirror, which is made up of 18 segments, is folded into three inside the rocket, and unfolded on its way to L2. The antennae and the solar cells also need to unfold.

Unlike the Hubble, the Webb needs to be kept extremely cool to do its work. It has a cryo-cooler to help with that, but it also has an enormous sunshade. This sunshade is five layers, and very large.

We need all of these components to deploy for the Webb to do its thing. And nothing like this has been tried before.

The Webb’s launch is only 7 months away. That’s really close, considering the project almost got cancelled. There’s a cornucopia of science to be done once it’s working.

But we’re not there yet, and we’ll have to go through the nerve-wracking launch and deployment before we can really get excited.

Rise of the Super Telescopes: The Thirty Meter Telescope

An artist's illustration of the Thirty Meter Telescope at its preferred location at Mauna Kea, Hawaii. Image Courtesy TMT International Observatory

As Carl Sagan said, “Understanding is Ecstasy.” But in order to understand the Universe, we need better and better ways to observe it. And that means one thing: big, huge, enormous telescopes.

In this series, we’ll look at six Super Telescopes being built:

The Thirty Meter Telescope

The Thirty Meter Telescope (TMT) is being built by an international group of countries and institutions, like a lot of Super Telescopes are. In fact, they’re proud of pointing out that the international consortium behind the TMT represents almost half of the world’s population; China, India, the USA, Japan, and Canada. The project needs that many partners to absorb the cost; an estimated $1.5 billion.

The heart of any of the world’s Super Telescopes is the primary mirror, and the TMT is no different. The primary mirror for the TMT is, obviously, 30 meters in diameter. It’s a segmented design consisting of 492 smaller mirrors, each one a 1.4 meter hexagon.

The light collecting capability of the TMT will be 10 times that of the Keck Telescope, and more than 144 times that of the Hubble Space Telescope.

But the TMT is more than just an enormous ‘light bucket.’ It also excels with other capabilities that define a super telescope’s effectiveness. One of those is what’s called diffraction-limited spatial resolution (DLSR).

An illustration of the segmented primary mirror of the Thirty Meter Telescope. Image Courtesy TMT International Observatory

When a telescope is pointed at distant objects that appear close together, the light from both can scatter enough to make the two objects appear as one. Diffraction-limited spatial resolution means that when a ‘scope is observing a star or other object, none of the light from that object is scattered by defects in the telescope. The TMT will more easily distinguish objects that are close to each other. When it comes to DLSR, the TMT will exceed the Keck by a factor of 3, and will exceed the Hubble by a factor of 10 at some wavelengths.

Crucial to the function of large, segmented mirrors like the TMT is active optics. By controlling the shape and position of each segment, active optics allows the primary mirror to compensate for changes in wind, temperature, or mechanical stress on the telescope. Without active optics, and its sister technology adaptive optics, which compensates for atmospheric disturbance, any telescope larger than about 8 meters would not function properly.

The TMT will operate in the near-ultraviolet, visible, and near-infrared wavelengths. It will be smaller than the European Extremely Large Telescope (E-ELT), which will have a 39 meter primary mirror. The E-ELT will operate in the optical and infrared wavelengths.

The world’s Super Telescopes are behemoths. Not just in the size of their mirrors, but in their mass. The TMT’s moving mass will be about 1,420 tonnes. Moving the TMT quickly is part of the design of the TMT, because it must respond quickly when something like a supernova is spotted. The detailed science case calls for the TMT to acquire a new target within 5 to 10 minutes.

This requires a complex computer system to coordinate the science instruments, the mirrors, the active optics, and the adaptive optics. This was one of the initial challenges of the TMT project. It will allow the TMT to respond to transient phenomena like supernovae when spotted by other telescopes like the Large Synoptic Survey Telescope.

The Science

The TMT will investigate most of the important questions in astronomy and cosmology today. Here’s an overview of major topics that the TMT will address:

  • The Nature of Dark Matter
  • The Physics of Extreme Objects like Neutron Stars
  • Early galaxies and Cosmic Reionization
  • Galaxy Formation
  • Super-Massive Black Holes
  • Exploration of the Milky Way and Nearby Galaxies
  • The Birth and Early Lives of Stars and Planets
  • Time Domain Science: Supernovae and Gamma Ray Bursts
  • Exo-planets
  • Our Solar System

This is a comprehensive list of topics, to be sure. It leaves very little out, and is a testament to the power and effectiveness of the TMT.

The raw power of the TMT is not in question. Once in operation it will advance our understanding of the Universe on multiple fronts. But the actual location of the TMT could still be in question.

Where Will the TMT Be Built?

The original location for the TMT was Mauna Kea, the 4,200 meter summit in Hawaii. Mauna Kea is an excellent location, and is the home of several telescopes, most notably the Keck Observatory, the Gemini Telescope, the Subaru Telescope, the Canada-France-Hawaii Telescope, and the James Clerk Maxwell Telescope. Mauna Kea is also the site of the westernmost antenna of the Very Long Baseline Array.

The top of Mauna Kea is a prime site for telescopes, as shown in this image. Image Courtesy Mauna Kea Observatories

The dispute between some of the Hawaiian people and the TMT has been well-documented elsewhere, but the basic complaint about the TMT is that the top of Mauna Kea is sacred land, and they would like the TMT to be built elsewhere.

The organizations behind the TMT would still like it to be built at Mauna Kea, and a legal process is unfolding around the dispute. During that process, they identified several possible alternate sites for the telescope, including La Palma in the Canary Islands. Universe Today contacted TMT Observatory Scientist Christophe Dumas, PhD., about the possible relocation of the TMT to another site.

Dr. Dumas told us that “Mauna Kea remains the preferred location for the TMT because of its superb observing conditions, and because of the synergy with other TMT partner facilities already present on the mountain. Its very high elevation of almost 14,000 feet makes it the premier astronomical site in the northern hemisphere. The sky above Mauna Kea is very stable, which allows very sharp images to be obtained. It has also excellent transparency, low light pollution and stable cold temperatures that improves sensitivity for observations in the infrared.”

The preferred secondary site at La Palma is home to over 10 other telescopes, but would relocation to the Canary Islands affect the science done by the TMT? Dr. Dumas says that the Canary Islands site is excellent as well, with similar atmospheric characteristics to Mauna Kea, including stability, transparency, darkness, and fraction of clear-nights.

The Gran Telescopio Canarias (Great Canary Telescope) is the largest ‘scope currently at La Palma. At 10m diameter, it would be dwarfed by the TMT. Image: By Pachango – Own work, CC BY-SA 3.0, https://commons.wikimedia.org/w/index.php?curid=6880933

As Dr. Dumas explains, “La Palma is at a lower elevation site and on average warmer than Mauna Kea. These two factors will reduce TMT sensitivity at some wavelengths in the infrared region of the spectrum.”

Dr. Dumas told Universe Today that this reduced sensitivity in the infrared can be overcome somewhat by scheduling different observing tasks. “This specific issue can be partly mitigated by implementing an adaptive scheduling of TMT observations, to match the execution of the most demanding infrared programs with the best atmospheric conditions above La Palma.”

Court Proceedings End

On March 3rd, 44 days of court hearings into the TMT wrapped up. In that time, 71 people testified for and against the TMT being constructed on Mauna Kea. Those against the telescope say that the site is sacred land and shouldn’t have any more telescope construction on it. Those for the TMT spoke in favor of the science that the TMT will deliver to everyone, and the education opportunities it will provide to Hawaiians.

Though construction has been delayed, and people have gone to court to have the project stopped, it seems like the TMT will definitely be built—somewhere. The funding is in place, the design is finalized, and manufacturing of the components is underway. The delays mean that the TMT’s first light is still uncertain, but once we get there, the TMT will be another game-changer, just like the world’s other Super Telescopes.

Towards A New Understanding Of Dark Matter

In February 2016, LIGO detected gravity waves for the first time. As this artist's illustration depicts, the gravitational waves were created by merging black holes. The third detection just announced was also created when two black holes merged. Credit: LIGO/A. Simonnet.
Artist's impression of merging binary black holes. Credit: LIGO/A. Simonnet.

Dark matter remains largely mysterious, but astrophysicists keep trying to crack open that mystery. Last year’s discovery of gravity waves by the Laser Interferometer Gravitational Wave Observatory (LIGO) may have opened up a new window into the dark matter mystery. Enter what are known as ‘primordial black holes.’

Theorists have predicted the existence of particles called Weakly Interacting Massive Particles (WIMPS). These WIMPs could be what dark matter is made of. But the problem is, there’s no experimental evidence to back it up. The mystery of dark matter is still an open case file.

When LIGO detected gravitational waves last year, it renewed interest in another theory attempting to explain dark matter. That theory says that dark matter could actually be in the form of Primordial Black Holes (PBHs), not the aforementioned WIMPS.

Primordial black holes are different than the black holes you’re probably thinking of. Those are called stellar black holes, and they form when a large enough star collapses in on itself at the end of its life. The size of these stellar black holes is limited by the size and evolution of the stars that they form from.

This artist’s drawing shows a stellar black hole as it pulls matter from a blue star beside it. Could the stellar black hole’s cousin, the primordial black hole, account for the dark matter in our Universe?
Credits: NASA/CXC/M.Weiss

Unlike stellar black holes, primordial black holes originated in high density fluctuations of matter during the first moments of the Universe. They can be much larger, or smaller, than stellar black holes. PBHs could be as small as asteroids or as large as 30 solar masses, even larger. They could also be more abundant, because they don’t require a large mass star to form.

When two of these PBHs larger than about 30 solar masses merge together, they would create the gravitational waves detected by LIGO. The theory says that these primordial black holes would be found in the halos of galaxies.

If there are enough of these intermediate sized PBHs in galactic halos, they would have an effect on light from distant quasars as it passes through the halo. This effect is called ‘micro-lensing’. The micro-lensing would concentrate the light and make the quasars appear brighter.

A depiction of quasar microlensing. The microlensing object in the foreground galaxy could be a star (as depicted), a primordial black hole, or any other compact object. Credit: NASA/Jason Cowan (Astronomy Technology Center).

The effect of this micro-lensing would be stronger the more mass a PBH has, or the more abundant the PBHs are in the galactic halo. We can’t see the black holes themselves, of course, but we can see the increased brightness of the quasars.

Working with this assumption, a team of astronomers at the Instituto de Astrofísica de Canarias examined the micro-lensing effect on quasars to estimate the numbers of primordial black holes of intermediate mass in galaxies.

“The black holes whose merging was detected by LIGO were probably formed by the collapse of stars, and were not primordial black holes.” -Evencio Mediavilla

The study looked at 24 quasars that are gravitationally lensed, and the results show that it is normal stars like our Sun that cause the micro-lensing effect on distant quasars. That rules out the existence of a large population of PBHs in the galactic halo. “This study implies “says Evencio Mediavilla, “that it is not at all probable that black holes with masses between 10 and 100 times the mass of the Sun make up a significant fraction of the dark matter”. For that reason the black holes whose merging was detected by LIGO were probably formed by the collapse of stars, and were not primordial black holes”.

Depending on you perspective, that either answers some of our questions about dark matter, or only deepens the mystery.

We may have to wait a long time before we know exactly what dark matter is. But the new telescopes being built around the world, like the European Extremely Large Telescope, the Giant Magellan Telescope, and the Large Synoptic Survey Telescope, promise to deepen our understanding of how dark matter behaves, and how it shapes the Universe.

It’s only a matter of time before the mystery of dark matter is solved.

Rise of the Super Telescopes: The Large Synoptic Survey Telescope

An artist's illustration of the Large Synoptic Survey Telescope with a simulated night sky. The team hopes to use the LSST to further refine their search for hard-surface supermassive objects. Image: Todd Mason, Mason Productions Inc. / LSST Corporation
An artist's illustration of the Large Synoptic Survey Telescope with a simulated night sky. The team hopes to use the LSST to further refine their search for hard-surface supermassive objects. Image: Todd Mason, Mason Productions Inc. / LSST Corporation

We humans have an insatiable hunger to understand the Universe. As Carl Sagan said, “Understanding is Ecstasy.” But to understand the Universe, we need better and better ways to observe it. And that means one thing: big, huge, enormous telescopes.

In this series we’ll look at 6 of the world’s Super Telescopes:

The Large Synoptic Survey Telescope

While the world’s other Super Telescopes rely on huge mirrors to do their work, the LSST is different. It’s a huge panoramic camera that will create an enormous moving image of the Universe. And its work will be guided by three words: wide, deep, and fast.

While other telescopes capture static images, the LSST will capture richly detailed images of the entire available night sky, over and over. This will allow astronomers to basically “watch” the movement of objects in the sky, night after night. And the imagery will be available to anyone.

The LSST is being built by a group of institutions in the US, and even got some money from Bill Gates. It will be situated atop Cerro Pachon, a peak in Northern Chile. The Gemini South and Southern Astrophysical Research Telescopes are also situated there.

The Camera Inside the ‘Scope

At the heart of the LSST is its enormous digital camera. It weighs over three tons, and the sensor is segmented in a similar way that other Super Telescopes have segmented mirrors. The LSST’s camera is made up of 189 segments, which together create a camera sensor about 2 ft. in diameter, behind a lens that is over 5 ft. in diameter.

Each image that the LSST captures is 40 times larger than the full moon, and will measure 3.2 gigapixels. The camera will capture one of these wide-field images every 20 seconds, all night long. Every few nights, the LSST will give us an image of the entire available night sky, and it will do that for 10 years.

“The LSST survey will open a movie-like window on objects that change brightness, or move, on timescales ranging from 10 seconds to 10 years.” – LSST: FROM SCIENCE DRIVERS TO REFERENCE DESIGN AND ANTICIPATED DATA PRODUCTS

The LSST will capture a vast, movie-like image of over 40 billion objects. This will range from distant, enormous galaxies all the way down to Potentially Hazardous Objects as small as 140 meters in diameter.

The primary-tertiay mirror at its construction facility. Image: LSST

There’s a whole other side to the LSST which is a little more challenging. We get the idea of an in-depth, moving, detailed image of the sky. That’s intuitively easy to engage with. But there’s another side, the data mining challenge.

The Data Challenge

The whole endeavour will create an enormous amount of data. Over 15 terabytes will have to be processed every night. Over its 10 year lifespan, it will capture 60 petabytes of data.

Once data is captured by the LSST, it will travel via two dedicated 40 GB lines to the Data Processing and Archive Center. That Center is a super-computing facility that will manage all the data and make it available to users. But when it comes to handling the data, that’s just the tip of the iceberg.

“LSST is a new way to observe, and gaining knowledge from the Big Data LSST delivers is indeed a challenge.” – Suzanne H. Jacoby, LSST

The sheer amount of data created by the LSST is a challenge that the team behind it saw coming. They knew they would have to build the capacity of the scientific community in advance, in order to get the most out of the LSST.

Handling all of the data from the LSST requires its own infrastructure. Image: LSST

As Suzanne Jacoby, from the LSST team, told Universe today, “To prepare the science community for LSST Operations, the LSST Corporation has undertaken an “Enabling Science” effort which funds the LSST Data Science Fellowship Program (DSFP). This two-year program is designed to supplement existing graduate school curriculum and explores topics including statistics, machine learning, information theory, and scalable programming.”

The Science

The Nature of Dark Matter and Understanding Dark Energy

Contributing to our understanding Dark Energy and Dark Matter is a goal of all of the Super Telescopes. The LSST will map several billion galaxies through time and space. It will help us understand how Dark Energy behaves over time, and how Dark Matter affects the development of cosmic structure.

Cataloging the Solar System

The raw imaging power of the LSST will be a game-changer for mapping and cataloguing our Solar System. It’s thought that the LSST could detect between 60-90% of all potentially hazardous asteroids (PHAs) larger than 140 meters in diameter, as far away as the main asteroid belt. This will not only contribute to NASA’s goal of identifying threats to Earth posed by asteroids, but will help us understand how planets formed and how our Solar System evolved.

Exploring the Changing Sky

The repeated imaging of the night sky, at great depth and with excellent image quality, should tell us a lot about supernovae, variable stars, and possible other events we haven’t even discovered yet. There are always surprising results whenever we build a new telescope or send a probe to a new destination. The LSST will probably be no different.

Milky Way Structure & Formation

The LSST will give us an unprecedented look at the Milky Way. It will survey over half of the sky, and will do so repeatedly. Hundreds of times, in fact. The end result will be an enormously detailed look at the motion of millions of stars in our galaxy.

Open Access

Perhaps the best part of the whole LSST project is that the all of the data will be available to everyone. Anyone with a computer and an internet connection will be able to access LSST’s movie of the Universe. It’s warm and fuzzy, to be sure, to have the results of large science endeavours like this available to anyone. But there’s more to it. The LSST team suspects that the majority of the discoveries resulting from its rich data will come from unaffiliated astronomers, students, and even amateurs.

It was designed from the ground up in this way, and there will be no delay or proprietary barriers when it comes to public data access. In fact, Google has signed on as a partner with LSST because of the desire for public access to the data. We’ve seen what Google has done with Google Earth and Google Sky. What will they come up with for Google LSST?

The Sloan Digital Sky Survey (SDSS), a kind of predecessor to the LSST, was modelled in the same way. All of its data was available to astronomers not affiliated with it, and out of over 6000 papers that refer to SDSS data, the large majority of them were published by astronomers not affiliated with SDSS.

First Light

We’ll have to wait a while for all of this to come our way, though. First light for the LSST won’t be until 2021, and it will begin its 10 year run in 2022. At that time, be ready for a whole new look at our Universe. The LSST will be a game-changer.