Jump to content

Earth Is Surrounded by a 1000 Light Year Wide Bubble Carved Out by Supernovas


Recommended Posts

Posted
The cosmic void is surrounded by multiple star-forming regions created by the explosions. Earth is slap bang in the middle of a 1,000 light-year-wide bubble with a dense surface birthing thousands of baby stars. 

AVvXsEi3OpX_GjA4AYlSoecRyNh1AH-9vX_yX3F47tmz9KKsEa9Akm13RMbEwdnpABfux28jJfnk0EU8oe2sBgsVH_ocwyUB9jqfeBo90UpzrkTDNT5L2UUpkP9Wno58pzQA7C0gh5CNAI9Qpr6GsPU1eHTyYr5TvoKNft1BbIvZ5nCSvnRwRSIzdP_-GIHy0A=w640-h360

Researchers have long wondered what created this "superbubble." Now, a new study suggests that at least 15 powerful star explosions inflated this cosmic bubble. 

Astronomers in the 1970s first discovered the gigantic void, known as the Local Bubble, after realizing that no stars had formed inside the blob for around 14 million years. The only stars inside the bubble either existed before the bubble emerged or formed outside the void and are now passing through; the sun is one such trespasser. 

This setup had suggested that several supernovas were responsible for this void. Those stellar explosions, the researchers said, would have blasted the materials needed to make new stars, such as hydrogen gas, to the edge of a huge area in space, leaving behind the Local Bubble that's surrounded by a frenzy of star births. 

In a new study, published online Jan. 12 in the journal Nature, researchers accurately mapped the star-forming regions surrounding the Local Bubble and, in doing so, calculated how fast the superbubble is expanding. 

This allowed the team to work out exactly how many supernovas were needed to carve out the gigantic cosmic void and better understand how star-forming regions are created across the Milky Way.

 

View the full article

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.

Guest
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

  • Similar Topics

    • By NASA
      NASA’s Jet Propulsion Laboratory used radar data taken by ESA’s Sentinel-1A satellite before and after the 2015 eruption of the Calbuco volcano in Chile to create this inter-ferogram showing land deformation. The color bands west of the volcano indicate land sinking. NISAR will produce similar images.ESA/NASA/JPL-Caltech A SAR image — like ones NISAR will produce — shows land cover on Mount Okmok on Alaska’s Umnak Island . Created with data taken in August 2011 by NASA’s UAVSAR instrument, it is an example of polarimetry, which measures return waves’ orientation relative to that of transmitted signals.NASA/JPL-Caltech Data from NASA’s Magellan spacecraft, which launched in 1989, was used to create this image of Crater Isabella, a 108-mile-wide (175-kilometer-wide) impact crater on Venus’ surface. NISAR will use the same basic SAR principles to measure properties and characteristics of Earth’s solid surfaces.NASA/JPL-Caltech Set to launch within a few months, NISAR will use a technique called synthetic aperture radar to produce incredibly detailed maps of surface change on our planet.
      When NASA and the Indian Space Research Organization’s (ISRO) new Earth satellite NISAR (NASA-ISRO Synthetic Aperture Radar) launches in coming months, it will capture images of Earth’s surface so detailed they will show how much small plots of land and ice are moving, down to fractions of an inch. Imaging nearly all of Earth’s solid surfaces twice every 12 days, it will see the flex of Earth’s crust before and after natural disasters such as earthquakes; it will monitor the motion of glaciers and ice sheets; and it will track ecosystem changes, including forest growth and deforestation.  
      The mission’s extraordinary capabilities come from the technique noted in its name: synthetic aperture radar, or SAR. Pioneered by NASA for use in space, SAR combines multiple measurements, taken as a radar flies overhead, to sharpen the scene below. It works like conventional radar, which uses microwaves to detect distant surfaces and objects, but steps up the data processing to reveal properties and characteristics at high resolution.
      To get such detail without SAR, radar satellites would need antennas too enormous to launch, much less operate. At 39 feet (12 meters) wide when deployed, NISAR’s radar antenna reflector is as wide as a city bus is long. Yet it would have to be 12 miles (19 kilometers) in diameter for the mission’s L-band instrument, using traditional radar techniques, to image pixels of Earth down to 30 feet (10 meters) across.
      Synthetic aperture radar “allows us to refine things very accurately,” said Charles Elachi, who led NASA spaceborne SAR missions before serving as director of NASA’s Jet Propulsion Laboratory in Southern California from 2001 to 2016. “The NISAR mission will open a whole new realm to learn about our planet as a dynamic system.”
      Data from NASA’s Magellan spacecraft, which launched in 1989, was used to create this image of Crater Isabella, a 108-mile-wide (175-kilometer-wide) impact crater on Venus’ surface. NISAR will use the same basic SAR principles to measure properties and characteristics of Earth’s solid surfaces.NASA/JPL-Caltech How SAR Works
      Elachi arrived at JPL in 1971 after graduating from Caltech, joining a group of engineers developing a radar to study Venus’ surface. Then, as now, radar’s allure was simple: It could collect measurements day and night and see through clouds. The team’s work led to the Magellan mission to Venus in 1989 and several NASA space shuttle radar missions.
      An orbiting radar operates on the same principles as one tracking planes at an airport. The spaceborne antenna emits microwave pulses toward Earth. When the pulses hit something — a volcanic cone, for example — they scatter. The antenna receives those signals that echo back to the instrument, which measures their strength, change in frequency, how long they took to return, and if they bounced off of multiple surfaces, such as buildings.
      This information can help detect the presence of an object or surface, its distance away, and its speed, but the resolution is too low to generate a clear picture. First conceived at Goodyear Aircraft Corp. in 1952, SAR addresses that issue.
      “It’s a technique to create high-resolution images from a low-resolution system,” said Paul Rosen, NISAR’s project scientist at JPL.
      As the radar travels, its antenna continuously transmits microwaves and receives echoes from the surface. Because the instrument is moving relative to Earth, there are slight changes in frequency in the return signals. Called the Doppler shift, it’s the same effect that causes a siren’s pitch to rise as a fire engine approaches then fall as it departs.
      Computer processing of those signals is like a camera lens redirecting and focusing light to produce a sharp photograph. With SAR, the spacecraft’s path forms the “lens,” and the processing adjusts for the Doppler shifts, allowing the echoes to be aggregated into a single, focused image.
      Using SAR
      One type of SAR-based visualization is an interferogram, a composite of two images taken at separate times that reveals the differences by measuring the change in the delay of echoes. Though they may look like modern art to the untrained eye, the multicolor concentric bands of interferograms show how far land surfaces have moved: The closer the bands, the greater the motion. Seismologists use these visualizations to measure land deformation from earthquakes.
      Another type of SAR analysis, called polarimetry, measures the vertical or horizontal orientation of return waves relative to that of transmitted signals. Waves bouncing off linear structures like buildings tend to return in the same orientation, while those bouncing off irregular features, like tree canopies, return in another orientation. By mapping the differences and the strength of the return signals, researchers can identify an area’s land cover, which is useful for studying deforestation and flooding.
      Such analyses are examples of ways NISAR will help researchers better understand processes that affect billions of lives.
      “This mission packs in a wide range of science toward a common goal of studying our changing planet and the impacts of natural hazards,” said Deepak Putrevu, co-lead of the ISRO science team at the Space Applications Centre in Ahmedabad, India.
      Learn more about NISAR at:
      https://nisar.jpl.nasa.gov
      News Media Contacts
      Andrew Wang / Jane J. Lee
      Jet Propulsion Laboratory, Pasadena, Calif.
      626-379-6874 / 818-354-0307
      andrew.wang@jpl.nasa.gov / jane.j.lee@jpl.nasa.gov
      2025-006
      Share
      Details
      Last Updated Jan 21, 2025 Related Terms
      NISAR (NASA-ISRO Synthetic Aperture Radar) Earth Earth Science Earth Science Division Jet Propulsion Laboratory Explore More
      4 min read NASA Scientists, Engineers Receive Presidential Early Career Awards 
      Article 4 days ago 6 min read NASA International Space Apps Challenge Announces 2024 Global Winners
      Article 5 days ago 3 min read NASA Scientists Find New Human-Caused Shifts in Global Water Cycle
      Article 5 days ago Keep Exploring Discover More Topics From NASA
      Missions
      Humans in Space
      Climate Change
      Solar System
      View the full article
    • By European Space Agency
      Image: This Copernicus Sentinel-2 image captures the borders between North and South Dakota and Minnesota blanketed with snow and ice. View the full article
    • By NASA
      NASA, ESA, and M. Wong (University of California – Berkeley); Processing: Gladys Kober (NASA/Catholic University of America) This NASA Hubble Space Telescope image shows the planet Jupiter in a color composite of ultraviolet wavelengths. Released on Nov. 3, 2023, in honor of Jupiter reaching opposition, which occurs when the planet and the Sun are in opposite sides of the sky, this view of the gas giant planet includes the iconic, massive storm called the “Great Red Spot.” Though the storm appears red to the human eye, in this ultraviolet image it appears darker because high altitude haze particles absorb light at these wavelengths. The reddish, wavy polar hazes are absorbing slightly less of this light due to differences in either particle size, composition, or altitude.
      Learn more about Hubble and how this type of data can help us learn more about our universe.
      Image credit: NASA, ESA, and M. Wong (University of California – Berkeley); Processing: Gladys Kober (NASA/Catholic University of America)
      View the full article
    • By Space Force
      The Space Force Year in Review is a collection of photos that showcase the men and women of the U.S. Space Force participating in activities and operations necessary to develop and protect U.S. interests in space.

      View the full article
    • By Space Force
      The Space Force Year in Review is a collection of photos that showcase the men and women of the U.S. Space Force participating in activities and operations necessary to develop and protect U.S. interests in space.

      View the full article
  • Check out these Videos

×
×
  • Create New...