Learning Library

← Back to Library

NASA’s Geospatial Foundation Model

Key Points

  • Foundation models are large‑scale neural networks pretrained on massive datasets that can transfer learned knowledge to new tasks through fine‑tuning with relatively few labeled examples.
  • NASA archives roughly 70 PB of Earth‑science satellite imagery (projected to hit ~300 PB by 2030), providing an unparalleled reservoir of data for climate‑related research.
  • In partnership with IBM, NASA released the open‑source “IBM NASA Geospatial” foundation model on Hugging Face, which leverages transformer architecture to compress raw satellite images into useful representations for many downstream tasks.
  • By extracting structure from raw imagery, the geospatial foundation model dramatically reduces the need for time‑intensive human annotation, speeding up analysis of crops, forests, and other land‑cover features.

Full Transcript

# NASA’s Geospatial Foundation Model **Source:** [https://www.youtube.com/watch?v=QPQy7jUpmyA](https://www.youtube.com/watch?v=QPQy7jUpmyA) **Duration:** 00:05:13 ## Summary - Foundation models are large‑scale neural networks pretrained on massive datasets that can transfer learned knowledge to new tasks through fine‑tuning with relatively few labeled examples. - NASA archives roughly 70 PB of Earth‑science satellite imagery (projected to hit ~300 PB by 2030), providing an unparalleled reservoir of data for climate‑related research. - In partnership with IBM, NASA released the open‑source “IBM NASA Geospatial” foundation model on Hugging Face, which leverages transformer architecture to compress raw satellite images into useful representations for many downstream tasks. - By extracting structure from raw imagery, the geospatial foundation model dramatically reduces the need for time‑intensive human annotation, speeding up analysis of crops, forests, and other land‑cover features. ## Sections - [00:00:00](https://www.youtube.com/watch?v=QPQy7jUpmyA&t=0s) **NASA Data Fuels Foundation Models** - The speaker defines foundation models, highlights the thousands of open‑source versions on Huggingface, and proposes NASA’s massive Earth‑science dataset as a valuable resource for training and fine‑tuning these models. - [00:03:08](https://www.youtube.com/watch?v=QPQy7jUpmyA&t=188s) **Foundation Models Transform Satellite Analysis** - The speaker explains how foundation models automate satellite image labeling, enable fine‑tuned flood and wildfire mapping, and can be repurposed for tasks like deforestation tracking, crop‑yield prediction, and greenhouse‑gas monitoring, dramatically expanding the utility of NASA Earth‑science data. ## Full Transcript
0:00If you head over to Huggingface, you will find literally thousands of foundation models available for download. 0:06And that's just the open source ones. 0:08So this does beg the question, why are there so many foundation models? 0:14Well, to help answer that, we're going to look to NASA. 0:19But first, we should probably define what a foundation model actually is. 0:23And look, I have a whole video on that topic. 0:25So for now, let's just say that foundation models, which is what this represents here, 0:30are large scale neural networks trained on vast amounts of data, 0:35and they serve as a base or a "foundation" for a multitude of applications. 0:40And a foundation model can apply information. 0:43It's learned about one situation to a different situation it was not trained on. And we call that transfer learning. 0:53Pre-trained a foundation model, and you can teach it an entirely new task with a limited set of hand-labeled examples. 1:01So if we pick a foundation model that has ingested the right data and we provide the right fine tuning, 1:07we can put it to work in our own specific applications. 1:11Which brings us to NASA. 1:14If you're looking for huge amounts of data, well, look no further than NASA. 1:18Today we are sitting on about 70 petabytes of earth science data captured from satellite images, 1:27which which sounds like a lot, 1:29but by 2030, with the launch of a dozen or so new space missions, 1:33that number is expected to be closer to 300 petabytes of data. 1:40So we have a vast, vast amount of data. 1:45and we may be able to use that to provide insights to, well, all sorts of climate-related discoveries. 1:53But how can we possibly utilize it? 1:55Well, through a foundation model, of course! 1:59Now, for the last six months, NASA has been working with IBM to create an AI foundation model for Earth observations. 2:05And now you and, well, anybody who wants it can download the whole thing. 2:11It's called the "IBM NASA Geospatial model," 2:14and it's an open source model available on Huggingface. 2:18And this geospatial foundation model really does help us to answer the question of why are there so many foundation models? 2:25Look, underpinning all foundation models is the concept of a transformer. 2:33That's an AI architecture that can turn heaps of raw data, be that text or audio, or in this case, 2:39satellite images into a compressed representation that captures the data's basic structure. That represents this. 2:50And then we can use this with a foundation model for a wide variety of tasks with some extra label data and tuning. 2:57Now, look, 2:58traditionally, analyzing satellite data like this has been a tedious process 3:03because of the time required for human experts to annotate features. 3:08So in each satellite image we label-- let's see this group of pixels, that's crops 3:14--and then we would say this group of pixels. 3:16Yeah, that's trees and so forth. 3:19And a human is having to go through this takes a lot of time. 3:23So foundation models can cut out a lot of this manual effort by extracting the structure of raw natural images 3:29so that fewer label examples are needed. 3:32Then the foundation model has been fine tuned to allow users to map the extent of past US floods and wildfires. 3:41Why do that? 3:42Because these measurements then can be used to predict future areas of risk. 3:47So we have a flood and wildfire prediction model. 3:52Pretty cool. 3:53But look, foundation models are well, foundational. 3:57We can take that model and apply our own fine tuning to build upon the model to perform different tasks entirely. 4:04So with additional fine tuning, our flood and wildfire prediction model 4:10can be redeployed for tasks like tracking deforestation or predicting crop yields, 4:16or even looking at detecting and monitoring greenhouse gases. 4:21In fact, Clark University are adapting this very model for other applications, 4:25including time series segmentation and similarity search. 4:29So in this case, foundation models are multiplying the usefulness of NASA data 4:35where fine tuning can adapt these models to new use cases. 4:39And look, that's just NASA Earth science data. 4:42Those thousands of open source foundation models that I mentioned at the beginning, 4:46they are trained and tuned on a wide variety of other data, 4:51like code generation or foundation models related to a specific industry. 4:56So by selecting the right foundation model and adapting it, 5:01we can put that model to work in new ways to meet our needs. 5:05And that is why there are so many foundation models available 5:09and why there are so many more to come.