Machine Learning

Foundation Model

A large-scale AI model pre-trained on broad data that can be adapted to a wide range of downstream tasks. In geospatial AI, foundation models trained on massive satellite imagery datasets can perform tasks like land cover classification, change detection, and feature extraction with minimal task-specific training.

Overview

A foundation model in earth observation is a large neural network pre-trained on massive unlabeled satellite imagery using self-supervised learning, then adapted to downstream tasks with minimal labeled data. This represents a paradigm shift from training task-specific models from scratch — a single model learns general-purpose Earth surface representations applicable to land cover mapping, change detection, crop classification, and flood monitoring.

How It Works

EO foundation models typically use Vision Transformer (ViT) architectures trained with Masked Autoencoding (MAE) — learning to reconstruct randomly masked patches of satellite imagery. Prithvi-EO-2.0 (NASA/IBM, 600M parameters) was pre-trained on 4.2 million HLS samples spanning seven years. The Clay Foundation Model uses MAE and DINO objectives with location/time metadata. Once pre-trained, models can be fine-tuned with as few as 100 labeled samples.

Key Facts

  • Prithvi-EO-2.0 has 600M parameters pre-trained on 4.2 million HLS samples.
  • Clay Foundation Model is fully open source with permissive licenses.
  • Self-supervised pre-training requires no labeled data during initial training.
  • Outperforms task-specific models on GEO-Bench benchmarks.

Applications

Multi-Task Analysis

A single model fine-tuned for classification, change detection, and crop monitoring — replacing dozens of specialized models.

Rapid Disaster Response

Quickly adapted to new disaster scenarios with minimal labeled data.

Global Environmental Monitoring

Generalizing across geographic regions without region-specific retraining.

Few-Shot Inference

Producing useful predictions for novel tasks with little or no labeled data.

Limitations & Considerations

Require enormous computational resources for pre-training. Effectiveness on SAR and hyperspectral data still being validated. Evaluation frameworks are still maturing. The black-box nature can be a barrier for regulatory applications.

History & Background

Emerged 2022-2023 following success of BERT/GPT in NLP. IBM/NASA released Prithvi-EO-1.0 in August 2023. Prithvi-EO-2.0 released December 2024 with AGU Open Science Prize. STAC and COG enable the data pipelines feeding these models.

Analyze Foundation Model data with LYRASENSE

Use our agentic notebook environment to work with satellite data and compute indices like Foundation Model — no setup required.