Foundation Models
Copernicus Foundation Models
These models are trained on some of the largest and most diverse satellite datasets ever assembled, such as SSL4EO-S12-ML, SSL4EO-S, Kuro Siwo, and FoMo-Bench. Together, these datasets span tens of terabytes and cover a wide variety of modalities and use cases, including multispectral imagery, SAR data, climate grids, and detailed forest inventories.
Using innovative self-supervised and cross-modal learning techniques, the project has developed several cutting-edge model families:
-
SoftCon and SupCon enhance representation learning using multi-label land cover annotations.
-
FoMo-Net enables training across multiple remote sensing modalities with a unified architecture.
-
MindTheModalityGap aligns satellite imagery with natural language using vision-language models like CLIP, opening the door to zero-shot EO classification.
By leveraging these models, ThinkingEarth enables more accurate, scalable, and accessible AI tools for environmental monitoring, climate action, and global sustainability challenges.
Data sets and benchmarks
Foundations Models
Pushing further, work on remote sensing vision-language models adapts the CLIP framework to EO, allowing AI to connect images with natural language in new ways — and enabling powerful zero-shot classification. These advancements mark a major leap toward flexible, intelligent systems that can help us monitor and understand our planet at scale.