Recent Advances in AI/ML with Applications in Earth Sciences and Hydrology

Listen here on Spotify | Listen here on Apple Podcast

Episode released on February 12, 2026 
Episode recorded on January 12, 2026


Alex SunAlex Sun discusses the evolution of AI/ML over the past decades and resources for hydrology. 

Alex Sun is a Data Scientist at the Department of Energy National Energy Technology Lab and previously worked as a Senior Research Scientist at the University of Texas at Austin, focusing on applications of AI/ML to solve hydrologic problems using remote sensing, modeling, and monitoring data at global to regional scales. 

Highlights | Transcript

General Applications and Terminology

  1. Artificial intelligence and machine learning are pervasive in our work and lives, as exemplified by self-driving vehicles that combine sensors (LiDAR, cameras, GPS), deep learning, and real-time decision-making to replicate human driving behavior (TeslaWaymo).
  2. Terminology:
    1. Artificial Intelligence: generic term referring to the use of computers to imitate human-like behaviors.
    2. Machine Learning: subset of AI that trains models to learn or reproduce human thinking to solve real-world problems.
  3. AI/ML evolved markedly over past decades:
    1. 1990s: statistical and ML era, learning from data: Bayesian networks, Support Vector Machine, decision trees (e.g., Random Forest, etc.)
    2. Early 2000s: neural networks and deep learning
    3. 2010s on: deep learning breakthroughs: Convolutional Neural Networks (CNNs)
    4. Mid 2010s: word embeddings, Long-short term memory (LSTM) renaissance in hydrology (originally introduced in 1997, Hochreiter and Schmidhuber, Neural Comp., 1997)
    5. 2017 – 2020: 2017: Transformer models. Attention is all you Need paper by Vaswani et al., 2017, Attention based transformers formed main support for language, vision, and multimodal models.
    6. 2020s: Multimodal AI (text and image); ChatGPT (Chat Generative Pre-trained Transformer), conversational AI
    7. 2024 onwards: Agentic, Multimodal, and Infrastructure AI: Attention (plan, use tools, act), Multimodal models (text, image, social media), Foundation models (domain specific)
  4. Deep learning advanced substantially in early 2010s: e.g., convolutional neural networks (CNNs) (Goodfellow et al., Deep Learning, MIT Press, 2016).
    1. CNNs allowed automated extraction of features from images and complex datasets (LeCun et al., Deep Learning, Nature, 2015).
    2. Hierarchical model layers learn compact representations of high-dimensional data, reducing manual predictor selection.
  5. Foundation models power today’s generative AI
    Large-capacity models trained on massive datasets enable “emergent behavior” of the foundation models.
  6. Classic machine learning models are still highly relevant
    1. Methods such as random forests and support vector machines are still widely used. They are particularly useful when data or computing resources are limited.
    2. LSTM models are very valuable for hydrology (Fig. 1)
      Long short-term memory networks mirror hydrologic storage and memory processes, making them well-suited for rainfall–runoff modeling (Kratzert et al., LSTM, HESS, 2022).
  7. AI development is tightly coupled to hardware advances:
    1. Graphics Processing Units (GPU) developed by NVIDIA.
    2. Tensor Processing Units (TPU) by Google.
      GPUs and TPUs are critical for training large-scale models.
  8. Latent representations enable generalization across regions. High quality representation learning requires large-volume, cross-domain training data. Deep learning can map heterogeneous systems (e.g., river basins) into a shared low-dimensional space for transferable modeling.
    1. https://www.nature.com/articles/s41467-018-07210-0
  9. Physics-informed ML improves interpretability and realism (Fig. 2). 
    Embedding physical constraints (e.g., conservation laws) into data-driven models helps ensure physically meaningful results (Karniadakis et al., Physics Informed Machine Learning, Nature, 2021).
  10. Ensemble modeling increases ability to sample extreme events and improves uncertainty estimation. Using multiple models trained on the same data often yields more robust predictions than relying on a single “best” algorithm.
  11. AI enhances real-time hazard management by fusing multisensory, multimodal data. 
    Flood forecasting and evacuation planning increasingly rely on AI to integrate sensor data, remote sensing, and social media feeds (e.g., Google, Crisis Response) (Figs. 3 and 4).
  12. Remote sensing provides big data in Earth science, such as Landsat data, MODIS, and Sentinel.
    1. AI/ML helps fuse remote sensing products with different spatial and temporal resolution, e.g., Landsat high spatial low temporal resolution with MODIS high temporal low spatial resolution (NASAs Harmonized Landsat Sentinel 2, HLS) (Fig. 5).
  13. Global foundation models enable applications
    Models such as Google’s DeepMind AlphaEarth create high-resolution (10 m) annual (from 2017) global coverage of land by fusing Landsat 8/9, Sentinel-1 (SAR), Sentine-2 (optical) Earth representations that can be applied for environmental analysis.
  14. AI resource use is a growing concern
    Training and operating large models requires substantial electricity and water, raising sustainability questions for data centers.  10 to 50 medium size large language model queries would use up about a half a liter of water and maybe 50 watt hours of electricity.
  15. Domain expertise is critical for reliable application of AI/ML, including validating AI outputs, interpreting results, providing ethical oversight for AI-generated content and even decisions.

 

Improvement of the runoff simulation during the learning process of the LSTM.
Figure 1. Improvement of the runoff simulation during the learning process of the LSTM. Visualized are the observed discharge and LSTM output after various epochs for the basin 13337000 of the CAMELs data set from 1 October 1983 until 30 September 1986 (Kratzert et al., HESS, 2018).

 

 

Description of physics informed neural network (PINN).
Figure 2. Description of physics informed neural network (PINN). Neural network (NN) u(xtθ) with θ the set of trainable weights and biases b, and σ denotes a nonlinear activation function. Specify the measurement data {xitiui} for and the residual points {xjtj} for the PDE. Specify the loss L in Eq. (3) by summing the weighted losses of the data and PDE. Train the NN to find the best parameters θ* by minimizing the loss L. (Karniadakis et al., Nature, 2021).

 

 

 

Googles Crisis Reponse areas, wildfires, combining FireSat
Figure 3. Googles Crisis Reponse areas, wildfires, combining FireSat (constallation of ! 50 low earth orbiting satellites) provides 5 x 5 m resolution data to detect fires at 2 – 3 acre area (https://sites.research.google/gr/wildfires/firesat/). 

  

Google Crisis Area Flood Hub showing risk level, event present, inundation probability, and inundation history.
Figure 4. Google Crisis Area Flood Hub showing risk level, event present, inundation probability, and inundation history (link). 

 

 

Figure 5. NASA Harmonized Landsat 9 (left hand side) and ESA Sentinel 2 (right hand side) satellites.
Figure 5. NASA Harmonized Landsat 9 (left hand side) and ESA Sentinel 2 (right hand side) satellites (https://hls.gsfc.nasa.gov/).

 

 

Thank you to our podcast partners

Image
BEG
Image
NAE_2023_width150
Image
JSG