Earthly Machine Learning

Amirpasha

“Earthly Machine Learning (EML)” offers AI-generated insights into cutting-edge machine learning research in weather and climate sciences. Powered by Google NotebookLM, each episode distils the essence of a standout paper, helping you decide if it’s worth a deeper look. Stay updated on the ML innovations shaping our understanding of Earth. It may contain hallucinations.

  1. قبل ٢٣ ساعة

    Machine learning for the physics of climate

    Machine learning for the physics of climate Citation: Bracco, A., Brajard, J., Dijkstra, H. A., Hassanzadeh, P., Lessig, C., & Monteleoni, C. (2025). Machine learning for the physics of climate. Nature Reviews Physics, 7, 6–20. https://doi.org/10.1038/s42254-024-00776-3 Main Takeaways: Breaking the El Niño Spring Barrier: For decades, forecasts of the El Niño Southern Oscillation hit a hard wall at roughly 6 months lead time — a limit known as the spring predictability barrier. Convolutional neural networks trained on a mix of climate model and reanalysis data have shattered this ceiling, delivering skillful forecasts at 17 months out, with newer architectures pushing to 21–24 months. ML models can also now anticipate which type of El Niño will develop (eastern vs. central Pacific), which matters enormously because the two flavors produce very different regional impacts around the world.Weather Forecasting at a Fraction of the Cost: A new generation of ML weather emulators — Pangu-Weather, GraphCast, FourCastNet, FuXi, NeuralGCM — now match or beat the European Centre's flagship physics-based forecasting system on most variables, including hurricane tracks, while running orders of magnitude faster. They achieve this with surprisingly compressed state representations: roughly 10 vertical atmospheric levels and 0.25° horizontal resolution, compared to 100+ levels and 0.1° in conventional models. The catch is that these models can violate basic physics — geostrophic balance, energy conservation, the butterfly effect — which currently blocks naive extension to climate timescales.Hybrid Models Are Eating the Climate Stack: Pure ML works for short-range forecasts, but for climate-length runs the field is converging on hybrid architectures that pair a traditional dynamical core with neural-network parameterizations of sub-grid processes like clouds, turbulence, and gravity waves. Google's NeuralGCM exemplifies the approach and already reduces biases in tropical cyclone frequency and tracks. A telling case study on the quasi-biennial oscillation showed that an offline-trained neural network produced unstable, unphysical results — but retraining just two layers online, coupled to the model, recovered the correct physics. Offline-only or online-only training each fail in characteristic ways; the mix is what works.The Data Wall Is the Real Bottleneck: Climate ML has less than 50 years of dense satellite-era observations to work with, and those observations are heavily biased toward the atmosphere and ocean surface — a single, spatiotemporally correlated realization of one climate. This limits how confidently ML models can extrapolate to warmer, unseen climates, which is exactly what climate projection requires. The path forward involves three parallel bets: hybrid physics-ML models that bake in conservation laws, large-scale "foundation models" for weather and climate trained across simulations and observations together (efforts like ClimaX and AtmoRep are early examples), and rare-event sampling strategies to handle the extremes that matter most for adaptation policy but are by definition underrepresented in any training set.

    ٢٠ د
  2. ٢٧ أبريل

    Atmospheric Transport Modeling of CO2 With Neural Networks

    Citation: Benson, V., Bastos, A., Reimers, C., Winkler, A. J., Yang, F., & Reichstein, M. (2025). Atmospheric transport modeling of CO2 with neural networks. Journal of Advances in Modeling Earth Systems, 17, e2024MS004655. https://doi.org/10.1029/2024MS004655 Main Takeaways: A New Benchmark for AI Carbon Tracking: The authors introduce CarbonBench, the first systematic benchmark dataset designed specifically for training and evaluating machine learning emulators of Eulerian atmospheric transport. Built from CarbonTracker CT2022 inversions and ObsPack station observations, it ships at three resolutions (the coarsest being 5.625° × 10 vertical levels × 6h) and is engineered to plug directly into modern deep learning pipelines — opening atmospheric carbon modeling to the broader ML community.SwinTransformer Wins, Decisively: Of the four architectures tested (UNet, GraphCast, SFNO, and SwinTransformer), the SwinTransformer reaches near-perfect emulation with a 90-day R² above 0.99 and stays stable in physically plausible forward runs for over three years — a regime where neural PDE solvers typically blow up. At measurement stations, it actually captures the seasonal cycle in Svalbard better than TM5, the conventional model it was trained to emulate, possibly due to differences in boundary layer transport near the poles.Physics Tricks Were the Unlock: Out of the box, the neural networks were unstable — especially the mesh-based UNet and GraphCast. Two simple physics-aware adjustments fixed this across all four architectures: centering the CO2 input field at each timestep to remove the covariate shift from steadily rising atmospheric CO2 (called CentFlux), and a post-hoc mass fixer that rescales predicted mass to match the surface flux budget. The result is mass conservation with RMSE of just 0.00058 PgC against a total atmospheric carbon mass of ~865 PgC — effectively negligible.Speed Isn't the Selling Point (Yet): Unlike AI weather models, which famously outpace numerical forecasting by orders of magnitude, the SwinTransformer is not significantly faster than TM5 at this resolution — about 1.5 seconds for a 30-day run on an A40 GPU versus a few minutes for TM5 on 24 CPUs. The real promise lies elsewhere: the networks are fully differentiable (useful for inverse modeling of surface fluxes), natively support batched ensembles, and scale better to high resolution where conventional solvers become prohibitively expensive — exactly the regime where current CO2 inversions struggle most.

    ٢١ د
  3. ٢٠ أبريل

    On the foundations of Earth foundation models

    Citation: Zhu, X. X., Xiong, Z., Wang, Y., Stewart, A. J., Heidler, K., Wang, Y., Yuan, Z., Dujardin, T., Xu, Q., & Shi, Y. (2026). On the foundations of Earth foundation models. Communications Earth & Environment, 7, 103. https://doi.org/10.1038/s43247-025-03127-x Main Takeaways: Current Earth AI Models Are Missing the Point: Researchers have identified eleven features that an ideal Earth foundation model must have — including geolocation awareness, multi-sensor integration, physical consistency, and carbon minimization — yet no existing model comes close to checking all eleven boxes. Most models focus on only one or two features, leaving a major gap between what we have and what we actually need to tackle real-world climate and environmental challenges. The Data Situation Is More Lopsided Than You'd Think: There are now over 1,000 active remote sensing satellites generating nearly 100 petabytes of open satellite data — but labeled datasets used to train AI models account for less than 0.1% of that archive. This massive imbalance is precisely why self-supervised foundation models, which can learn from unlabeled data, are so critical for Earth science going forward. Weather AI Is Already Dramatically More Efficient — But Incomplete: Models like FourCastNet can generate a week-long global weather forecast in under two seconds on a single GPU, using roughly 12,000 times less energy than traditional forecasting systems. Despite this leap in efficiency, major gaps remain: models struggle beyond two-week forecasts, long-term climate projections drift due to incomplete energy balance, and connecting fine-scale satellite imagery with coarse climate models remains largely unsolved. What Comes After the Ideal Model: Once a true Earth foundation model exists, the authors argue the most exciting frontier is using it to build an "Earth Embedding" — a compact, unified representation of our entire planet that researchers worldwide could query without ever touching raw satellite data. Beyond that, challenges like machine unlearning (making models forget sensitive imagery), adversarial defenses, and continual learning as the climate itself changes will define the next generation of Earth AI research.

    ١٨ د
  4. ١٤ أبريل

    Whose weather is it? A fairness framework for data-driven weather forecasting

    Citation: Olivetti, L., & Messori, G. (2025). Whose weather is it? A fairness framework for data-driven weather forecasting. Environmental Research Letters, 20, 121006. https://doi.org/10.1088/1748-9326/ae21f5 Main Takeaways: AI Weather Models Aren't Fair to Everyone: The latest generation of AI-powered weather forecasts improves predictions globally — but not equally. Using ECMWF's AIFS model as a case study, the authors show that wealthier and more densely populated areas consistently receive a higher share of forecast improvements compared to poorer and more rural regions, violating basic fairness criteria borrowed from the algorithmic fairness literature. Two Measurable Fairness Tests — Both Failed: The paper proposes two concrete criteria: statistical parity(improvement rates should be similar across income groups) and conditional independence (a region's GDP or population density should not predict whether it benefits from the new model). Across nearly all tested variables and forecast lead times, AIFS fails both tests at the 0.01 significance level — meaning the disparity is not a statistical fluke. Extreme Weather Is Where the Gap Hurts Most: For standard temperature and wind forecasts, gaps between rich and poor regions are modest. But for cold extremes, the fairness gap is especially pronounced — precisely the events where accurate early warnings matter most for vulnerable populations with fewer resources to adapt. Fixing It Is Technically Feasible: Unlike traditional physics-based models, AI weather models offer genuine design levers for fairness. The authors describe two practical approaches: adding penalty terms to the loss function (such as the Hilbert–Schmidt Independence Criterion) to reduce associations with protected variables, and using geographically adaptive weighting that iteratively compensates for emerging performance gaps — without necessarily sacrificing global accuracy.

    ٢٢ د
  5. ٧ مارس

    Learning predictable and informative dynamical drivers of extreme precipitation using variational autoencoders

    Citation: Spuler, F. R., Kretschmer, M., Balmaseda, M. A., Kovalchuk, Y., & Shepherd, T. G. (2025). Learning predictable and informative dynamical drivers of extreme precipitation using variational autoencoders. Weather and Climate Dynamics, 6, 995–1014. https://doi.org/10.5194/wcd-6-995-2025 Main Takeaways: Innovative Machine Learning Approach: The study introduces the Categorical Mixture Model Variational Autoencoder (CMM-VAE), a novel generative machine learning method designed to identify probabilistic atmospheric circulation regimes by combining targeted dimensionality reduction and probabilistic clustering into a single model.Resolving a Major Forecasting Trade-off: Traditionally, atmospheric regimes are either highly predictable globally but locally uninformative, or highly informative for local impacts but lacking in subseasonal predictability. CMM-VAE resolves this trade-off, successfully identifying patterns that predict local extremes without sacrificing forecast skill at subseasonal lead times.Targeted Application for Moroccan Rainfall: When applied to extreme winter precipitation in Morocco, the CMM-VAE method successfully disentangled a distinct, highly impactful weather pattern—a Scandinavian blocking coupled with a localized cut-off low—that traditional linear clustering methods failed to isolate.Linkages to Global Climate Drivers: The weather regimes identified by the model remain physically interpretable and show clear, predictable teleconnections to large-scale, low-frequency climate drivers, notably the Madden-Julian Oscillation (MJO) and the Stratospheric Polar Vortex (SPV).Enhancing Early Warning Systems: By providing a better representation of regional dynamical drivers, this framework offers significant potential to improve subseasonal-to-seasonal (S2S) forecasts, statistical downscaling, and early-warning systems for severe, localized weather impacts.

    ١٨ د
  6. ٢٨ فبراير

    Green and intelligent: the role of AI in the climate transition

    Green and intelligent: the role of AI in the climate transition Citation: Stern, N., Romani, M., Pierfederici, R., Braun, M., Barraclough, D., Lingeswaran, S., Weirich-Benet, E., & Niemann, N. (2025). Green and intelligent: the role of AI in the climate transition. https://doi.org/10.1038/s44168-025-00252-3. Main Takeaways: Five Key Areas for Climate Action: Artificial Intelligence can accelerate the net-zero transition across five primary avenues: transforming complex economic systems, innovating technology discovery and resource efficiency, nudging consumer behavior toward sustainable choices, modeling climate systems for better policy, and managing adaptation and resilience.Significant Emissions Reduction Potential: By applying AI to just three major sectors—power, food (specifically meat and dairy), and mobility (light road vehicles)—global emissions could be reduced by 3.2 to 5.4 GtCO2e annually by 2035.Net-Positive Climate Impact: The emissions savings generated by AI in these three sectors alone would more than offset the projected 0.4 to 1.6 GtCO2e increase in emissions caused by the energy consumption of all global AI activities and data centers.Closing the Emissions Gap: Harnessing AI to improve the efficiency and market adoption of low-carbon solutions could push global progress 36% closer to aligning with an ambitious emissions reduction trajectory by 2035.The Critical Role of Government: Relying solely on market forces to govern AI is risky; an "active state" is essential to direct AI toward public goods, regulate its environmental footprint (like mandating renewable energy for data centers), and ensure equitable deployment so the Global South is not left behind.

    ١٨ د
  7. ٢٦ يناير

    Climate Knowledge in Large Language Models

    Climate Knowledge in Large Language Models Kuznetsov, I., Grassi, J., Pantiukhin, D., Shapkin, B., Jung, T., & Koldunov, N. (2025). Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research. LLMs have an internal "map" of the climate, but it is fuzzy: Without access to external tools, Large Language Models (LLMs) can recall the general structure of Earth’s climate—correctly identifying that the tropics are warm and high latitudes are cold. However, their specific numeric predictions are often inaccurate, with average errors ranging from 3°C to 6°C compared to historical weather data. Location names matter more than coordinates: The study found that providing geographic context—such as the country, region, or city name—alongside coordinates reduced prediction errors by an average of 27%. This suggests models rely heavily on text associations with place names rather than possessing a precise spatial understanding of latitude and longitude. Performance struggles with altitude and local trends: Models perform significantly worse in mountainous regions, with errors spiking sharply at elevations above 1500 meters. Furthermore, while LLMs can estimate the global average magnitude of warming, they fail to accurately reproduce the specific local patterns of temperature change that are essential for understanding regional climate dynamics. Caution is needed for scientific use: The results highlight that while LLMs encode a static snapshot of climatological averages, they lack true physical understanding and struggle with dynamic trends. Consequently, they should not be relied upon as standalone climate databases; reliable applications require connecting them to external, authoritative data sources.

    ١٢ د
  8. ١١ يناير

    Artificial Intelligence for Atmospheric Sciences: A Research Roadmap

    Artificial Intelligence for Atmospheric Sciences: A Research Roadmap Citation: Zaidan, M. A., Motlagh, N. H., Nurmi, P., Hussein, T., Kulmala, M., Petäjä, T., & Tarkoma, S. (2025). Artificial Intelligence for Atmospheric Sciences: A Research Roadmap. Revolutionizing Environmental Monitoring: The paper illustrates how AI is transforming atmospheric sciences by bridging the gap between computer science and environmental research. It details how AI processes massive datasets generated by diverse sources—including satellite imagery, ground-based research stations, and low-cost IoT sensors—to improve our understanding of air quality, extreme weather events, and climate change. Optimizing Infrastructure and Prediction: Current AI applications are already enhancing operational meteorology and Earth system modeling. By utilizing techniques like deep learning and neural networks, researchers can automate sensor calibration, detect anomalies in real-time, and simulate complex climate scenarios with greater speed and efficiency than traditional physical models allow. A Roadmap for Future Hardware: To handle the escalating demand for data, the authors propose a hardware roadmap that includes self-sustaining and biodegradable sensor networks, CubeSat constellations for high-resolution monitoring, and the adoption of cutting-edge computing paradigms like quantum, neuromorphic, and DNA-based molecular computing. Next-Generation AI Methodologies: The paper argues for the adoption of advanced AI techniques such as Foundation Models and Generative AI (including Digital Twins of Earth) to predict complex atmospheric phenomena. Crucially, it emphasizes the need for Explainable AI (XAI) and Physics-Informed Machine Learning to solve the "black box" problem, ensuring that AI predictions abide by physical laws and are transparent enough for scientists and policymakers to trust. From Data to Action: Beyond observation, the research highlights the shift toward actionable insights. This includes automated feedback loops (such as smart HVAC systems responding to air quality data), the integration of citizen science to augment data collection, and the establishment of robust ethical frameworks to manage data privacy and governance in global monitoring networks.

    ١٤ د

حول

“Earthly Machine Learning (EML)” offers AI-generated insights into cutting-edge machine learning research in weather and climate sciences. Powered by Google NotebookLM, each episode distils the essence of a standout paper, helping you decide if it’s worth a deeper look. Stay updated on the ML innovations shaping our understanding of Earth. It may contain hallucinations.