Can We Trust LLMs for Complex Earth System Model Analysis? Silent Failure and Evidence from Module-Grounded Benchmarking

EGUsphere (preprint)

Large language models (LLMs) are becoming increasingly capable of complex scientific scripting, but this growing robustness creates a paradox: the more trustworthy their outputs appear, the more easily scientifically incorrect results can pass unnoticed.

Authors: Zhou, T., Qian, Y., Leung, L. R.

DOI: 10.5194/egusphere-2026-2237

Leave a Comment