AI image-based method for a robust automatic real-time water level monitoring: a long-term application case

Other authors

Universitat Politècnica de Catalunya. Departament d'Enginyeria Civil i Ambiental

Universitat Politècnica de Catalunya. Geo2Aqua - Monitoring, modelling and geomatics for hydro-geomorphological processes

Publication date

2026-02-12



Abstract

The study presents a robust, automated camera gauge for long-term river water level monitoring operating in near real-time. The system employs artificial intelligence (AI) for the image-based segmentation of water bodies and the identification of ground control points (GCPs), combined with photogrammetric techniques, to determine water levels from surveillance camera data acquired every 15¿min. The method was tested at four locations over a period of more than 2.5 years. During this period almost 218¿000 images were processed. The results demonstrate a high performance, with mean absolute errors ranging from 0.96 to 2.66¿cm in comparison to official gauge references. The camera gauge demonstrates resilience to adverse weather and lighting conditions, achieving an image utilisation rate of above 95¿% throughout the entire period. The integration of infrared illumination enabled 24/7 monitoring capabilities. Key factors influencing absolute error were identified as camera calibration, GCP stability, and vegetation changes. The low-cost, non-invasive approach advances hydrological monitoring capabilities, particularly for flood detection and mitigation in ungauged or remote areas, enhancing image-based techniques for robust, long-term environmental monitoring with frequent, near real-time updates.


Peer Reviewed


Postprint (published version)

Document Type

Article

Language

English

Publisher

European Geosciences Union (EGU)

Related items

https://hess.copernicus.org/articles/30/797/2026/

Recommended citation

This citation was generated automatically.

Rights

http://creativecommons.org/licenses/by/4.0/

Open Access

Attribution 4.0 International

This item appears in the following Collection(s)

E-prints [72399]