- [ONLINE] H13K-VR8836: Diffusion-Based Convolutional Learning and Multi-View Data Fusion for Flood Inundation Mapping
-
OnlineOnline
Author(s):Generic 'disconnected' Message
Ankita Agrawal, Indian Institute of Technology Roorkee (First Author, Presenting Author)
Balasubramanian Raman, Indian Institute of Technology Roorkee (IITR)
Aparajita Khan, Indian Institute of Technology (BHU) Varanasi
This study presents a new method to improve flood mapping using satellite images from two different sensors: radar (Sentinel-1) and optical (Sentinel-2). Each has its strengths—radar works in cloudy weather, and optical gives clear visual details. By combining them, the method gets more accurate results.The process involves:
- Extracting important features from both types of data.
- Using a special model (called a diffusion U-Net) to reduce noise and improve the accuracy of flood boundaries.
- Refining the final flood map using another U-Net model for precise, pixel-level classification.
To make the model work better, we applied various data augmentation techniques. We tested the approach on well-known datasets and found it performs better than existing methods, especially in tough conditions. The result is a lightweight, robust, and effective system for real-world flood monitoring.
Scientific DisciplineSuggested ItinerariesNeighborhoodTypeWhere to Watch - Extracting important features from both types of data.
Enter Note
Go to previous page in this tab
Session
