Satellite-to-Street: Synthesizing Post-Disaster Views from Satellite Imagery via Generative Vision Models
arXiv cs.CV / 3/24/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes “Satellite-to-Street View” synthesis to generate ground-level, post-disaster street perspectives from satellite imagery, aiming to improve situational awareness when ground data is unavailable.
- It introduces two generative strategies: a VLM-guided method and a damage-sensitive Mixture-of-Experts (MoE) approach, designed to better align generated views with real disaster conditions.
- The authors benchmark their methods against general-purpose baselines like Pix2Pix and ControlNet using a new Structure-Aware Evaluation Framework combining pixel quality, ResNet-based semantic consistency, and a VLM-as-a-Judge perceptual alignment step.
- Experiments on 300 disaster scenarios show a realism–fidelity trade-off: diffusion/control methods can look realistic but may hallucinate structural details that are critical for reliable damage assessment.
- Quantitatively, ControlNet attains the best semantic accuracy (0.71), while VLM-enhanced and MoE approaches tend to produce more texturally plausible outputs at the cost of semantic clarity.
Related Articles

Black Hat Asia
AI Business

"The Agent Didn't Decide Wrong. The Instructions Were Conflicting — and Nobody Noticed."
Dev.to
Top 5 LLM Gateway Alternatives After the LiteLLM Supply Chain Attack
Dev.to

Stop Counting Prompts — Start Reflecting on AI Fluency
Dev.to

Reliable Function Calling in Deeply Recursive Union Types: Fixing Qwen Models' Double-Stringify Bug
Dev.to