DST-Net: A Dual-Stream Transformer with Illumination-Independent Feature Guidance and Multi-Scale Spatial Convolution for Low-Light Image Enhancement
arXiv cs.CV / 3/18/2026
📰 NewsModels & Research
Key Points
- DST-Net presents a Dual-Stream Transformer for low-light image enhancement that leverages illumination-agnostic priors and a multi-scale spatial fusion mechanism to improve quality while preserving details.
- A feature extraction module combines Difference of Gaussians (DoG), LAB color space transformations, and VGG-16 to obtain texture priors that guide the enhancement without destroying intrinsic signal information.
- The dual-stream architecture uses a cross-modal attention mechanism to dynamically rectify degraded signal representations and perform iterative enhancement via differentiable curve estimation.
- The Multi-Scale Spatial Fusion Block (MSFB) employs pseudo-3D and 3D gradient operator convolutions to recover high-frequency edges and capture inter-channel spatial correlations, achieving PSNR of 25.64 dB on the LOL dataset and robust cross-scene generalization on LSRW.
Related Articles
Next-Generation LLM Inference Technology: From Flash-MoE to Gemini Flash-Lite, and Local GPU Utilization
Dev.to
The Wave of Open-Source AI and Investment in Security: Trends from Qwen, MS, and Google
Dev.to
Implementing Deep Q-Learning (DQN) from Scratch Using RLax JAX Haiku and Optax to Train a CartPole Reinforcement Learning Agent
MarkTechPost
[D] Training a classifier entirely in SQL (no iterative optimization)
Reddit r/MachineLearning
LLM failure modes map surprisingly well onto ADHD cognitive science. Six parallels from independent research.
Reddit r/artificial