Can Geometric Deep Learning lead eliminate the need of "Brute Force" pre-training [D]

Reddit r/MachineLearning / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The post argues that Geometric Deep Learning (using grids, graphs, groups, and manifolds) may reduce reliance on brute-force large-scale pretraining by encoding symmetries directly into model architectures.
  • It contrasts this with conventional deep learning, which often must learn invariances such as rotation and permutation purely from massive data and compute.
  • The author questions whether, when a model is architecturally unable to break certain symmetries, it would need far fewer examples to learn the corresponding invariances.
  • A central question raised is whether large pretraining datasets are primarily compensating for missing inductive biases, and whether “getting the geometry right” could lower the need for huge datasets.
  • The author seeks clarification from experts, noting they have not yet deeply reviewed recent GDL advancements and may be missing foundational details.

I’ve been reading about Geometric Deep Learning lately (the whole grids, graphs, groups, manifolds idea), and something clicked that i wanted to get a clarity on, i don't think i'm an expert at GDL or anything mentioned here, so i can most definitely be wrong at a fundamental level as well,

A lot of modern deep learning feels like we're throwing massive data and compute and we just hope the model learns the right invariances.

But doesn't GDL kind of flips that?

Instead of learning invariances (like rotation, permutation, etc.), you can build them directly into the architecture using symmetry and geometry. So it got me wondering, if a model literally cannot break a symmetry (like confusing a rotated cat for something else), does it even need tons of examples to learn that, Like why show it 10,000 rotated cats if rotation invariance is already guaranteed?

Which leads to a bigger question:

Are we doing massive-scale pretraining mostly because our architectures are missing the right inductive biases, And if we get the geometry right, does the need for huge datasets actually go down?

it feels like a shift from learning everything from the data to encode what must be true, learn the rest to me

still haven't read the recent advancements in GDL to comment enough, thought i should ask experts here

submitted by /u/Amdidev317
[link] [comments]