AI Navigate

MAcPNN: Mutual Assisted Learning on Data Streams with Temporal Dependence

arXiv cs.LG / 3/11/2026

Signals & Early TrendsModels & Research

Key Points

  • The paper introduces Mutual Assisted Learning, a novel machine learning paradigm for IoT devices based on Vygotsky's Sociocultural Theory, allowing autonomous edge devices to request help from peers when facing concept drift.
  • This paradigm reduces communication overhead compared to traditional Federated Learning by only activating connections when performance degradation occurs, enhancing efficiency.
  • Each device utilizes a Continuous Progressive Neural Network (cPNN) adapted for streaming data with temporal dependence, and model quantization is applied to lower memory usage.
  • Experimental results on both synthetic and real-world data streams demonstrate that MAcPNN effectively improves performance in handling dynamic IoT data analytics scenarios.
  • The approach addresses challenges such as continuous learning, concept drift, temporal dependence, and knowledge reuse across decentralized IoT nodes without a central orchestrator.

Computer Science > Machine Learning

arXiv:2603.08972 (cs)
[Submitted on 9 Mar 2026]

Title:MAcPNN: Mutual Assisted Learning on Data Streams with Temporal Dependence

View a PDF of the paper titled MAcPNN: Mutual Assisted Learning on Data Streams with Temporal Dependence, by Federico Giannini and 1 other authors
View PDF HTML (experimental)
Abstract:Internet of Things (IoT) Analytics often involves applying machine learning (ML) models on data streams. In such scenarios, traditional ML paradigms face obstacles related to continuous learning while dealing with concept drifts, temporal dependence, and avoiding forgetting. Moreover, in IoT, different edge devices build up a network. When learning models on those devices, connecting them could be useful in improving performance and reusing others' knowledge. This work proposes Mutual Assisted Learning, a learning paradigm grounded on Vygotsky's popular Sociocultural Theory of Cognitive Development. Each device is autonomous and does not need a central orchestrator. Whenever it degrades its performance due to a concept drift, it asks for assistance from others and decides whether their knowledge is useful for solving the new problem. This way, the number of connections is drastically reduced compared to the classical Federated Learning approaches, where the devices communicate at each training round. Every device is equipped with a Continuous Progressive Neural Network (cPNN) to handle the dynamic nature of data streams. We call this implementation Mutual Assisted cPNN (MAcPNN). To implement it, we allow cPNNs for single data point predictions and apply quantization to reduce the memory footprint. Experimental results prove the effectiveness of MAcPNN in boosting performance on synthetic and real data streams.
Subjects: Machine Learning (cs.LG)
MSC classes: 68T05, 68T07
ACM classes: I.2.6; I.2.7
Cite as: arXiv:2603.08972 [cs.LG]
  (or arXiv:2603.08972v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2603.08972
Focus to learn more
arXiv-issued DOI via DataCite
Journal reference: Proc. IEEE Big Data 2024, pp. 890-899
Related DOI: https://doi.org/10.1109/BigData62323.2024.10825150
Focus to learn more
DOI(s) linking to related resources

Submission history

From: Federico Giannini [view email]
[v1] Mon, 9 Mar 2026 22:03:37 UTC (994 KB)
Full-text links:

Access Paper:

Current browse context:
cs.LG
< prev   |   next >
Change to browse by:
cs

References & Citations

export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo
Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle
Bibliographic Explorer (What is the Explorer?)
Connected Papers Toggle
Connected Papers (What is Connected Papers?)
Litmaps Toggle
Litmaps (What is Litmaps?)
scite.ai Toggle
scite Smart Citations (What are Smart Citations?)
Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle
alphaXiv (What is alphaXiv?)
Links to Code Toggle
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub Toggle
DagsHub (What is DagsHub?)
GotitPub Toggle
Gotit.pub (What is GotitPub?)
Huggingface Toggle
Hugging Face (What is Huggingface?)
Links to Code Toggle
Papers with Code (What is Papers with Code?)
ScienceCast Toggle
ScienceCast (What is ScienceCast?)
Demos

Demos

Replicate Toggle
Replicate (What is Replicate?)
Spaces Toggle
Hugging Face Spaces (What is Spaces?)
Spaces Toggle
TXYZ.AI (What is TXYZ.AI?)
Related Papers

Recommenders and Search Tools

Link to Influence Flower
Influence Flower (What are Influence Flowers?)
Core recommender toggle
CORE Recommender (What is CORE?)
IArxiv recommender toggle
IArxiv Recommender (What is IArxiv?)
About arXivLabs

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.