AI Navigate

myMNIST: Benchmark of PETNN, KAN, and Classical Deep Learning Models for Burmese Handwritten Digit Recognition

arXiv cs.CL / 3/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents the first systematic benchmark on myMNIST (BHDD) evaluating 11 architectures across classical DL models, FastKAN, EfficientKAN, an energy-based model (JEM), and PETNN variants, to establish baselines for Burmese handwritten digit recognition.
  • The CNN baseline achieves the best overall performance with F1 = 0.9959 and Accuracy = 0.9970, setting a strong reference for this dataset.
  • PETNN variants (GELU) closely follow with F1 = 0.9955 and Accuracy = 0.9966, outperforming LSTM, GRU, Transformer, and KAN variants in this benchmark.
  • JEM, representing energy-based modeling, is competitive with F1 = 0.9944 and Accuracy = 0.9958, demonstrating viability of energy-inspired approaches on regional scripts.
  • The study provides reproducible baselines, highlights PETNN’s strong performance relative to classical and Transformer-based models, and releases the benchmark to foster future research on Myanmar script recognition.

Abstract

We present the first systematic benchmark on myMNIST (formerly BHDD), a publicly available Burmese handwritten digit dataset important for Myanmar NLP/AI research. We evaluate eleven architectures spanning classical deep learning models (Multi-Layer Perceptron, Convolutional Neural Network, Long Short-Term Memory, Gated Recurrent Unit, Transformer), recent alternatives (FastKAN, EfficientKAN), an energy-based model (JEM), and physics-inspired PETNN variants (Sigmoid, GELU, SiLU). Using Precision, Recall, F1-Score, and Accuracy as evaluation metrics, our results show that the CNN remains a strong baseline, achieving the best overall scores (F1 = 0.9959, Accuracy = 0.9970). The PETNN (GELU) model closely follows (F1 = 0.9955, Accuracy = 0.9966), outperforming LSTM, GRU, Transformer, and KAN variants. JEM, representing energy-based modeling, performs competitively (F1 = 0.9944, Accuracy = 0.9958). KAN-based models (FastKAN, EfficientKAN) trail the top performers but provide a meaningful alternative baseline (Accuracy ~0.992). These findings (i) establish reproducible baselines for myMNIST across diverse modeling paradigms, (ii) highlight PETNN's strong performance relative to classical and Transformer-based models, and (iii) quantify the gap between energy-inspired PETNNs and a true energy-based model (JEM). We release this benchmark to facilitate future research on Myanmar digit recognition and to encourage broader evaluation of emerging architectures on regional scripts.