Hi Everybody! I just wanted to share an update on a project I’ve been working on called BULaMU, a family of language models trained (20M, 47M, and 110M parameters) trained entirely from scratch for a low resource language, Luganda. The models are small and compute-efficient enough to run offline on a phone without requiring a GPU or internet connection. I recently built an Android app called E.A.S.T. (Expanding Access to Systems of Learning and Intelligence) that allows you to interact with the models directly on-device. It is available on my GitHub page. This is part of a broader effort to make artificial intelligence more accessible to speakers of low-resource languages and to people using low-power, low-cost devices.
Demo: https://x.com/mwebazarick/status/2038384599320170760?s=46
GitHub: https://github.com/mwebazarick/EAST
Huggingface: https://huggingface.co/datasets/mwebazarick/BULaMU
Model Whitepaper: https://zenodo.org/records/17271688
[link] [comments]




