| Hi r/LocalLLaMA ! I built a 5M Llama model with HF Transformers on 2x T4 in Kaggle to see, if it is able to be as good as my previous Apex 350M model (https://huggingface.co/LH-Tech-AI/Apex-1.6-Instruct-350M). Link to the research site: https://lh-tech.de/ai/sub-5m-research.html It came out, that if you optimize the model enough and train it on much data it can be nearly as good as a 70 times heavier model (like Apex 350M; GPT-2 architecture). Tell me what you think about it! Spark v5 coming soon... Expect it to be good 😃 [link] [comments] |
I built a 5M model to see if it outperforms my 350M model...
Reddit r/LocalLLaMA / 4/30/2026
💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The author trained a small 5M-parameter Llama model using Hugging Face Transformers on Kaggle with two NVIDIA T4 GPUs to compare against a prior 350M-parameter model.
Related Articles

Black Hat USA
AI Business

Building a Local AI Agent (Part 2): Six UX and UI Design Challenges
Dev.to

The Prompt Caching Mistake That's Costing You 70% More Than You Need to Pay
Dev.to

We Built a DNS-Based Discovery Protocol for AI Agents — Here's How It Works
Dev.to

Your first business opportunity in 3 commands: /register_directory in @biznode_bot, wait for matches, then /my_pulse to view...
Dev.to