AI Navigate

[Release] Apex-1: A 350M Tiny-LLM trained locally on an RTX 5060 Ti 16GB

Reddit r/LocalLLaMA / 3/11/2026

📰 NewsTools & Practical UsageModels & Research

Key Points

  • Apex-1 is a lightweight 350 million parameter language model designed for speed and efficiency on edge devices with limited hardware capability.
  • The model architecture is based on nanoGPT and Transformer, pretrained on a subset of FineWeb-Edu dataset and finetuned with Alpaca-Cleaned data for improved instruction following.
  • Apex-1 supports ONNX and PyTorch formats, making it suitable for mobile and web deployment environments that normally cannot handle large LLMs.
  • The developer trained the model locally on consumer-grade hardware (an RTX 5060 Ti 16GB GPU), demonstrating that decent small LLMs can be created without large-scale infrastructure.
  • Future plans include Apex 1.5 and a dedicated code-focused version, with the developer seeking feedback and benchmarking from the community.

Hey everyone!

I wanted to share my latest project: Apex-1, a lightweight 350M parameter model designed for speed and efficiency on edge devices.

The Goal: I wanted to see how much "world knowledge" and instruction-following I could cram into a tiny model using consumer hardware and high-quality data.

Key Info:

  • Architecture: Based on nanoGPT / Transformer.
  • Dataset: Pre-trained on a subset of FineWeb-Edu (10BT) for reasoning and knowledge.
  • Finetuning: Alpaca-Cleaned for better instruction following.
  • Format: Weights available as ONNX (perfect for mobile/web) and standard PyTorch.

It’s great for basic summarization, simple Q&A, and running on hardware that usually can't handle LLMs.

Check it out here:https://huggingface.co/LH-Tech-AI/Apex-1-Instruct-350M

This is just the beginning – Apex 1.5 and a dedicated Code version are already in the pipeline. I'd love to get some feedback or see your benchmarks!

submitted by /u/LH-Tech_AI
[link] [comments]