Hey, r/LocalLLaMA !
I'm back with a new model: Apex 1.6 Instruct 350M
This is basically something like Apex 1, Apex 1.5 or Apex 1.5 Coder, but it's my most powerful chat model this march!
Why?
Because I changed the ratio of instruction data to pretraining data in the finetuning script to 2:1 - so the ratio is 2x Alpaca-Cleaned to 1x Fineweb-Edu-10BT.
This increased the world knowledge again a bit compared to Apex 1.5 Coder (which was already a huge leap better than Apex 1 and Apex 1.5 :D)!
You can download the code and the weights here on HF: https://huggingface.co/LH-Tech-AI/Apex-1.6-Instruct-350M/
And you can use it in the GGUF format for example in Ollama, LM Studio or llama.cpp.
Example of usage in Ollama:
ollama run hf.co/LH-Tech-AI/Apex-1.6-Instruct-350M
Here's a overview that compares Apex 1.5 Coder with the brand new Apex 1.6:
| Category | Apex 1.5 Coder | Apex 1.6 | Summary |
|---|---|---|---|
| AI definition | Precise but boring | Much more complex sentences, more interesting, uses lists and better structure. | 1.6 seems to be more educated |
| Logic (train from Munich to Berlin - how long does it take) | Correct (4 hours) but very short answer → could be guessed! | Wrong! | 1.5 is winning here |
| Python Code | Completely wrong! | Uses markdown blocks, but the code was wrong | 1.6 is MUCH better! |
| Flight (NY-LDN) | Thinks that it’s a 1,5 hour flight and it would cost $20,000! | Explains why taking the bus is good?! | Both are hardly hallucinating. |
| Humor (joke) | Gives a definition of robots! | Tries to describe robots poetically… | 1.6 is better. |
| Explanation (FFT) | Technically wrong! | Technically almost correct. | 1.6 is more helpful. |
Have fun with my new model! :D
Coming soon: Axiom 1 Coder Instruct 350M - a coding and math logic model based on the base model of Apex 1... Stay tuned! Axiom 1 Coder will focus on fixing the logic issues seen in 1.6 by using Orca-Math and a massive HTML structure boost.
[link] [comments]




