| Last time I posted on how this model has performed in creating the webapp based on provided research paper. I got so much love to see people has appreciated the post and of-course the potential of this MOE model. I am sharing details on how I used this model to create webapp just using prompt and step by step guiding it. Later I converted my guidance steps into skills using same qwen-code cli with this model, that helped to add more examples. Here is github repo where I have added the research-webapp-skill that you all can use and validate potential of this model on different papers. I have added examples in the repo research-webapp-skill/examples at main · statisticalplumber/research-webapp-skill Below is the command that I use to run this model on 16GB VRAM RTX 5080 Laptop I have tried gemma4 26b moe, its not able to make app where qwen is keeping hold of context even at 70 80K. I tried latest jinja template of gemma4 and latest models from unsloth but still its not able to pull this task. Again, I might be doing somewhere wrong, as I like this model too which I am using running at llama-server native UI for other tasks. Thanks [link] [comments] |
Qwen3.5 35b is sure still one the best local model (pulling above its weight) - More Details
Reddit r/LocalLLaMA / 4/16/2026
💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- A contributor shares their experience using the Qwen3.5 35B MoE model to build a research-paper-driven web app via prompt-based step-by-step guidance and then converting those steps into reusable “skills” using the qwen-code CLI.
- They provide a GitHub repository (research-webapp-skill) with downloadable examples so others can validate the model’s ability to generate/structure applications from different papers.
- The post includes a concrete llama-server command setup for running the Qwen3.5 35B GGUF model on a laptop with 16GB VRAM (RTX 5080), including chat-template and generation/context parameters.
- The author reports trying Gemma4 26B MoE but not achieving similar context retention for this web-app task, suggesting Qwen3.5 35B performs notably better for their specific workflow.
- The overall takeaway is a practical “local LLM benchmark by usage,” emphasizing local deployability and task effectiveness rather than new model releases.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat USA
AI Business

Black Hat Asia
AI Business
Are gamers being used as free labeling labor? The rise of "Simulators" that look like AI training grounds [D]
Reddit r/MachineLearning

I built a trading intelligence MCP server in 2 days — here's how
Dev.to

Voice-Controlled AI Agent Using Whisper and Local LLM
Dev.to