I'm trying to run small models on my poor laptop lol

Reddit r/LocalLLaMA / 4/10/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • A Reddit user asks whether small LLMs can be run on their laptop given constraints of an Intel i5 11th gen CPU, 24 GB RAM, and a desire to keep model RAM usage under about 4 GB while targeting 10–12 tokens per second.
  • They want an on-device “Jarvis”-style assistant for daily tasks such as remembering appointments, reading emails, interpreting information, and answering basic programming questions.
  • The post is essentially a feasibility question about model size/performance tradeoffs and suitable options for local inference on limited hardware.
  • It reflects ongoing community interest in running local LLMs (“LocalLLaMA”) and selecting models that balance speed and memory footprint.

my current specs are

Intel i5 11th generation

24 GB RAM

I would like some model with 12~10 tokens /s

and at maximum of 4 GB RAM usage

is there any model that attends my constraints?

😂😂

I want to have my own Jarvis to help me with my daily basis tasks, for example: remember some appointment, read my emails, interpret, some basic programming questions

submitted by /u/BreakfastSecure6504
[link] [comments]