Recently made a post about the 35b MOE. Now the dense 27b variant has been released.


  • venusaur@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 days ago

    Thanks! I’m experimenting with my laptop with 16GB RAM and no GPU/VRAM. I installed llama.cpp and am testing Gemma 7b Q5 but it’s not answering prompts correctly. It’s analyzing the prompt and not answering the question, or it gives me a poem haha. Trying to figure it out.

    Any lightweight model you recommend for just chat experimenting for now? Can they connect to the internet?

    • SuspciousCarrot78@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      2 days ago

      I’ll never not recommend Qwen3-4B 2507 instruct…because despite being ancient in AI terms (so, 8 months lol) it’s solid. Notably, the base models in Jan are all Qwen 3-4 variants.

      Most models can search the web, if they have access to web searching tool.