Recently made a post about the 35b MOE. Now the dense 27b variant has been released.


  • venusaur@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 days ago

    Thank you!!! This is awesome!

    When using llama.cpp, does it pass your prompts through a web server to process? Any privacy concerns?

    Sounds like I’m looking at a few grand to run something decent. I’ll need to do more research before I commit to that big of a purchase, but your machine sounds nice!

    Are there any small models you recommend that can run on 16GB DDR4 and an i7? No dedicated graphics card with separate VRAM. Maybe I’ll just experiment with something v small first.

    Thanks again!

    • SuspciousCarrot78@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      4 days ago

      No - it absolutely does NOT pass to a clandestine web-server. llama.cpp has thousands of eyes on the code; there’d be an uproar if there was any sneaky bullshit telemetry inbuilt.

      PS: llama.cpp has its own Web-ui front end in built (think: chatGPT but local on your machine) that’s really, really nice and worth considering as your daily chat front end.

      Small models in the 16GB range: sure. What would you like to do with your LLM? General use or something specific?

      • venusaur@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 days ago

        Thanks. I don’t understand web UI enough. Thought it had to be hosted somewhere. I’ll try it out.

        I just want to use it for general web search, data processing and maybe some light automation. In the beginning I just wanna understand how it all works and how to set it up so small model is fine.

        • SuspciousCarrot78@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 days ago

          The web-ui is the thing you type on :) You host it yourself. llama.cpp is the back end runner…it just so happens that it now has a in-built front end too. You can see it below

          https://github.com/ggml-org/llama.cpp/discussions/16938

          (Most things run llama.cpp underneath btw and then slap something else on top)

          Probably you’re going be better served with Jan.ai until your up on your feet; it’s a little friendlier / less cryptic when starting out. Jan has both llama.cpp AND a different web-ui and stuff on top. All of it always on your machine.

          https://www.jan.ai/docs/desktop/quickstart

          As I recall, Jan has has a few one-touch install models (older but pretty decent ones; worth trying when just starting out)

          • venusaur@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            4 days ago

            Got it. I’ve built simple webpages and opened those files in my browser to preview them, but didn’t know those would connect to anything.

            I’ll check out Jan and Llama and see which works for me.

            Thanks!!