They support Claude, ChatGPT, Gemini, HuggingChat, and Mistral.

  • Lojcs@lemm.ee
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    9 hours ago

    Last time I tried using a local llm (about a year ago) it generated only a couple words per second and the answers were barely relevant. Also I don’t see how a local llm can fulfill the glorified search engine role that people use llms for.

    • ocassionallyaduck@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      3 hours ago

      Try again. Simplified models take the large ones and pare them down in terms of memory requirements, and can be run off the CPU even. The “smol” model I mentioned is real, and hyperfast.

      Llama 3.2 is pretty solid as well.

    • TheDorkfromYork@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 hours ago

      They’re fast and high quality now. ChatGPT is the best, but local llms are great, even with 10gb of vram.