• @brucethemoose
    link
    English
    1
    edit-2
    1 day ago

    Locally? Arcee 14B and the 14B Deepseek distill are currently the best models that fill fit.

    I’d recommend hosting them with TabbyAPI instead of ollama, as they will be much faster and more VRAM efficient. But this is more fuss.

    Honestly, I would just try free APIs like Gemini, Groq, and such through open web ui, or use really cheap APIs like openrouter. Newer 14B models are okay, but they’re definitely lacking that “encyclopedic intelligence” larger models have.