• Scrubbles
    link
    fedilink
    English
    297 months ago

    Great, but it’s restrictive only letting you use openai and google. I’m already hosting oogabooga text generation, let me use that

    • @[email protected]
      link
      fedilink
      English
      197 months ago

      I believe that’s because those two APIs support function calling, open source support is still coming along.

      • Scrubbles
        link
        fedilink
        English
        97 months ago

        Ah that makes sense. That’s when I’d start using it myself. Self hosted models and audio

      • wagesj45
        link
        fedilink
        37 months ago

        Mistral Instruct v0.3 added in function calling, but I don’t know if its method for implementation is the same/compatible. Also, it is fairly new and wasn’t released all that long ago. Hopefully we’ll get there soon. :)

        • @[email protected]
          link
          fedilink
          English
          27 months ago

          I saw a few others, but the ones I looked at were basically instruct layers where you’d need to add your own parser. I didn’t find anything (in my 3 minutes of searching) that offers an openai chat completions endpoint, which is probably the main stopper.

          • wagesj45
            link
            fedilink
            17 months ago

            Looking at the documentation it looks like it relies on Mistral’s python tooling to work. I’m fairly dumb, so I don’t know if the tool suggestion coming from Mistral is from some kind of separate neural net or as some kind of special response you have to parse (or that their client parses for you?).

  • @[email protected]
    link
    fedilink
    English
    107 months ago

    Okay but when can we use the weather forecast on our dashboards? Functionality was retired with no replacement

  • muppeth
    link
    fedilink
    English
    17 months ago

    Ok. Npw Its definitely time to migrate my instance to something more powerful then my raspberry pi

  • @bushvin
    link
    English
    -17 months ago

    Oh cool, implementing mediocre algorithms. What could possibly go wrong?

    • @warmaster
      link
      English
      7
      edit-2
      7 months ago

      Local LLMs have been supported via the Ollama integration since Home Assistant 2024.4. Ollama and the major open source LLM models are not tuned for tool calling, so this has to be built from scratch and was not done in time for this release. We’re collaborating with NVIDIA to get this working – they showed a prototype last week.

      Are all Ollama-supported algos mediocre? Which ones would be better?