• @filisterOP
    link
    English
    429 days ago

    That’s a very toxic attitude.

    Inference is in principle the process of generation of the AI response. So when you run locally and LLM you are using your GPU only for inference.