• @Larry
    link
    English
    37 months ago

    I tried llamas when they were initially released, and it seems like training took garbage amounts of GPU. Did that change?

    • @Womble
      link
      English
      27 months ago

      Look into quantised models (like gguf format) these significantly reduce the amout of memory needed and speed up computation time at the expense of some quality. If you have 16GB of rm or more you can run decent models locally without any gpu, though your speed will be more like 1 word a second than chatgpt speeds